1 /******************************************************************************
3 Copyright (c) 2001-2013, Intel Corporation
6 Redistribution and use in source and binary forms, with or without
7 modification, are permitted provided that the following conditions are met:
9 1. Redistributions of source code must retain the above copyright notice,
10 this list of conditions and the following disclaimer.
12 2. Redistributions in binary form must reproduce the above copyright
13 notice, this list of conditions and the following disclaimer in the
14 documentation and/or other materials provided with the distribution.
16 3. Neither the name of the Intel Corporation nor the names of its
17 contributors may be used to endorse or promote products derived from
18 this software without specific prior written permission.
20 THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
21 AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22 IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23 ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
24 LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25 CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26 SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27 INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28 CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29 ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
30 POSSIBILITY OF SUCH DAMAGE.
32 ******************************************************************************/
35 #include "e1000_api.h"
38 static s32 e1000_acquire_nvm_i210(struct e1000_hw *hw);
39 static void e1000_release_nvm_i210(struct e1000_hw *hw);
40 static s32 e1000_get_hw_semaphore_i210(struct e1000_hw *hw);
41 static s32 e1000_write_nvm_srwr(struct e1000_hw *hw, u16 offset, u16 words,
43 static s32 e1000_pool_flash_update_done_i210(struct e1000_hw *hw);
44 static s32 e1000_valid_led_default_i210(struct e1000_hw *hw, u16 *data);
47 * e1000_acquire_nvm_i210 - Request for access to EEPROM
48 * @hw: pointer to the HW structure
50 * Acquire the necessary semaphores for exclusive access to the EEPROM.
51 * Set the EEPROM access request bit and wait for EEPROM access grant bit.
52 * Return successful if access grant bit set, else clear the request for
53 * EEPROM access and return -E1000_ERR_NVM (-1).
55 static s32 e1000_acquire_nvm_i210(struct e1000_hw *hw)
59 DEBUGFUNC("e1000_acquire_nvm_i210");
61 ret_val = e1000_acquire_swfw_sync_i210(hw, E1000_SWFW_EEP_SM);
67 * e1000_release_nvm_i210 - Release exclusive access to EEPROM
68 * @hw: pointer to the HW structure
70 * Stop any current commands to the EEPROM and clear the EEPROM request bit,
71 * then release the semaphores acquired.
73 static void e1000_release_nvm_i210(struct e1000_hw *hw)
75 DEBUGFUNC("e1000_release_nvm_i210");
77 e1000_release_swfw_sync_i210(hw, E1000_SWFW_EEP_SM);
81 * e1000_acquire_swfw_sync_i210 - Acquire SW/FW semaphore
82 * @hw: pointer to the HW structure
83 * @mask: specifies which semaphore to acquire
85 * Acquire the SW/FW semaphore to access the PHY or NVM. The mask
86 * will also specify which port we're acquiring the lock for.
88 s32 e1000_acquire_swfw_sync_i210(struct e1000_hw *hw, u16 mask)
92 u32 fwmask = mask << 16;
93 s32 ret_val = E1000_SUCCESS;
94 s32 i = 0, timeout = 200; /* FIXME: find real value to use here */
96 DEBUGFUNC("e1000_acquire_swfw_sync_i210");
99 if (e1000_get_hw_semaphore_i210(hw)) {
100 ret_val = -E1000_ERR_SWFW_SYNC;
104 swfw_sync = E1000_READ_REG(hw, E1000_SW_FW_SYNC);
105 if (!(swfw_sync & (fwmask | swmask)))
109 * Firmware currently using resource (fwmask)
110 * or other software thread using resource (swmask)
112 e1000_put_hw_semaphore_generic(hw);
118 DEBUGOUT("Driver can't access resource, SW_FW_SYNC timeout.\n");
119 ret_val = -E1000_ERR_SWFW_SYNC;
124 E1000_WRITE_REG(hw, E1000_SW_FW_SYNC, swfw_sync);
126 e1000_put_hw_semaphore_generic(hw);
133 * e1000_release_swfw_sync_i210 - Release SW/FW semaphore
134 * @hw: pointer to the HW structure
135 * @mask: specifies which semaphore to acquire
137 * Release the SW/FW semaphore used to access the PHY or NVM. The mask
138 * will also specify which port we're releasing the lock for.
140 void e1000_release_swfw_sync_i210(struct e1000_hw *hw, u16 mask)
144 DEBUGFUNC("e1000_release_swfw_sync_i210");
146 while (e1000_get_hw_semaphore_i210(hw) != E1000_SUCCESS)
149 swfw_sync = E1000_READ_REG(hw, E1000_SW_FW_SYNC);
151 E1000_WRITE_REG(hw, E1000_SW_FW_SYNC, swfw_sync);
153 e1000_put_hw_semaphore_generic(hw);
157 * e1000_get_hw_semaphore_i210 - Acquire hardware semaphore
158 * @hw: pointer to the HW structure
160 * Acquire the HW semaphore to access the PHY or NVM
162 static s32 e1000_get_hw_semaphore_i210(struct e1000_hw *hw)
165 s32 timeout = hw->nvm.word_size + 1;
168 DEBUGFUNC("e1000_get_hw_semaphore_i210");
170 /* Get the SW semaphore */
171 while (i < timeout) {
172 swsm = E1000_READ_REG(hw, E1000_SWSM);
173 if (!(swsm & E1000_SWSM_SMBI))
181 /* In rare circumstances, the SW semaphore may already be held
182 * unintentionally. Clear the semaphore once before giving up.
184 if (hw->dev_spec._82575.clear_semaphore_once) {
185 hw->dev_spec._82575.clear_semaphore_once = FALSE;
186 e1000_put_hw_semaphore_generic(hw);
187 for (i = 0; i < timeout; i++) {
188 swsm = E1000_READ_REG(hw, E1000_SWSM);
189 if (!(swsm & E1000_SWSM_SMBI))
196 /* If we do not have the semaphore here, we have to give up. */
198 DEBUGOUT("Driver can't access device - SMBI bit is set.\n");
199 return -E1000_ERR_NVM;
203 /* Get the FW semaphore. */
204 for (i = 0; i < timeout; i++) {
205 swsm = E1000_READ_REG(hw, E1000_SWSM);
206 E1000_WRITE_REG(hw, E1000_SWSM, swsm | E1000_SWSM_SWESMBI);
208 /* Semaphore acquired if bit latched */
209 if (E1000_READ_REG(hw, E1000_SWSM) & E1000_SWSM_SWESMBI)
216 /* Release semaphores */
217 e1000_put_hw_semaphore_generic(hw);
218 DEBUGOUT("Driver can't access the NVM\n");
219 return -E1000_ERR_NVM;
222 return E1000_SUCCESS;
226 * e1000_read_nvm_srrd_i210 - Reads Shadow Ram using EERD register
227 * @hw: pointer to the HW structure
228 * @offset: offset of word in the Shadow Ram to read
229 * @words: number of words to read
230 * @data: word read from the Shadow Ram
232 * Reads a 16 bit word from the Shadow Ram using the EERD register.
233 * Uses necessary synchronization semaphores.
235 s32 e1000_read_nvm_srrd_i210(struct e1000_hw *hw, u16 offset, u16 words,
238 s32 status = E1000_SUCCESS;
241 DEBUGFUNC("e1000_read_nvm_srrd_i210");
243 /* We cannot hold synchronization semaphores for too long,
244 * because of forceful takeover procedure. However it is more efficient
245 * to read in bursts than synchronizing access for each word. */
246 for (i = 0; i < words; i += E1000_EERD_EEWR_MAX_COUNT) {
247 count = (words - i) / E1000_EERD_EEWR_MAX_COUNT > 0 ?
248 E1000_EERD_EEWR_MAX_COUNT : (words - i);
249 if (hw->nvm.ops.acquire(hw) == E1000_SUCCESS) {
250 status = e1000_read_nvm_eerd(hw, offset, count,
252 hw->nvm.ops.release(hw);
254 status = E1000_ERR_SWFW_SYNC;
257 if (status != E1000_SUCCESS)
265 * e1000_write_nvm_srwr_i210 - Write to Shadow RAM using EEWR
266 * @hw: pointer to the HW structure
267 * @offset: offset within the Shadow RAM to be written to
268 * @words: number of words to write
269 * @data: 16 bit word(s) to be written to the Shadow RAM
271 * Writes data to Shadow RAM at offset using EEWR register.
273 * If e1000_update_nvm_checksum is not called after this function , the
274 * data will not be committed to FLASH and also Shadow RAM will most likely
275 * contain an invalid checksum.
277 * If error code is returned, data and Shadow RAM may be inconsistent - buffer
280 s32 e1000_write_nvm_srwr_i210(struct e1000_hw *hw, u16 offset, u16 words,
283 s32 status = E1000_SUCCESS;
286 DEBUGFUNC("e1000_write_nvm_srwr_i210");
288 /* We cannot hold synchronization semaphores for too long,
289 * because of forceful takeover procedure. However it is more efficient
290 * to write in bursts than synchronizing access for each word. */
291 for (i = 0; i < words; i += E1000_EERD_EEWR_MAX_COUNT) {
292 count = (words - i) / E1000_EERD_EEWR_MAX_COUNT > 0 ?
293 E1000_EERD_EEWR_MAX_COUNT : (words - i);
294 if (hw->nvm.ops.acquire(hw) == E1000_SUCCESS) {
295 status = e1000_write_nvm_srwr(hw, offset, count,
297 hw->nvm.ops.release(hw);
299 status = E1000_ERR_SWFW_SYNC;
302 if (status != E1000_SUCCESS)
310 * e1000_write_nvm_srwr - Write to Shadow Ram using EEWR
311 * @hw: pointer to the HW structure
312 * @offset: offset within the Shadow Ram to be written to
313 * @words: number of words to write
314 * @data: 16 bit word(s) to be written to the Shadow Ram
316 * Writes data to Shadow Ram at offset using EEWR register.
318 * If e1000_update_nvm_checksum is not called after this function , the
319 * Shadow Ram will most likely contain an invalid checksum.
321 static s32 e1000_write_nvm_srwr(struct e1000_hw *hw, u16 offset, u16 words,
324 struct e1000_nvm_info *nvm = &hw->nvm;
326 u32 attempts = 100000;
327 s32 ret_val = E1000_SUCCESS;
329 DEBUGFUNC("e1000_write_nvm_srwr");
332 * A check for invalid values: offset too large, too many words,
333 * too many words for the offset, and not enough words.
335 if ((offset >= nvm->word_size) || (words > (nvm->word_size - offset)) ||
337 DEBUGOUT("nvm parameter(s) out of bounds\n");
338 ret_val = -E1000_ERR_NVM;
342 for (i = 0; i < words; i++) {
343 eewr = ((offset+i) << E1000_NVM_RW_ADDR_SHIFT) |
344 (data[i] << E1000_NVM_RW_REG_DATA) |
345 E1000_NVM_RW_REG_START;
347 E1000_WRITE_REG(hw, E1000_SRWR, eewr);
349 for (k = 0; k < attempts; k++) {
350 if (E1000_NVM_RW_REG_DONE &
351 E1000_READ_REG(hw, E1000_SRWR)) {
352 ret_val = E1000_SUCCESS;
358 if (ret_val != E1000_SUCCESS) {
359 DEBUGOUT("Shadow RAM write EEWR timed out\n");
368 /** e1000_read_invm_word_i210 - Reads OTP
369 * @hw: pointer to the HW structure
370 * @address: the word address (aka eeprom offset) to read
371 * @data: pointer to the data read
373 * Reads 16-bit words from the OTP. Return error when the word is not
376 static s32 e1000_read_invm_word_i210(struct e1000_hw *hw, u8 address, u16 *data)
378 s32 status = -E1000_ERR_INVM_VALUE_NOT_FOUND;
381 u8 record_type, word_address;
383 DEBUGFUNC("e1000_read_invm_word_i210");
385 for (i = 0; i < E1000_INVM_SIZE; i++) {
386 invm_dword = E1000_READ_REG(hw, E1000_INVM_DATA_REG(i));
387 /* Get record type */
388 record_type = INVM_DWORD_TO_RECORD_TYPE(invm_dword);
389 if (record_type == E1000_INVM_UNINITIALIZED_STRUCTURE)
391 if (record_type == E1000_INVM_CSR_AUTOLOAD_STRUCTURE)
392 i += E1000_INVM_CSR_AUTOLOAD_DATA_SIZE_IN_DWORDS;
393 if (record_type == E1000_INVM_RSA_KEY_SHA256_STRUCTURE)
394 i += E1000_INVM_RSA_KEY_SHA256_DATA_SIZE_IN_DWORDS;
395 if (record_type == E1000_INVM_WORD_AUTOLOAD_STRUCTURE) {
396 word_address = INVM_DWORD_TO_WORD_ADDRESS(invm_dword);
397 if (word_address == address) {
398 *data = INVM_DWORD_TO_WORD_DATA(invm_dword);
399 DEBUGOUT2("Read INVM Word 0x%02x = %x",
401 status = E1000_SUCCESS;
406 if (status != E1000_SUCCESS)
407 DEBUGOUT1("Requested word 0x%02x not found in OTP\n", address);
411 /** e1000_read_invm_i210 - Read invm wrapper function for I210/I211
412 * @hw: pointer to the HW structure
413 * @address: the word address (aka eeprom offset) to read
414 * @data: pointer to the data read
416 * Wrapper function to return data formerly found in the NVM.
418 static s32 e1000_read_invm_i210(struct e1000_hw *hw, u16 offset,
419 u16 E1000_UNUSEDARG words, u16 *data)
421 s32 ret_val = E1000_SUCCESS;
423 DEBUGFUNC("e1000_read_invm_i210");
425 /* Only the MAC addr is required to be present in the iNVM */
428 ret_val = e1000_read_invm_word_i210(hw, (u8)offset, &data[0]);
429 ret_val |= e1000_read_invm_word_i210(hw, (u8)offset+1,
431 ret_val |= e1000_read_invm_word_i210(hw, (u8)offset+2,
433 if (ret_val != E1000_SUCCESS)
434 DEBUGOUT("MAC Addr not found in iNVM\n");
436 case NVM_INIT_CTRL_2:
437 ret_val = e1000_read_invm_word_i210(hw, (u8)offset, data);
438 if (ret_val != E1000_SUCCESS) {
439 *data = NVM_INIT_CTRL_2_DEFAULT_I211;
440 ret_val = E1000_SUCCESS;
443 case NVM_INIT_CTRL_4:
444 ret_val = e1000_read_invm_word_i210(hw, (u8)offset, data);
445 if (ret_val != E1000_SUCCESS) {
446 *data = NVM_INIT_CTRL_4_DEFAULT_I211;
447 ret_val = E1000_SUCCESS;
451 ret_val = e1000_read_invm_word_i210(hw, (u8)offset, data);
452 if (ret_val != E1000_SUCCESS) {
453 *data = NVM_LED_1_CFG_DEFAULT_I211;
454 ret_val = E1000_SUCCESS;
457 case NVM_LED_0_2_CFG:
458 ret_val = e1000_read_invm_word_i210(hw, (u8)offset, data);
459 if (ret_val != E1000_SUCCESS) {
460 *data = NVM_LED_0_2_CFG_DEFAULT_I211;
461 ret_val = E1000_SUCCESS;
464 case NVM_ID_LED_SETTINGS:
465 ret_val = e1000_read_invm_word_i210(hw, (u8)offset, data);
466 if (ret_val != E1000_SUCCESS) {
467 *data = ID_LED_RESERVED_FFFF;
468 ret_val = E1000_SUCCESS;
472 *data = hw->subsystem_device_id;
475 *data = hw->subsystem_vendor_id;
478 *data = hw->device_id;
481 *data = hw->vendor_id;
484 DEBUGOUT1("NVM word 0x%02x is not mapped.\n", offset);
485 *data = NVM_RESERVED_WORD;
492 * e1000_read_invm_version - Reads iNVM version and image type
493 * @hw: pointer to the HW structure
494 * @invm_ver: version structure for the version read
496 * Reads iNVM version and image type.
498 s32 e1000_read_invm_version(struct e1000_hw *hw,
499 struct e1000_fw_version *invm_ver)
502 u32 *next_record = NULL;
505 u32 invm_blocks = E1000_INVM_SIZE - (E1000_INVM_ULT_BYTES_SIZE /
506 E1000_INVM_RECORD_SIZE_IN_BYTES);
507 u32 buffer[E1000_INVM_SIZE];
508 s32 status = -E1000_ERR_INVM_VALUE_NOT_FOUND;
511 DEBUGFUNC("e1000_read_invm_version");
513 /* Read iNVM memory */
514 for (i = 0; i < E1000_INVM_SIZE; i++) {
515 invm_dword = E1000_READ_REG(hw, E1000_INVM_DATA_REG(i));
516 buffer[i] = invm_dword;
519 /* Read version number */
520 for (i = 1; i < invm_blocks; i++) {
521 record = &buffer[invm_blocks - i];
522 next_record = &buffer[invm_blocks - i + 1];
524 /* Check if we have first version location used */
525 if ((i == 1) && ((*record & E1000_INVM_VER_FIELD_ONE) == 0)) {
527 status = E1000_SUCCESS;
530 /* Check if we have second version location used */
532 ((*record & E1000_INVM_VER_FIELD_TWO) == 0)) {
533 version = (*record & E1000_INVM_VER_FIELD_ONE) >> 3;
534 status = E1000_SUCCESS;
538 * Check if we have odd version location
539 * used and it is the last one used
541 else if ((((*record & E1000_INVM_VER_FIELD_ONE) == 0) &&
542 ((*record & 0x3) == 0)) || (((*record & 0x3) != 0) &&
544 version = (*next_record & E1000_INVM_VER_FIELD_TWO)
546 status = E1000_SUCCESS;
550 * Check if we have even version location
551 * used and it is the last one used
553 else if (((*record & E1000_INVM_VER_FIELD_TWO) == 0) &&
554 ((*record & 0x3) == 0)) {
555 version = (*record & E1000_INVM_VER_FIELD_ONE) >> 3;
556 status = E1000_SUCCESS;
561 if (status == E1000_SUCCESS) {
562 invm_ver->invm_major = (version & E1000_INVM_MAJOR_MASK)
563 >> E1000_INVM_MAJOR_SHIFT;
564 invm_ver->invm_minor = version & E1000_INVM_MINOR_MASK;
566 /* Read Image Type */
567 for (i = 1; i < invm_blocks; i++) {
568 record = &buffer[invm_blocks - i];
569 next_record = &buffer[invm_blocks - i + 1];
571 /* Check if we have image type in first location used */
572 if ((i == 1) && ((*record & E1000_INVM_IMGTYPE_FIELD) == 0)) {
573 invm_ver->invm_img_type = 0;
574 status = E1000_SUCCESS;
577 /* Check if we have image type in first location used */
578 else if ((((*record & 0x3) == 0) &&
579 ((*record & E1000_INVM_IMGTYPE_FIELD) == 0)) ||
580 ((((*record & 0x3) != 0) && (i != 1)))) {
581 invm_ver->invm_img_type =
582 (*next_record & E1000_INVM_IMGTYPE_FIELD) >> 23;
583 status = E1000_SUCCESS;
591 * e1000_validate_nvm_checksum_i210 - Validate EEPROM checksum
592 * @hw: pointer to the HW structure
594 * Calculates the EEPROM checksum by reading/adding each word of the EEPROM
595 * and then verifies that the sum of the EEPROM is equal to 0xBABA.
597 s32 e1000_validate_nvm_checksum_i210(struct e1000_hw *hw)
599 s32 status = E1000_SUCCESS;
600 s32 (*read_op_ptr)(struct e1000_hw *, u16, u16, u16 *);
602 DEBUGFUNC("e1000_validate_nvm_checksum_i210");
604 if (hw->nvm.ops.acquire(hw) == E1000_SUCCESS) {
607 * Replace the read function with semaphore grabbing with
608 * the one that skips this for a while.
609 * We have semaphore taken already here.
611 read_op_ptr = hw->nvm.ops.read;
612 hw->nvm.ops.read = e1000_read_nvm_eerd;
614 status = e1000_validate_nvm_checksum_generic(hw);
616 /* Revert original read operation. */
617 hw->nvm.ops.read = read_op_ptr;
619 hw->nvm.ops.release(hw);
621 status = E1000_ERR_SWFW_SYNC;
629 * e1000_update_nvm_checksum_i210 - Update EEPROM checksum
630 * @hw: pointer to the HW structure
632 * Updates the EEPROM checksum by reading/adding each word of the EEPROM
633 * up to the checksum. Then calculates the EEPROM checksum and writes the
634 * value to the EEPROM. Next commit EEPROM data onto the Flash.
636 s32 e1000_update_nvm_checksum_i210(struct e1000_hw *hw)
642 DEBUGFUNC("e1000_update_nvm_checksum_i210");
645 * Read the first word from the EEPROM. If this times out or fails, do
646 * not continue or we could be in for a very long wait while every
649 ret_val = e1000_read_nvm_eerd(hw, 0, 1, &nvm_data);
650 if (ret_val != E1000_SUCCESS) {
651 DEBUGOUT("EEPROM read failed\n");
655 if (hw->nvm.ops.acquire(hw) == E1000_SUCCESS) {
657 * Do not use hw->nvm.ops.write, hw->nvm.ops.read
658 * because we do not want to take the synchronization
659 * semaphores twice here.
662 for (i = 0; i < NVM_CHECKSUM_REG; i++) {
663 ret_val = e1000_read_nvm_eerd(hw, i, 1, &nvm_data);
665 hw->nvm.ops.release(hw);
666 DEBUGOUT("NVM Read Error while updating checksum.\n");
669 checksum += nvm_data;
671 checksum = (u16) NVM_SUM - checksum;
672 ret_val = e1000_write_nvm_srwr(hw, NVM_CHECKSUM_REG, 1,
674 if (ret_val != E1000_SUCCESS) {
675 hw->nvm.ops.release(hw);
676 DEBUGOUT("NVM Write Error while updating checksum.\n");
680 hw->nvm.ops.release(hw);
682 ret_val = e1000_update_flash_i210(hw);
684 ret_val = E1000_ERR_SWFW_SYNC;
691 * e1000_get_flash_presence_i210 - Check if flash device is detected.
692 * @hw: pointer to the HW structure
695 bool e1000_get_flash_presence_i210(struct e1000_hw *hw)
698 bool ret_val = FALSE;
700 DEBUGFUNC("e1000_get_flash_presence_i210");
702 eec = E1000_READ_REG(hw, E1000_EECD);
704 if (eec & E1000_EECD_FLASH_DETECTED_I210)
711 * e1000_update_flash_i210 - Commit EEPROM to the flash
712 * @hw: pointer to the HW structure
715 s32 e1000_update_flash_i210(struct e1000_hw *hw)
720 DEBUGFUNC("e1000_update_flash_i210");
722 ret_val = e1000_pool_flash_update_done_i210(hw);
723 if (ret_val == -E1000_ERR_NVM) {
724 DEBUGOUT("Flash update time out\n");
728 flup = E1000_READ_REG(hw, E1000_EECD) | E1000_EECD_FLUPD_I210;
729 E1000_WRITE_REG(hw, E1000_EECD, flup);
731 ret_val = e1000_pool_flash_update_done_i210(hw);
732 if (ret_val == E1000_SUCCESS)
733 DEBUGOUT("Flash update complete\n");
735 DEBUGOUT("Flash update time out\n");
742 * e1000_pool_flash_update_done_i210 - Pool FLUDONE status.
743 * @hw: pointer to the HW structure
746 s32 e1000_pool_flash_update_done_i210(struct e1000_hw *hw)
748 s32 ret_val = -E1000_ERR_NVM;
751 DEBUGFUNC("e1000_pool_flash_update_done_i210");
753 for (i = 0; i < E1000_FLUDONE_ATTEMPTS; i++) {
754 reg = E1000_READ_REG(hw, E1000_EECD);
755 if (reg & E1000_EECD_FLUDONE_I210) {
756 ret_val = E1000_SUCCESS;
766 * e1000_init_nvm_params_i210 - Initialize i210 NVM function pointers
767 * @hw: pointer to the HW structure
769 * Initialize the i210/i211 NVM parameters and function pointers.
771 static s32 e1000_init_nvm_params_i210(struct e1000_hw *hw)
774 struct e1000_nvm_info *nvm = &hw->nvm;
776 DEBUGFUNC("e1000_init_nvm_params_i210");
778 ret_val = e1000_init_nvm_params_82575(hw);
779 nvm->ops.acquire = e1000_acquire_nvm_i210;
780 nvm->ops.release = e1000_release_nvm_i210;
781 nvm->ops.valid_led_default = e1000_valid_led_default_i210;
782 if (e1000_get_flash_presence_i210(hw)) {
783 hw->nvm.type = e1000_nvm_flash_hw;
784 nvm->ops.read = e1000_read_nvm_srrd_i210;
785 nvm->ops.write = e1000_write_nvm_srwr_i210;
786 nvm->ops.validate = e1000_validate_nvm_checksum_i210;
787 nvm->ops.update = e1000_update_nvm_checksum_i210;
789 hw->nvm.type = e1000_nvm_invm;
790 nvm->ops.read = e1000_read_invm_i210;
791 nvm->ops.write = e1000_null_write_nvm;
792 nvm->ops.validate = e1000_null_ops_generic;
793 nvm->ops.update = e1000_null_ops_generic;
799 * e1000_init_function_pointers_i210 - Init func ptrs.
800 * @hw: pointer to the HW structure
802 * Called to initialize all function pointers and parameters.
804 void e1000_init_function_pointers_i210(struct e1000_hw *hw)
806 e1000_init_function_pointers_82575(hw);
807 hw->nvm.ops.init_params = e1000_init_nvm_params_i210;
813 * e1000_valid_led_default_i210 - Verify a valid default LED config
814 * @hw: pointer to the HW structure
815 * @data: pointer to the NVM (EEPROM)
817 * Read the EEPROM for the current default LED configuration. If the
818 * LED configuration is not valid, set to a valid LED configuration.
820 static s32 e1000_valid_led_default_i210(struct e1000_hw *hw, u16 *data)
824 DEBUGFUNC("e1000_valid_led_default_i210");
826 ret_val = hw->nvm.ops.read(hw, NVM_ID_LED_SETTINGS, 1, data);
828 DEBUGOUT("NVM Read Error\n");
832 if (*data == ID_LED_RESERVED_0000 || *data == ID_LED_RESERVED_FFFF) {
833 switch (hw->phy.media_type) {
834 case e1000_media_type_internal_serdes:
835 *data = ID_LED_DEFAULT_I210_SERDES;
837 case e1000_media_type_copper:
839 *data = ID_LED_DEFAULT_I210;
848 * __e1000_access_xmdio_reg - Read/write XMDIO register
849 * @hw: pointer to the HW structure
850 * @address: XMDIO address to program
851 * @dev_addr: device address to program
852 * @data: pointer to value to read/write from/to the XMDIO address
853 * @read: boolean flag to indicate read or write
855 static s32 __e1000_access_xmdio_reg(struct e1000_hw *hw, u16 address,
856 u8 dev_addr, u16 *data, bool read)
860 DEBUGFUNC("__e1000_access_xmdio_reg");
862 ret_val = hw->phy.ops.write_reg(hw, E1000_MMDAC, dev_addr);
866 ret_val = hw->phy.ops.write_reg(hw, E1000_MMDAAD, address);
870 ret_val = hw->phy.ops.write_reg(hw, E1000_MMDAC, E1000_MMDAC_FUNC_DATA |
876 ret_val = hw->phy.ops.read_reg(hw, E1000_MMDAAD, data);
878 ret_val = hw->phy.ops.write_reg(hw, E1000_MMDAAD, *data);
882 /* Recalibrate the device back to 0 */
883 ret_val = hw->phy.ops.write_reg(hw, E1000_MMDAC, 0);
891 * e1000_read_xmdio_reg - Read XMDIO register
892 * @hw: pointer to the HW structure
893 * @addr: XMDIO address to program
894 * @dev_addr: device address to program
895 * @data: value to be read from the EMI address
897 s32 e1000_read_xmdio_reg(struct e1000_hw *hw, u16 addr, u8 dev_addr, u16 *data)
899 DEBUGFUNC("e1000_read_xmdio_reg");
901 return __e1000_access_xmdio_reg(hw, addr, dev_addr, data, TRUE);
905 * e1000_write_xmdio_reg - Write XMDIO register
906 * @hw: pointer to the HW structure
907 * @addr: XMDIO address to program
908 * @dev_addr: device address to program
909 * @data: value to be written to the XMDIO address
911 s32 e1000_write_xmdio_reg(struct e1000_hw *hw, u16 addr, u8 dev_addr, u16 data)
913 DEBUGFUNC("e1000_read_xmdio_reg");
915 return __e1000_access_xmdio_reg(hw, addr, dev_addr, &data, FALSE);
919 * e1000_pll_workaround_i210
920 * @hw: pointer to the HW structure
922 * Works around an errata in the PLL circuit where it occasionally
923 * provides the wrong clock frequency after power up.
925 static s32 e1000_pll_workaround_i210(struct e1000_hw *hw)
928 u32 wuc, mdicnfg, ctrl, ctrl_ext, reg_val;
929 u16 nvm_word, phy_word, pci_word, tmp_nvm;
932 /* Get and set needed register values */
933 wuc = E1000_READ_REG(hw, E1000_WUC);
934 mdicnfg = E1000_READ_REG(hw, E1000_MDICNFG);
935 reg_val = mdicnfg & ~E1000_MDICNFG_EXT_MDIO;
936 E1000_WRITE_REG(hw, E1000_MDICNFG, reg_val);
938 /* Get data from NVM, or set default */
939 ret_val = e1000_read_invm_word_i210(hw, E1000_INVM_AUTOLOAD,
941 if (ret_val != E1000_SUCCESS)
942 nvm_word = E1000_INVM_DEFAULT_AL;
943 tmp_nvm = nvm_word | E1000_INVM_PLL_WO_VAL;
944 for (i = 0; i < E1000_MAX_PLL_TRIES; i++) {
945 /* check current state directly from internal PHY */
946 e1000_read_phy_reg_gs40g(hw, (E1000_PHY_PLL_FREQ_PAGE |
947 E1000_PHY_PLL_FREQ_REG), &phy_word);
948 if ((phy_word & E1000_PHY_PLL_UNCONF)
949 != E1000_PHY_PLL_UNCONF) {
950 ret_val = E1000_SUCCESS;
953 ret_val = -E1000_ERR_PHY;
955 /* directly reset the internal PHY */
956 ctrl = E1000_READ_REG(hw, E1000_CTRL);
957 E1000_WRITE_REG(hw, E1000_CTRL, ctrl|E1000_CTRL_PHY_RST);
959 ctrl_ext = E1000_READ_REG(hw, E1000_CTRL_EXT);
960 ctrl_ext |= (E1000_CTRL_EXT_PHYPDEN | E1000_CTRL_EXT_SDLPE);
961 E1000_WRITE_REG(hw, E1000_CTRL_EXT, ctrl_ext);
963 E1000_WRITE_REG(hw, E1000_WUC, 0);
964 reg_val = (E1000_INVM_AUTOLOAD << 4) | (tmp_nvm << 16);
965 E1000_WRITE_REG(hw, E1000_EEARBC_I210, reg_val);
967 e1000_read_pci_cfg(hw, E1000_PCI_PMCSR, &pci_word);
968 pci_word |= E1000_PCI_PMCSR_D3;
969 e1000_write_pci_cfg(hw, E1000_PCI_PMCSR, &pci_word);
971 pci_word &= ~E1000_PCI_PMCSR_D3;
972 e1000_write_pci_cfg(hw, E1000_PCI_PMCSR, &pci_word);
973 reg_val = (E1000_INVM_AUTOLOAD << 4) | (nvm_word << 16);
974 E1000_WRITE_REG(hw, E1000_EEARBC_I210, reg_val);
976 /* restore WUC register */
977 E1000_WRITE_REG(hw, E1000_WUC, wuc);
979 /* restore MDICNFG setting */
980 E1000_WRITE_REG(hw, E1000_MDICNFG, mdicnfg);
985 * e1000_init_hw_i210 - Init hw for I210/I211
986 * @hw: pointer to the HW structure
988 * Called to initialize hw for i210 hw family.
990 s32 e1000_init_hw_i210(struct e1000_hw *hw)
994 DEBUGFUNC("e1000_init_hw_i210");
995 if ((hw->mac.type >= e1000_i210) &&
996 !(e1000_get_flash_presence_i210(hw))) {
997 ret_val = e1000_pll_workaround_i210(hw);
998 if (ret_val != E1000_SUCCESS)
1001 ret_val = e1000_init_hw_82575(hw);