1// SPDX-License-Identifier: GPL-2.0 2/* Copyright(c) 1999 - 2018 Intel Corporation. */ 3 4#include <linux/pci.h> 5#include <linux/delay.h> 6#include <linux/sched.h> 7 8#include "ixgbe.h" 9#include "ixgbe_phy.h" 10#include "ixgbe_mbx.h" 11 12#define IXGBE_82599_MAX_TX_QUEUES 128 13#define IXGBE_82599_MAX_RX_QUEUES 128 14#define IXGBE_82599_RAR_ENTRIES 128 15#define IXGBE_82599_MC_TBL_SIZE 128 16#define IXGBE_82599_VFT_TBL_SIZE 128 17#define IXGBE_82599_RX_PB_SIZE 512 18 19static void ixgbe_disable_tx_laser_multispeed_fiber(struct ixgbe_hw *hw); 20static void ixgbe_enable_tx_laser_multispeed_fiber(struct ixgbe_hw *hw); 21static void ixgbe_flap_tx_laser_multispeed_fiber(struct ixgbe_hw *hw); 22static void 23ixgbe_set_hard_rate_select_speed(struct ixgbe_hw *, ixgbe_link_speed); 24static s32 ixgbe_setup_mac_link_smartspeed(struct ixgbe_hw *hw, 25 ixgbe_link_speed speed, 26 bool autoneg_wait_to_complete); 27static void ixgbe_stop_mac_link_on_d3_82599(struct ixgbe_hw *hw); 28static s32 ixgbe_start_mac_link_82599(struct ixgbe_hw *hw, 29 bool autoneg_wait_to_complete); 30static s32 ixgbe_setup_mac_link_82599(struct ixgbe_hw *hw, 31 ixgbe_link_speed speed, 32 bool autoneg_wait_to_complete); 33static s32 ixgbe_setup_copper_link_82599(struct ixgbe_hw *hw, 34 ixgbe_link_speed speed, 35 bool autoneg_wait_to_complete); 36static s32 ixgbe_verify_fw_version_82599(struct ixgbe_hw *hw); 37static s32 ixgbe_read_i2c_byte_82599(struct ixgbe_hw *hw, u8 byte_offset, 38 u8 dev_addr, u8 *data); 39static s32 ixgbe_write_i2c_byte_82599(struct ixgbe_hw *hw, u8 byte_offset, 40 u8 dev_addr, u8 data); 41static s32 ixgbe_reset_pipeline_82599(struct ixgbe_hw *hw); 42static bool ixgbe_verify_lesm_fw_enabled_82599(struct ixgbe_hw *hw); 43 44bool ixgbe_mng_enabled(struct ixgbe_hw *hw) 45{ 46 u32 fwsm, manc, factps; 47 48 fwsm = IXGBE_READ_REG(hw, IXGBE_FWSM(hw)); 49 if ((fwsm & IXGBE_FWSM_MODE_MASK) != IXGBE_FWSM_FW_MODE_PT) 50 return false; 51 52 manc = IXGBE_READ_REG(hw, IXGBE_MANC); 53 if (!(manc & IXGBE_MANC_RCV_TCO_EN)) 54 return false; 55 56 factps = IXGBE_READ_REG(hw, IXGBE_FACTPS(hw)); 57 if (factps & IXGBE_FACTPS_MNGCG) 58 return false; 59 60 return true; 61} 62 63static void ixgbe_init_mac_link_ops_82599(struct ixgbe_hw *hw) 64{ 65 struct ixgbe_mac_info *mac = &hw->mac; 66 67 /* enable the laser control functions for SFP+ fiber 68 * and MNG not enabled 69 */ 70 if ((mac->ops.get_media_type(hw) == ixgbe_media_type_fiber) && 71 !ixgbe_mng_enabled(hw)) { 72 mac->ops.disable_tx_laser = 73 &ixgbe_disable_tx_laser_multispeed_fiber; 74 mac->ops.enable_tx_laser = 75 &ixgbe_enable_tx_laser_multispeed_fiber; 76 mac->ops.flap_tx_laser = &ixgbe_flap_tx_laser_multispeed_fiber; 77 } else { 78 mac->ops.disable_tx_laser = NULL; 79 mac->ops.enable_tx_laser = NULL; 80 mac->ops.flap_tx_laser = NULL; 81 } 82 83 if (hw->phy.multispeed_fiber) { 84 /* Set up dual speed SFP+ support */ 85 mac->ops.setup_link = &ixgbe_setup_mac_link_multispeed_fiber; 86 mac->ops.setup_mac_link = ixgbe_setup_mac_link_82599; 87 mac->ops.set_rate_select_speed = 88 ixgbe_set_hard_rate_select_speed; 89 } else { 90 if ((mac->ops.get_media_type(hw) == 91 ixgbe_media_type_backplane) && 92 (hw->phy.smart_speed == ixgbe_smart_speed_auto || 93 hw->phy.smart_speed == ixgbe_smart_speed_on) && 94 !ixgbe_verify_lesm_fw_enabled_82599(hw)) 95 mac->ops.setup_link = &ixgbe_setup_mac_link_smartspeed; 96 else 97 mac->ops.setup_link = &ixgbe_setup_mac_link_82599; 98 } 99} 100 101static s32 ixgbe_setup_sfp_modules_82599(struct ixgbe_hw *hw) 102{ 103 s32 ret_val; 104 u16 list_offset, data_offset, data_value; 105 106 if (hw->phy.sfp_type != ixgbe_sfp_type_unknown) { 107 ixgbe_init_mac_link_ops_82599(hw); 108 109 hw->phy.ops.reset = NULL; 110 111 ret_val = ixgbe_get_sfp_init_sequence_offsets(hw, &list_offset, 112 &data_offset); 113 if (ret_val) 114 return ret_val; 115 116 /* PHY config will finish before releasing the semaphore */ 117 ret_val = hw->mac.ops.acquire_swfw_sync(hw, 118 IXGBE_GSSR_MAC_CSR_SM); 119 if (ret_val) 120 return -EBUSY; 121 122 if (hw->eeprom.ops.read(hw, ++data_offset, &data_value)) 123 goto setup_sfp_err; 124 while (data_value != 0xffff) { 125 IXGBE_WRITE_REG(hw, IXGBE_CORECTL, data_value); 126 IXGBE_WRITE_FLUSH(hw); 127 if (hw->eeprom.ops.read(hw, ++data_offset, &data_value)) 128 goto setup_sfp_err; 129 } 130 131 /* Release the semaphore */ 132 hw->mac.ops.release_swfw_sync(hw, IXGBE_GSSR_MAC_CSR_SM); 133 /* 134 * Delay obtaining semaphore again to allow FW access, 135 * semaphore_delay is in ms usleep_range needs us. 136 */ 137 usleep_range(hw->eeprom.semaphore_delay * 1000, 138 hw->eeprom.semaphore_delay * 2000); 139 140 /* Restart DSP and set SFI mode */ 141 ret_val = hw->mac.ops.prot_autoc_write(hw, 142 hw->mac.orig_autoc | IXGBE_AUTOC_LMS_10G_SERIAL, 143 false); 144 145 if (ret_val) { 146 hw_dbg(hw, " sfp module setup not complete\n"); 147 return -EIO; 148 } 149 } 150 151 return 0; 152 153setup_sfp_err: 154 /* Release the semaphore */ 155 hw->mac.ops.release_swfw_sync(hw, IXGBE_GSSR_MAC_CSR_SM); 156 /* Delay obtaining semaphore again to allow FW access, 157 * semaphore_delay is in ms usleep_range needs us. 158 */ 159 usleep_range(hw->eeprom.semaphore_delay * 1000, 160 hw->eeprom.semaphore_delay * 2000); 161 hw_err(hw, "eeprom read at offset %d failed\n", data_offset); 162 return -EIO; 163} 164 165/** 166 * prot_autoc_read_82599 - Hides MAC differences needed for AUTOC read 167 * @hw: pointer to hardware structure 168 * @locked: Return the if we locked for this read. 169 * @reg_val: Value we read from AUTOC 170 * 171 * For this part (82599) we need to wrap read-modify-writes with a possible 172 * FW/SW lock. It is assumed this lock will be freed with the next 173 * prot_autoc_write_82599(). Note, that locked can only be true in cases 174 * where this function doesn't return an error. 175 **/ 176static s32 prot_autoc_read_82599(struct ixgbe_hw *hw, bool *locked, 177 u32 *reg_val) 178{ 179 s32 ret_val; 180 181 *locked = false; 182 /* If LESM is on then we need to hold the SW/FW semaphore. */ 183 if (ixgbe_verify_lesm_fw_enabled_82599(hw)) { 184 ret_val = hw->mac.ops.acquire_swfw_sync(hw, 185 IXGBE_GSSR_MAC_CSR_SM); 186 if (ret_val) 187 return -EBUSY; 188 189 *locked = true; 190 } 191 192 *reg_val = IXGBE_READ_REG(hw, IXGBE_AUTOC); 193 return 0; 194} 195 196/** 197 * prot_autoc_write_82599 - Hides MAC differences needed for AUTOC write 198 * @hw: pointer to hardware structure 199 * @autoc: value to write to AUTOC 200 * @locked: bool to indicate whether the SW/FW lock was already taken by 201 * previous proc_autoc_read_82599. 202 * 203 * This part (82599) may need to hold a the SW/FW lock around all writes to 204 * AUTOC. Likewise after a write we need to do a pipeline reset. 205 **/ 206static s32 prot_autoc_write_82599(struct ixgbe_hw *hw, u32 autoc, bool locked) 207{ 208 s32 ret_val = 0; 209 210 /* Blocked by MNG FW so bail */ 211 if (ixgbe_check_reset_blocked(hw)) 212 goto out; 213 214 /* We only need to get the lock if: 215 * - We didn't do it already (in the read part of a read-modify-write) 216 * - LESM is enabled. 217 */ 218 if (!locked && ixgbe_verify_lesm_fw_enabled_82599(hw)) { 219 ret_val = hw->mac.ops.acquire_swfw_sync(hw, 220 IXGBE_GSSR_MAC_CSR_SM); 221 if (ret_val) 222 return -EBUSY; 223 224 locked = true; 225 } 226 227 IXGBE_WRITE_REG(hw, IXGBE_AUTOC, autoc); 228 ret_val = ixgbe_reset_pipeline_82599(hw); 229 230out: 231 /* Free the SW/FW semaphore as we either grabbed it here or 232 * already had it when this function was called. 233 */ 234 if (locked) 235 hw->mac.ops.release_swfw_sync(hw, IXGBE_GSSR_MAC_CSR_SM); 236 237 return ret_val; 238} 239 240static s32 ixgbe_get_invariants_82599(struct ixgbe_hw *hw) 241{ 242 struct ixgbe_mac_info *mac = &hw->mac; 243 244 ixgbe_init_mac_link_ops_82599(hw); 245 246 mac->mcft_size = IXGBE_82599_MC_TBL_SIZE; 247 mac->vft_size = IXGBE_82599_VFT_TBL_SIZE; 248 mac->num_rar_entries = IXGBE_82599_RAR_ENTRIES; 249 mac->rx_pb_size = IXGBE_82599_RX_PB_SIZE; 250 mac->max_rx_queues = IXGBE_82599_MAX_RX_QUEUES; 251 mac->max_tx_queues = IXGBE_82599_MAX_TX_QUEUES; 252 mac->max_msix_vectors = ixgbe_get_pcie_msix_count_generic(hw); 253 254 return 0; 255} 256 257/** 258 * ixgbe_init_phy_ops_82599 - PHY/SFP specific init 259 * @hw: pointer to hardware structure 260 * 261 * Initialize any function pointers that were not able to be 262 * set during get_invariants because the PHY/SFP type was 263 * not known. Perform the SFP init if necessary. 264 * 265 **/ 266static s32 ixgbe_init_phy_ops_82599(struct ixgbe_hw *hw) 267{ 268 struct ixgbe_mac_info *mac = &hw->mac; 269 struct ixgbe_phy_info *phy = &hw->phy; 270 s32 ret_val; 271 u32 esdp; 272 273 if (hw->device_id == IXGBE_DEV_ID_82599_QSFP_SF_QP) { 274 /* Store flag indicating I2C bus access control unit. */ 275 hw->phy.qsfp_shared_i2c_bus = true; 276 277 /* Initialize access to QSFP+ I2C bus */ 278 esdp = IXGBE_READ_REG(hw, IXGBE_ESDP); 279 esdp |= IXGBE_ESDP_SDP0_DIR; 280 esdp &= ~IXGBE_ESDP_SDP1_DIR; 281 esdp &= ~IXGBE_ESDP_SDP0; 282 esdp &= ~IXGBE_ESDP_SDP0_NATIVE; 283 esdp &= ~IXGBE_ESDP_SDP1_NATIVE; 284 IXGBE_WRITE_REG(hw, IXGBE_ESDP, esdp); 285 IXGBE_WRITE_FLUSH(hw); 286 287 phy->ops.read_i2c_byte = &ixgbe_read_i2c_byte_82599; 288 phy->ops.write_i2c_byte = &ixgbe_write_i2c_byte_82599; 289 } 290 291 /* Identify the PHY or SFP module */ 292 ret_val = phy->ops.identify(hw); 293 294 /* Setup function pointers based on detected SFP module and speeds */ 295 ixgbe_init_mac_link_ops_82599(hw); 296 297 /* If copper media, overwrite with copper function pointers */ 298 if (mac->ops.get_media_type(hw) == ixgbe_media_type_copper) { 299 mac->ops.setup_link = &ixgbe_setup_copper_link_82599; 300 mac->ops.get_link_capabilities = 301 &ixgbe_get_copper_link_capabilities_generic; 302 } 303 304 /* Set necessary function pointers based on phy type */ 305 switch (hw->phy.type) { 306 case ixgbe_phy_tn: 307 phy->ops.check_link = &ixgbe_check_phy_link_tnx; 308 phy->ops.setup_link = &ixgbe_setup_phy_link_tnx; 309 break; 310 default: 311 break; 312 } 313 314 return ret_val; 315} 316 317/** 318 * ixgbe_get_link_capabilities_82599 - Determines link capabilities 319 * @hw: pointer to hardware structure 320 * @speed: pointer to link speed 321 * @autoneg: true when autoneg or autotry is enabled 322 * 323 * Determines the link capabilities by reading the AUTOC register. 324 **/ 325static s32 ixgbe_get_link_capabilities_82599(struct ixgbe_hw *hw, 326 ixgbe_link_speed *speed, 327 bool *autoneg) 328{ 329 u32 autoc = 0; 330 331 /* Determine 1G link capabilities off of SFP+ type */ 332 if (hw->phy.sfp_type == ixgbe_sfp_type_1g_cu_core0 || 333 hw->phy.sfp_type == ixgbe_sfp_type_1g_cu_core1 || 334 hw->phy.sfp_type == ixgbe_sfp_type_1g_lx_core0 || 335 hw->phy.sfp_type == ixgbe_sfp_type_1g_lx_core1 || 336 hw->phy.sfp_type == ixgbe_sfp_type_1g_sx_core0 || 337 hw->phy.sfp_type == ixgbe_sfp_type_1g_sx_core1) { 338 *speed = IXGBE_LINK_SPEED_1GB_FULL; 339 *autoneg = true; 340 return 0; 341 } 342 343 /* 344 * Determine link capabilities based on the stored value of AUTOC, 345 * which represents EEPROM defaults. If AUTOC value has not been 346 * stored, use the current register value. 347 */ 348 if (hw->mac.orig_link_settings_stored) 349 autoc = hw->mac.orig_autoc; 350 else 351 autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC); 352 353 switch (autoc & IXGBE_AUTOC_LMS_MASK) { 354 case IXGBE_AUTOC_LMS_1G_LINK_NO_AN: 355 *speed = IXGBE_LINK_SPEED_1GB_FULL; 356 *autoneg = false; 357 break; 358 359 case IXGBE_AUTOC_LMS_10G_LINK_NO_AN: 360 *speed = IXGBE_LINK_SPEED_10GB_FULL; 361 *autoneg = false; 362 break; 363 364 case IXGBE_AUTOC_LMS_1G_AN: 365 *speed = IXGBE_LINK_SPEED_1GB_FULL; 366 *autoneg = true; 367 break; 368 369 case IXGBE_AUTOC_LMS_10G_SERIAL: 370 *speed = IXGBE_LINK_SPEED_10GB_FULL; 371 *autoneg = false; 372 break; 373 374 case IXGBE_AUTOC_LMS_KX4_KX_KR: 375 case IXGBE_AUTOC_LMS_KX4_KX_KR_1G_AN: 376 *speed = IXGBE_LINK_SPEED_UNKNOWN; 377 if (autoc & IXGBE_AUTOC_KR_SUPP) 378 *speed |= IXGBE_LINK_SPEED_10GB_FULL; 379 if (autoc & IXGBE_AUTOC_KX4_SUPP) 380 *speed |= IXGBE_LINK_SPEED_10GB_FULL; 381 if (autoc & IXGBE_AUTOC_KX_SUPP) 382 *speed |= IXGBE_LINK_SPEED_1GB_FULL; 383 *autoneg = true; 384 break; 385 386 case IXGBE_AUTOC_LMS_KX4_KX_KR_SGMII: 387 *speed = IXGBE_LINK_SPEED_100_FULL; 388 if (autoc & IXGBE_AUTOC_KR_SUPP) 389 *speed |= IXGBE_LINK_SPEED_10GB_FULL; 390 if (autoc & IXGBE_AUTOC_KX4_SUPP) 391 *speed |= IXGBE_LINK_SPEED_10GB_FULL; 392 if (autoc & IXGBE_AUTOC_KX_SUPP) 393 *speed |= IXGBE_LINK_SPEED_1GB_FULL; 394 *autoneg = true; 395 break; 396 397 case IXGBE_AUTOC_LMS_SGMII_1G_100M: 398 *speed = IXGBE_LINK_SPEED_1GB_FULL | IXGBE_LINK_SPEED_100_FULL; 399 *autoneg = false; 400 break; 401 402 default: 403 return -EIO; 404 } 405 406 if (hw->phy.multispeed_fiber) { 407 *speed |= IXGBE_LINK_SPEED_10GB_FULL | 408 IXGBE_LINK_SPEED_1GB_FULL; 409 410 /* QSFP must not enable auto-negotiation */ 411 if (hw->phy.media_type == ixgbe_media_type_fiber_qsfp) 412 *autoneg = false; 413 else 414 *autoneg = true; 415 } 416 417 return 0; 418} 419 420/** 421 * ixgbe_get_media_type_82599 - Get media type 422 * @hw: pointer to hardware structure 423 * 424 * Returns the media type (fiber, copper, backplane) 425 **/ 426static enum ixgbe_media_type ixgbe_get_media_type_82599(struct ixgbe_hw *hw) 427{ 428 /* Detect if there is a copper PHY attached. */ 429 switch (hw->phy.type) { 430 case ixgbe_phy_cu_unknown: 431 case ixgbe_phy_tn: 432 return ixgbe_media_type_copper; 433 434 default: 435 break; 436 } 437 438 switch (hw->device_id) { 439 case IXGBE_DEV_ID_82599_KX4: 440 case IXGBE_DEV_ID_82599_KX4_MEZZ: 441 case IXGBE_DEV_ID_82599_COMBO_BACKPLANE: 442 case IXGBE_DEV_ID_82599_KR: 443 case IXGBE_DEV_ID_82599_BACKPLANE_FCOE: 444 case IXGBE_DEV_ID_82599_XAUI_LOM: 445 /* Default device ID is mezzanine card KX/KX4 */ 446 return ixgbe_media_type_backplane; 447 448 case IXGBE_DEV_ID_82599_SFP: 449 case IXGBE_DEV_ID_82599_SFP_FCOE: 450 case IXGBE_DEV_ID_82599_SFP_EM: 451 case IXGBE_DEV_ID_82599_SFP_SF2: 452 case IXGBE_DEV_ID_82599_SFP_SF_QP: 453 case IXGBE_DEV_ID_82599EN_SFP: 454 return ixgbe_media_type_fiber; 455 456 case IXGBE_DEV_ID_82599_CX4: 457 return ixgbe_media_type_cx4; 458 459 case IXGBE_DEV_ID_82599_T3_LOM: 460 return ixgbe_media_type_copper; 461 462 case IXGBE_DEV_ID_82599_LS: 463 return ixgbe_media_type_fiber_lco; 464 465 case IXGBE_DEV_ID_82599_QSFP_SF_QP: 466 return ixgbe_media_type_fiber_qsfp; 467 468 default: 469 return ixgbe_media_type_unknown; 470 } 471} 472 473/** 474 * ixgbe_stop_mac_link_on_d3_82599 - Disables link on D3 475 * @hw: pointer to hardware structure 476 * 477 * Disables link, should be called during D3 power down sequence. 478 * 479 **/ 480static void ixgbe_stop_mac_link_on_d3_82599(struct ixgbe_hw *hw) 481{ 482 u32 autoc2_reg; 483 u16 ee_ctrl_2 = 0; 484 485 hw->eeprom.ops.read(hw, IXGBE_EEPROM_CTRL_2, &ee_ctrl_2); 486 487 if (!ixgbe_mng_present(hw) && !hw->wol_enabled && 488 ee_ctrl_2 & IXGBE_EEPROM_CCD_BIT) { 489 autoc2_reg = IXGBE_READ_REG(hw, IXGBE_AUTOC2); 490 autoc2_reg |= IXGBE_AUTOC2_LINK_DISABLE_ON_D3_MASK; 491 IXGBE_WRITE_REG(hw, IXGBE_AUTOC2, autoc2_reg); 492 } 493} 494 495/** 496 * ixgbe_start_mac_link_82599 - Setup MAC link settings 497 * @hw: pointer to hardware structure 498 * @autoneg_wait_to_complete: true when waiting for completion is needed 499 * 500 * Configures link settings based on values in the ixgbe_hw struct. 501 * Restarts the link. Performs autonegotiation if needed. 502 **/ 503static s32 ixgbe_start_mac_link_82599(struct ixgbe_hw *hw, 504 bool autoneg_wait_to_complete) 505{ 506 u32 autoc_reg; 507 u32 links_reg; 508 u32 i; 509 s32 status = 0; 510 bool got_lock = false; 511 512 if (ixgbe_verify_lesm_fw_enabled_82599(hw)) { 513 status = hw->mac.ops.acquire_swfw_sync(hw, 514 IXGBE_GSSR_MAC_CSR_SM); 515 if (status) 516 return status; 517 518 got_lock = true; 519 } 520 521 /* Restart link */ 522 ixgbe_reset_pipeline_82599(hw); 523 524 if (got_lock) 525 hw->mac.ops.release_swfw_sync(hw, IXGBE_GSSR_MAC_CSR_SM); 526 527 /* Only poll for autoneg to complete if specified to do so */ 528 if (autoneg_wait_to_complete) { 529 autoc_reg = IXGBE_READ_REG(hw, IXGBE_AUTOC); 530 if ((autoc_reg & IXGBE_AUTOC_LMS_MASK) == 531 IXGBE_AUTOC_LMS_KX4_KX_KR || 532 (autoc_reg & IXGBE_AUTOC_LMS_MASK) == 533 IXGBE_AUTOC_LMS_KX4_KX_KR_1G_AN || 534 (autoc_reg & IXGBE_AUTOC_LMS_MASK) == 535 IXGBE_AUTOC_LMS_KX4_KX_KR_SGMII) { 536 links_reg = 0; /* Just in case Autoneg time = 0 */ 537 for (i = 0; i < IXGBE_AUTO_NEG_TIME; i++) { 538 links_reg = IXGBE_READ_REG(hw, IXGBE_LINKS); 539 if (links_reg & IXGBE_LINKS_KX_AN_COMP) 540 break; 541 msleep(100); 542 } 543 if (!(links_reg & IXGBE_LINKS_KX_AN_COMP)) { 544 status = -EIO; 545 hw_dbg(hw, "Autoneg did not complete.\n"); 546 } 547 } 548 } 549 550 /* Add delay to filter out noises during initial link setup */ 551 msleep(50); 552 553 return status; 554} 555 556/** 557 * ixgbe_disable_tx_laser_multispeed_fiber - Disable Tx laser 558 * @hw: pointer to hardware structure 559 * 560 * The base drivers may require better control over SFP+ module 561 * PHY states. This includes selectively shutting down the Tx 562 * laser on the PHY, effectively halting physical link. 563 **/ 564static void ixgbe_disable_tx_laser_multispeed_fiber(struct ixgbe_hw *hw) 565{ 566 u32 esdp_reg = IXGBE_READ_REG(hw, IXGBE_ESDP); 567 568 /* Blocked by MNG FW so bail */ 569 if (ixgbe_check_reset_blocked(hw)) 570 return; 571 572 /* Disable tx laser; allow 100us to go dark per spec */ 573 esdp_reg |= IXGBE_ESDP_SDP3; 574 IXGBE_WRITE_REG(hw, IXGBE_ESDP, esdp_reg); 575 IXGBE_WRITE_FLUSH(hw); 576 udelay(100); 577} 578 579/** 580 * ixgbe_enable_tx_laser_multispeed_fiber - Enable Tx laser 581 * @hw: pointer to hardware structure 582 * 583 * The base drivers may require better control over SFP+ module 584 * PHY states. This includes selectively turning on the Tx 585 * laser on the PHY, effectively starting physical link. 586 **/ 587static void ixgbe_enable_tx_laser_multispeed_fiber(struct ixgbe_hw *hw) 588{ 589 u32 esdp_reg = IXGBE_READ_REG(hw, IXGBE_ESDP); 590 591 /* Enable tx laser; allow 100ms to light up */ 592 esdp_reg &= ~IXGBE_ESDP_SDP3; 593 IXGBE_WRITE_REG(hw, IXGBE_ESDP, esdp_reg); 594 IXGBE_WRITE_FLUSH(hw); 595 msleep(100); 596} 597 598/** 599 * ixgbe_flap_tx_laser_multispeed_fiber - Flap Tx laser 600 * @hw: pointer to hardware structure 601 * 602 * When the driver changes the link speeds that it can support, 603 * it sets autotry_restart to true to indicate that we need to 604 * initiate a new autotry session with the link partner. To do 605 * so, we set the speed then disable and re-enable the tx laser, to 606 * alert the link partner that it also needs to restart autotry on its 607 * end. This is consistent with true clause 37 autoneg, which also 608 * involves a loss of signal. 609 **/ 610static void ixgbe_flap_tx_laser_multispeed_fiber(struct ixgbe_hw *hw) 611{ 612 /* Blocked by MNG FW so bail */ 613 if (ixgbe_check_reset_blocked(hw)) 614 return; 615 616 if (hw->mac.autotry_restart) { 617 ixgbe_disable_tx_laser_multispeed_fiber(hw); 618 ixgbe_enable_tx_laser_multispeed_fiber(hw); 619 hw->mac.autotry_restart = false; 620 } 621} 622 623/** 624 * ixgbe_set_hard_rate_select_speed - Set module link speed 625 * @hw: pointer to hardware structure 626 * @speed: link speed to set 627 * 628 * Set module link speed via RS0/RS1 rate select pins. 629 */ 630static void 631ixgbe_set_hard_rate_select_speed(struct ixgbe_hw *hw, ixgbe_link_speed speed) 632{ 633 u32 esdp_reg = IXGBE_READ_REG(hw, IXGBE_ESDP); 634 635 switch (speed) { 636 case IXGBE_LINK_SPEED_10GB_FULL: 637 esdp_reg |= (IXGBE_ESDP_SDP5_DIR | IXGBE_ESDP_SDP5); 638 break; 639 case IXGBE_LINK_SPEED_1GB_FULL: 640 esdp_reg &= ~IXGBE_ESDP_SDP5; 641 esdp_reg |= IXGBE_ESDP_SDP5_DIR; 642 break; 643 default: 644 hw_dbg(hw, "Invalid fixed module speed\n"); 645 return; 646 } 647 648 IXGBE_WRITE_REG(hw, IXGBE_ESDP, esdp_reg); 649 IXGBE_WRITE_FLUSH(hw); 650} 651 652/** 653 * ixgbe_setup_mac_link_smartspeed - Set MAC link speed using SmartSpeed 654 * @hw: pointer to hardware structure 655 * @speed: new link speed 656 * @autoneg_wait_to_complete: true when waiting for completion is needed 657 * 658 * Implements the Intel SmartSpeed algorithm. 659 **/ 660static s32 ixgbe_setup_mac_link_smartspeed(struct ixgbe_hw *hw, 661 ixgbe_link_speed speed, 662 bool autoneg_wait_to_complete) 663{ 664 s32 status = 0; 665 ixgbe_link_speed link_speed = IXGBE_LINK_SPEED_UNKNOWN; 666 s32 i, j; 667 bool link_up = false; 668 u32 autoc_reg = IXGBE_READ_REG(hw, IXGBE_AUTOC); 669 670 /* Set autoneg_advertised value based on input link speed */ 671 hw->phy.autoneg_advertised = 0; 672 673 if (speed & IXGBE_LINK_SPEED_10GB_FULL) 674 hw->phy.autoneg_advertised |= IXGBE_LINK_SPEED_10GB_FULL; 675 676 if (speed & IXGBE_LINK_SPEED_1GB_FULL) 677 hw->phy.autoneg_advertised |= IXGBE_LINK_SPEED_1GB_FULL; 678 679 if (speed & IXGBE_LINK_SPEED_100_FULL) 680 hw->phy.autoneg_advertised |= IXGBE_LINK_SPEED_100_FULL; 681 682 /* 683 * Implement Intel SmartSpeed algorithm. SmartSpeed will reduce the 684 * autoneg advertisement if link is unable to be established at the 685 * highest negotiated rate. This can sometimes happen due to integrity 686 * issues with the physical media connection. 687 */ 688 689 /* First, try to get link with full advertisement */ 690 hw->phy.smart_speed_active = false; 691 for (j = 0; j < IXGBE_SMARTSPEED_MAX_RETRIES; j++) { 692 status = ixgbe_setup_mac_link_82599(hw, speed, 693 autoneg_wait_to_complete); 694 if (status != 0) 695 goto out; 696 697 /* 698 * Wait for the controller to acquire link. Per IEEE 802.3ap, 699 * Section 73.10.2, we may have to wait up to 500ms if KR is 700 * attempted, or 200ms if KX/KX4/BX/BX4 is attempted, per 701 * Table 9 in the AN MAS. 702 */ 703 for (i = 0; i < 5; i++) { 704 mdelay(100); 705 706 /* If we have link, just jump out */ 707 status = hw->mac.ops.check_link(hw, &link_speed, 708 &link_up, false); 709 if (status != 0) 710 goto out; 711 712 if (link_up) 713 goto out; 714 } 715 } 716 717 /* 718 * We didn't get link. If we advertised KR plus one of KX4/KX 719 * (or BX4/BX), then disable KR and try again. 720 */ 721 if (((autoc_reg & IXGBE_AUTOC_KR_SUPP) == 0) || 722 ((autoc_reg & IXGBE_AUTOC_KX4_KX_SUPP_MASK) == 0)) 723 goto out; 724 725 /* Turn SmartSpeed on to disable KR support */ 726 hw->phy.smart_speed_active = true; 727 status = ixgbe_setup_mac_link_82599(hw, speed, 728 autoneg_wait_to_complete); 729 if (status != 0) 730 goto out; 731 732 /* 733 * Wait for the controller to acquire link. 600ms will allow for 734 * the AN link_fail_inhibit_timer as well for multiple cycles of 735 * parallel detect, both 10g and 1g. This allows for the maximum 736 * connect attempts as defined in the AN MAS table 73-7. 737 */ 738 for (i = 0; i < 6; i++) { 739 mdelay(100); 740 741 /* If we have link, just jump out */ 742 status = hw->mac.ops.check_link(hw, &link_speed, 743 &link_up, false); 744 if (status != 0) 745 goto out; 746 747 if (link_up) 748 goto out; 749 } 750 751 /* We didn't get link. Turn SmartSpeed back off. */ 752 hw->phy.smart_speed_active = false; 753 status = ixgbe_setup_mac_link_82599(hw, speed, 754 autoneg_wait_to_complete); 755 756out: 757 if (link_up && (link_speed == IXGBE_LINK_SPEED_1GB_FULL)) 758 hw_dbg(hw, "Smartspeed has downgraded the link speed from the maximum advertised\n"); 759 return status; 760} 761 762/** 763 * ixgbe_setup_mac_link_82599 - Set MAC link speed 764 * @hw: pointer to hardware structure 765 * @speed: new link speed 766 * @autoneg_wait_to_complete: true when waiting for completion is needed 767 * 768 * Set the link speed in the AUTOC register and restarts link. 769 **/ 770static s32 ixgbe_setup_mac_link_82599(struct ixgbe_hw *hw, 771 ixgbe_link_speed speed, 772 bool autoneg_wait_to_complete) 773{ 774 bool autoneg = false; 775 s32 status; 776 u32 pma_pmd_1g, link_mode, links_reg, i; 777 u32 autoc2 = IXGBE_READ_REG(hw, IXGBE_AUTOC2); 778 u32 pma_pmd_10g_serial = autoc2 & IXGBE_AUTOC2_10G_SERIAL_PMA_PMD_MASK; 779 ixgbe_link_speed link_capabilities = IXGBE_LINK_SPEED_UNKNOWN; 780 781 /* holds the value of AUTOC register at this current point in time */ 782 u32 current_autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC); 783 /* holds the cached value of AUTOC register */ 784 u32 orig_autoc = 0; 785 /* temporary variable used for comparison purposes */ 786 u32 autoc = current_autoc; 787 788 /* Check to see if speed passed in is supported. */ 789 status = hw->mac.ops.get_link_capabilities(hw, &link_capabilities, 790 &autoneg); 791 if (status) 792 return status; 793 794 speed &= link_capabilities; 795 796 if (speed == IXGBE_LINK_SPEED_UNKNOWN) 797 return -EINVAL; 798 799 /* Use stored value (EEPROM defaults) of AUTOC to find KR/KX4 support*/ 800 if (hw->mac.orig_link_settings_stored) 801 orig_autoc = hw->mac.orig_autoc; 802 else 803 orig_autoc = autoc; 804 805 link_mode = autoc & IXGBE_AUTOC_LMS_MASK; 806 pma_pmd_1g = autoc & IXGBE_AUTOC_1G_PMA_PMD_MASK; 807 808 if (link_mode == IXGBE_AUTOC_LMS_KX4_KX_KR || 809 link_mode == IXGBE_AUTOC_LMS_KX4_KX_KR_1G_AN || 810 link_mode == IXGBE_AUTOC_LMS_KX4_KX_KR_SGMII) { 811 /* Set KX4/KX/KR support according to speed requested */ 812 autoc &= ~(IXGBE_AUTOC_KX4_KX_SUPP_MASK | IXGBE_AUTOC_KR_SUPP); 813 if (speed & IXGBE_LINK_SPEED_10GB_FULL) { 814 if (orig_autoc & IXGBE_AUTOC_KX4_SUPP) 815 autoc |= IXGBE_AUTOC_KX4_SUPP; 816 if ((orig_autoc & IXGBE_AUTOC_KR_SUPP) && 817 (hw->phy.smart_speed_active == false)) 818 autoc |= IXGBE_AUTOC_KR_SUPP; 819 } 820 if (speed & IXGBE_LINK_SPEED_1GB_FULL) 821 autoc |= IXGBE_AUTOC_KX_SUPP; 822 } else if ((pma_pmd_1g == IXGBE_AUTOC_1G_SFI) && 823 (link_mode == IXGBE_AUTOC_LMS_1G_LINK_NO_AN || 824 link_mode == IXGBE_AUTOC_LMS_1G_AN)) { 825 /* Switch from 1G SFI to 10G SFI if requested */ 826 if ((speed == IXGBE_LINK_SPEED_10GB_FULL) && 827 (pma_pmd_10g_serial == IXGBE_AUTOC2_10G_SFI)) { 828 autoc &= ~IXGBE_AUTOC_LMS_MASK; 829 autoc |= IXGBE_AUTOC_LMS_10G_SERIAL; 830 } 831 } else if ((pma_pmd_10g_serial == IXGBE_AUTOC2_10G_SFI) && 832 (link_mode == IXGBE_AUTOC_LMS_10G_SERIAL)) { 833 /* Switch from 10G SFI to 1G SFI if requested */ 834 if ((speed == IXGBE_LINK_SPEED_1GB_FULL) && 835 (pma_pmd_1g == IXGBE_AUTOC_1G_SFI)) { 836 autoc &= ~IXGBE_AUTOC_LMS_MASK; 837 if (autoneg) 838 autoc |= IXGBE_AUTOC_LMS_1G_AN; 839 else 840 autoc |= IXGBE_AUTOC_LMS_1G_LINK_NO_AN; 841 } 842 } 843 844 if (autoc != current_autoc) { 845 /* Restart link */ 846 status = hw->mac.ops.prot_autoc_write(hw, autoc, false); 847 if (status) 848 return status; 849 850 /* Only poll for autoneg to complete if specified to do so */ 851 if (autoneg_wait_to_complete) { 852 if (link_mode == IXGBE_AUTOC_LMS_KX4_KX_KR || 853 link_mode == IXGBE_AUTOC_LMS_KX4_KX_KR_1G_AN || 854 link_mode == IXGBE_AUTOC_LMS_KX4_KX_KR_SGMII) { 855 links_reg = 0; /*Just in case Autoneg time=0*/ 856 for (i = 0; i < IXGBE_AUTO_NEG_TIME; i++) { 857 links_reg = 858 IXGBE_READ_REG(hw, IXGBE_LINKS); 859 if (links_reg & IXGBE_LINKS_KX_AN_COMP) 860 break; 861 msleep(100); 862 } 863 if (!(links_reg & IXGBE_LINKS_KX_AN_COMP)) { 864 status = -EIO; 865 hw_dbg(hw, "Autoneg did not complete.\n"); 866 } 867 } 868 } 869 870 /* Add delay to filter out noises during initial link setup */ 871 msleep(50); 872 } 873 874 return status; 875} 876 877/** 878 * ixgbe_setup_copper_link_82599 - Set the PHY autoneg advertised field 879 * @hw: pointer to hardware structure 880 * @speed: new link speed 881 * @autoneg_wait_to_complete: true if waiting is needed to complete 882 * 883 * Restarts link on PHY and MAC based on settings passed in. 884 **/ 885static s32 ixgbe_setup_copper_link_82599(struct ixgbe_hw *hw, 886 ixgbe_link_speed speed, 887 bool autoneg_wait_to_complete) 888{ 889 s32 status; 890 891 /* Setup the PHY according to input speed */ 892 status = hw->phy.ops.setup_link_speed(hw, speed, 893 autoneg_wait_to_complete); 894 /* Set up MAC */ 895 ixgbe_start_mac_link_82599(hw, autoneg_wait_to_complete); 896 897 return status; 898} 899 900/** 901 * ixgbe_reset_hw_82599 - Perform hardware reset 902 * @hw: pointer to hardware structure 903 * 904 * Resets the hardware by resetting the transmit and receive units, masks 905 * and clears all interrupts, perform a PHY reset, and perform a link (MAC) 906 * reset. 907 **/ 908static s32 ixgbe_reset_hw_82599(struct ixgbe_hw *hw) 909{ 910 ixgbe_link_speed link_speed; 911 s32 status; 912 u32 ctrl, i, autoc, autoc2; 913 u32 curr_lms; 914 bool link_up = false; 915 916 /* Call adapter stop to disable tx/rx and clear interrupts */ 917 status = hw->mac.ops.stop_adapter(hw); 918 if (status) 919 return status; 920 921 /* flush pending Tx transactions */ 922 ixgbe_clear_tx_pending(hw); 923 924 /* PHY ops must be identified and initialized prior to reset */ 925 926 /* Identify PHY and related function pointers */ 927 status = hw->phy.ops.init(hw); 928 929 if (status == -EOPNOTSUPP) 930 return status; 931 932 /* Setup SFP module if there is one present. */ 933 if (hw->phy.sfp_setup_needed) { 934 status = hw->mac.ops.setup_sfp(hw); 935 hw->phy.sfp_setup_needed = false; 936 } 937 938 if (status == -EOPNOTSUPP) 939 return status; 940 941 /* Reset PHY */ 942 if (hw->phy.reset_disable == false && hw->phy.ops.reset != NULL) 943 hw->phy.ops.reset(hw); 944 945 /* remember AUTOC from before we reset */ 946 curr_lms = IXGBE_READ_REG(hw, IXGBE_AUTOC) & IXGBE_AUTOC_LMS_MASK; 947 948mac_reset_top: 949 /* 950 * Issue global reset to the MAC. Needs to be SW reset if link is up. 951 * If link reset is used when link is up, it might reset the PHY when 952 * mng is using it. If link is down or the flag to force full link 953 * reset is set, then perform link reset. 954 */ 955 ctrl = IXGBE_CTRL_LNK_RST; 956 if (!hw->force_full_reset) { 957 hw->mac.ops.check_link(hw, &link_speed, &link_up, false); 958 if (link_up) 959 ctrl = IXGBE_CTRL_RST; 960 } 961 962 ctrl |= IXGBE_READ_REG(hw, IXGBE_CTRL); 963 IXGBE_WRITE_REG(hw, IXGBE_CTRL, ctrl); 964 IXGBE_WRITE_FLUSH(hw); 965 usleep_range(1000, 1200); 966 967 /* Poll for reset bit to self-clear indicating reset is complete */ 968 for (i = 0; i < 10; i++) { 969 ctrl = IXGBE_READ_REG(hw, IXGBE_CTRL); 970 if (!(ctrl & IXGBE_CTRL_RST_MASK)) 971 break; 972 udelay(1); 973 } 974 975 if (ctrl & IXGBE_CTRL_RST_MASK) { 976 status = -EIO; 977 hw_dbg(hw, "Reset polling failed to complete.\n"); 978 } 979 980 msleep(50); 981 982 /* 983 * Double resets are required for recovery from certain error 984 * conditions. Between resets, it is necessary to stall to allow time 985 * for any pending HW events to complete. 986 */ 987 if (hw->mac.flags & IXGBE_FLAGS_DOUBLE_RESET_REQUIRED) { 988 hw->mac.flags &= ~IXGBE_FLAGS_DOUBLE_RESET_REQUIRED; 989 goto mac_reset_top; 990 } 991 992 /* 993 * Store the original AUTOC/AUTOC2 values if they have not been 994 * stored off yet. Otherwise restore the stored original 995 * values since the reset operation sets back to defaults. 996 */ 997 autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC); 998 autoc2 = IXGBE_READ_REG(hw, IXGBE_AUTOC2); 999 1000 /* Enable link if disabled in NVM */ 1001 if (autoc2 & IXGBE_AUTOC2_LINK_DISABLE_MASK) { 1002 autoc2 &= ~IXGBE_AUTOC2_LINK_DISABLE_MASK; 1003 IXGBE_WRITE_REG(hw, IXGBE_AUTOC2, autoc2); 1004 IXGBE_WRITE_FLUSH(hw); 1005 } 1006 1007 if (hw->mac.orig_link_settings_stored == false) { 1008 hw->mac.orig_autoc = autoc; 1009 hw->mac.orig_autoc2 = autoc2; 1010 hw->mac.orig_link_settings_stored = true; 1011 } else { 1012 1013 /* If MNG FW is running on a multi-speed device that 1014 * doesn't autoneg with out driver support we need to 1015 * leave LMS in the state it was before we MAC reset. 1016 * Likewise if we support WoL we don't want change the 1017 * LMS state either. 1018 */ 1019 if ((hw->phy.multispeed_fiber && ixgbe_mng_enabled(hw)) || 1020 hw->wol_enabled) 1021 hw->mac.orig_autoc = 1022 (hw->mac.orig_autoc & ~IXGBE_AUTOC_LMS_MASK) | 1023 curr_lms; 1024 1025 if (autoc != hw->mac.orig_autoc) { 1026 status = hw->mac.ops.prot_autoc_write(hw, 1027 hw->mac.orig_autoc, 1028 false); 1029 if (status) 1030 return status; 1031 } 1032 1033 if ((autoc2 & IXGBE_AUTOC2_UPPER_MASK) != 1034 (hw->mac.orig_autoc2 & IXGBE_AUTOC2_UPPER_MASK)) { 1035 autoc2 &= ~IXGBE_AUTOC2_UPPER_MASK; 1036 autoc2 |= (hw->mac.orig_autoc2 & 1037 IXGBE_AUTOC2_UPPER_MASK); 1038 IXGBE_WRITE_REG(hw, IXGBE_AUTOC2, autoc2); 1039 } 1040 } 1041 1042 /* Store the permanent mac address */ 1043 hw->mac.ops.get_mac_addr(hw, hw->mac.perm_addr); 1044 1045 /* 1046 * Store MAC address from RAR0, clear receive address registers, and 1047 * clear the multicast table. Also reset num_rar_entries to 128, 1048 * since we modify this value when programming the SAN MAC address. 1049 */ 1050 hw->mac.num_rar_entries = IXGBE_82599_RAR_ENTRIES; 1051 hw->mac.ops.init_rx_addrs(hw); 1052 1053 /* Store the permanent SAN mac address */ 1054 hw->mac.ops.get_san_mac_addr(hw, hw->mac.san_addr); 1055 1056 /* Add the SAN MAC address to the RAR only if it's a valid address */ 1057 if (is_valid_ether_addr(hw->mac.san_addr)) { 1058 /* Save the SAN MAC RAR index */ 1059 hw->mac.san_mac_rar_index = hw->mac.num_rar_entries - 1; 1060 1061 hw->mac.ops.set_rar(hw, hw->mac.san_mac_rar_index, 1062 hw->mac.san_addr, 0, IXGBE_RAH_AV); 1063 1064 /* clear VMDq pool/queue selection for this RAR */ 1065 hw->mac.ops.clear_vmdq(hw, hw->mac.san_mac_rar_index, 1066 IXGBE_CLEAR_VMDQ_ALL); 1067 1068 /* Reserve the last RAR for the SAN MAC address */ 1069 hw->mac.num_rar_entries--; 1070 } 1071 1072 /* Store the alternative WWNN/WWPN prefix */ 1073 hw->mac.ops.get_wwn_prefix(hw, &hw->mac.wwnn_prefix, 1074 &hw->mac.wwpn_prefix); 1075 1076 return status; 1077} 1078 1079/** 1080 * ixgbe_fdir_check_cmd_complete - poll to check whether FDIRCMD is complete 1081 * @hw: pointer to hardware structure 1082 * @fdircmd: current value of FDIRCMD register 1083 */ 1084static s32 ixgbe_fdir_check_cmd_complete(struct ixgbe_hw *hw, u32 *fdircmd) 1085{ 1086 int i; 1087 1088 for (i = 0; i < IXGBE_FDIRCMD_CMD_POLL; i++) { 1089 *fdircmd = IXGBE_READ_REG(hw, IXGBE_FDIRCMD); 1090 if (!(*fdircmd & IXGBE_FDIRCMD_CMD_MASK)) 1091 return 0; 1092 udelay(10); 1093 } 1094 1095 return -EIO; 1096} 1097 1098/** 1099 * ixgbe_reinit_fdir_tables_82599 - Reinitialize Flow Director tables. 1100 * @hw: pointer to hardware structure 1101 **/ 1102s32 ixgbe_reinit_fdir_tables_82599(struct ixgbe_hw *hw) 1103{ 1104 int i; 1105 u32 fdirctrl = IXGBE_READ_REG(hw, IXGBE_FDIRCTRL); 1106 u32 fdircmd; 1107 s32 err; 1108 1109 fdirctrl &= ~IXGBE_FDIRCTRL_INIT_DONE; 1110 1111 /* 1112 * Before starting reinitialization process, 1113 * FDIRCMD.CMD must be zero. 1114 */ 1115 err = ixgbe_fdir_check_cmd_complete(hw, &fdircmd); 1116 if (err) { 1117 hw_dbg(hw, "Flow Director previous command did not complete, aborting table re-initialization.\n"); 1118 return err; 1119 } 1120 1121 IXGBE_WRITE_REG(hw, IXGBE_FDIRFREE, 0); 1122 IXGBE_WRITE_FLUSH(hw); 1123 /* 1124 * 82599 adapters flow director init flow cannot be restarted, 1125 * Workaround 82599 silicon errata by performing the following steps 1126 * before re-writing the FDIRCTRL control register with the same value. 1127 * - write 1 to bit 8 of FDIRCMD register & 1128 * - write 0 to bit 8 of FDIRCMD register 1129 */ 1130 IXGBE_WRITE_REG(hw, IXGBE_FDIRCMD, 1131 (IXGBE_READ_REG(hw, IXGBE_FDIRCMD) | 1132 IXGBE_FDIRCMD_CLEARHT)); 1133 IXGBE_WRITE_FLUSH(hw); 1134 IXGBE_WRITE_REG(hw, IXGBE_FDIRCMD, 1135 (IXGBE_READ_REG(hw, IXGBE_FDIRCMD) & 1136 ~IXGBE_FDIRCMD_CLEARHT)); 1137 IXGBE_WRITE_FLUSH(hw); 1138 /* 1139 * Clear FDIR Hash register to clear any leftover hashes 1140 * waiting to be programmed. 1141 */ 1142 IXGBE_WRITE_REG(hw, IXGBE_FDIRHASH, 0x00); 1143 IXGBE_WRITE_FLUSH(hw); 1144 1145 IXGBE_WRITE_REG(hw, IXGBE_FDIRCTRL, fdirctrl); 1146 IXGBE_WRITE_FLUSH(hw); 1147 1148 /* Poll init-done after we write FDIRCTRL register */ 1149 for (i = 0; i < IXGBE_FDIR_INIT_DONE_POLL; i++) { 1150 if (IXGBE_READ_REG(hw, IXGBE_FDIRCTRL) & 1151 IXGBE_FDIRCTRL_INIT_DONE) 1152 break; 1153 usleep_range(1000, 2000); 1154 } 1155 if (i >= IXGBE_FDIR_INIT_DONE_POLL) { 1156 hw_dbg(hw, "Flow Director Signature poll time exceeded!\n"); 1157 return -EIO; 1158 } 1159 1160 /* Clear FDIR statistics registers (read to clear) */ 1161 IXGBE_READ_REG(hw, IXGBE_FDIRUSTAT); 1162 IXGBE_READ_REG(hw, IXGBE_FDIRFSTAT); 1163 IXGBE_READ_REG(hw, IXGBE_FDIRMATCH); 1164 IXGBE_READ_REG(hw, IXGBE_FDIRMISS); 1165 IXGBE_READ_REG(hw, IXGBE_FDIRLEN); 1166 1167 return 0; 1168} 1169 1170/** 1171 * ixgbe_fdir_enable_82599 - Initialize Flow Director control registers 1172 * @hw: pointer to hardware structure 1173 * @fdirctrl: value to write to flow director control register 1174 **/ 1175static void ixgbe_fdir_enable_82599(struct ixgbe_hw *hw, u32 fdirctrl) 1176{ 1177 int i; 1178 1179 /* Prime the keys for hashing */ 1180 IXGBE_WRITE_REG(hw, IXGBE_FDIRHKEY, IXGBE_ATR_BUCKET_HASH_KEY); 1181 IXGBE_WRITE_REG(hw, IXGBE_FDIRSKEY, IXGBE_ATR_SIGNATURE_HASH_KEY); 1182 1183 /* 1184 * Poll init-done after we write the register. Estimated times: 1185 * 10G: PBALLOC = 11b, timing is 60us 1186 * 1G: PBALLOC = 11b, timing is 600us 1187 * 100M: PBALLOC = 11b, timing is 6ms 1188 * 1189 * Multiple these timings by 4 if under full Rx load 1190 * 1191 * So we'll poll for IXGBE_FDIR_INIT_DONE_POLL times, sleeping for 1192 * 1 msec per poll time. If we're at line rate and drop to 100M, then 1193 * this might not finish in our poll time, but we can live with that 1194 * for now. 1195 */ 1196 IXGBE_WRITE_REG(hw, IXGBE_FDIRCTRL, fdirctrl); 1197 IXGBE_WRITE_FLUSH(hw); 1198 for (i = 0; i < IXGBE_FDIR_INIT_DONE_POLL; i++) { 1199 if (IXGBE_READ_REG(hw, IXGBE_FDIRCTRL) & 1200 IXGBE_FDIRCTRL_INIT_DONE) 1201 break; 1202 usleep_range(1000, 2000); 1203 } 1204 1205 if (i >= IXGBE_FDIR_INIT_DONE_POLL) 1206 hw_dbg(hw, "Flow Director poll time exceeded!\n"); 1207} 1208 1209/** 1210 * ixgbe_init_fdir_signature_82599 - Initialize Flow Director signature filters 1211 * @hw: pointer to hardware structure 1212 * @fdirctrl: value to write to flow director control register, initially 1213 * contains just the value of the Rx packet buffer allocation 1214 **/ 1215s32 ixgbe_init_fdir_signature_82599(struct ixgbe_hw *hw, u32 fdirctrl) 1216{ 1217 /* 1218 * Continue setup of fdirctrl register bits: 1219 * Move the flexible bytes to use the ethertype - shift 6 words 1220 * Set the maximum length per hash bucket to 0xA filters 1221 * Send interrupt when 64 filters are left 1222 */ 1223 fdirctrl |= (0x6 << IXGBE_FDIRCTRL_FLEX_SHIFT) | 1224 (0xA << IXGBE_FDIRCTRL_MAX_LENGTH_SHIFT) | 1225 (4 << IXGBE_FDIRCTRL_FULL_THRESH_SHIFT); 1226 1227 /* write hashes and fdirctrl register, poll for completion */ 1228 ixgbe_fdir_enable_82599(hw, fdirctrl); 1229 1230 return 0; 1231} 1232 1233/** 1234 * ixgbe_init_fdir_perfect_82599 - Initialize Flow Director perfect filters 1235 * @hw: pointer to hardware structure 1236 * @fdirctrl: value to write to flow director control register, initially 1237 * contains just the value of the Rx packet buffer allocation 1238 **/ 1239s32 ixgbe_init_fdir_perfect_82599(struct ixgbe_hw *hw, u32 fdirctrl) 1240{ 1241 /* 1242 * Continue setup of fdirctrl register bits: 1243 * Turn perfect match filtering on 1244 * Initialize the drop queue 1245 * Move the flexible bytes to use the ethertype - shift 6 words 1246 * Set the maximum length per hash bucket to 0xA filters 1247 * Send interrupt when 64 (0x4 * 16) filters are left 1248 */ 1249 fdirctrl |= IXGBE_FDIRCTRL_PERFECT_MATCH | 1250 (IXGBE_FDIR_DROP_QUEUE << IXGBE_FDIRCTRL_DROP_Q_SHIFT) | 1251 (0x6 << IXGBE_FDIRCTRL_FLEX_SHIFT) | 1252 (0xA << IXGBE_FDIRCTRL_MAX_LENGTH_SHIFT) | 1253 (4 << IXGBE_FDIRCTRL_FULL_THRESH_SHIFT); 1254 1255 /* write hashes and fdirctrl register, poll for completion */ 1256 ixgbe_fdir_enable_82599(hw, fdirctrl); 1257 1258 return 0; 1259} 1260 1261/* 1262 * These defines allow us to quickly generate all of the necessary instructions 1263 * in the function below by simply calling out IXGBE_COMPUTE_SIG_HASH_ITERATION 1264 * for values 0 through 15 1265 */ 1266#define IXGBE_ATR_COMMON_HASH_KEY \ 1267 (IXGBE_ATR_BUCKET_HASH_KEY & IXGBE_ATR_SIGNATURE_HASH_KEY) 1268#define IXGBE_COMPUTE_SIG_HASH_ITERATION(_n) \ 1269do { \ 1270 u32 n = (_n); \ 1271 if (IXGBE_ATR_COMMON_HASH_KEY & BIT(n)) \ 1272 common_hash ^= lo_hash_dword >> n; \ 1273 else if (IXGBE_ATR_BUCKET_HASH_KEY & BIT(n)) \ 1274 bucket_hash ^= lo_hash_dword >> n; \ 1275 else if (IXGBE_ATR_SIGNATURE_HASH_KEY & BIT(n)) \ 1276 sig_hash ^= lo_hash_dword << (16 - n); \ 1277 if (IXGBE_ATR_COMMON_HASH_KEY & BIT(n + 16)) \ 1278 common_hash ^= hi_hash_dword >> n; \ 1279 else if (IXGBE_ATR_BUCKET_HASH_KEY & BIT(n + 16)) \ 1280 bucket_hash ^= hi_hash_dword >> n; \ 1281 else if (IXGBE_ATR_SIGNATURE_HASH_KEY & BIT(n + 16)) \ 1282 sig_hash ^= hi_hash_dword << (16 - n); \ 1283} while (0) 1284 1285/** 1286 * ixgbe_atr_compute_sig_hash_82599 - Compute the signature hash 1287 * @input: input bitstream to compute the hash on 1288 * @common: compressed common input dword 1289 * 1290 * This function is almost identical to the function above but contains 1291 * several optimizations such as unwinding all of the loops, letting the 1292 * compiler work out all of the conditional ifs since the keys are static 1293 * defines, and computing two keys at once since the hashed dword stream 1294 * will be the same for both keys. 1295 **/ 1296static u32 ixgbe_atr_compute_sig_hash_82599(union ixgbe_atr_hash_dword input, 1297 union ixgbe_atr_hash_dword common) 1298{ 1299 u32 hi_hash_dword, lo_hash_dword, flow_vm_vlan; 1300 u32 sig_hash = 0, bucket_hash = 0, common_hash = 0; 1301 1302 /* record the flow_vm_vlan bits as they are a key part to the hash */ 1303 flow_vm_vlan = ntohl(input.dword); 1304 1305 /* generate common hash dword */ 1306 hi_hash_dword = ntohl(common.dword); 1307 1308 /* low dword is word swapped version of common */ 1309 lo_hash_dword = (hi_hash_dword >> 16) | (hi_hash_dword << 16); 1310 1311 /* apply flow ID/VM pool/VLAN ID bits to hash words */ 1312 hi_hash_dword ^= flow_vm_vlan ^ (flow_vm_vlan >> 16); 1313 1314 /* Process bits 0 and 16 */ 1315 IXGBE_COMPUTE_SIG_HASH_ITERATION(0); 1316 1317 /* 1318 * apply flow ID/VM pool/VLAN ID bits to lo hash dword, we had to 1319 * delay this because bit 0 of the stream should not be processed 1320 * so we do not add the vlan until after bit 0 was processed 1321 */ 1322 lo_hash_dword ^= flow_vm_vlan ^ (flow_vm_vlan << 16); 1323 1324 /* Process remaining 30 bit of the key */ 1325 IXGBE_COMPUTE_SIG_HASH_ITERATION(1); 1326 IXGBE_COMPUTE_SIG_HASH_ITERATION(2); 1327 IXGBE_COMPUTE_SIG_HASH_ITERATION(3); 1328 IXGBE_COMPUTE_SIG_HASH_ITERATION(4); 1329 IXGBE_COMPUTE_SIG_HASH_ITERATION(5); 1330 IXGBE_COMPUTE_SIG_HASH_ITERATION(6); 1331 IXGBE_COMPUTE_SIG_HASH_ITERATION(7); 1332 IXGBE_COMPUTE_SIG_HASH_ITERATION(8); 1333 IXGBE_COMPUTE_SIG_HASH_ITERATION(9); 1334 IXGBE_COMPUTE_SIG_HASH_ITERATION(10); 1335 IXGBE_COMPUTE_SIG_HASH_ITERATION(11); 1336 IXGBE_COMPUTE_SIG_HASH_ITERATION(12); 1337 IXGBE_COMPUTE_SIG_HASH_ITERATION(13); 1338 IXGBE_COMPUTE_SIG_HASH_ITERATION(14); 1339 IXGBE_COMPUTE_SIG_HASH_ITERATION(15); 1340 1341 /* combine common_hash result with signature and bucket hashes */ 1342 bucket_hash ^= common_hash; 1343 bucket_hash &= IXGBE_ATR_HASH_MASK; 1344 1345 sig_hash ^= common_hash << 16; 1346 sig_hash &= IXGBE_ATR_HASH_MASK << 16; 1347 1348 /* return completed signature hash */ 1349 return sig_hash ^ bucket_hash; 1350} 1351 1352/** 1353 * ixgbe_atr_add_signature_filter_82599 - Adds a signature hash filter 1354 * @hw: pointer to hardware structure 1355 * @input: unique input dword 1356 * @common: compressed common input dword 1357 * @queue: queue index to direct traffic to 1358 * 1359 * Note that the tunnel bit in input must not be set when the hardware 1360 * tunneling support does not exist. 1361 **/ 1362s32 ixgbe_fdir_add_signature_filter_82599(struct ixgbe_hw *hw, 1363 union ixgbe_atr_hash_dword input, 1364 union ixgbe_atr_hash_dword common, 1365 u8 queue) 1366{ 1367 u64 fdirhashcmd; 1368 u8 flow_type; 1369 bool tunnel; 1370 u32 fdircmd; 1371 1372 /* 1373 * Get the flow_type in order to program FDIRCMD properly 1374 * lowest 2 bits are FDIRCMD.L4TYPE, third lowest bit is FDIRCMD.IPV6 1375 */ 1376 tunnel = !!(input.formatted.flow_type & IXGBE_ATR_L4TYPE_TUNNEL_MASK); 1377 flow_type = input.formatted.flow_type & 1378 (IXGBE_ATR_L4TYPE_TUNNEL_MASK - 1); 1379 switch (flow_type) { 1380 case IXGBE_ATR_FLOW_TYPE_TCPV4: 1381 case IXGBE_ATR_FLOW_TYPE_UDPV4: 1382 case IXGBE_ATR_FLOW_TYPE_SCTPV4: 1383 case IXGBE_ATR_FLOW_TYPE_TCPV6: 1384 case IXGBE_ATR_FLOW_TYPE_UDPV6: 1385 case IXGBE_ATR_FLOW_TYPE_SCTPV6: 1386 break; 1387 default: 1388 hw_dbg(hw, " Error on flow type input\n"); 1389 return -EIO; 1390 } 1391 1392 /* configure FDIRCMD register */ 1393 fdircmd = IXGBE_FDIRCMD_CMD_ADD_FLOW | IXGBE_FDIRCMD_FILTER_UPDATE | 1394 IXGBE_FDIRCMD_LAST | IXGBE_FDIRCMD_QUEUE_EN; 1395 fdircmd |= (u32)flow_type << IXGBE_FDIRCMD_FLOW_TYPE_SHIFT; 1396 fdircmd |= (u32)queue << IXGBE_FDIRCMD_RX_QUEUE_SHIFT; 1397 if (tunnel) 1398 fdircmd |= IXGBE_FDIRCMD_TUNNEL_FILTER; 1399 1400 /* 1401 * The lower 32-bits of fdirhashcmd is for FDIRHASH, the upper 32-bits 1402 * is for FDIRCMD. Then do a 64-bit register write from FDIRHASH. 1403 */ 1404 fdirhashcmd = (u64)fdircmd << 32; 1405 fdirhashcmd |= ixgbe_atr_compute_sig_hash_82599(input, common); 1406 IXGBE_WRITE_REG64(hw, IXGBE_FDIRHASH, fdirhashcmd); 1407 1408 hw_dbg(hw, "Tx Queue=%x hash=%x\n", queue, (u32)fdirhashcmd); 1409 1410 return 0; 1411} 1412 1413#define IXGBE_COMPUTE_BKT_HASH_ITERATION(_n) \ 1414do { \ 1415 u32 n = (_n); \ 1416 if (IXGBE_ATR_BUCKET_HASH_KEY & BIT(n)) \ 1417 bucket_hash ^= lo_hash_dword >> n; \ 1418 if (IXGBE_ATR_BUCKET_HASH_KEY & BIT(n + 16)) \ 1419 bucket_hash ^= hi_hash_dword >> n; \ 1420} while (0) 1421 1422/** 1423 * ixgbe_atr_compute_perfect_hash_82599 - Compute the perfect filter hash 1424 * @input: input bitstream to compute the hash on 1425 * @input_mask: mask for the input bitstream 1426 * 1427 * This function serves two main purposes. First it applies the input_mask 1428 * to the atr_input resulting in a cleaned up atr_input data stream. 1429 * Secondly it computes the hash and stores it in the bkt_hash field at 1430 * the end of the input byte stream. This way it will be available for 1431 * future use without needing to recompute the hash. 1432 **/ 1433void ixgbe_atr_compute_perfect_hash_82599(union ixgbe_atr_input *input, 1434 union ixgbe_atr_input *input_mask) 1435{ 1436 1437 u32 hi_hash_dword, lo_hash_dword, flow_vm_vlan; 1438 u32 bucket_hash = 0; 1439 __be32 hi_dword = 0; 1440 int i; 1441 1442 /* Apply masks to input data */ 1443 for (i = 0; i <= 10; i++) 1444 input->dword_stream[i] &= input_mask->dword_stream[i]; 1445 1446 /* record the flow_vm_vlan bits as they are a key part to the hash */ 1447 flow_vm_vlan = ntohl(input->dword_stream[0]); 1448 1449 /* generate common hash dword */ 1450 for (i = 1; i <= 10; i++) 1451 hi_dword ^= input->dword_stream[i]; 1452 hi_hash_dword = ntohl(hi_dword); 1453 1454 /* low dword is word swapped version of common */ 1455 lo_hash_dword = (hi_hash_dword >> 16) | (hi_hash_dword << 16); 1456 1457 /* apply flow ID/VM pool/VLAN ID bits to hash words */ 1458 hi_hash_dword ^= flow_vm_vlan ^ (flow_vm_vlan >> 16); 1459 1460 /* Process bits 0 and 16 */ 1461 IXGBE_COMPUTE_BKT_HASH_ITERATION(0); 1462 1463 /* 1464 * apply flow ID/VM pool/VLAN ID bits to lo hash dword, we had to 1465 * delay this because bit 0 of the stream should not be processed 1466 * so we do not add the vlan until after bit 0 was processed 1467 */ 1468 lo_hash_dword ^= flow_vm_vlan ^ (flow_vm_vlan << 16); 1469 1470 /* Process remaining 30 bit of the key */ 1471 for (i = 1; i <= 15; i++) 1472 IXGBE_COMPUTE_BKT_HASH_ITERATION(i); 1473 1474 /* 1475 * Limit hash to 13 bits since max bucket count is 8K. 1476 * Store result at the end of the input stream. 1477 */ 1478 input->formatted.bkt_hash = (__force __be16)(bucket_hash & 0x1FFF); 1479} 1480 1481/** 1482 * ixgbe_get_fdirtcpm_82599 - generate a tcp port from atr_input_masks 1483 * @input_mask: mask to be bit swapped 1484 * 1485 * The source and destination port masks for flow director are bit swapped 1486 * in that bit 15 effects bit 0, 14 effects 1, 13, 2 etc. In order to 1487 * generate a correctly swapped value we need to bit swap the mask and that 1488 * is what is accomplished by this function. 1489 **/ 1490static u32 ixgbe_get_fdirtcpm_82599(union ixgbe_atr_input *input_mask) 1491{ 1492 u32 mask = ntohs(input_mask->formatted.dst_port); 1493 1494 mask <<= IXGBE_FDIRTCPM_DPORTM_SHIFT; 1495 mask |= ntohs(input_mask->formatted.src_port); 1496 mask = ((mask & 0x55555555) << 1) | ((mask & 0xAAAAAAAA) >> 1); 1497 mask = ((mask & 0x33333333) << 2) | ((mask & 0xCCCCCCCC) >> 2); 1498 mask = ((mask & 0x0F0F0F0F) << 4) | ((mask & 0xF0F0F0F0) >> 4); 1499 return ((mask & 0x00FF00FF) << 8) | ((mask & 0xFF00FF00) >> 8); 1500} 1501 1502/* 1503 * These two macros are meant to address the fact that we have registers 1504 * that are either all or in part big-endian. As a result on big-endian 1505 * systems we will end up byte swapping the value to little-endian before 1506 * it is byte swapped again and written to the hardware in the original 1507 * big-endian format. 1508 */ 1509#define IXGBE_STORE_AS_BE32(_value) \ 1510 (((u32)(_value) >> 24) | (((u32)(_value) & 0x00FF0000) >> 8) | \ 1511 (((u32)(_value) & 0x0000FF00) << 8) | ((u32)(_value) << 24)) 1512 1513#define IXGBE_WRITE_REG_BE32(a, reg, value) \ 1514 IXGBE_WRITE_REG((a), (reg), IXGBE_STORE_AS_BE32(ntohl(value))) 1515 1516#define IXGBE_STORE_AS_BE16(_value) \ 1517 ntohs(((u16)(_value) >> 8) | ((u16)(_value) << 8)) 1518 1519s32 ixgbe_fdir_set_input_mask_82599(struct ixgbe_hw *hw, 1520 union ixgbe_atr_input *input_mask) 1521{ 1522 /* mask IPv6 since it is currently not supported */ 1523 u32 fdirm = IXGBE_FDIRM_DIPv6; 1524 u32 fdirtcpm; 1525 1526 /* 1527 * Program the relevant mask registers. If src/dst_port or src/dst_addr 1528 * are zero, then assume a full mask for that field. Also assume that 1529 * a VLAN of 0 is unspecified, so mask that out as well. L4type 1530 * cannot be masked out in this implementation. 1531 * 1532 * This also assumes IPv4 only. IPv6 masking isn't supported at this 1533 * point in time. 1534 */ 1535 1536 /* verify bucket hash is cleared on hash generation */ 1537 if (input_mask->formatted.bkt_hash) 1538 hw_dbg(hw, " bucket hash should always be 0 in mask\n"); 1539 1540 /* Program FDIRM and verify partial masks */ 1541 switch (input_mask->formatted.vm_pool & 0x7F) { 1542 case 0x0: 1543 fdirm |= IXGBE_FDIRM_POOL; 1544 case 0x7F: 1545 break; 1546 default: 1547 hw_dbg(hw, " Error on vm pool mask\n"); 1548 return -EIO; 1549 } 1550 1551 switch (input_mask->formatted.flow_type & IXGBE_ATR_L4TYPE_MASK) { 1552 case 0x0: 1553 fdirm |= IXGBE_FDIRM_L4P; 1554 if (input_mask->formatted.dst_port || 1555 input_mask->formatted.src_port) { 1556 hw_dbg(hw, " Error on src/dst port mask\n"); 1557 return -EIO; 1558 } 1559 case IXGBE_ATR_L4TYPE_MASK: 1560 break; 1561 default: 1562 hw_dbg(hw, " Error on flow type mask\n"); 1563 return -EIO; 1564 } 1565 1566 switch (ntohs(input_mask->formatted.vlan_id) & 0xEFFF) { 1567 case 0x0000: 1568 /* mask VLAN ID */ 1569 fdirm |= IXGBE_FDIRM_VLANID; 1570 fallthrough; 1571 case 0x0FFF: 1572 /* mask VLAN priority */ 1573 fdirm |= IXGBE_FDIRM_VLANP; 1574 break; 1575 case 0xE000: 1576 /* mask VLAN ID only */ 1577 fdirm |= IXGBE_FDIRM_VLANID; 1578 fallthrough; 1579 case 0xEFFF: 1580 /* no VLAN fields masked */ 1581 break; 1582 default: 1583 hw_dbg(hw, " Error on VLAN mask\n"); 1584 return -EIO; 1585 } 1586 1587 switch ((__force u16)input_mask->formatted.flex_bytes & 0xFFFF) { 1588 case 0x0000: 1589 /* Mask Flex Bytes */ 1590 fdirm |= IXGBE_FDIRM_FLEX; 1591 fallthrough; 1592 case 0xFFFF: 1593 break; 1594 default: 1595 hw_dbg(hw, " Error on flexible byte mask\n"); 1596 return -EIO; 1597 } 1598 1599 /* Now mask VM pool and destination IPv6 - bits 5 and 2 */ 1600 IXGBE_WRITE_REG(hw, IXGBE_FDIRM, fdirm); 1601 1602 /* store the TCP/UDP port masks, bit reversed from port layout */ 1603 fdirtcpm = ixgbe_get_fdirtcpm_82599(input_mask); 1604 1605 /* write both the same so that UDP and TCP use the same mask */ 1606 IXGBE_WRITE_REG(hw, IXGBE_FDIRTCPM, ~fdirtcpm); 1607 IXGBE_WRITE_REG(hw, IXGBE_FDIRUDPM, ~fdirtcpm); 1608 1609 /* also use it for SCTP */ 1610 switch (hw->mac.type) { 1611 case ixgbe_mac_X550: 1612 case ixgbe_mac_X550EM_x: 1613 case ixgbe_mac_x550em_a: 1614 IXGBE_WRITE_REG(hw, IXGBE_FDIRSCTPM, ~fdirtcpm); 1615 break; 1616 default: 1617 break; 1618 } 1619 1620 /* store source and destination IP masks (big-enian) */ 1621 IXGBE_WRITE_REG_BE32(hw, IXGBE_FDIRSIP4M, 1622 ~input_mask->formatted.src_ip[0]); 1623 IXGBE_WRITE_REG_BE32(hw, IXGBE_FDIRDIP4M, 1624 ~input_mask->formatted.dst_ip[0]); 1625 1626 return 0; 1627} 1628 1629s32 ixgbe_fdir_write_perfect_filter_82599(struct ixgbe_hw *hw, 1630 union ixgbe_atr_input *input, 1631 u16 soft_id, u8 queue) 1632{ 1633 u32 fdirport, fdirvlan, fdirhash, fdircmd; 1634 s32 err; 1635 1636 /* currently IPv6 is not supported, must be programmed with 0 */ 1637 IXGBE_WRITE_REG_BE32(hw, IXGBE_FDIRSIPv6(0), 1638 input->formatted.src_ip[0]); 1639 IXGBE_WRITE_REG_BE32(hw, IXGBE_FDIRSIPv6(1), 1640 input->formatted.src_ip[1]); 1641 IXGBE_WRITE_REG_BE32(hw, IXGBE_FDIRSIPv6(2), 1642 input->formatted.src_ip[2]); 1643 1644 /* record the source address (big-endian) */ 1645 IXGBE_WRITE_REG_BE32(hw, IXGBE_FDIRIPSA, input->formatted.src_ip[0]); 1646 1647 /* record the first 32 bits of the destination address (big-endian) */ 1648 IXGBE_WRITE_REG_BE32(hw, IXGBE_FDIRIPDA, input->formatted.dst_ip[0]); 1649 1650 /* record source and destination port (little-endian)*/ 1651 fdirport = ntohs(input->formatted.dst_port); 1652 fdirport <<= IXGBE_FDIRPORT_DESTINATION_SHIFT; 1653 fdirport |= ntohs(input->formatted.src_port); 1654 IXGBE_WRITE_REG(hw, IXGBE_FDIRPORT, fdirport); 1655 1656 /* record vlan (little-endian) and flex_bytes(big-endian) */ 1657 fdirvlan = IXGBE_STORE_AS_BE16((__force u16)input->formatted.flex_bytes); 1658 fdirvlan <<= IXGBE_FDIRVLAN_FLEX_SHIFT; 1659 fdirvlan |= ntohs(input->formatted.vlan_id); 1660 IXGBE_WRITE_REG(hw, IXGBE_FDIRVLAN, fdirvlan); 1661 1662 /* configure FDIRHASH register */ 1663 fdirhash = (__force u32)input->formatted.bkt_hash; 1664 fdirhash |= soft_id << IXGBE_FDIRHASH_SIG_SW_INDEX_SHIFT; 1665 IXGBE_WRITE_REG(hw, IXGBE_FDIRHASH, fdirhash); 1666 1667 /* 1668 * flush all previous writes to make certain registers are 1669 * programmed prior to issuing the command 1670 */ 1671 IXGBE_WRITE_FLUSH(hw); 1672 1673 /* configure FDIRCMD register */ 1674 fdircmd = IXGBE_FDIRCMD_CMD_ADD_FLOW | IXGBE_FDIRCMD_FILTER_UPDATE | 1675 IXGBE_FDIRCMD_LAST | IXGBE_FDIRCMD_QUEUE_EN; 1676 if (queue == IXGBE_FDIR_DROP_QUEUE) 1677 fdircmd |= IXGBE_FDIRCMD_DROP; 1678 fdircmd |= input->formatted.flow_type << IXGBE_FDIRCMD_FLOW_TYPE_SHIFT; 1679 fdircmd |= (u32)queue << IXGBE_FDIRCMD_RX_QUEUE_SHIFT; 1680 fdircmd |= (u32)input->formatted.vm_pool << IXGBE_FDIRCMD_VT_POOL_SHIFT; 1681 1682 IXGBE_WRITE_REG(hw, IXGBE_FDIRCMD, fdircmd); 1683 err = ixgbe_fdir_check_cmd_complete(hw, &fdircmd); 1684 if (err) { 1685 hw_dbg(hw, "Flow Director command did not complete!\n"); 1686 return err; 1687 } 1688 1689 return 0; 1690} 1691 1692s32 ixgbe_fdir_erase_perfect_filter_82599(struct ixgbe_hw *hw, 1693 union ixgbe_atr_input *input, 1694 u16 soft_id) 1695{ 1696 u32 fdirhash; 1697 u32 fdircmd; 1698 s32 err; 1699 1700 /* configure FDIRHASH register */ 1701 fdirhash = (__force u32)input->formatted.bkt_hash; 1702 fdirhash |= soft_id << IXGBE_FDIRHASH_SIG_SW_INDEX_SHIFT; 1703 IXGBE_WRITE_REG(hw, IXGBE_FDIRHASH, fdirhash); 1704 1705 /* flush hash to HW */ 1706 IXGBE_WRITE_FLUSH(hw); 1707 1708 /* Query if filter is present */ 1709 IXGBE_WRITE_REG(hw, IXGBE_FDIRCMD, IXGBE_FDIRCMD_CMD_QUERY_REM_FILT); 1710 1711 err = ixgbe_fdir_check_cmd_complete(hw, &fdircmd); 1712 if (err) { 1713 hw_dbg(hw, "Flow Director command did not complete!\n"); 1714 return err; 1715 } 1716 1717 /* if filter exists in hardware then remove it */ 1718 if (fdircmd & IXGBE_FDIRCMD_FILTER_VALID) { 1719 IXGBE_WRITE_REG(hw, IXGBE_FDIRHASH, fdirhash); 1720 IXGBE_WRITE_FLUSH(hw); 1721 IXGBE_WRITE_REG(hw, IXGBE_FDIRCMD, 1722 IXGBE_FDIRCMD_CMD_REMOVE_FLOW); 1723 } 1724 1725 return 0; 1726} 1727 1728/** 1729 * ixgbe_read_analog_reg8_82599 - Reads 8 bit Omer analog register 1730 * @hw: pointer to hardware structure 1731 * @reg: analog register to read 1732 * @val: read value 1733 * 1734 * Performs read operation to Omer analog register specified. 1735 **/ 1736static s32 ixgbe_read_analog_reg8_82599(struct ixgbe_hw *hw, u32 reg, u8 *val) 1737{ 1738 u32 core_ctl; 1739 1740 IXGBE_WRITE_REG(hw, IXGBE_CORECTL, IXGBE_CORECTL_WRITE_CMD | 1741 (reg << 8)); 1742 IXGBE_WRITE_FLUSH(hw); 1743 udelay(10); 1744 core_ctl = IXGBE_READ_REG(hw, IXGBE_CORECTL); 1745 *val = (u8)core_ctl; 1746 1747 return 0; 1748} 1749 1750/** 1751 * ixgbe_write_analog_reg8_82599 - Writes 8 bit Omer analog register 1752 * @hw: pointer to hardware structure 1753 * @reg: atlas register to write 1754 * @val: value to write 1755 * 1756 * Performs write operation to Omer analog register specified. 1757 **/ 1758static s32 ixgbe_write_analog_reg8_82599(struct ixgbe_hw *hw, u32 reg, u8 val) 1759{ 1760 u32 core_ctl; 1761 1762 core_ctl = (reg << 8) | val; 1763 IXGBE_WRITE_REG(hw, IXGBE_CORECTL, core_ctl); 1764 IXGBE_WRITE_FLUSH(hw); 1765 udelay(10); 1766 1767 return 0; 1768} 1769 1770/** 1771 * ixgbe_start_hw_82599 - Prepare hardware for Tx/Rx 1772 * @hw: pointer to hardware structure 1773 * 1774 * Starts the hardware using the generic start_hw function 1775 * and the generation start_hw function. 1776 * Then performs revision-specific operations, if any. 1777 **/ 1778static s32 ixgbe_start_hw_82599(struct ixgbe_hw *hw) 1779{ 1780 s32 ret_val = 0; 1781 1782 ret_val = ixgbe_start_hw_generic(hw); 1783 if (ret_val) 1784 return ret_val; 1785 1786 ret_val = ixgbe_start_hw_gen2(hw); 1787 if (ret_val) 1788 return ret_val; 1789 1790 /* We need to run link autotry after the driver loads */ 1791 hw->mac.autotry_restart = true; 1792 1793 return ixgbe_verify_fw_version_82599(hw); 1794} 1795 1796/** 1797 * ixgbe_identify_phy_82599 - Get physical layer module 1798 * @hw: pointer to hardware structure 1799 * 1800 * Determines the physical layer module found on the current adapter. 1801 * If PHY already detected, maintains current PHY type in hw struct, 1802 * otherwise executes the PHY detection routine. 1803 **/ 1804static s32 ixgbe_identify_phy_82599(struct ixgbe_hw *hw) 1805{ 1806 s32 status; 1807 1808 /* Detect PHY if not unknown - returns success if already detected. */ 1809 status = ixgbe_identify_phy_generic(hw); 1810 if (status) { 1811 /* 82599 10GBASE-T requires an external PHY */ 1812 if (hw->mac.ops.get_media_type(hw) == ixgbe_media_type_copper) 1813 return status; 1814 status = ixgbe_identify_module_generic(hw); 1815 } 1816 1817 /* Set PHY type none if no PHY detected */ 1818 if (hw->phy.type == ixgbe_phy_unknown) { 1819 hw->phy.type = ixgbe_phy_none; 1820 status = 0; 1821 } 1822 1823 /* Return error if SFP module has been detected but is not supported */ 1824 if (hw->phy.type == ixgbe_phy_sfp_unsupported) 1825 return -EOPNOTSUPP; 1826 1827 return status; 1828} 1829 1830/** 1831 * ixgbe_enable_rx_dma_82599 - Enable the Rx DMA unit on 82599 1832 * @hw: pointer to hardware structure 1833 * @regval: register value to write to RXCTRL 1834 * 1835 * Enables the Rx DMA unit for 82599 1836 **/ 1837static s32 ixgbe_enable_rx_dma_82599(struct ixgbe_hw *hw, u32 regval) 1838{ 1839 /* 1840 * Workaround for 82599 silicon errata when enabling the Rx datapath. 1841 * If traffic is incoming before we enable the Rx unit, it could hang 1842 * the Rx DMA unit. Therefore, make sure the security engine is 1843 * completely disabled prior to enabling the Rx unit. 1844 */ 1845 hw->mac.ops.disable_rx_buff(hw); 1846 1847 if (regval & IXGBE_RXCTRL_RXEN) 1848 hw->mac.ops.enable_rx(hw); 1849 else 1850 hw->mac.ops.disable_rx(hw); 1851 1852 hw->mac.ops.enable_rx_buff(hw); 1853 1854 return 0; 1855} 1856 1857/** 1858 * ixgbe_verify_fw_version_82599 - verify fw version for 82599 1859 * @hw: pointer to hardware structure 1860 * 1861 * Verifies that installed the firmware version is 0.6 or higher 1862 * for SFI devices. All 82599 SFI devices should have version 0.6 or higher. 1863 * 1864 * Return: -EACCES if the FW is not present or if the FW version is 1865 * not supported. 1866 **/ 1867static s32 ixgbe_verify_fw_version_82599(struct ixgbe_hw *hw) 1868{ 1869 u16 fw_offset, fw_ptp_cfg_offset; 1870 s32 status = -EACCES; 1871 u16 offset; 1872 u16 fw_version = 0; 1873 1874 /* firmware check is only necessary for SFI devices */ 1875 if (hw->phy.media_type != ixgbe_media_type_fiber) 1876 return 0; 1877 1878 /* get the offset to the Firmware Module block */ 1879 offset = IXGBE_FW_PTR; 1880 if (hw->eeprom.ops.read(hw, offset, &fw_offset)) 1881 goto fw_version_err; 1882 1883 if (fw_offset == 0 || fw_offset == 0xFFFF) 1884 return -EACCES; 1885 1886 /* get the offset to the Pass Through Patch Configuration block */ 1887 offset = fw_offset + IXGBE_FW_PASSTHROUGH_PATCH_CONFIG_PTR; 1888 if (hw->eeprom.ops.read(hw, offset, &fw_ptp_cfg_offset)) 1889 goto fw_version_err; 1890 1891 if (fw_ptp_cfg_offset == 0 || fw_ptp_cfg_offset == 0xFFFF) 1892 return -EACCES; 1893 1894 /* get the firmware version */ 1895 offset = fw_ptp_cfg_offset + IXGBE_FW_PATCH_VERSION_4; 1896 if (hw->eeprom.ops.read(hw, offset, &fw_version)) 1897 goto fw_version_err; 1898 1899 if (fw_version > 0x5) 1900 status = 0; 1901 1902 return status; 1903 1904fw_version_err: 1905 hw_err(hw, "eeprom read at offset %d failed\n", offset); 1906 return -EACCES; 1907} 1908 1909/** 1910 * ixgbe_verify_lesm_fw_enabled_82599 - Checks LESM FW module state. 1911 * @hw: pointer to hardware structure 1912 * 1913 * Returns true if the LESM FW module is present and enabled. Otherwise 1914 * returns false. Smart Speed must be disabled if LESM FW module is enabled. 1915 **/ 1916static bool ixgbe_verify_lesm_fw_enabled_82599(struct ixgbe_hw *hw) 1917{ 1918 u16 fw_offset, fw_lesm_param_offset, fw_lesm_state; 1919 s32 status; 1920 1921 /* get the offset to the Firmware Module block */ 1922 status = hw->eeprom.ops.read(hw, IXGBE_FW_PTR, &fw_offset); 1923 1924 if (status || fw_offset == 0 || fw_offset == 0xFFFF) 1925 return false; 1926 1927 /* get the offset to the LESM Parameters block */ 1928 status = hw->eeprom.ops.read(hw, (fw_offset + 1929 IXGBE_FW_LESM_PARAMETERS_PTR), 1930 &fw_lesm_param_offset); 1931 1932 if (status || 1933 fw_lesm_param_offset == 0 || fw_lesm_param_offset == 0xFFFF) 1934 return false; 1935 1936 /* get the lesm state word */ 1937 status = hw->eeprom.ops.read(hw, (fw_lesm_param_offset + 1938 IXGBE_FW_LESM_STATE_1), 1939 &fw_lesm_state); 1940 1941 if (!status && (fw_lesm_state & IXGBE_FW_LESM_STATE_ENABLED)) 1942 return true; 1943 1944 return false; 1945} 1946 1947/** 1948 * ixgbe_read_eeprom_buffer_82599 - Read EEPROM word(s) using 1949 * fastest available method 1950 * 1951 * @hw: pointer to hardware structure 1952 * @offset: offset of word in EEPROM to read 1953 * @words: number of words 1954 * @data: word(s) read from the EEPROM 1955 * 1956 * Retrieves 16 bit word(s) read from EEPROM 1957 **/ 1958static s32 ixgbe_read_eeprom_buffer_82599(struct ixgbe_hw *hw, u16 offset, 1959 u16 words, u16 *data) 1960{ 1961 struct ixgbe_eeprom_info *eeprom = &hw->eeprom; 1962 1963 /* If EEPROM is detected and can be addressed using 14 bits, 1964 * use EERD otherwise use bit bang 1965 */ 1966 if (eeprom->type == ixgbe_eeprom_spi && 1967 offset + (words - 1) <= IXGBE_EERD_MAX_ADDR) 1968 return ixgbe_read_eerd_buffer_generic(hw, offset, words, data); 1969 1970 return ixgbe_read_eeprom_buffer_bit_bang_generic(hw, offset, words, 1971 data); 1972} 1973 1974/** 1975 * ixgbe_read_eeprom_82599 - Read EEPROM word using 1976 * fastest available method 1977 * 1978 * @hw: pointer to hardware structure 1979 * @offset: offset of word in the EEPROM to read 1980 * @data: word read from the EEPROM 1981 * 1982 * Reads a 16 bit word from the EEPROM 1983 **/ 1984static s32 ixgbe_read_eeprom_82599(struct ixgbe_hw *hw, 1985 u16 offset, u16 *data) 1986{ 1987 struct ixgbe_eeprom_info *eeprom = &hw->eeprom; 1988 1989 /* 1990 * If EEPROM is detected and can be addressed using 14 bits, 1991 * use EERD otherwise use bit bang 1992 */ 1993 if (eeprom->type == ixgbe_eeprom_spi && offset <= IXGBE_EERD_MAX_ADDR) 1994 return ixgbe_read_eerd_generic(hw, offset, data); 1995 1996 return ixgbe_read_eeprom_bit_bang_generic(hw, offset, data); 1997} 1998 1999/** 2000 * ixgbe_reset_pipeline_82599 - perform pipeline reset 2001 * 2002 * @hw: pointer to hardware structure 2003 * 2004 * Reset pipeline by asserting Restart_AN together with LMS change to ensure 2005 * full pipeline reset. Note - We must hold the SW/FW semaphore before writing 2006 * to AUTOC, so this function assumes the semaphore is held. 2007 **/ 2008static s32 ixgbe_reset_pipeline_82599(struct ixgbe_hw *hw) 2009{ 2010 s32 ret_val; 2011 u32 anlp1_reg = 0; 2012 u32 i, autoc_reg, autoc2_reg; 2013 2014 /* Enable link if disabled in NVM */ 2015 autoc2_reg = IXGBE_READ_REG(hw, IXGBE_AUTOC2); 2016 if (autoc2_reg & IXGBE_AUTOC2_LINK_DISABLE_MASK) { 2017 autoc2_reg &= ~IXGBE_AUTOC2_LINK_DISABLE_MASK; 2018 IXGBE_WRITE_REG(hw, IXGBE_AUTOC2, autoc2_reg); 2019 IXGBE_WRITE_FLUSH(hw); 2020 } 2021 2022 autoc_reg = IXGBE_READ_REG(hw, IXGBE_AUTOC); 2023 autoc_reg |= IXGBE_AUTOC_AN_RESTART; 2024 2025 /* Write AUTOC register with toggled LMS[2] bit and Restart_AN */ 2026 IXGBE_WRITE_REG(hw, IXGBE_AUTOC, 2027 autoc_reg ^ (0x4 << IXGBE_AUTOC_LMS_SHIFT)); 2028 2029 /* Wait for AN to leave state 0 */ 2030 for (i = 0; i < 10; i++) { 2031 usleep_range(4000, 8000); 2032 anlp1_reg = IXGBE_READ_REG(hw, IXGBE_ANLP1); 2033 if (anlp1_reg & IXGBE_ANLP1_AN_STATE_MASK) 2034 break; 2035 } 2036 2037 if (!(anlp1_reg & IXGBE_ANLP1_AN_STATE_MASK)) { 2038 hw_dbg(hw, "auto negotiation not completed\n"); 2039 ret_val = -EIO; 2040 goto reset_pipeline_out; 2041 } 2042 2043 ret_val = 0; 2044 2045reset_pipeline_out: 2046 /* Write AUTOC register with original LMS field and Restart_AN */ 2047 IXGBE_WRITE_REG(hw, IXGBE_AUTOC, autoc_reg); 2048 IXGBE_WRITE_FLUSH(hw); 2049 2050 return ret_val; 2051} 2052 2053/** 2054 * ixgbe_read_i2c_byte_82599 - Reads 8 bit word over I2C 2055 * @hw: pointer to hardware structure 2056 * @byte_offset: byte offset to read 2057 * @dev_addr: address to read from 2058 * @data: value read 2059 * 2060 * Performs byte read operation to SFP module's EEPROM over I2C interface at 2061 * a specified device address. 2062 **/ 2063static s32 ixgbe_read_i2c_byte_82599(struct ixgbe_hw *hw, u8 byte_offset, 2064 u8 dev_addr, u8 *data) 2065{ 2066 u32 esdp; 2067 s32 status; 2068 s32 timeout = 200; 2069 2070 if (hw->phy.qsfp_shared_i2c_bus == true) { 2071 /* Acquire I2C bus ownership. */ 2072 esdp = IXGBE_READ_REG(hw, IXGBE_ESDP); 2073 esdp |= IXGBE_ESDP_SDP0; 2074 IXGBE_WRITE_REG(hw, IXGBE_ESDP, esdp); 2075 IXGBE_WRITE_FLUSH(hw); 2076 2077 while (timeout) { 2078 esdp = IXGBE_READ_REG(hw, IXGBE_ESDP); 2079 if (esdp & IXGBE_ESDP_SDP1) 2080 break; 2081 2082 usleep_range(5000, 10000); 2083 timeout--; 2084 } 2085 2086 if (!timeout) { 2087 hw_dbg(hw, "Driver can't access resource, acquiring I2C bus timeout.\n"); 2088 status = -EIO; 2089 goto release_i2c_access; 2090 } 2091 } 2092 2093 status = ixgbe_read_i2c_byte_generic(hw, byte_offset, dev_addr, data); 2094 2095release_i2c_access: 2096 if (hw->phy.qsfp_shared_i2c_bus == true) { 2097 /* Release I2C bus ownership. */ 2098 esdp = IXGBE_READ_REG(hw, IXGBE_ESDP); 2099 esdp &= ~IXGBE_ESDP_SDP0; 2100 IXGBE_WRITE_REG(hw, IXGBE_ESDP, esdp); 2101 IXGBE_WRITE_FLUSH(hw); 2102 } 2103 2104 return status; 2105} 2106 2107/** 2108 * ixgbe_write_i2c_byte_82599 - Writes 8 bit word over I2C 2109 * @hw: pointer to hardware structure 2110 * @byte_offset: byte offset to write 2111 * @dev_addr: address to write to 2112 * @data: value to write 2113 * 2114 * Performs byte write operation to SFP module's EEPROM over I2C interface at 2115 * a specified device address. 2116 **/ 2117static s32 ixgbe_write_i2c_byte_82599(struct ixgbe_hw *hw, u8 byte_offset, 2118 u8 dev_addr, u8 data) 2119{ 2120 u32 esdp; 2121 s32 status; 2122 s32 timeout = 200; 2123 2124 if (hw->phy.qsfp_shared_i2c_bus == true) { 2125 /* Acquire I2C bus ownership. */ 2126 esdp = IXGBE_READ_REG(hw, IXGBE_ESDP); 2127 esdp |= IXGBE_ESDP_SDP0; 2128 IXGBE_WRITE_REG(hw, IXGBE_ESDP, esdp); 2129 IXGBE_WRITE_FLUSH(hw); 2130 2131 while (timeout) { 2132 esdp = IXGBE_READ_REG(hw, IXGBE_ESDP); 2133 if (esdp & IXGBE_ESDP_SDP1) 2134 break; 2135 2136 usleep_range(5000, 10000); 2137 timeout--; 2138 } 2139 2140 if (!timeout) { 2141 hw_dbg(hw, "Driver can't access resource, acquiring I2C bus timeout.\n"); 2142 status = -EIO; 2143 goto release_i2c_access; 2144 } 2145 } 2146 2147 status = ixgbe_write_i2c_byte_generic(hw, byte_offset, dev_addr, data); 2148 2149release_i2c_access: 2150 if (hw->phy.qsfp_shared_i2c_bus == true) { 2151 /* Release I2C bus ownership. */ 2152 esdp = IXGBE_READ_REG(hw, IXGBE_ESDP); 2153 esdp &= ~IXGBE_ESDP_SDP0; 2154 IXGBE_WRITE_REG(hw, IXGBE_ESDP, esdp); 2155 IXGBE_WRITE_FLUSH(hw); 2156 } 2157 2158 return status; 2159} 2160 2161static const struct ixgbe_mac_operations mac_ops_82599 = { 2162 .init_hw = &ixgbe_init_hw_generic, 2163 .reset_hw = &ixgbe_reset_hw_82599, 2164 .start_hw = &ixgbe_start_hw_82599, 2165 .clear_hw_cntrs = &ixgbe_clear_hw_cntrs_generic, 2166 .get_media_type = &ixgbe_get_media_type_82599, 2167 .enable_rx_dma = &ixgbe_enable_rx_dma_82599, 2168 .disable_rx_buff = &ixgbe_disable_rx_buff_generic, 2169 .enable_rx_buff = &ixgbe_enable_rx_buff_generic, 2170 .get_mac_addr = &ixgbe_get_mac_addr_generic, 2171 .get_san_mac_addr = &ixgbe_get_san_mac_addr_generic, 2172 .get_device_caps = &ixgbe_get_device_caps_generic, 2173 .get_wwn_prefix = &ixgbe_get_wwn_prefix_generic, 2174 .stop_adapter = &ixgbe_stop_adapter_generic, 2175 .get_bus_info = &ixgbe_get_bus_info_generic, 2176 .set_lan_id = &ixgbe_set_lan_id_multi_port_pcie, 2177 .read_analog_reg8 = &ixgbe_read_analog_reg8_82599, 2178 .write_analog_reg8 = &ixgbe_write_analog_reg8_82599, 2179 .stop_link_on_d3 = &ixgbe_stop_mac_link_on_d3_82599, 2180 .setup_link = &ixgbe_setup_mac_link_82599, 2181 .set_rxpba = &ixgbe_set_rxpba_generic, 2182 .check_link = &ixgbe_check_mac_link_generic, 2183 .get_link_capabilities = &ixgbe_get_link_capabilities_82599, 2184 .led_on = &ixgbe_led_on_generic, 2185 .led_off = &ixgbe_led_off_generic, 2186 .init_led_link_act = ixgbe_init_led_link_act_generic, 2187 .blink_led_start = &ixgbe_blink_led_start_generic, 2188 .blink_led_stop = &ixgbe_blink_led_stop_generic, 2189 .set_rar = &ixgbe_set_rar_generic, 2190 .clear_rar = &ixgbe_clear_rar_generic, 2191 .set_vmdq = &ixgbe_set_vmdq_generic, 2192 .set_vmdq_san_mac = &ixgbe_set_vmdq_san_mac_generic, 2193 .clear_vmdq = &ixgbe_clear_vmdq_generic, 2194 .init_rx_addrs = &ixgbe_init_rx_addrs_generic, 2195 .update_mc_addr_list = &ixgbe_update_mc_addr_list_generic, 2196 .enable_mc = &ixgbe_enable_mc_generic, 2197 .disable_mc = &ixgbe_disable_mc_generic, 2198 .clear_vfta = &ixgbe_clear_vfta_generic, 2199 .set_vfta = &ixgbe_set_vfta_generic, 2200 .fc_enable = &ixgbe_fc_enable_generic, 2201 .setup_fc = ixgbe_setup_fc_generic, 2202 .fc_autoneg = ixgbe_fc_autoneg, 2203 .set_fw_drv_ver = &ixgbe_set_fw_drv_ver_generic, 2204 .init_uta_tables = &ixgbe_init_uta_tables_generic, 2205 .setup_sfp = &ixgbe_setup_sfp_modules_82599, 2206 .set_mac_anti_spoofing = &ixgbe_set_mac_anti_spoofing, 2207 .set_vlan_anti_spoofing = &ixgbe_set_vlan_anti_spoofing, 2208 .acquire_swfw_sync = &ixgbe_acquire_swfw_sync, 2209 .release_swfw_sync = &ixgbe_release_swfw_sync, 2210 .init_swfw_sync = NULL, 2211 .get_thermal_sensor_data = &ixgbe_get_thermal_sensor_data_generic, 2212 .init_thermal_sensor_thresh = &ixgbe_init_thermal_sensor_thresh_generic, 2213 .prot_autoc_read = &prot_autoc_read_82599, 2214 .prot_autoc_write = &prot_autoc_write_82599, 2215 .enable_rx = &ixgbe_enable_rx_generic, 2216 .disable_rx = &ixgbe_disable_rx_generic, 2217}; 2218 2219static const struct ixgbe_eeprom_operations eeprom_ops_82599 = { 2220 .init_params = &ixgbe_init_eeprom_params_generic, 2221 .read = &ixgbe_read_eeprom_82599, 2222 .read_buffer = &ixgbe_read_eeprom_buffer_82599, 2223 .write = &ixgbe_write_eeprom_generic, 2224 .write_buffer = &ixgbe_write_eeprom_buffer_bit_bang_generic, 2225 .calc_checksum = &ixgbe_calc_eeprom_checksum_generic, 2226 .validate_checksum = &ixgbe_validate_eeprom_checksum_generic, 2227 .update_checksum = &ixgbe_update_eeprom_checksum_generic, 2228}; 2229 2230static const struct ixgbe_phy_operations phy_ops_82599 = { 2231 .identify = &ixgbe_identify_phy_82599, 2232 .identify_sfp = &ixgbe_identify_module_generic, 2233 .init = &ixgbe_init_phy_ops_82599, 2234 .reset = &ixgbe_reset_phy_generic, 2235 .read_reg = &ixgbe_read_phy_reg_generic, 2236 .write_reg = &ixgbe_write_phy_reg_generic, 2237 .setup_link = &ixgbe_setup_phy_link_generic, 2238 .setup_link_speed = &ixgbe_setup_phy_link_speed_generic, 2239 .read_i2c_byte = &ixgbe_read_i2c_byte_generic, 2240 .write_i2c_byte = &ixgbe_write_i2c_byte_generic, 2241 .read_i2c_sff8472 = &ixgbe_read_i2c_sff8472_generic, 2242 .read_i2c_eeprom = &ixgbe_read_i2c_eeprom_generic, 2243 .write_i2c_eeprom = &ixgbe_write_i2c_eeprom_generic, 2244 .check_overtemp = &ixgbe_tn_check_overtemp, 2245}; 2246 2247const struct ixgbe_info ixgbe_82599_info = { 2248 .mac = ixgbe_mac_82599EB, 2249 .get_invariants = &ixgbe_get_invariants_82599, 2250 .mac_ops = &mac_ops_82599, 2251 .eeprom_ops = &eeprom_ops_82599, 2252 .phy_ops = &phy_ops_82599, 2253 .mbx_ops = &mbx_ops_generic, 2254 .mvals = ixgbe_mvals_8259X, 2255}; 2256