Deleted Added
full compact
ixgbe_82598.c (302408) ixgbe_82598.c (320897)
1/******************************************************************************
2
1/******************************************************************************
2
3 Copyright (c) 2001-2015, Intel Corporation
3 Copyright (c) 2001-2017, Intel Corporation
4 All rights reserved.
4 All rights reserved.
5
6 Redistribution and use in source and binary forms, with or without
5
6 Redistribution and use in source and binary forms, with or without
7 modification, are permitted provided that the following conditions are met:
7 modification, are permitted provided that the following conditions are met:
8
9 1. Redistributions of source code must retain the above copyright notice,
8
9 1. Redistributions of source code must retain the above copyright notice,
10 this list of conditions and the following disclaimer.
10 this list of conditions and the following disclaimer.
11
12 2. Redistributions in binary form must reproduce the above copyright
13 notice, this list of conditions and the following disclaimer in the
11
12 2. Redistributions in binary form must reproduce the above copyright
13 notice, this list of conditions and the following disclaimer in the
14 documentation and/or other materials provided with the distribution.
14 documentation and/or other materials provided with the distribution.
15
16 3. Neither the name of the Intel Corporation nor the names of its
17 contributors may be used to endorse or promote products derived from
15
16 3. Neither the name of the Intel Corporation nor the names of its
17 contributors may be used to endorse or promote products derived from
18 this software without specific prior written permission.
18 this software without specific prior written permission.
19
19
20 THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
20 THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
21 AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22 IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23 ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
24 LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25 CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26 SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27 INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28 CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
21 AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22 IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23 ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
24 LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25 CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26 SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27 INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28 CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29 ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
30 POSSIBILITY OF SUCH DAMAGE.
31
32******************************************************************************/
29 ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
30 POSSIBILITY OF SUCH DAMAGE.
31
32******************************************************************************/
33/*$FreeBSD: stable/11/sys/dev/ixgbe/ixgbe_82598.c 292674 2015-12-23 22:45:17Z sbruno $*/
33/*$FreeBSD: stable/11/sys/dev/ixgbe/ixgbe_82598.c 320897 2017-07-11 21:25:07Z erj $*/
34
35#include "ixgbe_type.h"
36#include "ixgbe_82598.h"
37#include "ixgbe_api.h"
38#include "ixgbe_common.h"
39#include "ixgbe_phy.h"
40
41#define IXGBE_82598_MAX_TX_QUEUES 32
42#define IXGBE_82598_MAX_RX_QUEUES 64
43#define IXGBE_82598_RAR_ENTRIES 16
44#define IXGBE_82598_MC_TBL_SIZE 128
45#define IXGBE_82598_VFT_TBL_SIZE 128
46#define IXGBE_82598_RX_PB_SIZE 512
47
48static s32 ixgbe_get_link_capabilities_82598(struct ixgbe_hw *hw,
49 ixgbe_link_speed *speed,
50 bool *autoneg);
51static enum ixgbe_media_type ixgbe_get_media_type_82598(struct ixgbe_hw *hw);
52static s32 ixgbe_start_mac_link_82598(struct ixgbe_hw *hw,
53 bool autoneg_wait_to_complete);
54static s32 ixgbe_check_mac_link_82598(struct ixgbe_hw *hw,
55 ixgbe_link_speed *speed, bool *link_up,
56 bool link_up_wait_to_complete);
57static s32 ixgbe_setup_mac_link_82598(struct ixgbe_hw *hw,
58 ixgbe_link_speed speed,
59 bool autoneg_wait_to_complete);
60static s32 ixgbe_setup_copper_link_82598(struct ixgbe_hw *hw,
61 ixgbe_link_speed speed,
62 bool autoneg_wait_to_complete);
63static s32 ixgbe_reset_hw_82598(struct ixgbe_hw *hw);
64static s32 ixgbe_clear_vmdq_82598(struct ixgbe_hw *hw, u32 rar, u32 vmdq);
65static s32 ixgbe_clear_vfta_82598(struct ixgbe_hw *hw);
66static void ixgbe_set_rxpba_82598(struct ixgbe_hw *hw, int num_pb,
67 u32 headroom, int strategy);
68static s32 ixgbe_read_i2c_sff8472_82598(struct ixgbe_hw *hw, u8 byte_offset,
69 u8 *sff8472_data);
70/**
71 * ixgbe_set_pcie_completion_timeout - set pci-e completion timeout
72 * @hw: pointer to the HW structure
73 *
74 * The defaults for 82598 should be in the range of 50us to 50ms,
75 * however the hardware default for these parts is 500us to 1ms which is less
76 * than the 10ms recommended by the pci-e spec. To address this we need to
77 * increase the value to either 10ms to 250ms for capability version 1 config,
78 * or 16ms to 55ms for version 2.
79 **/
80void ixgbe_set_pcie_completion_timeout(struct ixgbe_hw *hw)
81{
82 u32 gcr = IXGBE_READ_REG(hw, IXGBE_GCR);
83 u16 pcie_devctl2;
84
85 /* only take action if timeout value is defaulted to 0 */
86 if (gcr & IXGBE_GCR_CMPL_TMOUT_MASK)
87 goto out;
88
89 /*
90 * if capababilities version is type 1 we can write the
91 * timeout of 10ms to 250ms through the GCR register
92 */
93 if (!(gcr & IXGBE_GCR_CAP_VER2)) {
94 gcr |= IXGBE_GCR_CMPL_TMOUT_10ms;
95 goto out;
96 }
97
98 /*
99 * for version 2 capabilities we need to write the config space
100 * directly in order to set the completion timeout value for
101 * 16ms to 55ms
102 */
103 pcie_devctl2 = IXGBE_READ_PCIE_WORD(hw, IXGBE_PCI_DEVICE_CONTROL2);
104 pcie_devctl2 |= IXGBE_PCI_DEVICE_CONTROL2_16ms;
105 IXGBE_WRITE_PCIE_WORD(hw, IXGBE_PCI_DEVICE_CONTROL2, pcie_devctl2);
106out:
107 /* disable completion timeout resend */
108 gcr &= ~IXGBE_GCR_CMPL_TMOUT_RESEND;
109 IXGBE_WRITE_REG(hw, IXGBE_GCR, gcr);
110}
111
112/**
113 * ixgbe_init_ops_82598 - Inits func ptrs and MAC type
114 * @hw: pointer to hardware structure
115 *
116 * Initialize the function pointers and assign the MAC type for 82598.
117 * Does not touch the hardware.
118 **/
119s32 ixgbe_init_ops_82598(struct ixgbe_hw *hw)
120{
121 struct ixgbe_mac_info *mac = &hw->mac;
122 struct ixgbe_phy_info *phy = &hw->phy;
123 s32 ret_val;
124
125 DEBUGFUNC("ixgbe_init_ops_82598");
126
127 ret_val = ixgbe_init_phy_ops_generic(hw);
128 ret_val = ixgbe_init_ops_generic(hw);
129
130 /* PHY */
131 phy->ops.init = ixgbe_init_phy_ops_82598;
132
133 /* MAC */
134 mac->ops.start_hw = ixgbe_start_hw_82598;
135 mac->ops.enable_relaxed_ordering = ixgbe_enable_relaxed_ordering_82598;
136 mac->ops.reset_hw = ixgbe_reset_hw_82598;
137 mac->ops.get_media_type = ixgbe_get_media_type_82598;
138 mac->ops.get_supported_physical_layer =
139 ixgbe_get_supported_physical_layer_82598;
140 mac->ops.read_analog_reg8 = ixgbe_read_analog_reg8_82598;
141 mac->ops.write_analog_reg8 = ixgbe_write_analog_reg8_82598;
142 mac->ops.set_lan_id = ixgbe_set_lan_id_multi_port_pcie_82598;
143 mac->ops.enable_rx_dma = ixgbe_enable_rx_dma_82598;
144
145 /* RAR, Multicast, VLAN */
146 mac->ops.set_vmdq = ixgbe_set_vmdq_82598;
147 mac->ops.clear_vmdq = ixgbe_clear_vmdq_82598;
148 mac->ops.set_vfta = ixgbe_set_vfta_82598;
149 mac->ops.set_vlvf = NULL;
150 mac->ops.clear_vfta = ixgbe_clear_vfta_82598;
151
152 /* Flow Control */
153 mac->ops.fc_enable = ixgbe_fc_enable_82598;
154
155 mac->mcft_size = IXGBE_82598_MC_TBL_SIZE;
156 mac->vft_size = IXGBE_82598_VFT_TBL_SIZE;
157 mac->num_rar_entries = IXGBE_82598_RAR_ENTRIES;
158 mac->rx_pb_size = IXGBE_82598_RX_PB_SIZE;
159 mac->max_rx_queues = IXGBE_82598_MAX_RX_QUEUES;
160 mac->max_tx_queues = IXGBE_82598_MAX_TX_QUEUES;
161 mac->max_msix_vectors = ixgbe_get_pcie_msix_count_generic(hw);
162
163 /* SFP+ Module */
164 phy->ops.read_i2c_eeprom = ixgbe_read_i2c_eeprom_82598;
165 phy->ops.read_i2c_sff8472 = ixgbe_read_i2c_sff8472_82598;
166
167 /* Link */
168 mac->ops.check_link = ixgbe_check_mac_link_82598;
169 mac->ops.setup_link = ixgbe_setup_mac_link_82598;
170 mac->ops.flap_tx_laser = NULL;
171 mac->ops.get_link_capabilities = ixgbe_get_link_capabilities_82598;
172 mac->ops.setup_rxpba = ixgbe_set_rxpba_82598;
173
174 /* Manageability interface */
175 mac->ops.set_fw_drv_ver = NULL;
176
177 mac->ops.get_rtrup2tc = NULL;
178
179 return ret_val;
180}
181
182/**
183 * ixgbe_init_phy_ops_82598 - PHY/SFP specific init
184 * @hw: pointer to hardware structure
185 *
186 * Initialize any function pointers that were not able to be
187 * set during init_shared_code because the PHY/SFP type was
188 * not known. Perform the SFP init if necessary.
189 *
190 **/
191s32 ixgbe_init_phy_ops_82598(struct ixgbe_hw *hw)
192{
193 struct ixgbe_mac_info *mac = &hw->mac;
194 struct ixgbe_phy_info *phy = &hw->phy;
195 s32 ret_val = IXGBE_SUCCESS;
196 u16 list_offset, data_offset;
197
198 DEBUGFUNC("ixgbe_init_phy_ops_82598");
199
200 /* Identify the PHY */
201 phy->ops.identify(hw);
202
203 /* Overwrite the link function pointers if copper PHY */
204 if (mac->ops.get_media_type(hw) == ixgbe_media_type_copper) {
205 mac->ops.setup_link = ixgbe_setup_copper_link_82598;
206 mac->ops.get_link_capabilities =
207 ixgbe_get_copper_link_capabilities_generic;
208 }
209
210 switch (hw->phy.type) {
211 case ixgbe_phy_tn:
212 phy->ops.setup_link = ixgbe_setup_phy_link_tnx;
213 phy->ops.check_link = ixgbe_check_phy_link_tnx;
214 phy->ops.get_firmware_version =
215 ixgbe_get_phy_firmware_version_tnx;
216 break;
217 case ixgbe_phy_nl:
218 phy->ops.reset = ixgbe_reset_phy_nl;
219
220 /* Call SFP+ identify routine to get the SFP+ module type */
221 ret_val = phy->ops.identify_sfp(hw);
222 if (ret_val != IXGBE_SUCCESS)
223 goto out;
224 else if (hw->phy.sfp_type == ixgbe_sfp_type_unknown) {
225 ret_val = IXGBE_ERR_SFP_NOT_SUPPORTED;
226 goto out;
227 }
228
229 /* Check to see if SFP+ module is supported */
230 ret_val = ixgbe_get_sfp_init_sequence_offsets(hw,
231 &list_offset,
232 &data_offset);
233 if (ret_val != IXGBE_SUCCESS) {
234 ret_val = IXGBE_ERR_SFP_NOT_SUPPORTED;
235 goto out;
236 }
237 break;
238 default:
239 break;
240 }
241
242out:
243 return ret_val;
244}
245
246/**
247 * ixgbe_start_hw_82598 - Prepare hardware for Tx/Rx
248 * @hw: pointer to hardware structure
249 *
250 * Starts the hardware using the generic start_hw function.
251 * Disables relaxed ordering Then set pcie completion timeout
252 *
253 **/
254s32 ixgbe_start_hw_82598(struct ixgbe_hw *hw)
255{
256 u32 regval;
257 u32 i;
258 s32 ret_val = IXGBE_SUCCESS;
259
260 DEBUGFUNC("ixgbe_start_hw_82598");
261
262 ret_val = ixgbe_start_hw_generic(hw);
263 if (ret_val)
264 return ret_val;
265
266 /* Disable relaxed ordering */
267 for (i = 0; ((i < hw->mac.max_tx_queues) &&
268 (i < IXGBE_DCA_MAX_QUEUES_82598)); i++) {
269 regval = IXGBE_READ_REG(hw, IXGBE_DCA_TXCTRL(i));
270 regval &= ~IXGBE_DCA_TXCTRL_DESC_WRO_EN;
271 IXGBE_WRITE_REG(hw, IXGBE_DCA_TXCTRL(i), regval);
272 }
273
274 for (i = 0; ((i < hw->mac.max_rx_queues) &&
275 (i < IXGBE_DCA_MAX_QUEUES_82598)); i++) {
276 regval = IXGBE_READ_REG(hw, IXGBE_DCA_RXCTRL(i));
277 regval &= ~(IXGBE_DCA_RXCTRL_DATA_WRO_EN |
278 IXGBE_DCA_RXCTRL_HEAD_WRO_EN);
279 IXGBE_WRITE_REG(hw, IXGBE_DCA_RXCTRL(i), regval);
280 }
281
282 /* set the completion timeout for interface */
283 ixgbe_set_pcie_completion_timeout(hw);
284
285 return ret_val;
286}
287
288/**
289 * ixgbe_get_link_capabilities_82598 - Determines link capabilities
290 * @hw: pointer to hardware structure
291 * @speed: pointer to link speed
292 * @autoneg: boolean auto-negotiation value
293 *
294 * Determines the link capabilities by reading the AUTOC register.
295 **/
296static s32 ixgbe_get_link_capabilities_82598(struct ixgbe_hw *hw,
297 ixgbe_link_speed *speed,
298 bool *autoneg)
299{
300 s32 status = IXGBE_SUCCESS;
301 u32 autoc = 0;
302
303 DEBUGFUNC("ixgbe_get_link_capabilities_82598");
304
305 /*
306 * Determine link capabilities based on the stored value of AUTOC,
307 * which represents EEPROM defaults. If AUTOC value has not been
308 * stored, use the current register value.
309 */
310 if (hw->mac.orig_link_settings_stored)
311 autoc = hw->mac.orig_autoc;
312 else
313 autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
314
315 switch (autoc & IXGBE_AUTOC_LMS_MASK) {
316 case IXGBE_AUTOC_LMS_1G_LINK_NO_AN:
317 *speed = IXGBE_LINK_SPEED_1GB_FULL;
318 *autoneg = FALSE;
319 break;
320
321 case IXGBE_AUTOC_LMS_10G_LINK_NO_AN:
322 *speed = IXGBE_LINK_SPEED_10GB_FULL;
323 *autoneg = FALSE;
324 break;
325
326 case IXGBE_AUTOC_LMS_1G_AN:
327 *speed = IXGBE_LINK_SPEED_1GB_FULL;
328 *autoneg = TRUE;
329 break;
330
331 case IXGBE_AUTOC_LMS_KX4_AN:
332 case IXGBE_AUTOC_LMS_KX4_AN_1G_AN:
333 *speed = IXGBE_LINK_SPEED_UNKNOWN;
334 if (autoc & IXGBE_AUTOC_KX4_SUPP)
335 *speed |= IXGBE_LINK_SPEED_10GB_FULL;
336 if (autoc & IXGBE_AUTOC_KX_SUPP)
337 *speed |= IXGBE_LINK_SPEED_1GB_FULL;
338 *autoneg = TRUE;
339 break;
340
341 default:
342 status = IXGBE_ERR_LINK_SETUP;
343 break;
344 }
345
346 return status;
347}
348
349/**
350 * ixgbe_get_media_type_82598 - Determines media type
351 * @hw: pointer to hardware structure
352 *
353 * Returns the media type (fiber, copper, backplane)
354 **/
355static enum ixgbe_media_type ixgbe_get_media_type_82598(struct ixgbe_hw *hw)
356{
357 enum ixgbe_media_type media_type;
358
359 DEBUGFUNC("ixgbe_get_media_type_82598");
360
361 /* Detect if there is a copper PHY attached. */
362 switch (hw->phy.type) {
363 case ixgbe_phy_cu_unknown:
364 case ixgbe_phy_tn:
365 media_type = ixgbe_media_type_copper;
366 goto out;
367 default:
368 break;
369 }
370
371 /* Media type for I82598 is based on device ID */
372 switch (hw->device_id) {
373 case IXGBE_DEV_ID_82598:
374 case IXGBE_DEV_ID_82598_BX:
375 /* Default device ID is mezzanine card KX/KX4 */
376 media_type = ixgbe_media_type_backplane;
377 break;
378 case IXGBE_DEV_ID_82598AF_DUAL_PORT:
379 case IXGBE_DEV_ID_82598AF_SINGLE_PORT:
380 case IXGBE_DEV_ID_82598_DA_DUAL_PORT:
381 case IXGBE_DEV_ID_82598_SR_DUAL_PORT_EM:
382 case IXGBE_DEV_ID_82598EB_XF_LR:
383 case IXGBE_DEV_ID_82598EB_SFP_LOM:
384 media_type = ixgbe_media_type_fiber;
385 break;
386 case IXGBE_DEV_ID_82598EB_CX4:
387 case IXGBE_DEV_ID_82598_CX4_DUAL_PORT:
388 media_type = ixgbe_media_type_cx4;
389 break;
390 case IXGBE_DEV_ID_82598AT:
391 case IXGBE_DEV_ID_82598AT2:
392 media_type = ixgbe_media_type_copper;
393 break;
394 default:
395 media_type = ixgbe_media_type_unknown;
396 break;
397 }
398out:
399 return media_type;
400}
401
402/**
403 * ixgbe_fc_enable_82598 - Enable flow control
404 * @hw: pointer to hardware structure
405 *
406 * Enable flow control according to the current settings.
407 **/
408s32 ixgbe_fc_enable_82598(struct ixgbe_hw *hw)
409{
410 s32 ret_val = IXGBE_SUCCESS;
411 u32 fctrl_reg;
412 u32 rmcs_reg;
413 u32 reg;
414 u32 fcrtl, fcrth;
415 u32 link_speed = 0;
416 int i;
417 bool link_up;
418
419 DEBUGFUNC("ixgbe_fc_enable_82598");
420
421 /* Validate the water mark configuration */
422 if (!hw->fc.pause_time) {
423 ret_val = IXGBE_ERR_INVALID_LINK_SETTINGS;
424 goto out;
425 }
426
427 /* Low water mark of zero causes XOFF floods */
428 for (i = 0; i < IXGBE_DCB_MAX_TRAFFIC_CLASS; i++) {
429 if ((hw->fc.current_mode & ixgbe_fc_tx_pause) &&
430 hw->fc.high_water[i]) {
431 if (!hw->fc.low_water[i] ||
432 hw->fc.low_water[i] >= hw->fc.high_water[i]) {
433 DEBUGOUT("Invalid water mark configuration\n");
434 ret_val = IXGBE_ERR_INVALID_LINK_SETTINGS;
435 goto out;
436 }
437 }
438 }
439
440 /*
441 * On 82598 having Rx FC on causes resets while doing 1G
442 * so if it's on turn it off once we know link_speed. For
443 * more details see 82598 Specification update.
444 */
445 hw->mac.ops.check_link(hw, &link_speed, &link_up, FALSE);
446 if (link_up && link_speed == IXGBE_LINK_SPEED_1GB_FULL) {
447 switch (hw->fc.requested_mode) {
448 case ixgbe_fc_full:
449 hw->fc.requested_mode = ixgbe_fc_tx_pause;
450 break;
451 case ixgbe_fc_rx_pause:
452 hw->fc.requested_mode = ixgbe_fc_none;
453 break;
454 default:
455 /* no change */
456 break;
457 }
458 }
459
460 /* Negotiate the fc mode to use */
461 ixgbe_fc_autoneg(hw);
462
463 /* Disable any previous flow control settings */
464 fctrl_reg = IXGBE_READ_REG(hw, IXGBE_FCTRL);
465 fctrl_reg &= ~(IXGBE_FCTRL_RFCE | IXGBE_FCTRL_RPFCE);
466
467 rmcs_reg = IXGBE_READ_REG(hw, IXGBE_RMCS);
468 rmcs_reg &= ~(IXGBE_RMCS_TFCE_PRIORITY | IXGBE_RMCS_TFCE_802_3X);
469
470 /*
471 * The possible values of fc.current_mode are:
472 * 0: Flow control is completely disabled
473 * 1: Rx flow control is enabled (we can receive pause frames,
474 * but not send pause frames).
475 * 2: Tx flow control is enabled (we can send pause frames but
476 * we do not support receiving pause frames).
477 * 3: Both Rx and Tx flow control (symmetric) are enabled.
478 * other: Invalid.
479 */
480 switch (hw->fc.current_mode) {
481 case ixgbe_fc_none:
482 /*
483 * Flow control is disabled by software override or autoneg.
484 * The code below will actually disable it in the HW.
485 */
486 break;
487 case ixgbe_fc_rx_pause:
488 /*
489 * Rx Flow control is enabled and Tx Flow control is
490 * disabled by software override. Since there really
491 * isn't a way to advertise that we are capable of RX
492 * Pause ONLY, we will advertise that we support both
493 * symmetric and asymmetric Rx PAUSE. Later, we will
494 * disable the adapter's ability to send PAUSE frames.
495 */
496 fctrl_reg |= IXGBE_FCTRL_RFCE;
497 break;
498 case ixgbe_fc_tx_pause:
499 /*
500 * Tx Flow control is enabled, and Rx Flow control is
501 * disabled by software override.
502 */
503 rmcs_reg |= IXGBE_RMCS_TFCE_802_3X;
504 break;
505 case ixgbe_fc_full:
506 /* Flow control (both Rx and Tx) is enabled by SW override. */
507 fctrl_reg |= IXGBE_FCTRL_RFCE;
508 rmcs_reg |= IXGBE_RMCS_TFCE_802_3X;
509 break;
510 default:
511 DEBUGOUT("Flow control param set incorrectly\n");
512 ret_val = IXGBE_ERR_CONFIG;
513 goto out;
514 break;
515 }
516
517 /* Set 802.3x based flow control settings. */
518 fctrl_reg |= IXGBE_FCTRL_DPF;
519 IXGBE_WRITE_REG(hw, IXGBE_FCTRL, fctrl_reg);
520 IXGBE_WRITE_REG(hw, IXGBE_RMCS, rmcs_reg);
521
522 /* Set up and enable Rx high/low water mark thresholds, enable XON. */
523 for (i = 0; i < IXGBE_DCB_MAX_TRAFFIC_CLASS; i++) {
524 if ((hw->fc.current_mode & ixgbe_fc_tx_pause) &&
525 hw->fc.high_water[i]) {
526 fcrtl = (hw->fc.low_water[i] << 10) | IXGBE_FCRTL_XONE;
527 fcrth = (hw->fc.high_water[i] << 10) | IXGBE_FCRTH_FCEN;
528 IXGBE_WRITE_REG(hw, IXGBE_FCRTL(i), fcrtl);
529 IXGBE_WRITE_REG(hw, IXGBE_FCRTH(i), fcrth);
530 } else {
531 IXGBE_WRITE_REG(hw, IXGBE_FCRTL(i), 0);
532 IXGBE_WRITE_REG(hw, IXGBE_FCRTH(i), 0);
533 }
534
535 }
536
537 /* Configure pause time (2 TCs per register) */
538 reg = hw->fc.pause_time * 0x00010001;
539 for (i = 0; i < (IXGBE_DCB_MAX_TRAFFIC_CLASS / 2); i++)
540 IXGBE_WRITE_REG(hw, IXGBE_FCTTV(i), reg);
541
542 /* Configure flow control refresh threshold value */
543 IXGBE_WRITE_REG(hw, IXGBE_FCRTV, hw->fc.pause_time / 2);
544
545out:
546 return ret_val;
547}
548
549/**
550 * ixgbe_start_mac_link_82598 - Configures MAC link settings
551 * @hw: pointer to hardware structure
552 *
553 * Configures link settings based on values in the ixgbe_hw struct.
554 * Restarts the link. Performs autonegotiation if needed.
555 **/
556static s32 ixgbe_start_mac_link_82598(struct ixgbe_hw *hw,
557 bool autoneg_wait_to_complete)
558{
559 u32 autoc_reg;
560 u32 links_reg;
561 u32 i;
562 s32 status = IXGBE_SUCCESS;
563
564 DEBUGFUNC("ixgbe_start_mac_link_82598");
565
566 /* Restart link */
567 autoc_reg = IXGBE_READ_REG(hw, IXGBE_AUTOC);
568 autoc_reg |= IXGBE_AUTOC_AN_RESTART;
569 IXGBE_WRITE_REG(hw, IXGBE_AUTOC, autoc_reg);
570
571 /* Only poll for autoneg to complete if specified to do so */
572 if (autoneg_wait_to_complete) {
573 if ((autoc_reg & IXGBE_AUTOC_LMS_MASK) ==
574 IXGBE_AUTOC_LMS_KX4_AN ||
575 (autoc_reg & IXGBE_AUTOC_LMS_MASK) ==
576 IXGBE_AUTOC_LMS_KX4_AN_1G_AN) {
577 links_reg = 0; /* Just in case Autoneg time = 0 */
578 for (i = 0; i < IXGBE_AUTO_NEG_TIME; i++) {
579 links_reg = IXGBE_READ_REG(hw, IXGBE_LINKS);
580 if (links_reg & IXGBE_LINKS_KX_AN_COMP)
581 break;
582 msec_delay(100);
583 }
584 if (!(links_reg & IXGBE_LINKS_KX_AN_COMP)) {
585 status = IXGBE_ERR_AUTONEG_NOT_COMPLETE;
586 DEBUGOUT("Autonegotiation did not complete.\n");
587 }
588 }
589 }
590
591 /* Add delay to filter out noises during initial link setup */
592 msec_delay(50);
593
594 return status;
595}
596
597/**
598 * ixgbe_validate_link_ready - Function looks for phy link
599 * @hw: pointer to hardware structure
600 *
601 * Function indicates success when phy link is available. If phy is not ready
602 * within 5 seconds of MAC indicating link, the function returns error.
603 **/
604static s32 ixgbe_validate_link_ready(struct ixgbe_hw *hw)
605{
606 u32 timeout;
607 u16 an_reg;
608
609 if (hw->device_id != IXGBE_DEV_ID_82598AT2)
610 return IXGBE_SUCCESS;
611
612 for (timeout = 0;
613 timeout < IXGBE_VALIDATE_LINK_READY_TIMEOUT; timeout++) {
614 hw->phy.ops.read_reg(hw, IXGBE_MDIO_AUTO_NEG_STATUS,
615 IXGBE_MDIO_AUTO_NEG_DEV_TYPE, &an_reg);
616
617 if ((an_reg & IXGBE_MII_AUTONEG_COMPLETE) &&
618 (an_reg & IXGBE_MII_AUTONEG_LINK_UP))
619 break;
620
621 msec_delay(100);
622 }
623
624 if (timeout == IXGBE_VALIDATE_LINK_READY_TIMEOUT) {
625 DEBUGOUT("Link was indicated but link is down\n");
626 return IXGBE_ERR_LINK_SETUP;
627 }
628
629 return IXGBE_SUCCESS;
630}
631
632/**
633 * ixgbe_check_mac_link_82598 - Get link/speed status
634 * @hw: pointer to hardware structure
635 * @speed: pointer to link speed
636 * @link_up: TRUE is link is up, FALSE otherwise
637 * @link_up_wait_to_complete: bool used to wait for link up or not
638 *
639 * Reads the links register to determine if link is up and the current speed
640 **/
641static s32 ixgbe_check_mac_link_82598(struct ixgbe_hw *hw,
642 ixgbe_link_speed *speed, bool *link_up,
643 bool link_up_wait_to_complete)
644{
645 u32 links_reg;
646 u32 i;
647 u16 link_reg, adapt_comp_reg;
648
649 DEBUGFUNC("ixgbe_check_mac_link_82598");
650
651 /*
652 * SERDES PHY requires us to read link status from undocumented
653 * register 0xC79F. Bit 0 set indicates link is up/ready; clear
654 * indicates link down. OxC00C is read to check that the XAUI lanes
655 * are active. Bit 0 clear indicates active; set indicates inactive.
656 */
657 if (hw->phy.type == ixgbe_phy_nl) {
658 hw->phy.ops.read_reg(hw, 0xC79F, IXGBE_TWINAX_DEV, &link_reg);
659 hw->phy.ops.read_reg(hw, 0xC79F, IXGBE_TWINAX_DEV, &link_reg);
660 hw->phy.ops.read_reg(hw, 0xC00C, IXGBE_TWINAX_DEV,
661 &adapt_comp_reg);
662 if (link_up_wait_to_complete) {
663 for (i = 0; i < hw->mac.max_link_up_time; i++) {
664 if ((link_reg & 1) &&
665 ((adapt_comp_reg & 1) == 0)) {
666 *link_up = TRUE;
667 break;
668 } else {
669 *link_up = FALSE;
670 }
671 msec_delay(100);
672 hw->phy.ops.read_reg(hw, 0xC79F,
673 IXGBE_TWINAX_DEV,
674 &link_reg);
675 hw->phy.ops.read_reg(hw, 0xC00C,
676 IXGBE_TWINAX_DEV,
677 &adapt_comp_reg);
678 }
679 } else {
680 if ((link_reg & 1) && ((adapt_comp_reg & 1) == 0))
681 *link_up = TRUE;
682 else
683 *link_up = FALSE;
684 }
685
686 if (*link_up == FALSE)
687 goto out;
688 }
689
690 links_reg = IXGBE_READ_REG(hw, IXGBE_LINKS);
691 if (link_up_wait_to_complete) {
692 for (i = 0; i < hw->mac.max_link_up_time; i++) {
693 if (links_reg & IXGBE_LINKS_UP) {
694 *link_up = TRUE;
695 break;
696 } else {
697 *link_up = FALSE;
698 }
699 msec_delay(100);
700 links_reg = IXGBE_READ_REG(hw, IXGBE_LINKS);
701 }
702 } else {
703 if (links_reg & IXGBE_LINKS_UP)
704 *link_up = TRUE;
705 else
706 *link_up = FALSE;
707 }
708
709 if (links_reg & IXGBE_LINKS_SPEED)
710 *speed = IXGBE_LINK_SPEED_10GB_FULL;
711 else
712 *speed = IXGBE_LINK_SPEED_1GB_FULL;
713
714 if ((hw->device_id == IXGBE_DEV_ID_82598AT2) && (*link_up == TRUE) &&
715 (ixgbe_validate_link_ready(hw) != IXGBE_SUCCESS))
716 *link_up = FALSE;
717
718out:
719 return IXGBE_SUCCESS;
720}
721
722/**
723 * ixgbe_setup_mac_link_82598 - Set MAC link speed
724 * @hw: pointer to hardware structure
725 * @speed: new link speed
726 * @autoneg_wait_to_complete: TRUE when waiting for completion is needed
727 *
728 * Set the link speed in the AUTOC register and restarts link.
729 **/
730static s32 ixgbe_setup_mac_link_82598(struct ixgbe_hw *hw,
731 ixgbe_link_speed speed,
732 bool autoneg_wait_to_complete)
733{
734 bool autoneg = FALSE;
735 s32 status = IXGBE_SUCCESS;
736 ixgbe_link_speed link_capabilities = IXGBE_LINK_SPEED_UNKNOWN;
737 u32 curr_autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
738 u32 autoc = curr_autoc;
739 u32 link_mode = autoc & IXGBE_AUTOC_LMS_MASK;
740
741 DEBUGFUNC("ixgbe_setup_mac_link_82598");
742
743 /* Check to see if speed passed in is supported. */
744 ixgbe_get_link_capabilities(hw, &link_capabilities, &autoneg);
745 speed &= link_capabilities;
746
747 if (speed == IXGBE_LINK_SPEED_UNKNOWN)
748 status = IXGBE_ERR_LINK_SETUP;
749
750 /* Set KX4/KX support according to speed requested */
751 else if (link_mode == IXGBE_AUTOC_LMS_KX4_AN ||
752 link_mode == IXGBE_AUTOC_LMS_KX4_AN_1G_AN) {
753 autoc &= ~IXGBE_AUTOC_KX4_KX_SUPP_MASK;
754 if (speed & IXGBE_LINK_SPEED_10GB_FULL)
755 autoc |= IXGBE_AUTOC_KX4_SUPP;
756 if (speed & IXGBE_LINK_SPEED_1GB_FULL)
757 autoc |= IXGBE_AUTOC_KX_SUPP;
758 if (autoc != curr_autoc)
759 IXGBE_WRITE_REG(hw, IXGBE_AUTOC, autoc);
760 }
761
762 if (status == IXGBE_SUCCESS) {
763 /*
764 * Setup and restart the link based on the new values in
765 * ixgbe_hw This will write the AUTOC register based on the new
766 * stored values
767 */
768 status = ixgbe_start_mac_link_82598(hw,
769 autoneg_wait_to_complete);
770 }
771
772 return status;
773}
774
775
776/**
777 * ixgbe_setup_copper_link_82598 - Set the PHY autoneg advertised field
778 * @hw: pointer to hardware structure
779 * @speed: new link speed
780 * @autoneg_wait_to_complete: TRUE if waiting is needed to complete
781 *
782 * Sets the link speed in the AUTOC register in the MAC and restarts link.
783 **/
784static s32 ixgbe_setup_copper_link_82598(struct ixgbe_hw *hw,
785 ixgbe_link_speed speed,
786 bool autoneg_wait_to_complete)
787{
788 s32 status;
789
790 DEBUGFUNC("ixgbe_setup_copper_link_82598");
791
792 /* Setup the PHY according to input speed */
793 status = hw->phy.ops.setup_link_speed(hw, speed,
794 autoneg_wait_to_complete);
795 /* Set up MAC */
796 ixgbe_start_mac_link_82598(hw, autoneg_wait_to_complete);
797
798 return status;
799}
800
801/**
802 * ixgbe_reset_hw_82598 - Performs hardware reset
803 * @hw: pointer to hardware structure
804 *
805 * Resets the hardware by resetting the transmit and receive units, masks and
806 * clears all interrupts, performing a PHY reset, and performing a link (MAC)
807 * reset.
808 **/
809static s32 ixgbe_reset_hw_82598(struct ixgbe_hw *hw)
810{
811 s32 status = IXGBE_SUCCESS;
812 s32 phy_status = IXGBE_SUCCESS;
813 u32 ctrl;
814 u32 gheccr;
815 u32 i;
816 u32 autoc;
817 u8 analog_val;
818
819 DEBUGFUNC("ixgbe_reset_hw_82598");
820
821 /* Call adapter stop to disable tx/rx and clear interrupts */
822 status = hw->mac.ops.stop_adapter(hw);
823 if (status != IXGBE_SUCCESS)
824 goto reset_hw_out;
825
826 /*
827 * Power up the Atlas Tx lanes if they are currently powered down.
828 * Atlas Tx lanes are powered down for MAC loopback tests, but
829 * they are not automatically restored on reset.
830 */
831 hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_LPBK, &analog_val);
832 if (analog_val & IXGBE_ATLAS_PDN_TX_REG_EN) {
833 /* Enable Tx Atlas so packets can be transmitted again */
834 hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_LPBK,
835 &analog_val);
836 analog_val &= ~IXGBE_ATLAS_PDN_TX_REG_EN;
837 hw->mac.ops.write_analog_reg8(hw, IXGBE_ATLAS_PDN_LPBK,
838 analog_val);
839
840 hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_10G,
841 &analog_val);
842 analog_val &= ~IXGBE_ATLAS_PDN_TX_10G_QL_ALL;
843 hw->mac.ops.write_analog_reg8(hw, IXGBE_ATLAS_PDN_10G,
844 analog_val);
845
846 hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_1G,
847 &analog_val);
848 analog_val &= ~IXGBE_ATLAS_PDN_TX_1G_QL_ALL;
849 hw->mac.ops.write_analog_reg8(hw, IXGBE_ATLAS_PDN_1G,
850 analog_val);
851
852 hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_AN,
853 &analog_val);
854 analog_val &= ~IXGBE_ATLAS_PDN_TX_AN_QL_ALL;
855 hw->mac.ops.write_analog_reg8(hw, IXGBE_ATLAS_PDN_AN,
856 analog_val);
857 }
858
859 /* Reset PHY */
860 if (hw->phy.reset_disable == FALSE) {
861 /* PHY ops must be identified and initialized prior to reset */
862
863 /* Init PHY and function pointers, perform SFP setup */
864 phy_status = hw->phy.ops.init(hw);
865 if (phy_status == IXGBE_ERR_SFP_NOT_SUPPORTED)
866 goto reset_hw_out;
867 if (phy_status == IXGBE_ERR_SFP_NOT_PRESENT)
868 goto mac_reset_top;
869
870 hw->phy.ops.reset(hw);
871 }
872
873mac_reset_top:
874 /*
875 * Issue global reset to the MAC. This needs to be a SW reset.
876 * If link reset is used, it might reset the MAC when mng is using it
877 */
878 ctrl = IXGBE_READ_REG(hw, IXGBE_CTRL) | IXGBE_CTRL_RST;
879 IXGBE_WRITE_REG(hw, IXGBE_CTRL, ctrl);
880 IXGBE_WRITE_FLUSH(hw);
881
882 /* Poll for reset bit to self-clear indicating reset is complete */
883 for (i = 0; i < 10; i++) {
884 usec_delay(1);
885 ctrl = IXGBE_READ_REG(hw, IXGBE_CTRL);
886 if (!(ctrl & IXGBE_CTRL_RST))
887 break;
888 }
889 if (ctrl & IXGBE_CTRL_RST) {
890 status = IXGBE_ERR_RESET_FAILED;
891 DEBUGOUT("Reset polling failed to complete.\n");
892 }
893
894 msec_delay(50);
895
896 /*
897 * Double resets are required for recovery from certain error
898 * conditions. Between resets, it is necessary to stall to allow time
899 * for any pending HW events to complete.
900 */
901 if (hw->mac.flags & IXGBE_FLAGS_DOUBLE_RESET_REQUIRED) {
902 hw->mac.flags &= ~IXGBE_FLAGS_DOUBLE_RESET_REQUIRED;
903 goto mac_reset_top;
904 }
905
906 gheccr = IXGBE_READ_REG(hw, IXGBE_GHECCR);
907 gheccr &= ~((1 << 21) | (1 << 18) | (1 << 9) | (1 << 6));
908 IXGBE_WRITE_REG(hw, IXGBE_GHECCR, gheccr);
909
910 /*
911 * Store the original AUTOC value if it has not been
912 * stored off yet. Otherwise restore the stored original
913 * AUTOC value since the reset operation sets back to deaults.
914 */
915 autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
916 if (hw->mac.orig_link_settings_stored == FALSE) {
917 hw->mac.orig_autoc = autoc;
918 hw->mac.orig_link_settings_stored = TRUE;
919 } else if (autoc != hw->mac.orig_autoc) {
920 IXGBE_WRITE_REG(hw, IXGBE_AUTOC, hw->mac.orig_autoc);
921 }
922
923 /* Store the permanent mac address */
924 hw->mac.ops.get_mac_addr(hw, hw->mac.perm_addr);
925
926 /*
927 * Store MAC address from RAR0, clear receive address registers, and
928 * clear the multicast table
929 */
930 hw->mac.ops.init_rx_addrs(hw);
931
932reset_hw_out:
933 if (phy_status != IXGBE_SUCCESS)
934 status = phy_status;
935
936 return status;
937}
938
939/**
940 * ixgbe_set_vmdq_82598 - Associate a VMDq set index with a rx address
941 * @hw: pointer to hardware struct
942 * @rar: receive address register index to associate with a VMDq index
943 * @vmdq: VMDq set index
944 **/
945s32 ixgbe_set_vmdq_82598(struct ixgbe_hw *hw, u32 rar, u32 vmdq)
946{
947 u32 rar_high;
948 u32 rar_entries = hw->mac.num_rar_entries;
949
950 DEBUGFUNC("ixgbe_set_vmdq_82598");
951
952 /* Make sure we are using a valid rar index range */
953 if (rar >= rar_entries) {
954 DEBUGOUT1("RAR index %d is out of range.\n", rar);
955 return IXGBE_ERR_INVALID_ARGUMENT;
956 }
957
958 rar_high = IXGBE_READ_REG(hw, IXGBE_RAH(rar));
959 rar_high &= ~IXGBE_RAH_VIND_MASK;
960 rar_high |= ((vmdq << IXGBE_RAH_VIND_SHIFT) & IXGBE_RAH_VIND_MASK);
961 IXGBE_WRITE_REG(hw, IXGBE_RAH(rar), rar_high);
962 return IXGBE_SUCCESS;
963}
964
965/**
966 * ixgbe_clear_vmdq_82598 - Disassociate a VMDq set index from an rx address
967 * @hw: pointer to hardware struct
968 * @rar: receive address register index to associate with a VMDq index
969 * @vmdq: VMDq clear index (not used in 82598, but elsewhere)
970 **/
971static s32 ixgbe_clear_vmdq_82598(struct ixgbe_hw *hw, u32 rar, u32 vmdq)
972{
973 u32 rar_high;
974 u32 rar_entries = hw->mac.num_rar_entries;
975
976 UNREFERENCED_1PARAMETER(vmdq);
977
978 /* Make sure we are using a valid rar index range */
979 if (rar >= rar_entries) {
980 DEBUGOUT1("RAR index %d is out of range.\n", rar);
981 return IXGBE_ERR_INVALID_ARGUMENT;
982 }
983
984 rar_high = IXGBE_READ_REG(hw, IXGBE_RAH(rar));
985 if (rar_high & IXGBE_RAH_VIND_MASK) {
986 rar_high &= ~IXGBE_RAH_VIND_MASK;
987 IXGBE_WRITE_REG(hw, IXGBE_RAH(rar), rar_high);
988 }
989
990 return IXGBE_SUCCESS;
991}
992
993/**
994 * ixgbe_set_vfta_82598 - Set VLAN filter table
995 * @hw: pointer to hardware structure
996 * @vlan: VLAN id to write to VLAN filter
997 * @vind: VMDq output index that maps queue to VLAN id in VFTA
998 * @vlan_on: boolean flag to turn on/off VLAN in VFTA
34
35#include "ixgbe_type.h"
36#include "ixgbe_82598.h"
37#include "ixgbe_api.h"
38#include "ixgbe_common.h"
39#include "ixgbe_phy.h"
40
41#define IXGBE_82598_MAX_TX_QUEUES 32
42#define IXGBE_82598_MAX_RX_QUEUES 64
43#define IXGBE_82598_RAR_ENTRIES 16
44#define IXGBE_82598_MC_TBL_SIZE 128
45#define IXGBE_82598_VFT_TBL_SIZE 128
46#define IXGBE_82598_RX_PB_SIZE 512
47
48static s32 ixgbe_get_link_capabilities_82598(struct ixgbe_hw *hw,
49 ixgbe_link_speed *speed,
50 bool *autoneg);
51static enum ixgbe_media_type ixgbe_get_media_type_82598(struct ixgbe_hw *hw);
52static s32 ixgbe_start_mac_link_82598(struct ixgbe_hw *hw,
53 bool autoneg_wait_to_complete);
54static s32 ixgbe_check_mac_link_82598(struct ixgbe_hw *hw,
55 ixgbe_link_speed *speed, bool *link_up,
56 bool link_up_wait_to_complete);
57static s32 ixgbe_setup_mac_link_82598(struct ixgbe_hw *hw,
58 ixgbe_link_speed speed,
59 bool autoneg_wait_to_complete);
60static s32 ixgbe_setup_copper_link_82598(struct ixgbe_hw *hw,
61 ixgbe_link_speed speed,
62 bool autoneg_wait_to_complete);
63static s32 ixgbe_reset_hw_82598(struct ixgbe_hw *hw);
64static s32 ixgbe_clear_vmdq_82598(struct ixgbe_hw *hw, u32 rar, u32 vmdq);
65static s32 ixgbe_clear_vfta_82598(struct ixgbe_hw *hw);
66static void ixgbe_set_rxpba_82598(struct ixgbe_hw *hw, int num_pb,
67 u32 headroom, int strategy);
68static s32 ixgbe_read_i2c_sff8472_82598(struct ixgbe_hw *hw, u8 byte_offset,
69 u8 *sff8472_data);
70/**
71 * ixgbe_set_pcie_completion_timeout - set pci-e completion timeout
72 * @hw: pointer to the HW structure
73 *
74 * The defaults for 82598 should be in the range of 50us to 50ms,
75 * however the hardware default for these parts is 500us to 1ms which is less
76 * than the 10ms recommended by the pci-e spec. To address this we need to
77 * increase the value to either 10ms to 250ms for capability version 1 config,
78 * or 16ms to 55ms for version 2.
79 **/
80void ixgbe_set_pcie_completion_timeout(struct ixgbe_hw *hw)
81{
82 u32 gcr = IXGBE_READ_REG(hw, IXGBE_GCR);
83 u16 pcie_devctl2;
84
85 /* only take action if timeout value is defaulted to 0 */
86 if (gcr & IXGBE_GCR_CMPL_TMOUT_MASK)
87 goto out;
88
89 /*
90 * if capababilities version is type 1 we can write the
91 * timeout of 10ms to 250ms through the GCR register
92 */
93 if (!(gcr & IXGBE_GCR_CAP_VER2)) {
94 gcr |= IXGBE_GCR_CMPL_TMOUT_10ms;
95 goto out;
96 }
97
98 /*
99 * for version 2 capabilities we need to write the config space
100 * directly in order to set the completion timeout value for
101 * 16ms to 55ms
102 */
103 pcie_devctl2 = IXGBE_READ_PCIE_WORD(hw, IXGBE_PCI_DEVICE_CONTROL2);
104 pcie_devctl2 |= IXGBE_PCI_DEVICE_CONTROL2_16ms;
105 IXGBE_WRITE_PCIE_WORD(hw, IXGBE_PCI_DEVICE_CONTROL2, pcie_devctl2);
106out:
107 /* disable completion timeout resend */
108 gcr &= ~IXGBE_GCR_CMPL_TMOUT_RESEND;
109 IXGBE_WRITE_REG(hw, IXGBE_GCR, gcr);
110}
111
112/**
113 * ixgbe_init_ops_82598 - Inits func ptrs and MAC type
114 * @hw: pointer to hardware structure
115 *
116 * Initialize the function pointers and assign the MAC type for 82598.
117 * Does not touch the hardware.
118 **/
119s32 ixgbe_init_ops_82598(struct ixgbe_hw *hw)
120{
121 struct ixgbe_mac_info *mac = &hw->mac;
122 struct ixgbe_phy_info *phy = &hw->phy;
123 s32 ret_val;
124
125 DEBUGFUNC("ixgbe_init_ops_82598");
126
127 ret_val = ixgbe_init_phy_ops_generic(hw);
128 ret_val = ixgbe_init_ops_generic(hw);
129
130 /* PHY */
131 phy->ops.init = ixgbe_init_phy_ops_82598;
132
133 /* MAC */
134 mac->ops.start_hw = ixgbe_start_hw_82598;
135 mac->ops.enable_relaxed_ordering = ixgbe_enable_relaxed_ordering_82598;
136 mac->ops.reset_hw = ixgbe_reset_hw_82598;
137 mac->ops.get_media_type = ixgbe_get_media_type_82598;
138 mac->ops.get_supported_physical_layer =
139 ixgbe_get_supported_physical_layer_82598;
140 mac->ops.read_analog_reg8 = ixgbe_read_analog_reg8_82598;
141 mac->ops.write_analog_reg8 = ixgbe_write_analog_reg8_82598;
142 mac->ops.set_lan_id = ixgbe_set_lan_id_multi_port_pcie_82598;
143 mac->ops.enable_rx_dma = ixgbe_enable_rx_dma_82598;
144
145 /* RAR, Multicast, VLAN */
146 mac->ops.set_vmdq = ixgbe_set_vmdq_82598;
147 mac->ops.clear_vmdq = ixgbe_clear_vmdq_82598;
148 mac->ops.set_vfta = ixgbe_set_vfta_82598;
149 mac->ops.set_vlvf = NULL;
150 mac->ops.clear_vfta = ixgbe_clear_vfta_82598;
151
152 /* Flow Control */
153 mac->ops.fc_enable = ixgbe_fc_enable_82598;
154
155 mac->mcft_size = IXGBE_82598_MC_TBL_SIZE;
156 mac->vft_size = IXGBE_82598_VFT_TBL_SIZE;
157 mac->num_rar_entries = IXGBE_82598_RAR_ENTRIES;
158 mac->rx_pb_size = IXGBE_82598_RX_PB_SIZE;
159 mac->max_rx_queues = IXGBE_82598_MAX_RX_QUEUES;
160 mac->max_tx_queues = IXGBE_82598_MAX_TX_QUEUES;
161 mac->max_msix_vectors = ixgbe_get_pcie_msix_count_generic(hw);
162
163 /* SFP+ Module */
164 phy->ops.read_i2c_eeprom = ixgbe_read_i2c_eeprom_82598;
165 phy->ops.read_i2c_sff8472 = ixgbe_read_i2c_sff8472_82598;
166
167 /* Link */
168 mac->ops.check_link = ixgbe_check_mac_link_82598;
169 mac->ops.setup_link = ixgbe_setup_mac_link_82598;
170 mac->ops.flap_tx_laser = NULL;
171 mac->ops.get_link_capabilities = ixgbe_get_link_capabilities_82598;
172 mac->ops.setup_rxpba = ixgbe_set_rxpba_82598;
173
174 /* Manageability interface */
175 mac->ops.set_fw_drv_ver = NULL;
176
177 mac->ops.get_rtrup2tc = NULL;
178
179 return ret_val;
180}
181
182/**
183 * ixgbe_init_phy_ops_82598 - PHY/SFP specific init
184 * @hw: pointer to hardware structure
185 *
186 * Initialize any function pointers that were not able to be
187 * set during init_shared_code because the PHY/SFP type was
188 * not known. Perform the SFP init if necessary.
189 *
190 **/
191s32 ixgbe_init_phy_ops_82598(struct ixgbe_hw *hw)
192{
193 struct ixgbe_mac_info *mac = &hw->mac;
194 struct ixgbe_phy_info *phy = &hw->phy;
195 s32 ret_val = IXGBE_SUCCESS;
196 u16 list_offset, data_offset;
197
198 DEBUGFUNC("ixgbe_init_phy_ops_82598");
199
200 /* Identify the PHY */
201 phy->ops.identify(hw);
202
203 /* Overwrite the link function pointers if copper PHY */
204 if (mac->ops.get_media_type(hw) == ixgbe_media_type_copper) {
205 mac->ops.setup_link = ixgbe_setup_copper_link_82598;
206 mac->ops.get_link_capabilities =
207 ixgbe_get_copper_link_capabilities_generic;
208 }
209
210 switch (hw->phy.type) {
211 case ixgbe_phy_tn:
212 phy->ops.setup_link = ixgbe_setup_phy_link_tnx;
213 phy->ops.check_link = ixgbe_check_phy_link_tnx;
214 phy->ops.get_firmware_version =
215 ixgbe_get_phy_firmware_version_tnx;
216 break;
217 case ixgbe_phy_nl:
218 phy->ops.reset = ixgbe_reset_phy_nl;
219
220 /* Call SFP+ identify routine to get the SFP+ module type */
221 ret_val = phy->ops.identify_sfp(hw);
222 if (ret_val != IXGBE_SUCCESS)
223 goto out;
224 else if (hw->phy.sfp_type == ixgbe_sfp_type_unknown) {
225 ret_val = IXGBE_ERR_SFP_NOT_SUPPORTED;
226 goto out;
227 }
228
229 /* Check to see if SFP+ module is supported */
230 ret_val = ixgbe_get_sfp_init_sequence_offsets(hw,
231 &list_offset,
232 &data_offset);
233 if (ret_val != IXGBE_SUCCESS) {
234 ret_val = IXGBE_ERR_SFP_NOT_SUPPORTED;
235 goto out;
236 }
237 break;
238 default:
239 break;
240 }
241
242out:
243 return ret_val;
244}
245
246/**
247 * ixgbe_start_hw_82598 - Prepare hardware for Tx/Rx
248 * @hw: pointer to hardware structure
249 *
250 * Starts the hardware using the generic start_hw function.
251 * Disables relaxed ordering Then set pcie completion timeout
252 *
253 **/
254s32 ixgbe_start_hw_82598(struct ixgbe_hw *hw)
255{
256 u32 regval;
257 u32 i;
258 s32 ret_val = IXGBE_SUCCESS;
259
260 DEBUGFUNC("ixgbe_start_hw_82598");
261
262 ret_val = ixgbe_start_hw_generic(hw);
263 if (ret_val)
264 return ret_val;
265
266 /* Disable relaxed ordering */
267 for (i = 0; ((i < hw->mac.max_tx_queues) &&
268 (i < IXGBE_DCA_MAX_QUEUES_82598)); i++) {
269 regval = IXGBE_READ_REG(hw, IXGBE_DCA_TXCTRL(i));
270 regval &= ~IXGBE_DCA_TXCTRL_DESC_WRO_EN;
271 IXGBE_WRITE_REG(hw, IXGBE_DCA_TXCTRL(i), regval);
272 }
273
274 for (i = 0; ((i < hw->mac.max_rx_queues) &&
275 (i < IXGBE_DCA_MAX_QUEUES_82598)); i++) {
276 regval = IXGBE_READ_REG(hw, IXGBE_DCA_RXCTRL(i));
277 regval &= ~(IXGBE_DCA_RXCTRL_DATA_WRO_EN |
278 IXGBE_DCA_RXCTRL_HEAD_WRO_EN);
279 IXGBE_WRITE_REG(hw, IXGBE_DCA_RXCTRL(i), regval);
280 }
281
282 /* set the completion timeout for interface */
283 ixgbe_set_pcie_completion_timeout(hw);
284
285 return ret_val;
286}
287
288/**
289 * ixgbe_get_link_capabilities_82598 - Determines link capabilities
290 * @hw: pointer to hardware structure
291 * @speed: pointer to link speed
292 * @autoneg: boolean auto-negotiation value
293 *
294 * Determines the link capabilities by reading the AUTOC register.
295 **/
296static s32 ixgbe_get_link_capabilities_82598(struct ixgbe_hw *hw,
297 ixgbe_link_speed *speed,
298 bool *autoneg)
299{
300 s32 status = IXGBE_SUCCESS;
301 u32 autoc = 0;
302
303 DEBUGFUNC("ixgbe_get_link_capabilities_82598");
304
305 /*
306 * Determine link capabilities based on the stored value of AUTOC,
307 * which represents EEPROM defaults. If AUTOC value has not been
308 * stored, use the current register value.
309 */
310 if (hw->mac.orig_link_settings_stored)
311 autoc = hw->mac.orig_autoc;
312 else
313 autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
314
315 switch (autoc & IXGBE_AUTOC_LMS_MASK) {
316 case IXGBE_AUTOC_LMS_1G_LINK_NO_AN:
317 *speed = IXGBE_LINK_SPEED_1GB_FULL;
318 *autoneg = FALSE;
319 break;
320
321 case IXGBE_AUTOC_LMS_10G_LINK_NO_AN:
322 *speed = IXGBE_LINK_SPEED_10GB_FULL;
323 *autoneg = FALSE;
324 break;
325
326 case IXGBE_AUTOC_LMS_1G_AN:
327 *speed = IXGBE_LINK_SPEED_1GB_FULL;
328 *autoneg = TRUE;
329 break;
330
331 case IXGBE_AUTOC_LMS_KX4_AN:
332 case IXGBE_AUTOC_LMS_KX4_AN_1G_AN:
333 *speed = IXGBE_LINK_SPEED_UNKNOWN;
334 if (autoc & IXGBE_AUTOC_KX4_SUPP)
335 *speed |= IXGBE_LINK_SPEED_10GB_FULL;
336 if (autoc & IXGBE_AUTOC_KX_SUPP)
337 *speed |= IXGBE_LINK_SPEED_1GB_FULL;
338 *autoneg = TRUE;
339 break;
340
341 default:
342 status = IXGBE_ERR_LINK_SETUP;
343 break;
344 }
345
346 return status;
347}
348
349/**
350 * ixgbe_get_media_type_82598 - Determines media type
351 * @hw: pointer to hardware structure
352 *
353 * Returns the media type (fiber, copper, backplane)
354 **/
355static enum ixgbe_media_type ixgbe_get_media_type_82598(struct ixgbe_hw *hw)
356{
357 enum ixgbe_media_type media_type;
358
359 DEBUGFUNC("ixgbe_get_media_type_82598");
360
361 /* Detect if there is a copper PHY attached. */
362 switch (hw->phy.type) {
363 case ixgbe_phy_cu_unknown:
364 case ixgbe_phy_tn:
365 media_type = ixgbe_media_type_copper;
366 goto out;
367 default:
368 break;
369 }
370
371 /* Media type for I82598 is based on device ID */
372 switch (hw->device_id) {
373 case IXGBE_DEV_ID_82598:
374 case IXGBE_DEV_ID_82598_BX:
375 /* Default device ID is mezzanine card KX/KX4 */
376 media_type = ixgbe_media_type_backplane;
377 break;
378 case IXGBE_DEV_ID_82598AF_DUAL_PORT:
379 case IXGBE_DEV_ID_82598AF_SINGLE_PORT:
380 case IXGBE_DEV_ID_82598_DA_DUAL_PORT:
381 case IXGBE_DEV_ID_82598_SR_DUAL_PORT_EM:
382 case IXGBE_DEV_ID_82598EB_XF_LR:
383 case IXGBE_DEV_ID_82598EB_SFP_LOM:
384 media_type = ixgbe_media_type_fiber;
385 break;
386 case IXGBE_DEV_ID_82598EB_CX4:
387 case IXGBE_DEV_ID_82598_CX4_DUAL_PORT:
388 media_type = ixgbe_media_type_cx4;
389 break;
390 case IXGBE_DEV_ID_82598AT:
391 case IXGBE_DEV_ID_82598AT2:
392 media_type = ixgbe_media_type_copper;
393 break;
394 default:
395 media_type = ixgbe_media_type_unknown;
396 break;
397 }
398out:
399 return media_type;
400}
401
402/**
403 * ixgbe_fc_enable_82598 - Enable flow control
404 * @hw: pointer to hardware structure
405 *
406 * Enable flow control according to the current settings.
407 **/
408s32 ixgbe_fc_enable_82598(struct ixgbe_hw *hw)
409{
410 s32 ret_val = IXGBE_SUCCESS;
411 u32 fctrl_reg;
412 u32 rmcs_reg;
413 u32 reg;
414 u32 fcrtl, fcrth;
415 u32 link_speed = 0;
416 int i;
417 bool link_up;
418
419 DEBUGFUNC("ixgbe_fc_enable_82598");
420
421 /* Validate the water mark configuration */
422 if (!hw->fc.pause_time) {
423 ret_val = IXGBE_ERR_INVALID_LINK_SETTINGS;
424 goto out;
425 }
426
427 /* Low water mark of zero causes XOFF floods */
428 for (i = 0; i < IXGBE_DCB_MAX_TRAFFIC_CLASS; i++) {
429 if ((hw->fc.current_mode & ixgbe_fc_tx_pause) &&
430 hw->fc.high_water[i]) {
431 if (!hw->fc.low_water[i] ||
432 hw->fc.low_water[i] >= hw->fc.high_water[i]) {
433 DEBUGOUT("Invalid water mark configuration\n");
434 ret_val = IXGBE_ERR_INVALID_LINK_SETTINGS;
435 goto out;
436 }
437 }
438 }
439
440 /*
441 * On 82598 having Rx FC on causes resets while doing 1G
442 * so if it's on turn it off once we know link_speed. For
443 * more details see 82598 Specification update.
444 */
445 hw->mac.ops.check_link(hw, &link_speed, &link_up, FALSE);
446 if (link_up && link_speed == IXGBE_LINK_SPEED_1GB_FULL) {
447 switch (hw->fc.requested_mode) {
448 case ixgbe_fc_full:
449 hw->fc.requested_mode = ixgbe_fc_tx_pause;
450 break;
451 case ixgbe_fc_rx_pause:
452 hw->fc.requested_mode = ixgbe_fc_none;
453 break;
454 default:
455 /* no change */
456 break;
457 }
458 }
459
460 /* Negotiate the fc mode to use */
461 ixgbe_fc_autoneg(hw);
462
463 /* Disable any previous flow control settings */
464 fctrl_reg = IXGBE_READ_REG(hw, IXGBE_FCTRL);
465 fctrl_reg &= ~(IXGBE_FCTRL_RFCE | IXGBE_FCTRL_RPFCE);
466
467 rmcs_reg = IXGBE_READ_REG(hw, IXGBE_RMCS);
468 rmcs_reg &= ~(IXGBE_RMCS_TFCE_PRIORITY | IXGBE_RMCS_TFCE_802_3X);
469
470 /*
471 * The possible values of fc.current_mode are:
472 * 0: Flow control is completely disabled
473 * 1: Rx flow control is enabled (we can receive pause frames,
474 * but not send pause frames).
475 * 2: Tx flow control is enabled (we can send pause frames but
476 * we do not support receiving pause frames).
477 * 3: Both Rx and Tx flow control (symmetric) are enabled.
478 * other: Invalid.
479 */
480 switch (hw->fc.current_mode) {
481 case ixgbe_fc_none:
482 /*
483 * Flow control is disabled by software override or autoneg.
484 * The code below will actually disable it in the HW.
485 */
486 break;
487 case ixgbe_fc_rx_pause:
488 /*
489 * Rx Flow control is enabled and Tx Flow control is
490 * disabled by software override. Since there really
491 * isn't a way to advertise that we are capable of RX
492 * Pause ONLY, we will advertise that we support both
493 * symmetric and asymmetric Rx PAUSE. Later, we will
494 * disable the adapter's ability to send PAUSE frames.
495 */
496 fctrl_reg |= IXGBE_FCTRL_RFCE;
497 break;
498 case ixgbe_fc_tx_pause:
499 /*
500 * Tx Flow control is enabled, and Rx Flow control is
501 * disabled by software override.
502 */
503 rmcs_reg |= IXGBE_RMCS_TFCE_802_3X;
504 break;
505 case ixgbe_fc_full:
506 /* Flow control (both Rx and Tx) is enabled by SW override. */
507 fctrl_reg |= IXGBE_FCTRL_RFCE;
508 rmcs_reg |= IXGBE_RMCS_TFCE_802_3X;
509 break;
510 default:
511 DEBUGOUT("Flow control param set incorrectly\n");
512 ret_val = IXGBE_ERR_CONFIG;
513 goto out;
514 break;
515 }
516
517 /* Set 802.3x based flow control settings. */
518 fctrl_reg |= IXGBE_FCTRL_DPF;
519 IXGBE_WRITE_REG(hw, IXGBE_FCTRL, fctrl_reg);
520 IXGBE_WRITE_REG(hw, IXGBE_RMCS, rmcs_reg);
521
522 /* Set up and enable Rx high/low water mark thresholds, enable XON. */
523 for (i = 0; i < IXGBE_DCB_MAX_TRAFFIC_CLASS; i++) {
524 if ((hw->fc.current_mode & ixgbe_fc_tx_pause) &&
525 hw->fc.high_water[i]) {
526 fcrtl = (hw->fc.low_water[i] << 10) | IXGBE_FCRTL_XONE;
527 fcrth = (hw->fc.high_water[i] << 10) | IXGBE_FCRTH_FCEN;
528 IXGBE_WRITE_REG(hw, IXGBE_FCRTL(i), fcrtl);
529 IXGBE_WRITE_REG(hw, IXGBE_FCRTH(i), fcrth);
530 } else {
531 IXGBE_WRITE_REG(hw, IXGBE_FCRTL(i), 0);
532 IXGBE_WRITE_REG(hw, IXGBE_FCRTH(i), 0);
533 }
534
535 }
536
537 /* Configure pause time (2 TCs per register) */
538 reg = hw->fc.pause_time * 0x00010001;
539 for (i = 0; i < (IXGBE_DCB_MAX_TRAFFIC_CLASS / 2); i++)
540 IXGBE_WRITE_REG(hw, IXGBE_FCTTV(i), reg);
541
542 /* Configure flow control refresh threshold value */
543 IXGBE_WRITE_REG(hw, IXGBE_FCRTV, hw->fc.pause_time / 2);
544
545out:
546 return ret_val;
547}
548
549/**
550 * ixgbe_start_mac_link_82598 - Configures MAC link settings
551 * @hw: pointer to hardware structure
552 *
553 * Configures link settings based on values in the ixgbe_hw struct.
554 * Restarts the link. Performs autonegotiation if needed.
555 **/
556static s32 ixgbe_start_mac_link_82598(struct ixgbe_hw *hw,
557 bool autoneg_wait_to_complete)
558{
559 u32 autoc_reg;
560 u32 links_reg;
561 u32 i;
562 s32 status = IXGBE_SUCCESS;
563
564 DEBUGFUNC("ixgbe_start_mac_link_82598");
565
566 /* Restart link */
567 autoc_reg = IXGBE_READ_REG(hw, IXGBE_AUTOC);
568 autoc_reg |= IXGBE_AUTOC_AN_RESTART;
569 IXGBE_WRITE_REG(hw, IXGBE_AUTOC, autoc_reg);
570
571 /* Only poll for autoneg to complete if specified to do so */
572 if (autoneg_wait_to_complete) {
573 if ((autoc_reg & IXGBE_AUTOC_LMS_MASK) ==
574 IXGBE_AUTOC_LMS_KX4_AN ||
575 (autoc_reg & IXGBE_AUTOC_LMS_MASK) ==
576 IXGBE_AUTOC_LMS_KX4_AN_1G_AN) {
577 links_reg = 0; /* Just in case Autoneg time = 0 */
578 for (i = 0; i < IXGBE_AUTO_NEG_TIME; i++) {
579 links_reg = IXGBE_READ_REG(hw, IXGBE_LINKS);
580 if (links_reg & IXGBE_LINKS_KX_AN_COMP)
581 break;
582 msec_delay(100);
583 }
584 if (!(links_reg & IXGBE_LINKS_KX_AN_COMP)) {
585 status = IXGBE_ERR_AUTONEG_NOT_COMPLETE;
586 DEBUGOUT("Autonegotiation did not complete.\n");
587 }
588 }
589 }
590
591 /* Add delay to filter out noises during initial link setup */
592 msec_delay(50);
593
594 return status;
595}
596
597/**
598 * ixgbe_validate_link_ready - Function looks for phy link
599 * @hw: pointer to hardware structure
600 *
601 * Function indicates success when phy link is available. If phy is not ready
602 * within 5 seconds of MAC indicating link, the function returns error.
603 **/
604static s32 ixgbe_validate_link_ready(struct ixgbe_hw *hw)
605{
606 u32 timeout;
607 u16 an_reg;
608
609 if (hw->device_id != IXGBE_DEV_ID_82598AT2)
610 return IXGBE_SUCCESS;
611
612 for (timeout = 0;
613 timeout < IXGBE_VALIDATE_LINK_READY_TIMEOUT; timeout++) {
614 hw->phy.ops.read_reg(hw, IXGBE_MDIO_AUTO_NEG_STATUS,
615 IXGBE_MDIO_AUTO_NEG_DEV_TYPE, &an_reg);
616
617 if ((an_reg & IXGBE_MII_AUTONEG_COMPLETE) &&
618 (an_reg & IXGBE_MII_AUTONEG_LINK_UP))
619 break;
620
621 msec_delay(100);
622 }
623
624 if (timeout == IXGBE_VALIDATE_LINK_READY_TIMEOUT) {
625 DEBUGOUT("Link was indicated but link is down\n");
626 return IXGBE_ERR_LINK_SETUP;
627 }
628
629 return IXGBE_SUCCESS;
630}
631
632/**
633 * ixgbe_check_mac_link_82598 - Get link/speed status
634 * @hw: pointer to hardware structure
635 * @speed: pointer to link speed
636 * @link_up: TRUE is link is up, FALSE otherwise
637 * @link_up_wait_to_complete: bool used to wait for link up or not
638 *
639 * Reads the links register to determine if link is up and the current speed
640 **/
641static s32 ixgbe_check_mac_link_82598(struct ixgbe_hw *hw,
642 ixgbe_link_speed *speed, bool *link_up,
643 bool link_up_wait_to_complete)
644{
645 u32 links_reg;
646 u32 i;
647 u16 link_reg, adapt_comp_reg;
648
649 DEBUGFUNC("ixgbe_check_mac_link_82598");
650
651 /*
652 * SERDES PHY requires us to read link status from undocumented
653 * register 0xC79F. Bit 0 set indicates link is up/ready; clear
654 * indicates link down. OxC00C is read to check that the XAUI lanes
655 * are active. Bit 0 clear indicates active; set indicates inactive.
656 */
657 if (hw->phy.type == ixgbe_phy_nl) {
658 hw->phy.ops.read_reg(hw, 0xC79F, IXGBE_TWINAX_DEV, &link_reg);
659 hw->phy.ops.read_reg(hw, 0xC79F, IXGBE_TWINAX_DEV, &link_reg);
660 hw->phy.ops.read_reg(hw, 0xC00C, IXGBE_TWINAX_DEV,
661 &adapt_comp_reg);
662 if (link_up_wait_to_complete) {
663 for (i = 0; i < hw->mac.max_link_up_time; i++) {
664 if ((link_reg & 1) &&
665 ((adapt_comp_reg & 1) == 0)) {
666 *link_up = TRUE;
667 break;
668 } else {
669 *link_up = FALSE;
670 }
671 msec_delay(100);
672 hw->phy.ops.read_reg(hw, 0xC79F,
673 IXGBE_TWINAX_DEV,
674 &link_reg);
675 hw->phy.ops.read_reg(hw, 0xC00C,
676 IXGBE_TWINAX_DEV,
677 &adapt_comp_reg);
678 }
679 } else {
680 if ((link_reg & 1) && ((adapt_comp_reg & 1) == 0))
681 *link_up = TRUE;
682 else
683 *link_up = FALSE;
684 }
685
686 if (*link_up == FALSE)
687 goto out;
688 }
689
690 links_reg = IXGBE_READ_REG(hw, IXGBE_LINKS);
691 if (link_up_wait_to_complete) {
692 for (i = 0; i < hw->mac.max_link_up_time; i++) {
693 if (links_reg & IXGBE_LINKS_UP) {
694 *link_up = TRUE;
695 break;
696 } else {
697 *link_up = FALSE;
698 }
699 msec_delay(100);
700 links_reg = IXGBE_READ_REG(hw, IXGBE_LINKS);
701 }
702 } else {
703 if (links_reg & IXGBE_LINKS_UP)
704 *link_up = TRUE;
705 else
706 *link_up = FALSE;
707 }
708
709 if (links_reg & IXGBE_LINKS_SPEED)
710 *speed = IXGBE_LINK_SPEED_10GB_FULL;
711 else
712 *speed = IXGBE_LINK_SPEED_1GB_FULL;
713
714 if ((hw->device_id == IXGBE_DEV_ID_82598AT2) && (*link_up == TRUE) &&
715 (ixgbe_validate_link_ready(hw) != IXGBE_SUCCESS))
716 *link_up = FALSE;
717
718out:
719 return IXGBE_SUCCESS;
720}
721
722/**
723 * ixgbe_setup_mac_link_82598 - Set MAC link speed
724 * @hw: pointer to hardware structure
725 * @speed: new link speed
726 * @autoneg_wait_to_complete: TRUE when waiting for completion is needed
727 *
728 * Set the link speed in the AUTOC register and restarts link.
729 **/
730static s32 ixgbe_setup_mac_link_82598(struct ixgbe_hw *hw,
731 ixgbe_link_speed speed,
732 bool autoneg_wait_to_complete)
733{
734 bool autoneg = FALSE;
735 s32 status = IXGBE_SUCCESS;
736 ixgbe_link_speed link_capabilities = IXGBE_LINK_SPEED_UNKNOWN;
737 u32 curr_autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
738 u32 autoc = curr_autoc;
739 u32 link_mode = autoc & IXGBE_AUTOC_LMS_MASK;
740
741 DEBUGFUNC("ixgbe_setup_mac_link_82598");
742
743 /* Check to see if speed passed in is supported. */
744 ixgbe_get_link_capabilities(hw, &link_capabilities, &autoneg);
745 speed &= link_capabilities;
746
747 if (speed == IXGBE_LINK_SPEED_UNKNOWN)
748 status = IXGBE_ERR_LINK_SETUP;
749
750 /* Set KX4/KX support according to speed requested */
751 else if (link_mode == IXGBE_AUTOC_LMS_KX4_AN ||
752 link_mode == IXGBE_AUTOC_LMS_KX4_AN_1G_AN) {
753 autoc &= ~IXGBE_AUTOC_KX4_KX_SUPP_MASK;
754 if (speed & IXGBE_LINK_SPEED_10GB_FULL)
755 autoc |= IXGBE_AUTOC_KX4_SUPP;
756 if (speed & IXGBE_LINK_SPEED_1GB_FULL)
757 autoc |= IXGBE_AUTOC_KX_SUPP;
758 if (autoc != curr_autoc)
759 IXGBE_WRITE_REG(hw, IXGBE_AUTOC, autoc);
760 }
761
762 if (status == IXGBE_SUCCESS) {
763 /*
764 * Setup and restart the link based on the new values in
765 * ixgbe_hw This will write the AUTOC register based on the new
766 * stored values
767 */
768 status = ixgbe_start_mac_link_82598(hw,
769 autoneg_wait_to_complete);
770 }
771
772 return status;
773}
774
775
776/**
777 * ixgbe_setup_copper_link_82598 - Set the PHY autoneg advertised field
778 * @hw: pointer to hardware structure
779 * @speed: new link speed
780 * @autoneg_wait_to_complete: TRUE if waiting is needed to complete
781 *
782 * Sets the link speed in the AUTOC register in the MAC and restarts link.
783 **/
784static s32 ixgbe_setup_copper_link_82598(struct ixgbe_hw *hw,
785 ixgbe_link_speed speed,
786 bool autoneg_wait_to_complete)
787{
788 s32 status;
789
790 DEBUGFUNC("ixgbe_setup_copper_link_82598");
791
792 /* Setup the PHY according to input speed */
793 status = hw->phy.ops.setup_link_speed(hw, speed,
794 autoneg_wait_to_complete);
795 /* Set up MAC */
796 ixgbe_start_mac_link_82598(hw, autoneg_wait_to_complete);
797
798 return status;
799}
800
801/**
802 * ixgbe_reset_hw_82598 - Performs hardware reset
803 * @hw: pointer to hardware structure
804 *
805 * Resets the hardware by resetting the transmit and receive units, masks and
806 * clears all interrupts, performing a PHY reset, and performing a link (MAC)
807 * reset.
808 **/
809static s32 ixgbe_reset_hw_82598(struct ixgbe_hw *hw)
810{
811 s32 status = IXGBE_SUCCESS;
812 s32 phy_status = IXGBE_SUCCESS;
813 u32 ctrl;
814 u32 gheccr;
815 u32 i;
816 u32 autoc;
817 u8 analog_val;
818
819 DEBUGFUNC("ixgbe_reset_hw_82598");
820
821 /* Call adapter stop to disable tx/rx and clear interrupts */
822 status = hw->mac.ops.stop_adapter(hw);
823 if (status != IXGBE_SUCCESS)
824 goto reset_hw_out;
825
826 /*
827 * Power up the Atlas Tx lanes if they are currently powered down.
828 * Atlas Tx lanes are powered down for MAC loopback tests, but
829 * they are not automatically restored on reset.
830 */
831 hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_LPBK, &analog_val);
832 if (analog_val & IXGBE_ATLAS_PDN_TX_REG_EN) {
833 /* Enable Tx Atlas so packets can be transmitted again */
834 hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_LPBK,
835 &analog_val);
836 analog_val &= ~IXGBE_ATLAS_PDN_TX_REG_EN;
837 hw->mac.ops.write_analog_reg8(hw, IXGBE_ATLAS_PDN_LPBK,
838 analog_val);
839
840 hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_10G,
841 &analog_val);
842 analog_val &= ~IXGBE_ATLAS_PDN_TX_10G_QL_ALL;
843 hw->mac.ops.write_analog_reg8(hw, IXGBE_ATLAS_PDN_10G,
844 analog_val);
845
846 hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_1G,
847 &analog_val);
848 analog_val &= ~IXGBE_ATLAS_PDN_TX_1G_QL_ALL;
849 hw->mac.ops.write_analog_reg8(hw, IXGBE_ATLAS_PDN_1G,
850 analog_val);
851
852 hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_AN,
853 &analog_val);
854 analog_val &= ~IXGBE_ATLAS_PDN_TX_AN_QL_ALL;
855 hw->mac.ops.write_analog_reg8(hw, IXGBE_ATLAS_PDN_AN,
856 analog_val);
857 }
858
859 /* Reset PHY */
860 if (hw->phy.reset_disable == FALSE) {
861 /* PHY ops must be identified and initialized prior to reset */
862
863 /* Init PHY and function pointers, perform SFP setup */
864 phy_status = hw->phy.ops.init(hw);
865 if (phy_status == IXGBE_ERR_SFP_NOT_SUPPORTED)
866 goto reset_hw_out;
867 if (phy_status == IXGBE_ERR_SFP_NOT_PRESENT)
868 goto mac_reset_top;
869
870 hw->phy.ops.reset(hw);
871 }
872
873mac_reset_top:
874 /*
875 * Issue global reset to the MAC. This needs to be a SW reset.
876 * If link reset is used, it might reset the MAC when mng is using it
877 */
878 ctrl = IXGBE_READ_REG(hw, IXGBE_CTRL) | IXGBE_CTRL_RST;
879 IXGBE_WRITE_REG(hw, IXGBE_CTRL, ctrl);
880 IXGBE_WRITE_FLUSH(hw);
881
882 /* Poll for reset bit to self-clear indicating reset is complete */
883 for (i = 0; i < 10; i++) {
884 usec_delay(1);
885 ctrl = IXGBE_READ_REG(hw, IXGBE_CTRL);
886 if (!(ctrl & IXGBE_CTRL_RST))
887 break;
888 }
889 if (ctrl & IXGBE_CTRL_RST) {
890 status = IXGBE_ERR_RESET_FAILED;
891 DEBUGOUT("Reset polling failed to complete.\n");
892 }
893
894 msec_delay(50);
895
896 /*
897 * Double resets are required for recovery from certain error
898 * conditions. Between resets, it is necessary to stall to allow time
899 * for any pending HW events to complete.
900 */
901 if (hw->mac.flags & IXGBE_FLAGS_DOUBLE_RESET_REQUIRED) {
902 hw->mac.flags &= ~IXGBE_FLAGS_DOUBLE_RESET_REQUIRED;
903 goto mac_reset_top;
904 }
905
906 gheccr = IXGBE_READ_REG(hw, IXGBE_GHECCR);
907 gheccr &= ~((1 << 21) | (1 << 18) | (1 << 9) | (1 << 6));
908 IXGBE_WRITE_REG(hw, IXGBE_GHECCR, gheccr);
909
910 /*
911 * Store the original AUTOC value if it has not been
912 * stored off yet. Otherwise restore the stored original
913 * AUTOC value since the reset operation sets back to deaults.
914 */
915 autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
916 if (hw->mac.orig_link_settings_stored == FALSE) {
917 hw->mac.orig_autoc = autoc;
918 hw->mac.orig_link_settings_stored = TRUE;
919 } else if (autoc != hw->mac.orig_autoc) {
920 IXGBE_WRITE_REG(hw, IXGBE_AUTOC, hw->mac.orig_autoc);
921 }
922
923 /* Store the permanent mac address */
924 hw->mac.ops.get_mac_addr(hw, hw->mac.perm_addr);
925
926 /*
927 * Store MAC address from RAR0, clear receive address registers, and
928 * clear the multicast table
929 */
930 hw->mac.ops.init_rx_addrs(hw);
931
932reset_hw_out:
933 if (phy_status != IXGBE_SUCCESS)
934 status = phy_status;
935
936 return status;
937}
938
939/**
940 * ixgbe_set_vmdq_82598 - Associate a VMDq set index with a rx address
941 * @hw: pointer to hardware struct
942 * @rar: receive address register index to associate with a VMDq index
943 * @vmdq: VMDq set index
944 **/
945s32 ixgbe_set_vmdq_82598(struct ixgbe_hw *hw, u32 rar, u32 vmdq)
946{
947 u32 rar_high;
948 u32 rar_entries = hw->mac.num_rar_entries;
949
950 DEBUGFUNC("ixgbe_set_vmdq_82598");
951
952 /* Make sure we are using a valid rar index range */
953 if (rar >= rar_entries) {
954 DEBUGOUT1("RAR index %d is out of range.\n", rar);
955 return IXGBE_ERR_INVALID_ARGUMENT;
956 }
957
958 rar_high = IXGBE_READ_REG(hw, IXGBE_RAH(rar));
959 rar_high &= ~IXGBE_RAH_VIND_MASK;
960 rar_high |= ((vmdq << IXGBE_RAH_VIND_SHIFT) & IXGBE_RAH_VIND_MASK);
961 IXGBE_WRITE_REG(hw, IXGBE_RAH(rar), rar_high);
962 return IXGBE_SUCCESS;
963}
964
965/**
966 * ixgbe_clear_vmdq_82598 - Disassociate a VMDq set index from an rx address
967 * @hw: pointer to hardware struct
968 * @rar: receive address register index to associate with a VMDq index
969 * @vmdq: VMDq clear index (not used in 82598, but elsewhere)
970 **/
971static s32 ixgbe_clear_vmdq_82598(struct ixgbe_hw *hw, u32 rar, u32 vmdq)
972{
973 u32 rar_high;
974 u32 rar_entries = hw->mac.num_rar_entries;
975
976 UNREFERENCED_1PARAMETER(vmdq);
977
978 /* Make sure we are using a valid rar index range */
979 if (rar >= rar_entries) {
980 DEBUGOUT1("RAR index %d is out of range.\n", rar);
981 return IXGBE_ERR_INVALID_ARGUMENT;
982 }
983
984 rar_high = IXGBE_READ_REG(hw, IXGBE_RAH(rar));
985 if (rar_high & IXGBE_RAH_VIND_MASK) {
986 rar_high &= ~IXGBE_RAH_VIND_MASK;
987 IXGBE_WRITE_REG(hw, IXGBE_RAH(rar), rar_high);
988 }
989
990 return IXGBE_SUCCESS;
991}
992
993/**
994 * ixgbe_set_vfta_82598 - Set VLAN filter table
995 * @hw: pointer to hardware structure
996 * @vlan: VLAN id to write to VLAN filter
997 * @vind: VMDq output index that maps queue to VLAN id in VFTA
998 * @vlan_on: boolean flag to turn on/off VLAN in VFTA
999 * @vlvf_bypass: boolean flag - unused
999 *
1000 * Turn on/off specified VLAN in the VLAN filter table.
1001 **/
1002s32 ixgbe_set_vfta_82598(struct ixgbe_hw *hw, u32 vlan, u32 vind,
1000 *
1001 * Turn on/off specified VLAN in the VLAN filter table.
1002 **/
1003s32 ixgbe_set_vfta_82598(struct ixgbe_hw *hw, u32 vlan, u32 vind,
1003 bool vlan_on)
1004 bool vlan_on, bool vlvf_bypass)
1004{
1005 u32 regindex;
1006 u32 bitindex;
1007 u32 bits;
1008 u32 vftabyte;
1009
1005{
1006 u32 regindex;
1007 u32 bitindex;
1008 u32 bits;
1009 u32 vftabyte;
1010
1011 UNREFERENCED_1PARAMETER(vlvf_bypass);
1012
1010 DEBUGFUNC("ixgbe_set_vfta_82598");
1011
1012 if (vlan > 4095)
1013 return IXGBE_ERR_PARAM;
1014
1015 /* Determine 32-bit word position in array */
1016 regindex = (vlan >> 5) & 0x7F; /* upper seven bits */
1017
1018 /* Determine the location of the (VMD) queue index */
1019 vftabyte = ((vlan >> 3) & 0x03); /* bits (4:3) indicating byte array */
1020 bitindex = (vlan & 0x7) << 2; /* lower 3 bits indicate nibble */
1021
1022 /* Set the nibble for VMD queue index */
1023 bits = IXGBE_READ_REG(hw, IXGBE_VFTAVIND(vftabyte, regindex));
1024 bits &= (~(0x0F << bitindex));
1025 bits |= (vind << bitindex);
1026 IXGBE_WRITE_REG(hw, IXGBE_VFTAVIND(vftabyte, regindex), bits);
1027
1028 /* Determine the location of the bit for this VLAN id */
1029 bitindex = vlan & 0x1F; /* lower five bits */
1030
1031 bits = IXGBE_READ_REG(hw, IXGBE_VFTA(regindex));
1032 if (vlan_on)
1033 /* Turn on this VLAN id */
1034 bits |= (1 << bitindex);
1035 else
1036 /* Turn off this VLAN id */
1037 bits &= ~(1 << bitindex);
1038 IXGBE_WRITE_REG(hw, IXGBE_VFTA(regindex), bits);
1039
1040 return IXGBE_SUCCESS;
1041}
1042
1043/**
1044 * ixgbe_clear_vfta_82598 - Clear VLAN filter table
1045 * @hw: pointer to hardware structure
1046 *
1047 * Clears the VLAN filer table, and the VMDq index associated with the filter
1048 **/
1049static s32 ixgbe_clear_vfta_82598(struct ixgbe_hw *hw)
1050{
1051 u32 offset;
1052 u32 vlanbyte;
1053
1054 DEBUGFUNC("ixgbe_clear_vfta_82598");
1055
1056 for (offset = 0; offset < hw->mac.vft_size; offset++)
1057 IXGBE_WRITE_REG(hw, IXGBE_VFTA(offset), 0);
1058
1059 for (vlanbyte = 0; vlanbyte < 4; vlanbyte++)
1060 for (offset = 0; offset < hw->mac.vft_size; offset++)
1061 IXGBE_WRITE_REG(hw, IXGBE_VFTAVIND(vlanbyte, offset),
1062 0);
1063
1064 return IXGBE_SUCCESS;
1065}
1066
1067/**
1068 * ixgbe_read_analog_reg8_82598 - Reads 8 bit Atlas analog register
1069 * @hw: pointer to hardware structure
1070 * @reg: analog register to read
1071 * @val: read value
1072 *
1073 * Performs read operation to Atlas analog register specified.
1074 **/
1075s32 ixgbe_read_analog_reg8_82598(struct ixgbe_hw *hw, u32 reg, u8 *val)
1076{
1077 u32 atlas_ctl;
1078
1079 DEBUGFUNC("ixgbe_read_analog_reg8_82598");
1080
1081 IXGBE_WRITE_REG(hw, IXGBE_ATLASCTL,
1082 IXGBE_ATLASCTL_WRITE_CMD | (reg << 8));
1083 IXGBE_WRITE_FLUSH(hw);
1084 usec_delay(10);
1085 atlas_ctl = IXGBE_READ_REG(hw, IXGBE_ATLASCTL);
1086 *val = (u8)atlas_ctl;
1087
1088 return IXGBE_SUCCESS;
1089}
1090
1091/**
1092 * ixgbe_write_analog_reg8_82598 - Writes 8 bit Atlas analog register
1093 * @hw: pointer to hardware structure
1094 * @reg: atlas register to write
1095 * @val: value to write
1096 *
1097 * Performs write operation to Atlas analog register specified.
1098 **/
1099s32 ixgbe_write_analog_reg8_82598(struct ixgbe_hw *hw, u32 reg, u8 val)
1100{
1101 u32 atlas_ctl;
1102
1103 DEBUGFUNC("ixgbe_write_analog_reg8_82598");
1104
1105 atlas_ctl = (reg << 8) | val;
1106 IXGBE_WRITE_REG(hw, IXGBE_ATLASCTL, atlas_ctl);
1107 IXGBE_WRITE_FLUSH(hw);
1108 usec_delay(10);
1109
1110 return IXGBE_SUCCESS;
1111}
1112
1113/**
1114 * ixgbe_read_i2c_phy_82598 - Reads 8 bit word over I2C interface.
1115 * @hw: pointer to hardware structure
1116 * @dev_addr: address to read from
1117 * @byte_offset: byte offset to read from dev_addr
1118 * @eeprom_data: value read
1119 *
1120 * Performs 8 byte read operation to SFP module's EEPROM over I2C interface.
1121 **/
1122static s32 ixgbe_read_i2c_phy_82598(struct ixgbe_hw *hw, u8 dev_addr,
1123 u8 byte_offset, u8 *eeprom_data)
1124{
1125 s32 status = IXGBE_SUCCESS;
1126 u16 sfp_addr = 0;
1127 u16 sfp_data = 0;
1128 u16 sfp_stat = 0;
1129 u16 gssr;
1130 u32 i;
1131
1132 DEBUGFUNC("ixgbe_read_i2c_phy_82598");
1133
1134 if (IXGBE_READ_REG(hw, IXGBE_STATUS) & IXGBE_STATUS_LAN_ID_1)
1135 gssr = IXGBE_GSSR_PHY1_SM;
1136 else
1137 gssr = IXGBE_GSSR_PHY0_SM;
1138
1139 if (hw->mac.ops.acquire_swfw_sync(hw, gssr) != IXGBE_SUCCESS)
1140 return IXGBE_ERR_SWFW_SYNC;
1141
1142 if (hw->phy.type == ixgbe_phy_nl) {
1143 /*
1144 * NetLogic phy SDA/SCL registers are at addresses 0xC30A to
1145 * 0xC30D. These registers are used to talk to the SFP+
1146 * module's EEPROM through the SDA/SCL (I2C) interface.
1147 */
1148 sfp_addr = (dev_addr << 8) + byte_offset;
1149 sfp_addr = (sfp_addr | IXGBE_I2C_EEPROM_READ_MASK);
1150 hw->phy.ops.write_reg_mdi(hw,
1151 IXGBE_MDIO_PMA_PMD_SDA_SCL_ADDR,
1152 IXGBE_MDIO_PMA_PMD_DEV_TYPE,
1153 sfp_addr);
1154
1155 /* Poll status */
1156 for (i = 0; i < 100; i++) {
1157 hw->phy.ops.read_reg_mdi(hw,
1158 IXGBE_MDIO_PMA_PMD_SDA_SCL_STAT,
1159 IXGBE_MDIO_PMA_PMD_DEV_TYPE,
1160 &sfp_stat);
1161 sfp_stat = sfp_stat & IXGBE_I2C_EEPROM_STATUS_MASK;
1162 if (sfp_stat != IXGBE_I2C_EEPROM_STATUS_IN_PROGRESS)
1163 break;
1164 msec_delay(10);
1165 }
1166
1167 if (sfp_stat != IXGBE_I2C_EEPROM_STATUS_PASS) {
1168 DEBUGOUT("EEPROM read did not pass.\n");
1169 status = IXGBE_ERR_SFP_NOT_PRESENT;
1170 goto out;
1171 }
1172
1173 /* Read data */
1174 hw->phy.ops.read_reg_mdi(hw, IXGBE_MDIO_PMA_PMD_SDA_SCL_DATA,
1175 IXGBE_MDIO_PMA_PMD_DEV_TYPE, &sfp_data);
1176
1177 *eeprom_data = (u8)(sfp_data >> 8);
1178 } else {
1179 status = IXGBE_ERR_PHY;
1180 }
1181
1182out:
1183 hw->mac.ops.release_swfw_sync(hw, gssr);
1184 return status;
1185}
1186
1187/**
1188 * ixgbe_read_i2c_eeprom_82598 - Reads 8 bit word over I2C interface.
1189 * @hw: pointer to hardware structure
1190 * @byte_offset: EEPROM byte offset to read
1191 * @eeprom_data: value read
1192 *
1193 * Performs 8 byte read operation to SFP module's EEPROM over I2C interface.
1194 **/
1195s32 ixgbe_read_i2c_eeprom_82598(struct ixgbe_hw *hw, u8 byte_offset,
1196 u8 *eeprom_data)
1197{
1198 return ixgbe_read_i2c_phy_82598(hw, IXGBE_I2C_EEPROM_DEV_ADDR,
1199 byte_offset, eeprom_data);
1200}
1201
1202/**
1203 * ixgbe_read_i2c_sff8472_82598 - Reads 8 bit word over I2C interface.
1204 * @hw: pointer to hardware structure
1205 * @byte_offset: byte offset at address 0xA2
1206 * @eeprom_data: value read
1207 *
1208 * Performs 8 byte read operation to SFP module's SFF-8472 data over I2C
1209 **/
1210static s32 ixgbe_read_i2c_sff8472_82598(struct ixgbe_hw *hw, u8 byte_offset,
1211 u8 *sff8472_data)
1212{
1213 return ixgbe_read_i2c_phy_82598(hw, IXGBE_I2C_EEPROM_DEV_ADDR2,
1214 byte_offset, sff8472_data);
1215}
1216
1217/**
1218 * ixgbe_get_supported_physical_layer_82598 - Returns physical layer type
1219 * @hw: pointer to hardware structure
1220 *
1221 * Determines physical layer capabilities of the current configuration.
1222 **/
1013 DEBUGFUNC("ixgbe_set_vfta_82598");
1014
1015 if (vlan > 4095)
1016 return IXGBE_ERR_PARAM;
1017
1018 /* Determine 32-bit word position in array */
1019 regindex = (vlan >> 5) & 0x7F; /* upper seven bits */
1020
1021 /* Determine the location of the (VMD) queue index */
1022 vftabyte = ((vlan >> 3) & 0x03); /* bits (4:3) indicating byte array */
1023 bitindex = (vlan & 0x7) << 2; /* lower 3 bits indicate nibble */
1024
1025 /* Set the nibble for VMD queue index */
1026 bits = IXGBE_READ_REG(hw, IXGBE_VFTAVIND(vftabyte, regindex));
1027 bits &= (~(0x0F << bitindex));
1028 bits |= (vind << bitindex);
1029 IXGBE_WRITE_REG(hw, IXGBE_VFTAVIND(vftabyte, regindex), bits);
1030
1031 /* Determine the location of the bit for this VLAN id */
1032 bitindex = vlan & 0x1F; /* lower five bits */
1033
1034 bits = IXGBE_READ_REG(hw, IXGBE_VFTA(regindex));
1035 if (vlan_on)
1036 /* Turn on this VLAN id */
1037 bits |= (1 << bitindex);
1038 else
1039 /* Turn off this VLAN id */
1040 bits &= ~(1 << bitindex);
1041 IXGBE_WRITE_REG(hw, IXGBE_VFTA(regindex), bits);
1042
1043 return IXGBE_SUCCESS;
1044}
1045
1046/**
1047 * ixgbe_clear_vfta_82598 - Clear VLAN filter table
1048 * @hw: pointer to hardware structure
1049 *
1050 * Clears the VLAN filer table, and the VMDq index associated with the filter
1051 **/
1052static s32 ixgbe_clear_vfta_82598(struct ixgbe_hw *hw)
1053{
1054 u32 offset;
1055 u32 vlanbyte;
1056
1057 DEBUGFUNC("ixgbe_clear_vfta_82598");
1058
1059 for (offset = 0; offset < hw->mac.vft_size; offset++)
1060 IXGBE_WRITE_REG(hw, IXGBE_VFTA(offset), 0);
1061
1062 for (vlanbyte = 0; vlanbyte < 4; vlanbyte++)
1063 for (offset = 0; offset < hw->mac.vft_size; offset++)
1064 IXGBE_WRITE_REG(hw, IXGBE_VFTAVIND(vlanbyte, offset),
1065 0);
1066
1067 return IXGBE_SUCCESS;
1068}
1069
1070/**
1071 * ixgbe_read_analog_reg8_82598 - Reads 8 bit Atlas analog register
1072 * @hw: pointer to hardware structure
1073 * @reg: analog register to read
1074 * @val: read value
1075 *
1076 * Performs read operation to Atlas analog register specified.
1077 **/
1078s32 ixgbe_read_analog_reg8_82598(struct ixgbe_hw *hw, u32 reg, u8 *val)
1079{
1080 u32 atlas_ctl;
1081
1082 DEBUGFUNC("ixgbe_read_analog_reg8_82598");
1083
1084 IXGBE_WRITE_REG(hw, IXGBE_ATLASCTL,
1085 IXGBE_ATLASCTL_WRITE_CMD | (reg << 8));
1086 IXGBE_WRITE_FLUSH(hw);
1087 usec_delay(10);
1088 atlas_ctl = IXGBE_READ_REG(hw, IXGBE_ATLASCTL);
1089 *val = (u8)atlas_ctl;
1090
1091 return IXGBE_SUCCESS;
1092}
1093
1094/**
1095 * ixgbe_write_analog_reg8_82598 - Writes 8 bit Atlas analog register
1096 * @hw: pointer to hardware structure
1097 * @reg: atlas register to write
1098 * @val: value to write
1099 *
1100 * Performs write operation to Atlas analog register specified.
1101 **/
1102s32 ixgbe_write_analog_reg8_82598(struct ixgbe_hw *hw, u32 reg, u8 val)
1103{
1104 u32 atlas_ctl;
1105
1106 DEBUGFUNC("ixgbe_write_analog_reg8_82598");
1107
1108 atlas_ctl = (reg << 8) | val;
1109 IXGBE_WRITE_REG(hw, IXGBE_ATLASCTL, atlas_ctl);
1110 IXGBE_WRITE_FLUSH(hw);
1111 usec_delay(10);
1112
1113 return IXGBE_SUCCESS;
1114}
1115
1116/**
1117 * ixgbe_read_i2c_phy_82598 - Reads 8 bit word over I2C interface.
1118 * @hw: pointer to hardware structure
1119 * @dev_addr: address to read from
1120 * @byte_offset: byte offset to read from dev_addr
1121 * @eeprom_data: value read
1122 *
1123 * Performs 8 byte read operation to SFP module's EEPROM over I2C interface.
1124 **/
1125static s32 ixgbe_read_i2c_phy_82598(struct ixgbe_hw *hw, u8 dev_addr,
1126 u8 byte_offset, u8 *eeprom_data)
1127{
1128 s32 status = IXGBE_SUCCESS;
1129 u16 sfp_addr = 0;
1130 u16 sfp_data = 0;
1131 u16 sfp_stat = 0;
1132 u16 gssr;
1133 u32 i;
1134
1135 DEBUGFUNC("ixgbe_read_i2c_phy_82598");
1136
1137 if (IXGBE_READ_REG(hw, IXGBE_STATUS) & IXGBE_STATUS_LAN_ID_1)
1138 gssr = IXGBE_GSSR_PHY1_SM;
1139 else
1140 gssr = IXGBE_GSSR_PHY0_SM;
1141
1142 if (hw->mac.ops.acquire_swfw_sync(hw, gssr) != IXGBE_SUCCESS)
1143 return IXGBE_ERR_SWFW_SYNC;
1144
1145 if (hw->phy.type == ixgbe_phy_nl) {
1146 /*
1147 * NetLogic phy SDA/SCL registers are at addresses 0xC30A to
1148 * 0xC30D. These registers are used to talk to the SFP+
1149 * module's EEPROM through the SDA/SCL (I2C) interface.
1150 */
1151 sfp_addr = (dev_addr << 8) + byte_offset;
1152 sfp_addr = (sfp_addr | IXGBE_I2C_EEPROM_READ_MASK);
1153 hw->phy.ops.write_reg_mdi(hw,
1154 IXGBE_MDIO_PMA_PMD_SDA_SCL_ADDR,
1155 IXGBE_MDIO_PMA_PMD_DEV_TYPE,
1156 sfp_addr);
1157
1158 /* Poll status */
1159 for (i = 0; i < 100; i++) {
1160 hw->phy.ops.read_reg_mdi(hw,
1161 IXGBE_MDIO_PMA_PMD_SDA_SCL_STAT,
1162 IXGBE_MDIO_PMA_PMD_DEV_TYPE,
1163 &sfp_stat);
1164 sfp_stat = sfp_stat & IXGBE_I2C_EEPROM_STATUS_MASK;
1165 if (sfp_stat != IXGBE_I2C_EEPROM_STATUS_IN_PROGRESS)
1166 break;
1167 msec_delay(10);
1168 }
1169
1170 if (sfp_stat != IXGBE_I2C_EEPROM_STATUS_PASS) {
1171 DEBUGOUT("EEPROM read did not pass.\n");
1172 status = IXGBE_ERR_SFP_NOT_PRESENT;
1173 goto out;
1174 }
1175
1176 /* Read data */
1177 hw->phy.ops.read_reg_mdi(hw, IXGBE_MDIO_PMA_PMD_SDA_SCL_DATA,
1178 IXGBE_MDIO_PMA_PMD_DEV_TYPE, &sfp_data);
1179
1180 *eeprom_data = (u8)(sfp_data >> 8);
1181 } else {
1182 status = IXGBE_ERR_PHY;
1183 }
1184
1185out:
1186 hw->mac.ops.release_swfw_sync(hw, gssr);
1187 return status;
1188}
1189
1190/**
1191 * ixgbe_read_i2c_eeprom_82598 - Reads 8 bit word over I2C interface.
1192 * @hw: pointer to hardware structure
1193 * @byte_offset: EEPROM byte offset to read
1194 * @eeprom_data: value read
1195 *
1196 * Performs 8 byte read operation to SFP module's EEPROM over I2C interface.
1197 **/
1198s32 ixgbe_read_i2c_eeprom_82598(struct ixgbe_hw *hw, u8 byte_offset,
1199 u8 *eeprom_data)
1200{
1201 return ixgbe_read_i2c_phy_82598(hw, IXGBE_I2C_EEPROM_DEV_ADDR,
1202 byte_offset, eeprom_data);
1203}
1204
1205/**
1206 * ixgbe_read_i2c_sff8472_82598 - Reads 8 bit word over I2C interface.
1207 * @hw: pointer to hardware structure
1208 * @byte_offset: byte offset at address 0xA2
1209 * @eeprom_data: value read
1210 *
1211 * Performs 8 byte read operation to SFP module's SFF-8472 data over I2C
1212 **/
1213static s32 ixgbe_read_i2c_sff8472_82598(struct ixgbe_hw *hw, u8 byte_offset,
1214 u8 *sff8472_data)
1215{
1216 return ixgbe_read_i2c_phy_82598(hw, IXGBE_I2C_EEPROM_DEV_ADDR2,
1217 byte_offset, sff8472_data);
1218}
1219
1220/**
1221 * ixgbe_get_supported_physical_layer_82598 - Returns physical layer type
1222 * @hw: pointer to hardware structure
1223 *
1224 * Determines physical layer capabilities of the current configuration.
1225 **/
1223u32 ixgbe_get_supported_physical_layer_82598(struct ixgbe_hw *hw)
1226u64 ixgbe_get_supported_physical_layer_82598(struct ixgbe_hw *hw)
1224{
1227{
1225 u32 physical_layer = IXGBE_PHYSICAL_LAYER_UNKNOWN;
1228 u64 physical_layer = IXGBE_PHYSICAL_LAYER_UNKNOWN;
1226 u32 autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
1227 u32 pma_pmd_10g = autoc & IXGBE_AUTOC_10G_PMA_PMD_MASK;
1228 u32 pma_pmd_1g = autoc & IXGBE_AUTOC_1G_PMA_PMD_MASK;
1229 u16 ext_ability = 0;
1230
1231 DEBUGFUNC("ixgbe_get_supported_physical_layer_82598");
1232
1233 hw->phy.ops.identify(hw);
1234
1235 /* Copper PHY must be checked before AUTOC LMS to determine correct
1236 * physical layer because 10GBase-T PHYs use LMS = KX4/KX */
1237 switch (hw->phy.type) {
1238 case ixgbe_phy_tn:
1239 case ixgbe_phy_cu_unknown:
1240 hw->phy.ops.read_reg(hw, IXGBE_MDIO_PHY_EXT_ABILITY,
1241 IXGBE_MDIO_PMA_PMD_DEV_TYPE, &ext_ability);
1242 if (ext_ability & IXGBE_MDIO_PHY_10GBASET_ABILITY)
1243 physical_layer |= IXGBE_PHYSICAL_LAYER_10GBASE_T;
1244 if (ext_ability & IXGBE_MDIO_PHY_1000BASET_ABILITY)
1245 physical_layer |= IXGBE_PHYSICAL_LAYER_1000BASE_T;
1246 if (ext_ability & IXGBE_MDIO_PHY_100BASETX_ABILITY)
1247 physical_layer |= IXGBE_PHYSICAL_LAYER_100BASE_TX;
1248 goto out;
1249 default:
1250 break;
1251 }
1252
1253 switch (autoc & IXGBE_AUTOC_LMS_MASK) {
1254 case IXGBE_AUTOC_LMS_1G_AN:
1255 case IXGBE_AUTOC_LMS_1G_LINK_NO_AN:
1256 if (pma_pmd_1g == IXGBE_AUTOC_1G_KX)
1257 physical_layer = IXGBE_PHYSICAL_LAYER_1000BASE_KX;
1258 else
1259 physical_layer = IXGBE_PHYSICAL_LAYER_1000BASE_BX;
1260 break;
1261 case IXGBE_AUTOC_LMS_10G_LINK_NO_AN:
1262 if (pma_pmd_10g == IXGBE_AUTOC_10G_CX4)
1263 physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_CX4;
1264 else if (pma_pmd_10g == IXGBE_AUTOC_10G_KX4)
1265 physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_KX4;
1266 else /* XAUI */
1267 physical_layer = IXGBE_PHYSICAL_LAYER_UNKNOWN;
1268 break;
1269 case IXGBE_AUTOC_LMS_KX4_AN:
1270 case IXGBE_AUTOC_LMS_KX4_AN_1G_AN:
1271 if (autoc & IXGBE_AUTOC_KX_SUPP)
1272 physical_layer |= IXGBE_PHYSICAL_LAYER_1000BASE_KX;
1273 if (autoc & IXGBE_AUTOC_KX4_SUPP)
1274 physical_layer |= IXGBE_PHYSICAL_LAYER_10GBASE_KX4;
1275 break;
1276 default:
1277 break;
1278 }
1279
1280 if (hw->phy.type == ixgbe_phy_nl) {
1281 hw->phy.ops.identify_sfp(hw);
1282
1283 switch (hw->phy.sfp_type) {
1284 case ixgbe_sfp_type_da_cu:
1285 physical_layer = IXGBE_PHYSICAL_LAYER_SFP_PLUS_CU;
1286 break;
1287 case ixgbe_sfp_type_sr:
1288 physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_SR;
1289 break;
1290 case ixgbe_sfp_type_lr:
1291 physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_LR;
1292 break;
1293 default:
1294 physical_layer = IXGBE_PHYSICAL_LAYER_UNKNOWN;
1295 break;
1296 }
1297 }
1298
1299 switch (hw->device_id) {
1300 case IXGBE_DEV_ID_82598_DA_DUAL_PORT:
1301 physical_layer = IXGBE_PHYSICAL_LAYER_SFP_PLUS_CU;
1302 break;
1303 case IXGBE_DEV_ID_82598AF_DUAL_PORT:
1304 case IXGBE_DEV_ID_82598AF_SINGLE_PORT:
1305 case IXGBE_DEV_ID_82598_SR_DUAL_PORT_EM:
1306 physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_SR;
1307 break;
1308 case IXGBE_DEV_ID_82598EB_XF_LR:
1309 physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_LR;
1310 break;
1311 default:
1312 break;
1313 }
1314
1315out:
1316 return physical_layer;
1317}
1318
1319/**
1320 * ixgbe_set_lan_id_multi_port_pcie_82598 - Set LAN id for PCIe multiple
1321 * port devices.
1322 * @hw: pointer to the HW structure
1323 *
1324 * Calls common function and corrects issue with some single port devices
1325 * that enable LAN1 but not LAN0.
1326 **/
1327void ixgbe_set_lan_id_multi_port_pcie_82598(struct ixgbe_hw *hw)
1328{
1329 struct ixgbe_bus_info *bus = &hw->bus;
1330 u16 pci_gen = 0;
1331 u16 pci_ctrl2 = 0;
1332
1333 DEBUGFUNC("ixgbe_set_lan_id_multi_port_pcie_82598");
1334
1335 ixgbe_set_lan_id_multi_port_pcie(hw);
1336
1337 /* check if LAN0 is disabled */
1338 hw->eeprom.ops.read(hw, IXGBE_PCIE_GENERAL_PTR, &pci_gen);
1339 if ((pci_gen != 0) && (pci_gen != 0xFFFF)) {
1340
1341 hw->eeprom.ops.read(hw, pci_gen + IXGBE_PCIE_CTRL2, &pci_ctrl2);
1342
1343 /* if LAN0 is completely disabled force function to 0 */
1344 if ((pci_ctrl2 & IXGBE_PCIE_CTRL2_LAN_DISABLE) &&
1345 !(pci_ctrl2 & IXGBE_PCIE_CTRL2_DISABLE_SELECT) &&
1346 !(pci_ctrl2 & IXGBE_PCIE_CTRL2_DUMMY_ENABLE)) {
1347
1348 bus->func = 0;
1349 }
1350 }
1351}
1352
1353/**
1354 * ixgbe_enable_relaxed_ordering_82598 - enable relaxed ordering
1355 * @hw: pointer to hardware structure
1356 *
1357 **/
1358void ixgbe_enable_relaxed_ordering_82598(struct ixgbe_hw *hw)
1359{
1360 u32 regval;
1361 u32 i;
1362
1363 DEBUGFUNC("ixgbe_enable_relaxed_ordering_82598");
1364
1365 /* Enable relaxed ordering */
1366 for (i = 0; ((i < hw->mac.max_tx_queues) &&
1367 (i < IXGBE_DCA_MAX_QUEUES_82598)); i++) {
1368 regval = IXGBE_READ_REG(hw, IXGBE_DCA_TXCTRL(i));
1369 regval |= IXGBE_DCA_TXCTRL_DESC_WRO_EN;
1370 IXGBE_WRITE_REG(hw, IXGBE_DCA_TXCTRL(i), regval);
1371 }
1372
1373 for (i = 0; ((i < hw->mac.max_rx_queues) &&
1374 (i < IXGBE_DCA_MAX_QUEUES_82598)); i++) {
1375 regval = IXGBE_READ_REG(hw, IXGBE_DCA_RXCTRL(i));
1376 regval |= IXGBE_DCA_RXCTRL_DATA_WRO_EN |
1377 IXGBE_DCA_RXCTRL_HEAD_WRO_EN;
1378 IXGBE_WRITE_REG(hw, IXGBE_DCA_RXCTRL(i), regval);
1379 }
1380
1381}
1382
1383/**
1384 * ixgbe_set_rxpba_82598 - Initialize RX packet buffer
1385 * @hw: pointer to hardware structure
1386 * @num_pb: number of packet buffers to allocate
1387 * @headroom: reserve n KB of headroom
1388 * @strategy: packet buffer allocation strategy
1389 **/
1390static void ixgbe_set_rxpba_82598(struct ixgbe_hw *hw, int num_pb,
1391 u32 headroom, int strategy)
1392{
1393 u32 rxpktsize = IXGBE_RXPBSIZE_64KB;
1394 u8 i = 0;
1395 UNREFERENCED_1PARAMETER(headroom);
1396
1397 if (!num_pb)
1398 return;
1399
1400 /* Setup Rx packet buffer sizes */
1401 switch (strategy) {
1402 case PBA_STRATEGY_WEIGHTED:
1403 /* Setup the first four at 80KB */
1404 rxpktsize = IXGBE_RXPBSIZE_80KB;
1405 for (; i < 4; i++)
1406 IXGBE_WRITE_REG(hw, IXGBE_RXPBSIZE(i), rxpktsize);
1407 /* Setup the last four at 48KB...don't re-init i */
1408 rxpktsize = IXGBE_RXPBSIZE_48KB;
1409 /* Fall Through */
1410 case PBA_STRATEGY_EQUAL:
1411 default:
1412 /* Divide the remaining Rx packet buffer evenly among the TCs */
1413 for (; i < IXGBE_MAX_PACKET_BUFFERS; i++)
1414 IXGBE_WRITE_REG(hw, IXGBE_RXPBSIZE(i), rxpktsize);
1415 break;
1416 }
1417
1418 /* Setup Tx packet buffer sizes */
1419 for (i = 0; i < IXGBE_MAX_PACKET_BUFFERS; i++)
1420 IXGBE_WRITE_REG(hw, IXGBE_TXPBSIZE(i), IXGBE_TXPBSIZE_40KB);
1421}
1422
1423/**
1424 * ixgbe_enable_rx_dma_82598 - Enable the Rx DMA unit
1425 * @hw: pointer to hardware structure
1426 * @regval: register value to write to RXCTRL
1427 *
1428 * Enables the Rx DMA unit
1429 **/
1430s32 ixgbe_enable_rx_dma_82598(struct ixgbe_hw *hw, u32 regval)
1431{
1432 DEBUGFUNC("ixgbe_enable_rx_dma_82598");
1433
1434 IXGBE_WRITE_REG(hw, IXGBE_RXCTRL, regval);
1435
1436 return IXGBE_SUCCESS;
1437}
1229 u32 autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
1230 u32 pma_pmd_10g = autoc & IXGBE_AUTOC_10G_PMA_PMD_MASK;
1231 u32 pma_pmd_1g = autoc & IXGBE_AUTOC_1G_PMA_PMD_MASK;
1232 u16 ext_ability = 0;
1233
1234 DEBUGFUNC("ixgbe_get_supported_physical_layer_82598");
1235
1236 hw->phy.ops.identify(hw);
1237
1238 /* Copper PHY must be checked before AUTOC LMS to determine correct
1239 * physical layer because 10GBase-T PHYs use LMS = KX4/KX */
1240 switch (hw->phy.type) {
1241 case ixgbe_phy_tn:
1242 case ixgbe_phy_cu_unknown:
1243 hw->phy.ops.read_reg(hw, IXGBE_MDIO_PHY_EXT_ABILITY,
1244 IXGBE_MDIO_PMA_PMD_DEV_TYPE, &ext_ability);
1245 if (ext_ability & IXGBE_MDIO_PHY_10GBASET_ABILITY)
1246 physical_layer |= IXGBE_PHYSICAL_LAYER_10GBASE_T;
1247 if (ext_ability & IXGBE_MDIO_PHY_1000BASET_ABILITY)
1248 physical_layer |= IXGBE_PHYSICAL_LAYER_1000BASE_T;
1249 if (ext_ability & IXGBE_MDIO_PHY_100BASETX_ABILITY)
1250 physical_layer |= IXGBE_PHYSICAL_LAYER_100BASE_TX;
1251 goto out;
1252 default:
1253 break;
1254 }
1255
1256 switch (autoc & IXGBE_AUTOC_LMS_MASK) {
1257 case IXGBE_AUTOC_LMS_1G_AN:
1258 case IXGBE_AUTOC_LMS_1G_LINK_NO_AN:
1259 if (pma_pmd_1g == IXGBE_AUTOC_1G_KX)
1260 physical_layer = IXGBE_PHYSICAL_LAYER_1000BASE_KX;
1261 else
1262 physical_layer = IXGBE_PHYSICAL_LAYER_1000BASE_BX;
1263 break;
1264 case IXGBE_AUTOC_LMS_10G_LINK_NO_AN:
1265 if (pma_pmd_10g == IXGBE_AUTOC_10G_CX4)
1266 physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_CX4;
1267 else if (pma_pmd_10g == IXGBE_AUTOC_10G_KX4)
1268 physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_KX4;
1269 else /* XAUI */
1270 physical_layer = IXGBE_PHYSICAL_LAYER_UNKNOWN;
1271 break;
1272 case IXGBE_AUTOC_LMS_KX4_AN:
1273 case IXGBE_AUTOC_LMS_KX4_AN_1G_AN:
1274 if (autoc & IXGBE_AUTOC_KX_SUPP)
1275 physical_layer |= IXGBE_PHYSICAL_LAYER_1000BASE_KX;
1276 if (autoc & IXGBE_AUTOC_KX4_SUPP)
1277 physical_layer |= IXGBE_PHYSICAL_LAYER_10GBASE_KX4;
1278 break;
1279 default:
1280 break;
1281 }
1282
1283 if (hw->phy.type == ixgbe_phy_nl) {
1284 hw->phy.ops.identify_sfp(hw);
1285
1286 switch (hw->phy.sfp_type) {
1287 case ixgbe_sfp_type_da_cu:
1288 physical_layer = IXGBE_PHYSICAL_LAYER_SFP_PLUS_CU;
1289 break;
1290 case ixgbe_sfp_type_sr:
1291 physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_SR;
1292 break;
1293 case ixgbe_sfp_type_lr:
1294 physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_LR;
1295 break;
1296 default:
1297 physical_layer = IXGBE_PHYSICAL_LAYER_UNKNOWN;
1298 break;
1299 }
1300 }
1301
1302 switch (hw->device_id) {
1303 case IXGBE_DEV_ID_82598_DA_DUAL_PORT:
1304 physical_layer = IXGBE_PHYSICAL_LAYER_SFP_PLUS_CU;
1305 break;
1306 case IXGBE_DEV_ID_82598AF_DUAL_PORT:
1307 case IXGBE_DEV_ID_82598AF_SINGLE_PORT:
1308 case IXGBE_DEV_ID_82598_SR_DUAL_PORT_EM:
1309 physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_SR;
1310 break;
1311 case IXGBE_DEV_ID_82598EB_XF_LR:
1312 physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_LR;
1313 break;
1314 default:
1315 break;
1316 }
1317
1318out:
1319 return physical_layer;
1320}
1321
1322/**
1323 * ixgbe_set_lan_id_multi_port_pcie_82598 - Set LAN id for PCIe multiple
1324 * port devices.
1325 * @hw: pointer to the HW structure
1326 *
1327 * Calls common function and corrects issue with some single port devices
1328 * that enable LAN1 but not LAN0.
1329 **/
1330void ixgbe_set_lan_id_multi_port_pcie_82598(struct ixgbe_hw *hw)
1331{
1332 struct ixgbe_bus_info *bus = &hw->bus;
1333 u16 pci_gen = 0;
1334 u16 pci_ctrl2 = 0;
1335
1336 DEBUGFUNC("ixgbe_set_lan_id_multi_port_pcie_82598");
1337
1338 ixgbe_set_lan_id_multi_port_pcie(hw);
1339
1340 /* check if LAN0 is disabled */
1341 hw->eeprom.ops.read(hw, IXGBE_PCIE_GENERAL_PTR, &pci_gen);
1342 if ((pci_gen != 0) && (pci_gen != 0xFFFF)) {
1343
1344 hw->eeprom.ops.read(hw, pci_gen + IXGBE_PCIE_CTRL2, &pci_ctrl2);
1345
1346 /* if LAN0 is completely disabled force function to 0 */
1347 if ((pci_ctrl2 & IXGBE_PCIE_CTRL2_LAN_DISABLE) &&
1348 !(pci_ctrl2 & IXGBE_PCIE_CTRL2_DISABLE_SELECT) &&
1349 !(pci_ctrl2 & IXGBE_PCIE_CTRL2_DUMMY_ENABLE)) {
1350
1351 bus->func = 0;
1352 }
1353 }
1354}
1355
1356/**
1357 * ixgbe_enable_relaxed_ordering_82598 - enable relaxed ordering
1358 * @hw: pointer to hardware structure
1359 *
1360 **/
1361void ixgbe_enable_relaxed_ordering_82598(struct ixgbe_hw *hw)
1362{
1363 u32 regval;
1364 u32 i;
1365
1366 DEBUGFUNC("ixgbe_enable_relaxed_ordering_82598");
1367
1368 /* Enable relaxed ordering */
1369 for (i = 0; ((i < hw->mac.max_tx_queues) &&
1370 (i < IXGBE_DCA_MAX_QUEUES_82598)); i++) {
1371 regval = IXGBE_READ_REG(hw, IXGBE_DCA_TXCTRL(i));
1372 regval |= IXGBE_DCA_TXCTRL_DESC_WRO_EN;
1373 IXGBE_WRITE_REG(hw, IXGBE_DCA_TXCTRL(i), regval);
1374 }
1375
1376 for (i = 0; ((i < hw->mac.max_rx_queues) &&
1377 (i < IXGBE_DCA_MAX_QUEUES_82598)); i++) {
1378 regval = IXGBE_READ_REG(hw, IXGBE_DCA_RXCTRL(i));
1379 regval |= IXGBE_DCA_RXCTRL_DATA_WRO_EN |
1380 IXGBE_DCA_RXCTRL_HEAD_WRO_EN;
1381 IXGBE_WRITE_REG(hw, IXGBE_DCA_RXCTRL(i), regval);
1382 }
1383
1384}
1385
1386/**
1387 * ixgbe_set_rxpba_82598 - Initialize RX packet buffer
1388 * @hw: pointer to hardware structure
1389 * @num_pb: number of packet buffers to allocate
1390 * @headroom: reserve n KB of headroom
1391 * @strategy: packet buffer allocation strategy
1392 **/
1393static void ixgbe_set_rxpba_82598(struct ixgbe_hw *hw, int num_pb,
1394 u32 headroom, int strategy)
1395{
1396 u32 rxpktsize = IXGBE_RXPBSIZE_64KB;
1397 u8 i = 0;
1398 UNREFERENCED_1PARAMETER(headroom);
1399
1400 if (!num_pb)
1401 return;
1402
1403 /* Setup Rx packet buffer sizes */
1404 switch (strategy) {
1405 case PBA_STRATEGY_WEIGHTED:
1406 /* Setup the first four at 80KB */
1407 rxpktsize = IXGBE_RXPBSIZE_80KB;
1408 for (; i < 4; i++)
1409 IXGBE_WRITE_REG(hw, IXGBE_RXPBSIZE(i), rxpktsize);
1410 /* Setup the last four at 48KB...don't re-init i */
1411 rxpktsize = IXGBE_RXPBSIZE_48KB;
1412 /* Fall Through */
1413 case PBA_STRATEGY_EQUAL:
1414 default:
1415 /* Divide the remaining Rx packet buffer evenly among the TCs */
1416 for (; i < IXGBE_MAX_PACKET_BUFFERS; i++)
1417 IXGBE_WRITE_REG(hw, IXGBE_RXPBSIZE(i), rxpktsize);
1418 break;
1419 }
1420
1421 /* Setup Tx packet buffer sizes */
1422 for (i = 0; i < IXGBE_MAX_PACKET_BUFFERS; i++)
1423 IXGBE_WRITE_REG(hw, IXGBE_TXPBSIZE(i), IXGBE_TXPBSIZE_40KB);
1424}
1425
1426/**
1427 * ixgbe_enable_rx_dma_82598 - Enable the Rx DMA unit
1428 * @hw: pointer to hardware structure
1429 * @regval: register value to write to RXCTRL
1430 *
1431 * Enables the Rx DMA unit
1432 **/
1433s32 ixgbe_enable_rx_dma_82598(struct ixgbe_hw *hw, u32 regval)
1434{
1435 DEBUGFUNC("ixgbe_enable_rx_dma_82598");
1436
1437 IXGBE_WRITE_REG(hw, IXGBE_RXCTRL, regval);
1438
1439 return IXGBE_SUCCESS;
1440}