mirror of https://gitee.com/openkylin/linux.git
738 lines
22 KiB
C
738 lines
22 KiB
C
/*******************************************************************************
|
|
|
|
Intel 10 Gigabit PCI Express Linux driver
|
|
Copyright(c) 1999 - 2011 Intel Corporation.
|
|
|
|
This program is free software; you can redistribute it and/or modify it
|
|
under the terms and conditions of the GNU General Public License,
|
|
version 2, as published by the Free Software Foundation.
|
|
|
|
This program is distributed in the hope it will be useful, but WITHOUT
|
|
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
more details.
|
|
|
|
You should have received a copy of the GNU General Public License along with
|
|
this program; if not, write to the Free Software Foundation, Inc.,
|
|
51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
|
|
|
|
The full GNU General Public License is included in this distribution in
|
|
the file called "COPYING".
|
|
|
|
Contact Information:
|
|
e1000-devel Mailing List <e1000-devel@lists.sourceforge.net>
|
|
Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
|
|
|
|
*******************************************************************************/
|
|
|
|
#include <linux/pci.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/sched.h>
|
|
|
|
#include "ixgbe.h"
|
|
#include "ixgbe_phy.h"
|
|
|
|
#define IXGBE_X540_MAX_TX_QUEUES 128
|
|
#define IXGBE_X540_MAX_RX_QUEUES 128
|
|
#define IXGBE_X540_RAR_ENTRIES 128
|
|
#define IXGBE_X540_MC_TBL_SIZE 128
|
|
#define IXGBE_X540_VFT_TBL_SIZE 128
|
|
|
|
static s32 ixgbe_update_flash_X540(struct ixgbe_hw *hw);
|
|
static s32 ixgbe_poll_flash_update_done_X540(struct ixgbe_hw *hw);
|
|
static s32 ixgbe_acquire_swfw_sync_X540(struct ixgbe_hw *hw, u16 mask);
|
|
static void ixgbe_release_swfw_sync_X540(struct ixgbe_hw *hw, u16 mask);
|
|
static s32 ixgbe_get_swfw_sync_semaphore(struct ixgbe_hw *hw);
|
|
static void ixgbe_release_swfw_sync_semaphore(struct ixgbe_hw *hw);
|
|
|
|
static enum ixgbe_media_type ixgbe_get_media_type_X540(struct ixgbe_hw *hw)
|
|
{
|
|
return ixgbe_media_type_copper;
|
|
}
|
|
|
|
static s32 ixgbe_get_invariants_X540(struct ixgbe_hw *hw)
|
|
{
|
|
struct ixgbe_mac_info *mac = &hw->mac;
|
|
|
|
/* Call PHY identify routine to get the phy type */
|
|
ixgbe_identify_phy_generic(hw);
|
|
|
|
mac->mcft_size = IXGBE_X540_MC_TBL_SIZE;
|
|
mac->vft_size = IXGBE_X540_VFT_TBL_SIZE;
|
|
mac->num_rar_entries = IXGBE_X540_RAR_ENTRIES;
|
|
mac->max_rx_queues = IXGBE_X540_MAX_RX_QUEUES;
|
|
mac->max_tx_queues = IXGBE_X540_MAX_TX_QUEUES;
|
|
mac->max_msix_vectors = ixgbe_get_pcie_msix_count_generic(hw);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* ixgbe_setup_mac_link_X540 - Set the auto advertised capabilitires
|
|
* @hw: pointer to hardware structure
|
|
* @speed: new link speed
|
|
* @autoneg: true if autonegotiation enabled
|
|
* @autoneg_wait_to_complete: true when waiting for completion is needed
|
|
**/
|
|
static s32 ixgbe_setup_mac_link_X540(struct ixgbe_hw *hw,
|
|
ixgbe_link_speed speed, bool autoneg,
|
|
bool autoneg_wait_to_complete)
|
|
{
|
|
return hw->phy.ops.setup_link_speed(hw, speed, autoneg,
|
|
autoneg_wait_to_complete);
|
|
}
|
|
|
|
/**
|
|
* ixgbe_reset_hw_X540 - Perform hardware reset
|
|
* @hw: pointer to hardware structure
|
|
*
|
|
* Resets the hardware by resetting the transmit and receive units, masks
|
|
* and clears all interrupts, perform a PHY reset, and perform a link (MAC)
|
|
* reset.
|
|
**/
|
|
static s32 ixgbe_reset_hw_X540(struct ixgbe_hw *hw)
|
|
{
|
|
ixgbe_link_speed link_speed;
|
|
s32 status = 0;
|
|
u32 ctrl;
|
|
u32 ctrl_ext;
|
|
u32 reset_bit;
|
|
u32 i;
|
|
u32 autoc;
|
|
u32 autoc2;
|
|
bool link_up = false;
|
|
|
|
/* Call adapter stop to disable tx/rx and clear interrupts */
|
|
hw->mac.ops.stop_adapter(hw);
|
|
|
|
/*
|
|
* Prevent the PCI-E bus from from hanging by disabling PCI-E master
|
|
* access and verify no pending requests before reset
|
|
*/
|
|
ixgbe_disable_pcie_master(hw);
|
|
|
|
mac_reset_top:
|
|
/*
|
|
* Issue global reset to the MAC. Needs to be SW reset if link is up.
|
|
* If link reset is used when link is up, it might reset the PHY when
|
|
* mng is using it. If link is down or the flag to force full link
|
|
* reset is set, then perform link reset.
|
|
*/
|
|
if (hw->force_full_reset) {
|
|
reset_bit = IXGBE_CTRL_LNK_RST;
|
|
} else {
|
|
hw->mac.ops.check_link(hw, &link_speed, &link_up, false);
|
|
if (!link_up)
|
|
reset_bit = IXGBE_CTRL_LNK_RST;
|
|
else
|
|
reset_bit = IXGBE_CTRL_RST;
|
|
}
|
|
|
|
ctrl = IXGBE_READ_REG(hw, IXGBE_CTRL);
|
|
IXGBE_WRITE_REG(hw, IXGBE_CTRL, (ctrl | reset_bit));
|
|
IXGBE_WRITE_FLUSH(hw);
|
|
|
|
/* Poll for reset bit to self-clear indicating reset is complete */
|
|
for (i = 0; i < 10; i++) {
|
|
udelay(1);
|
|
ctrl = IXGBE_READ_REG(hw, IXGBE_CTRL);
|
|
if (!(ctrl & reset_bit))
|
|
break;
|
|
}
|
|
if (ctrl & reset_bit) {
|
|
status = IXGBE_ERR_RESET_FAILED;
|
|
hw_dbg(hw, "Reset polling failed to complete.\n");
|
|
}
|
|
|
|
/*
|
|
* Double resets are required for recovery from certain error
|
|
* conditions. Between resets, it is necessary to stall to allow time
|
|
* for any pending HW events to complete. We use 1usec since that is
|
|
* what is needed for ixgbe_disable_pcie_master(). The second reset
|
|
* then clears out any effects of those events.
|
|
*/
|
|
if (hw->mac.flags & IXGBE_FLAGS_DOUBLE_RESET_REQUIRED) {
|
|
hw->mac.flags &= ~IXGBE_FLAGS_DOUBLE_RESET_REQUIRED;
|
|
udelay(1);
|
|
goto mac_reset_top;
|
|
}
|
|
|
|
/* Clear PF Reset Done bit so PF/VF Mail Ops can work */
|
|
ctrl_ext = IXGBE_READ_REG(hw, IXGBE_CTRL_EXT);
|
|
ctrl_ext |= IXGBE_CTRL_EXT_PFRSTD;
|
|
IXGBE_WRITE_REG(hw, IXGBE_CTRL_EXT, ctrl_ext);
|
|
|
|
msleep(50);
|
|
|
|
/* Set the Rx packet buffer size. */
|
|
IXGBE_WRITE_REG(hw, IXGBE_RXPBSIZE(0), 384 << IXGBE_RXPBSIZE_SHIFT);
|
|
|
|
/* Store the permanent mac address */
|
|
hw->mac.ops.get_mac_addr(hw, hw->mac.perm_addr);
|
|
|
|
/*
|
|
* Store the original AUTOC/AUTOC2 values if they have not been
|
|
* stored off yet. Otherwise restore the stored original
|
|
* values since the reset operation sets back to defaults.
|
|
*/
|
|
autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
|
|
autoc2 = IXGBE_READ_REG(hw, IXGBE_AUTOC2);
|
|
if (hw->mac.orig_link_settings_stored == false) {
|
|
hw->mac.orig_autoc = autoc;
|
|
hw->mac.orig_autoc2 = autoc2;
|
|
hw->mac.orig_link_settings_stored = true;
|
|
} else {
|
|
if (autoc != hw->mac.orig_autoc)
|
|
IXGBE_WRITE_REG(hw, IXGBE_AUTOC, (hw->mac.orig_autoc |
|
|
IXGBE_AUTOC_AN_RESTART));
|
|
|
|
if ((autoc2 & IXGBE_AUTOC2_UPPER_MASK) !=
|
|
(hw->mac.orig_autoc2 & IXGBE_AUTOC2_UPPER_MASK)) {
|
|
autoc2 &= ~IXGBE_AUTOC2_UPPER_MASK;
|
|
autoc2 |= (hw->mac.orig_autoc2 &
|
|
IXGBE_AUTOC2_UPPER_MASK);
|
|
IXGBE_WRITE_REG(hw, IXGBE_AUTOC2, autoc2);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Store MAC address from RAR0, clear receive address registers, and
|
|
* clear the multicast table. Also reset num_rar_entries to 128,
|
|
* since we modify this value when programming the SAN MAC address.
|
|
*/
|
|
hw->mac.num_rar_entries = IXGBE_X540_MAX_TX_QUEUES;
|
|
hw->mac.ops.init_rx_addrs(hw);
|
|
|
|
/* Store the permanent mac address */
|
|
hw->mac.ops.get_mac_addr(hw, hw->mac.perm_addr);
|
|
|
|
/* Store the permanent SAN mac address */
|
|
hw->mac.ops.get_san_mac_addr(hw, hw->mac.san_addr);
|
|
|
|
/* Add the SAN MAC address to the RAR only if it's a valid address */
|
|
if (ixgbe_validate_mac_addr(hw->mac.san_addr) == 0) {
|
|
hw->mac.ops.set_rar(hw, hw->mac.num_rar_entries - 1,
|
|
hw->mac.san_addr, 0, IXGBE_RAH_AV);
|
|
|
|
/* Reserve the last RAR for the SAN MAC address */
|
|
hw->mac.num_rar_entries--;
|
|
}
|
|
|
|
/* Store the alternative WWNN/WWPN prefix */
|
|
hw->mac.ops.get_wwn_prefix(hw, &hw->mac.wwnn_prefix,
|
|
&hw->mac.wwpn_prefix);
|
|
|
|
return status;
|
|
}
|
|
|
|
/**
|
|
* ixgbe_get_supported_physical_layer_X540 - Returns physical layer type
|
|
* @hw: pointer to hardware structure
|
|
*
|
|
* Determines physical layer capabilities of the current configuration.
|
|
**/
|
|
static u32 ixgbe_get_supported_physical_layer_X540(struct ixgbe_hw *hw)
|
|
{
|
|
u32 physical_layer = IXGBE_PHYSICAL_LAYER_UNKNOWN;
|
|
u16 ext_ability = 0;
|
|
|
|
hw->phy.ops.identify(hw);
|
|
|
|
hw->phy.ops.read_reg(hw, MDIO_PMA_EXTABLE, MDIO_MMD_PMAPMD,
|
|
&ext_ability);
|
|
if (ext_ability & MDIO_PMA_EXTABLE_10GBT)
|
|
physical_layer |= IXGBE_PHYSICAL_LAYER_10GBASE_T;
|
|
if (ext_ability & MDIO_PMA_EXTABLE_1000BT)
|
|
physical_layer |= IXGBE_PHYSICAL_LAYER_1000BASE_T;
|
|
if (ext_ability & MDIO_PMA_EXTABLE_100BTX)
|
|
physical_layer |= IXGBE_PHYSICAL_LAYER_100BASE_TX;
|
|
|
|
return physical_layer;
|
|
}
|
|
|
|
/**
|
|
* ixgbe_init_eeprom_params_X540 - Initialize EEPROM params
|
|
* @hw: pointer to hardware structure
|
|
*
|
|
* Initializes the EEPROM parameters ixgbe_eeprom_info within the
|
|
* ixgbe_hw struct in order to set up EEPROM access.
|
|
**/
|
|
static s32 ixgbe_init_eeprom_params_X540(struct ixgbe_hw *hw)
|
|
{
|
|
struct ixgbe_eeprom_info *eeprom = &hw->eeprom;
|
|
u32 eec;
|
|
u16 eeprom_size;
|
|
|
|
if (eeprom->type == ixgbe_eeprom_uninitialized) {
|
|
eeprom->semaphore_delay = 10;
|
|
eeprom->type = ixgbe_flash;
|
|
|
|
eec = IXGBE_READ_REG(hw, IXGBE_EEC);
|
|
eeprom_size = (u16)((eec & IXGBE_EEC_SIZE) >>
|
|
IXGBE_EEC_SIZE_SHIFT);
|
|
eeprom->word_size = 1 << (eeprom_size +
|
|
IXGBE_EEPROM_WORD_SIZE_SHIFT);
|
|
|
|
hw_dbg(hw, "Eeprom params: type = %d, size = %d\n",
|
|
eeprom->type, eeprom->word_size);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* ixgbe_read_eerd_X540 - Read EEPROM word using EERD
|
|
* @hw: pointer to hardware structure
|
|
* @offset: offset of word in the EEPROM to read
|
|
* @data: word read from the EERPOM
|
|
**/
|
|
static s32 ixgbe_read_eerd_X540(struct ixgbe_hw *hw, u16 offset, u16 *data)
|
|
{
|
|
s32 status;
|
|
|
|
if (hw->mac.ops.acquire_swfw_sync(hw, IXGBE_GSSR_EEP_SM) == 0)
|
|
status = ixgbe_read_eerd_generic(hw, offset, data);
|
|
else
|
|
status = IXGBE_ERR_SWFW_SYNC;
|
|
|
|
ixgbe_release_swfw_sync_X540(hw, IXGBE_GSSR_EEP_SM);
|
|
return status;
|
|
}
|
|
|
|
/**
|
|
* ixgbe_write_eewr_X540 - Write EEPROM word using EEWR
|
|
* @hw: pointer to hardware structure
|
|
* @offset: offset of word in the EEPROM to write
|
|
* @data: word write to the EEPROM
|
|
*
|
|
* Write a 16 bit word to the EEPROM using the EEWR register.
|
|
**/
|
|
static s32 ixgbe_write_eewr_X540(struct ixgbe_hw *hw, u16 offset, u16 data)
|
|
{
|
|
u32 eewr;
|
|
s32 status;
|
|
|
|
hw->eeprom.ops.init_params(hw);
|
|
|
|
if (offset >= hw->eeprom.word_size) {
|
|
status = IXGBE_ERR_EEPROM;
|
|
goto out;
|
|
}
|
|
|
|
eewr = (offset << IXGBE_EEPROM_RW_ADDR_SHIFT) |
|
|
(data << IXGBE_EEPROM_RW_REG_DATA) |
|
|
IXGBE_EEPROM_RW_REG_START;
|
|
|
|
if (hw->mac.ops.acquire_swfw_sync(hw, IXGBE_GSSR_EEP_SM) == 0) {
|
|
status = ixgbe_poll_eerd_eewr_done(hw, IXGBE_NVM_POLL_WRITE);
|
|
if (status != 0) {
|
|
hw_dbg(hw, "Eeprom write EEWR timed out\n");
|
|
goto out;
|
|
}
|
|
|
|
IXGBE_WRITE_REG(hw, IXGBE_EEWR, eewr);
|
|
|
|
status = ixgbe_poll_eerd_eewr_done(hw, IXGBE_NVM_POLL_WRITE);
|
|
if (status != 0) {
|
|
hw_dbg(hw, "Eeprom write EEWR timed out\n");
|
|
goto out;
|
|
}
|
|
} else {
|
|
status = IXGBE_ERR_SWFW_SYNC;
|
|
}
|
|
|
|
out:
|
|
ixgbe_release_swfw_sync_X540(hw, IXGBE_GSSR_EEP_SM);
|
|
return status;
|
|
}
|
|
|
|
/**
|
|
* ixgbe_calc_eeprom_checksum_X540 - Calculates and returns the checksum
|
|
* @hw: pointer to hardware structure
|
|
**/
|
|
static u16 ixgbe_calc_eeprom_checksum_X540(struct ixgbe_hw *hw)
|
|
{
|
|
u16 i;
|
|
u16 j;
|
|
u16 checksum = 0;
|
|
u16 length = 0;
|
|
u16 pointer = 0;
|
|
u16 word = 0;
|
|
|
|
/* Include 0x0-0x3F in the checksum */
|
|
for (i = 0; i < IXGBE_EEPROM_CHECKSUM; i++) {
|
|
if (hw->eeprom.ops.read(hw, i, &word) != 0) {
|
|
hw_dbg(hw, "EEPROM read failed\n");
|
|
break;
|
|
}
|
|
checksum += word;
|
|
}
|
|
|
|
/*
|
|
* Include all data from pointers 0x3, 0x6-0xE. This excludes the
|
|
* FW, PHY module, and PCIe Expansion/Option ROM pointers.
|
|
*/
|
|
for (i = IXGBE_PCIE_ANALOG_PTR; i < IXGBE_FW_PTR; i++) {
|
|
if (i == IXGBE_PHY_PTR || i == IXGBE_OPTION_ROM_PTR)
|
|
continue;
|
|
|
|
if (hw->eeprom.ops.read(hw, i, &pointer) != 0) {
|
|
hw_dbg(hw, "EEPROM read failed\n");
|
|
break;
|
|
}
|
|
|
|
/* Skip pointer section if the pointer is invalid. */
|
|
if (pointer == 0xFFFF || pointer == 0 ||
|
|
pointer >= hw->eeprom.word_size)
|
|
continue;
|
|
|
|
if (hw->eeprom.ops.read(hw, pointer, &length) != 0) {
|
|
hw_dbg(hw, "EEPROM read failed\n");
|
|
break;
|
|
}
|
|
|
|
/* Skip pointer section if length is invalid. */
|
|
if (length == 0xFFFF || length == 0 ||
|
|
(pointer + length) >= hw->eeprom.word_size)
|
|
continue;
|
|
|
|
for (j = pointer+1; j <= pointer+length; j++) {
|
|
if (hw->eeprom.ops.read(hw, j, &word) != 0) {
|
|
hw_dbg(hw, "EEPROM read failed\n");
|
|
break;
|
|
}
|
|
checksum += word;
|
|
}
|
|
}
|
|
|
|
checksum = (u16)IXGBE_EEPROM_SUM - checksum;
|
|
|
|
return checksum;
|
|
}
|
|
|
|
/**
|
|
* ixgbe_update_eeprom_checksum_X540 - Updates the EEPROM checksum and flash
|
|
* @hw: pointer to hardware structure
|
|
*
|
|
* After writing EEPROM to shadow RAM using EEWR register, software calculates
|
|
* checksum and updates the EEPROM and instructs the hardware to update
|
|
* the flash.
|
|
**/
|
|
static s32 ixgbe_update_eeprom_checksum_X540(struct ixgbe_hw *hw)
|
|
{
|
|
s32 status;
|
|
|
|
status = ixgbe_update_eeprom_checksum_generic(hw);
|
|
|
|
if (status)
|
|
status = ixgbe_update_flash_X540(hw);
|
|
|
|
return status;
|
|
}
|
|
|
|
/**
|
|
* ixgbe_update_flash_X540 - Instruct HW to copy EEPROM to Flash device
|
|
* @hw: pointer to hardware structure
|
|
*
|
|
* Set FLUP (bit 23) of the EEC register to instruct Hardware to copy
|
|
* EEPROM from shadow RAM to the flash device.
|
|
**/
|
|
static s32 ixgbe_update_flash_X540(struct ixgbe_hw *hw)
|
|
{
|
|
u32 flup;
|
|
s32 status = IXGBE_ERR_EEPROM;
|
|
|
|
status = ixgbe_poll_flash_update_done_X540(hw);
|
|
if (status == IXGBE_ERR_EEPROM) {
|
|
hw_dbg(hw, "Flash update time out\n");
|
|
goto out;
|
|
}
|
|
|
|
flup = IXGBE_READ_REG(hw, IXGBE_EEC) | IXGBE_EEC_FLUP;
|
|
IXGBE_WRITE_REG(hw, IXGBE_EEC, flup);
|
|
|
|
status = ixgbe_poll_flash_update_done_X540(hw);
|
|
if (status)
|
|
hw_dbg(hw, "Flash update complete\n");
|
|
else
|
|
hw_dbg(hw, "Flash update time out\n");
|
|
|
|
if (hw->revision_id == 0) {
|
|
flup = IXGBE_READ_REG(hw, IXGBE_EEC);
|
|
|
|
if (flup & IXGBE_EEC_SEC1VAL) {
|
|
flup |= IXGBE_EEC_FLUP;
|
|
IXGBE_WRITE_REG(hw, IXGBE_EEC, flup);
|
|
}
|
|
|
|
status = ixgbe_poll_flash_update_done_X540(hw);
|
|
if (status)
|
|
hw_dbg(hw, "Flash update complete\n");
|
|
else
|
|
hw_dbg(hw, "Flash update time out\n");
|
|
|
|
}
|
|
out:
|
|
return status;
|
|
}
|
|
|
|
/**
|
|
* ixgbe_poll_flash_update_done_X540 - Poll flash update status
|
|
* @hw: pointer to hardware structure
|
|
*
|
|
* Polls the FLUDONE (bit 26) of the EEC Register to determine when the
|
|
* flash update is done.
|
|
**/
|
|
static s32 ixgbe_poll_flash_update_done_X540(struct ixgbe_hw *hw)
|
|
{
|
|
u32 i;
|
|
u32 reg;
|
|
s32 status = IXGBE_ERR_EEPROM;
|
|
|
|
for (i = 0; i < IXGBE_FLUDONE_ATTEMPTS; i++) {
|
|
reg = IXGBE_READ_REG(hw, IXGBE_EEC);
|
|
if (reg & IXGBE_EEC_FLUDONE) {
|
|
status = 0;
|
|
break;
|
|
}
|
|
udelay(5);
|
|
}
|
|
return status;
|
|
}
|
|
|
|
/**
|
|
* ixgbe_acquire_swfw_sync_X540 - Acquire SWFW semaphore
|
|
* @hw: pointer to hardware structure
|
|
* @mask: Mask to specify which semaphore to acquire
|
|
*
|
|
* Acquires the SWFW semaphore thought the SW_FW_SYNC register for
|
|
* the specified function (CSR, PHY0, PHY1, NVM, Flash)
|
|
**/
|
|
static s32 ixgbe_acquire_swfw_sync_X540(struct ixgbe_hw *hw, u16 mask)
|
|
{
|
|
u32 swfw_sync;
|
|
u32 swmask = mask;
|
|
u32 fwmask = mask << 5;
|
|
u32 hwmask = 0;
|
|
u32 timeout = 200;
|
|
u32 i;
|
|
|
|
if (swmask == IXGBE_GSSR_EEP_SM)
|
|
hwmask = IXGBE_GSSR_FLASH_SM;
|
|
|
|
for (i = 0; i < timeout; i++) {
|
|
/*
|
|
* SW NVM semaphore bit is used for access to all
|
|
* SW_FW_SYNC bits (not just NVM)
|
|
*/
|
|
if (ixgbe_get_swfw_sync_semaphore(hw))
|
|
return IXGBE_ERR_SWFW_SYNC;
|
|
|
|
swfw_sync = IXGBE_READ_REG(hw, IXGBE_SWFW_SYNC);
|
|
if (!(swfw_sync & (fwmask | swmask | hwmask))) {
|
|
swfw_sync |= swmask;
|
|
IXGBE_WRITE_REG(hw, IXGBE_SWFW_SYNC, swfw_sync);
|
|
ixgbe_release_swfw_sync_semaphore(hw);
|
|
break;
|
|
} else {
|
|
/*
|
|
* Firmware currently using resource (fwmask),
|
|
* hardware currently using resource (hwmask),
|
|
* or other software thread currently using
|
|
* resource (swmask)
|
|
*/
|
|
ixgbe_release_swfw_sync_semaphore(hw);
|
|
msleep(5);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* If the resource is not released by the FW/HW the SW can assume that
|
|
* the FW/HW malfunctions. In that case the SW should sets the
|
|
* SW bit(s) of the requested resource(s) while ignoring the
|
|
* corresponding FW/HW bits in the SW_FW_SYNC register.
|
|
*/
|
|
if (i >= timeout) {
|
|
swfw_sync = IXGBE_READ_REG(hw, IXGBE_SWFW_SYNC);
|
|
if (swfw_sync & (fwmask | hwmask)) {
|
|
if (ixgbe_get_swfw_sync_semaphore(hw))
|
|
return IXGBE_ERR_SWFW_SYNC;
|
|
|
|
swfw_sync |= swmask;
|
|
IXGBE_WRITE_REG(hw, IXGBE_SWFW_SYNC, swfw_sync);
|
|
ixgbe_release_swfw_sync_semaphore(hw);
|
|
}
|
|
}
|
|
|
|
msleep(5);
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* ixgbe_release_swfw_sync_X540 - Release SWFW semaphore
|
|
* @hw: pointer to hardware structure
|
|
* @mask: Mask to specify which semaphore to release
|
|
*
|
|
* Releases the SWFW semaphore through the SW_FW_SYNC register
|
|
* for the specified function (CSR, PHY0, PHY1, EVM, Flash)
|
|
**/
|
|
static void ixgbe_release_swfw_sync_X540(struct ixgbe_hw *hw, u16 mask)
|
|
{
|
|
u32 swfw_sync;
|
|
u32 swmask = mask;
|
|
|
|
ixgbe_get_swfw_sync_semaphore(hw);
|
|
|
|
swfw_sync = IXGBE_READ_REG(hw, IXGBE_SWFW_SYNC);
|
|
swfw_sync &= ~swmask;
|
|
IXGBE_WRITE_REG(hw, IXGBE_SWFW_SYNC, swfw_sync);
|
|
|
|
ixgbe_release_swfw_sync_semaphore(hw);
|
|
msleep(5);
|
|
}
|
|
|
|
/**
|
|
* ixgbe_get_nvm_semaphore - Get hardware semaphore
|
|
* @hw: pointer to hardware structure
|
|
*
|
|
* Sets the hardware semaphores so SW/FW can gain control of shared resources
|
|
**/
|
|
static s32 ixgbe_get_swfw_sync_semaphore(struct ixgbe_hw *hw)
|
|
{
|
|
s32 status = IXGBE_ERR_EEPROM;
|
|
u32 timeout = 2000;
|
|
u32 i;
|
|
u32 swsm;
|
|
|
|
/* Get SMBI software semaphore between device drivers first */
|
|
for (i = 0; i < timeout; i++) {
|
|
/*
|
|
* If the SMBI bit is 0 when we read it, then the bit will be
|
|
* set and we have the semaphore
|
|
*/
|
|
swsm = IXGBE_READ_REG(hw, IXGBE_SWSM);
|
|
if (!(swsm & IXGBE_SWSM_SMBI)) {
|
|
status = 0;
|
|
break;
|
|
}
|
|
udelay(50);
|
|
}
|
|
|
|
/* Now get the semaphore between SW/FW through the REGSMP bit */
|
|
if (status) {
|
|
for (i = 0; i < timeout; i++) {
|
|
swsm = IXGBE_READ_REG(hw, IXGBE_SWFW_SYNC);
|
|
if (!(swsm & IXGBE_SWFW_REGSMP))
|
|
break;
|
|
|
|
udelay(50);
|
|
}
|
|
} else {
|
|
hw_dbg(hw, "Software semaphore SMBI between device drivers "
|
|
"not granted.\n");
|
|
}
|
|
|
|
return status;
|
|
}
|
|
|
|
/**
|
|
* ixgbe_release_nvm_semaphore - Release hardware semaphore
|
|
* @hw: pointer to hardware structure
|
|
*
|
|
* This function clears hardware semaphore bits.
|
|
**/
|
|
static void ixgbe_release_swfw_sync_semaphore(struct ixgbe_hw *hw)
|
|
{
|
|
u32 swsm;
|
|
|
|
/* Release both semaphores by writing 0 to the bits REGSMP and SMBI */
|
|
|
|
swsm = IXGBE_READ_REG(hw, IXGBE_SWSM);
|
|
swsm &= ~IXGBE_SWSM_SMBI;
|
|
IXGBE_WRITE_REG(hw, IXGBE_SWSM, swsm);
|
|
|
|
swsm = IXGBE_READ_REG(hw, IXGBE_SWFW_SYNC);
|
|
swsm &= ~IXGBE_SWFW_REGSMP;
|
|
IXGBE_WRITE_REG(hw, IXGBE_SWFW_SYNC, swsm);
|
|
|
|
IXGBE_WRITE_FLUSH(hw);
|
|
}
|
|
|
|
static struct ixgbe_mac_operations mac_ops_X540 = {
|
|
.init_hw = &ixgbe_init_hw_generic,
|
|
.reset_hw = &ixgbe_reset_hw_X540,
|
|
.start_hw = &ixgbe_start_hw_generic,
|
|
.clear_hw_cntrs = &ixgbe_clear_hw_cntrs_generic,
|
|
.get_media_type = &ixgbe_get_media_type_X540,
|
|
.get_supported_physical_layer =
|
|
&ixgbe_get_supported_physical_layer_X540,
|
|
.enable_rx_dma = &ixgbe_enable_rx_dma_generic,
|
|
.get_mac_addr = &ixgbe_get_mac_addr_generic,
|
|
.get_san_mac_addr = &ixgbe_get_san_mac_addr_generic,
|
|
.get_device_caps = NULL,
|
|
.get_wwn_prefix = &ixgbe_get_wwn_prefix_generic,
|
|
.stop_adapter = &ixgbe_stop_adapter_generic,
|
|
.get_bus_info = &ixgbe_get_bus_info_generic,
|
|
.set_lan_id = &ixgbe_set_lan_id_multi_port_pcie,
|
|
.read_analog_reg8 = NULL,
|
|
.write_analog_reg8 = NULL,
|
|
.setup_link = &ixgbe_setup_mac_link_X540,
|
|
.check_link = &ixgbe_check_mac_link_generic,
|
|
.get_link_capabilities = &ixgbe_get_copper_link_capabilities_generic,
|
|
.led_on = &ixgbe_led_on_generic,
|
|
.led_off = &ixgbe_led_off_generic,
|
|
.blink_led_start = &ixgbe_blink_led_start_generic,
|
|
.blink_led_stop = &ixgbe_blink_led_stop_generic,
|
|
.set_rar = &ixgbe_set_rar_generic,
|
|
.clear_rar = &ixgbe_clear_rar_generic,
|
|
.set_vmdq = &ixgbe_set_vmdq_generic,
|
|
.clear_vmdq = &ixgbe_clear_vmdq_generic,
|
|
.init_rx_addrs = &ixgbe_init_rx_addrs_generic,
|
|
.update_mc_addr_list = &ixgbe_update_mc_addr_list_generic,
|
|
.enable_mc = &ixgbe_enable_mc_generic,
|
|
.disable_mc = &ixgbe_disable_mc_generic,
|
|
.clear_vfta = &ixgbe_clear_vfta_generic,
|
|
.set_vfta = &ixgbe_set_vfta_generic,
|
|
.fc_enable = &ixgbe_fc_enable_generic,
|
|
.init_uta_tables = &ixgbe_init_uta_tables_generic,
|
|
.setup_sfp = NULL,
|
|
.set_mac_anti_spoofing = &ixgbe_set_mac_anti_spoofing,
|
|
.set_vlan_anti_spoofing = &ixgbe_set_vlan_anti_spoofing,
|
|
.acquire_swfw_sync = &ixgbe_acquire_swfw_sync_X540,
|
|
.release_swfw_sync = &ixgbe_release_swfw_sync_X540,
|
|
};
|
|
|
|
static struct ixgbe_eeprom_operations eeprom_ops_X540 = {
|
|
.init_params = &ixgbe_init_eeprom_params_X540,
|
|
.read = &ixgbe_read_eerd_X540,
|
|
.write = &ixgbe_write_eewr_X540,
|
|
.calc_checksum = &ixgbe_calc_eeprom_checksum_X540,
|
|
.validate_checksum = &ixgbe_validate_eeprom_checksum_generic,
|
|
.update_checksum = &ixgbe_update_eeprom_checksum_X540,
|
|
};
|
|
|
|
static struct ixgbe_phy_operations phy_ops_X540 = {
|
|
.identify = &ixgbe_identify_phy_generic,
|
|
.identify_sfp = &ixgbe_identify_sfp_module_generic,
|
|
.init = NULL,
|
|
.reset = NULL,
|
|
.read_reg = &ixgbe_read_phy_reg_generic,
|
|
.write_reg = &ixgbe_write_phy_reg_generic,
|
|
.setup_link = &ixgbe_setup_phy_link_generic,
|
|
.setup_link_speed = &ixgbe_setup_phy_link_speed_generic,
|
|
.read_i2c_byte = &ixgbe_read_i2c_byte_generic,
|
|
.write_i2c_byte = &ixgbe_write_i2c_byte_generic,
|
|
.read_i2c_eeprom = &ixgbe_read_i2c_eeprom_generic,
|
|
.write_i2c_eeprom = &ixgbe_write_i2c_eeprom_generic,
|
|
.check_overtemp = &ixgbe_tn_check_overtemp,
|
|
};
|
|
|
|
struct ixgbe_info ixgbe_X540_info = {
|
|
.mac = ixgbe_mac_X540,
|
|
.get_invariants = &ixgbe_get_invariants_X540,
|
|
.mac_ops = &mac_ops_X540,
|
|
.eeprom_ops = &eeprom_ops_X540,
|
|
.phy_ops = &phy_ops_X540,
|
|
.mbx_ops = &mbx_ops_generic,
|
|
};
|