cxgb4/chtls: Prefix adapter flags with CXGB4
Some of these macros were conflicting with global namespace, hence prefixing them with CXGB4. Signed-off-by: Arjun Vynipadath <arjun@chelsio.com> Signed-off-by: Vishal Kulkarni <vishal@chelsio.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
Родитель
29ca1c5a4b
Коммит
80f61f19e5
|
@ -616,7 +616,7 @@ int chtls_listen_start(struct chtls_dev *cdev, struct sock *sk)
|
|||
|
||||
pi = netdev_priv(ndev);
|
||||
adap = pi->adapter;
|
||||
if (!(adap->flags & FULL_INIT_DONE))
|
||||
if (!(adap->flags & CXGB4_FULL_INIT_DONE))
|
||||
return -EBADF;
|
||||
|
||||
if (listen_hash_find(cdev, sk) >= 0) /* already have it */
|
||||
|
|
|
@ -81,7 +81,7 @@ static int is_fw_attached(struct cudbg_init *pdbg_init)
|
|||
{
|
||||
struct adapter *padap = pdbg_init->adap;
|
||||
|
||||
if (!(padap->flags & FW_OK) || padap->use_bd)
|
||||
if (!(padap->flags & CXGB4_FW_OK) || padap->use_bd)
|
||||
return 0;
|
||||
|
||||
return 1;
|
||||
|
|
|
@ -606,18 +606,18 @@ struct dentry;
|
|||
struct work_struct;
|
||||
|
||||
enum { /* adapter flags */
|
||||
FULL_INIT_DONE = (1 << 0),
|
||||
DEV_ENABLED = (1 << 1),
|
||||
USING_MSI = (1 << 2),
|
||||
USING_MSIX = (1 << 3),
|
||||
FW_OK = (1 << 4),
|
||||
RSS_TNLALLLOOKUP = (1 << 5),
|
||||
USING_SOFT_PARAMS = (1 << 6),
|
||||
MASTER_PF = (1 << 7),
|
||||
FW_OFLD_CONN = (1 << 9),
|
||||
ROOT_NO_RELAXED_ORDERING = (1 << 10),
|
||||
SHUTTING_DOWN = (1 << 11),
|
||||
SGE_DBQ_TIMER = (1 << 12),
|
||||
CXGB4_FULL_INIT_DONE = (1 << 0),
|
||||
CXGB4_DEV_ENABLED = (1 << 1),
|
||||
CXGB4_USING_MSI = (1 << 2),
|
||||
CXGB4_USING_MSIX = (1 << 3),
|
||||
CXGB4_FW_OK = (1 << 4),
|
||||
CXGB4_RSS_TNLALLLOOKUP = (1 << 5),
|
||||
CXGB4_USING_SOFT_PARAMS = (1 << 6),
|
||||
CXGB4_MASTER_PF = (1 << 7),
|
||||
CXGB4_FW_OFLD_CONN = (1 << 9),
|
||||
CXGB4_ROOT_NO_RELAXED_ORDERING = (1 << 10),
|
||||
CXGB4_SHUTTING_DOWN = (1 << 11),
|
||||
CXGB4_SGE_DBQ_TIMER = (1 << 12),
|
||||
};
|
||||
|
||||
enum {
|
||||
|
|
|
@ -3143,7 +3143,7 @@ static int tid_info_show(struct seq_file *seq, void *v)
|
|||
seq_printf(seq, ", in use: %u/%u\n",
|
||||
atomic_read(&t->tids_in_use),
|
||||
atomic_read(&t->hash_tids_in_use));
|
||||
} else if (adap->flags & FW_OFLD_CONN) {
|
||||
} else if (adap->flags & CXGB4_FW_OFLD_CONN) {
|
||||
seq_printf(seq, "TID range: %u..%u/%u..%u",
|
||||
t->aftid_base,
|
||||
t->aftid_end,
|
||||
|
|
|
@ -875,7 +875,7 @@ static int set_sge_param(struct net_device *dev, struct ethtool_ringparam *e)
|
|||
e->rx_pending < MIN_FL_ENTRIES || e->tx_pending < MIN_TXQ_ENTRIES)
|
||||
return -EINVAL;
|
||||
|
||||
if (adapter->flags & FULL_INIT_DONE)
|
||||
if (adapter->flags & CXGB4_FULL_INIT_DONE)
|
||||
return -EBUSY;
|
||||
|
||||
for (i = 0; i < pi->nqsets; ++i) {
|
||||
|
@ -940,7 +940,7 @@ static int get_dbqtimer_tick(struct net_device *dev)
|
|||
struct port_info *pi = netdev_priv(dev);
|
||||
struct adapter *adap = pi->adapter;
|
||||
|
||||
if (!(adap->flags & SGE_DBQ_TIMER))
|
||||
if (!(adap->flags & CXGB4_SGE_DBQ_TIMER))
|
||||
return 0;
|
||||
|
||||
return adap->sge.dbqtimer_tick;
|
||||
|
@ -957,7 +957,7 @@ static int get_dbqtimer(struct net_device *dev)
|
|||
|
||||
txq = &adap->sge.ethtxq[pi->first_qset];
|
||||
|
||||
if (!(adap->flags & SGE_DBQ_TIMER))
|
||||
if (!(adap->flags & CXGB4_SGE_DBQ_TIMER))
|
||||
return 0;
|
||||
|
||||
/* all of the TX Queues use the same Timer Index */
|
||||
|
@ -979,7 +979,7 @@ static int set_dbqtimer_tick(struct net_device *dev, int usecs)
|
|||
u32 param, val;
|
||||
int ret;
|
||||
|
||||
if (!(adap->flags & SGE_DBQ_TIMER))
|
||||
if (!(adap->flags & CXGB4_SGE_DBQ_TIMER))
|
||||
return 0;
|
||||
|
||||
/* return early if it's the same Timer Tick we're already using */
|
||||
|
@ -1015,7 +1015,7 @@ static int set_dbqtimer(struct net_device *dev, int usecs)
|
|||
u32 param, val;
|
||||
int ret;
|
||||
|
||||
if (!(adap->flags & SGE_DBQ_TIMER))
|
||||
if (!(adap->flags & CXGB4_SGE_DBQ_TIMER))
|
||||
return 0;
|
||||
|
||||
/* Find the SGE Doorbell Timer Value that's closest to the requested
|
||||
|
@ -1042,7 +1042,7 @@ static int set_dbqtimer(struct net_device *dev, int usecs)
|
|||
return 0;
|
||||
|
||||
for (qix = 0; qix < pi->nqsets; qix++, txq++) {
|
||||
if (adap->flags & FULL_INIT_DONE) {
|
||||
if (adap->flags & CXGB4_FULL_INIT_DONE) {
|
||||
param =
|
||||
(FW_PARAMS_MNEM_V(FW_PARAMS_MNEM_DMAQ) |
|
||||
FW_PARAMS_PARAM_X_V(FW_PARAMS_PARAM_DMAQ_EQ_TIMERIX) |
|
||||
|
@ -1263,7 +1263,7 @@ static int set_flash(struct net_device *netdev, struct ethtool_flash *ef)
|
|||
* firmware image otherwise we'll try to do the entire job from the
|
||||
* host ... and we always "force" the operation in this path.
|
||||
*/
|
||||
if (adap->flags & FULL_INIT_DONE)
|
||||
if (adap->flags & CXGB4_FULL_INIT_DONE)
|
||||
mbox = adap->mbox;
|
||||
|
||||
ret = t4_fw_upgrade(adap, mbox, fw->data, fw->size, 1);
|
||||
|
@ -1342,7 +1342,7 @@ static int set_rss_table(struct net_device *dev, const u32 *p, const u8 *key,
|
|||
return 0;
|
||||
|
||||
/* Interface must be brought up atleast once */
|
||||
if (pi->adapter->flags & FULL_INIT_DONE) {
|
||||
if (pi->adapter->flags & CXGB4_FULL_INIT_DONE) {
|
||||
for (i = 0; i < pi->rss_size; i++)
|
||||
pi->rss[i] = p[i];
|
||||
|
||||
|
|
|
@ -74,7 +74,7 @@ int cxgb_fcoe_enable(struct net_device *netdev)
|
|||
if (is_t4(adap->params.chip))
|
||||
return -EINVAL;
|
||||
|
||||
if (!(adap->flags & FULL_INIT_DONE))
|
||||
if (!(adap->flags & CXGB4_FULL_INIT_DONE))
|
||||
return -EINVAL;
|
||||
|
||||
dev_info(adap->pdev_dev, "Enabling FCoE offload features\n");
|
||||
|
|
|
@ -524,7 +524,7 @@ static int del_filter_wr(struct adapter *adapter, int fidx)
|
|||
return -ENOMEM;
|
||||
|
||||
fwr = __skb_put(skb, len);
|
||||
t4_mk_filtdelwr(f->tid, fwr, (adapter->flags & SHUTTING_DOWN) ? -1
|
||||
t4_mk_filtdelwr(f->tid, fwr, (adapter->flags & CXGB4_SHUTTING_DOWN) ? -1
|
||||
: adapter->sge.fw_evtq.abs_id);
|
||||
|
||||
/* Mark the filter as "pending" and ship off the Filter Work Request.
|
||||
|
@ -1569,7 +1569,7 @@ int cxgb4_del_filter(struct net_device *dev, int filter_id,
|
|||
int ret;
|
||||
|
||||
/* If we are shutting down the adapter do not wait for completion */
|
||||
if (netdev2adap(dev)->flags & SHUTTING_DOWN)
|
||||
if (netdev2adap(dev)->flags & CXGB4_SHUTTING_DOWN)
|
||||
return __cxgb4_del_filter(dev, filter_id, fs, NULL);
|
||||
|
||||
init_completion(&ctx.completion);
|
||||
|
|
|
@ -651,12 +651,12 @@ out:
|
|||
|
||||
static void disable_msi(struct adapter *adapter)
|
||||
{
|
||||
if (adapter->flags & USING_MSIX) {
|
||||
if (adapter->flags & CXGB4_USING_MSIX) {
|
||||
pci_disable_msix(adapter->pdev);
|
||||
adapter->flags &= ~USING_MSIX;
|
||||
} else if (adapter->flags & USING_MSI) {
|
||||
adapter->flags &= ~CXGB4_USING_MSIX;
|
||||
} else if (adapter->flags & CXGB4_USING_MSI) {
|
||||
pci_disable_msi(adapter->pdev);
|
||||
adapter->flags &= ~USING_MSI;
|
||||
adapter->flags &= ~CXGB4_USING_MSI;
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -672,7 +672,7 @@ static irqreturn_t t4_nondata_intr(int irq, void *cookie)
|
|||
adap->swintr = 1;
|
||||
t4_write_reg(adap, MYPF_REG(PL_PF_INT_CAUSE_A), v);
|
||||
}
|
||||
if (adap->flags & MASTER_PF)
|
||||
if (adap->flags & CXGB4_MASTER_PF)
|
||||
t4_slow_intr_handler(adap);
|
||||
return IRQ_HANDLED;
|
||||
}
|
||||
|
@ -837,9 +837,9 @@ static void quiesce_rx(struct adapter *adap)
|
|||
/* Disable interrupt and napi handler */
|
||||
static void disable_interrupts(struct adapter *adap)
|
||||
{
|
||||
if (adap->flags & FULL_INIT_DONE) {
|
||||
if (adap->flags & CXGB4_FULL_INIT_DONE) {
|
||||
t4_intr_disable(adap);
|
||||
if (adap->flags & USING_MSIX) {
|
||||
if (adap->flags & CXGB4_USING_MSIX) {
|
||||
free_msix_queue_irqs(adap);
|
||||
free_irq(adap->msix_info[0].vec, adap);
|
||||
} else {
|
||||
|
@ -880,7 +880,7 @@ static int setup_fw_sge_queues(struct adapter *adap)
|
|||
bitmap_zero(s->starving_fl, s->egr_sz);
|
||||
bitmap_zero(s->txq_maperr, s->egr_sz);
|
||||
|
||||
if (adap->flags & USING_MSIX)
|
||||
if (adap->flags & CXGB4_USING_MSIX)
|
||||
adap->msi_idx = 1; /* vector 0 is for non-queue interrupts */
|
||||
else {
|
||||
err = t4_sge_alloc_rxq(adap, &s->intrq, false, adap->port[0], 0,
|
||||
|
@ -939,7 +939,7 @@ static int setup_sge_queues(struct adapter *adap)
|
|||
err = t4_sge_alloc_eth_txq(adap, t, dev,
|
||||
netdev_get_tx_queue(dev, j),
|
||||
q->rspq.cntxt_id,
|
||||
!!(adap->flags & SGE_DBQ_TIMER));
|
||||
!!(adap->flags & CXGB4_SGE_DBQ_TIMER));
|
||||
if (err)
|
||||
goto freeout;
|
||||
}
|
||||
|
@ -2280,7 +2280,7 @@ static int cxgb_up(struct adapter *adap)
|
|||
if (err)
|
||||
goto freeq;
|
||||
|
||||
if (adap->flags & USING_MSIX) {
|
||||
if (adap->flags & CXGB4_USING_MSIX) {
|
||||
name_msix_vecs(adap);
|
||||
err = request_irq(adap->msix_info[0].vec, t4_nondata_intr, 0,
|
||||
adap->msix_info[0].desc, adap);
|
||||
|
@ -2293,7 +2293,8 @@ static int cxgb_up(struct adapter *adap)
|
|||
}
|
||||
} else {
|
||||
err = request_irq(adap->pdev->irq, t4_intr_handler(adap),
|
||||
(adap->flags & USING_MSI) ? 0 : IRQF_SHARED,
|
||||
(adap->flags & CXGB4_USING_MSI) ? 0
|
||||
: IRQF_SHARED,
|
||||
adap->port[0]->name, adap);
|
||||
if (err)
|
||||
goto irq_err;
|
||||
|
@ -2302,7 +2303,7 @@ static int cxgb_up(struct adapter *adap)
|
|||
enable_rx(adap);
|
||||
t4_sge_start(adap);
|
||||
t4_intr_enable(adap);
|
||||
adap->flags |= FULL_INIT_DONE;
|
||||
adap->flags |= CXGB4_FULL_INIT_DONE;
|
||||
mutex_unlock(&uld_mutex);
|
||||
|
||||
notify_ulds(adap, CXGB4_STATE_UP);
|
||||
|
@ -2331,7 +2332,7 @@ static void cxgb_down(struct adapter *adapter)
|
|||
t4_sge_stop(adapter);
|
||||
t4_free_sge_resources(adapter);
|
||||
|
||||
adapter->flags &= ~FULL_INIT_DONE;
|
||||
adapter->flags &= ~CXGB4_FULL_INIT_DONE;
|
||||
}
|
||||
|
||||
/*
|
||||
|
@ -2345,7 +2346,7 @@ static int cxgb_open(struct net_device *dev)
|
|||
|
||||
netif_carrier_off(dev);
|
||||
|
||||
if (!(adapter->flags & FULL_INIT_DONE)) {
|
||||
if (!(adapter->flags & CXGB4_FULL_INIT_DONE)) {
|
||||
err = cxgb_up(adapter);
|
||||
if (err < 0)
|
||||
return err;
|
||||
|
@ -2950,7 +2951,7 @@ static void cxgb_netpoll(struct net_device *dev)
|
|||
struct port_info *pi = netdev_priv(dev);
|
||||
struct adapter *adap = pi->adapter;
|
||||
|
||||
if (adap->flags & USING_MSIX) {
|
||||
if (adap->flags & CXGB4_USING_MSIX) {
|
||||
int i;
|
||||
struct sge_eth_rxq *rx = &adap->sge.ethrxq[pi->first_qset];
|
||||
|
||||
|
@ -2977,7 +2978,7 @@ static int cxgb_set_tx_maxrate(struct net_device *dev, int index, u32 rate)
|
|||
if (index < 0 || index > pi->nqsets - 1)
|
||||
return -EINVAL;
|
||||
|
||||
if (!(adap->flags & FULL_INIT_DONE)) {
|
||||
if (!(adap->flags & CXGB4_FULL_INIT_DONE)) {
|
||||
dev_err(adap->pdev_dev,
|
||||
"Failed to rate limit on queue %d. Link Down?\n",
|
||||
index);
|
||||
|
@ -3078,7 +3079,7 @@ static int cxgb_setup_tc_block_cb(enum tc_setup_type type, void *type_data,
|
|||
struct port_info *pi = netdev2pinfo(dev);
|
||||
struct adapter *adap = netdev2adap(dev);
|
||||
|
||||
if (!(adap->flags & FULL_INIT_DONE)) {
|
||||
if (!(adap->flags & CXGB4_FULL_INIT_DONE)) {
|
||||
dev_err(adap->pdev_dev,
|
||||
"Failed to setup tc on port %d. Link Down?\n",
|
||||
pi->port_id);
|
||||
|
@ -4210,7 +4211,7 @@ static int adap_init0(struct adapter *adap)
|
|||
return ret;
|
||||
}
|
||||
if (ret == adap->mbox)
|
||||
adap->flags |= MASTER_PF;
|
||||
adap->flags |= CXGB4_MASTER_PF;
|
||||
|
||||
/*
|
||||
* If we're the Master PF Driver and the device is uninitialized,
|
||||
|
@ -4225,7 +4226,7 @@ static int adap_init0(struct adapter *adap)
|
|||
/* If firmware is too old (not supported by driver) force an update. */
|
||||
if (ret)
|
||||
state = DEV_STATE_UNINIT;
|
||||
if ((adap->flags & MASTER_PF) && state != DEV_STATE_INIT) {
|
||||
if ((adap->flags & CXGB4_MASTER_PF) && state != DEV_STATE_INIT) {
|
||||
struct fw_info *fw_info;
|
||||
struct fw_hdr *card_fw;
|
||||
const struct firmware *fw;
|
||||
|
@ -4287,7 +4288,7 @@ static int adap_init0(struct adapter *adap)
|
|||
ret);
|
||||
dev_info(adap->pdev_dev, "Coming up as %s: "\
|
||||
"Adapter already initialized\n",
|
||||
adap->flags & MASTER_PF ? "MASTER" : "SLAVE");
|
||||
adap->flags & CXGB4_MASTER_PF ? "MASTER" : "SLAVE");
|
||||
} else {
|
||||
dev_info(adap->pdev_dev, "Coming up as MASTER: "\
|
||||
"Initializing adapter\n");
|
||||
|
@ -4389,7 +4390,7 @@ static int adap_init0(struct adapter *adap)
|
|||
}
|
||||
|
||||
if (!ret)
|
||||
adap->flags |= SGE_DBQ_TIMER;
|
||||
adap->flags |= CXGB4_SGE_DBQ_TIMER;
|
||||
|
||||
if (is_bypass_device(adap->pdev->device))
|
||||
adap->params.bypass = 1;
|
||||
|
@ -4513,7 +4514,7 @@ static int adap_init0(struct adapter *adap)
|
|||
* offload connection through firmware work request
|
||||
*/
|
||||
if ((val[0] != val[1]) && (ret >= 0)) {
|
||||
adap->flags |= FW_OFLD_CONN;
|
||||
adap->flags |= CXGB4_FW_OFLD_CONN;
|
||||
adap->tids.aftid_base = val[0];
|
||||
adap->tids.aftid_end = val[1];
|
||||
}
|
||||
|
@ -4606,7 +4607,7 @@ static int adap_init0(struct adapter *adap)
|
|||
* 2. Server filter: This are special filters which are used
|
||||
* to redirect SYN packets to offload queue.
|
||||
*/
|
||||
if (adap->flags & FW_OFLD_CONN && !is_bypass(adap)) {
|
||||
if (adap->flags & CXGB4_FW_OFLD_CONN && !is_bypass(adap)) {
|
||||
adap->tids.sftid_base = adap->tids.ftid_base +
|
||||
DIV_ROUND_UP(adap->tids.nftids, 3);
|
||||
adap->tids.nsftids = adap->tids.nftids -
|
||||
|
@ -4785,7 +4786,7 @@ static int adap_init0(struct adapter *adap)
|
|||
adap->params.b_wnd);
|
||||
}
|
||||
t4_init_sge_params(adap);
|
||||
adap->flags |= FW_OK;
|
||||
adap->flags |= CXGB4_FW_OK;
|
||||
t4_init_tp_params(adap, true);
|
||||
return 0;
|
||||
|
||||
|
@ -4820,7 +4821,7 @@ static pci_ers_result_t eeh_err_detected(struct pci_dev *pdev,
|
|||
goto out;
|
||||
|
||||
rtnl_lock();
|
||||
adap->flags &= ~FW_OK;
|
||||
adap->flags &= ~CXGB4_FW_OK;
|
||||
notify_ulds(adap, CXGB4_STATE_START_RECOVERY);
|
||||
spin_lock(&adap->stats_lock);
|
||||
for_each_port(adap, i) {
|
||||
|
@ -4832,12 +4833,12 @@ static pci_ers_result_t eeh_err_detected(struct pci_dev *pdev,
|
|||
}
|
||||
spin_unlock(&adap->stats_lock);
|
||||
disable_interrupts(adap);
|
||||
if (adap->flags & FULL_INIT_DONE)
|
||||
if (adap->flags & CXGB4_FULL_INIT_DONE)
|
||||
cxgb_down(adap);
|
||||
rtnl_unlock();
|
||||
if ((adap->flags & DEV_ENABLED)) {
|
||||
if ((adap->flags & CXGB4_DEV_ENABLED)) {
|
||||
pci_disable_device(pdev);
|
||||
adap->flags &= ~DEV_ENABLED;
|
||||
adap->flags &= ~CXGB4_DEV_ENABLED;
|
||||
}
|
||||
out: return state == pci_channel_io_perm_failure ?
|
||||
PCI_ERS_RESULT_DISCONNECT : PCI_ERS_RESULT_NEED_RESET;
|
||||
|
@ -4855,13 +4856,13 @@ static pci_ers_result_t eeh_slot_reset(struct pci_dev *pdev)
|
|||
return PCI_ERS_RESULT_RECOVERED;
|
||||
}
|
||||
|
||||
if (!(adap->flags & DEV_ENABLED)) {
|
||||
if (!(adap->flags & CXGB4_DEV_ENABLED)) {
|
||||
if (pci_enable_device(pdev)) {
|
||||
dev_err(&pdev->dev, "Cannot reenable PCI "
|
||||
"device after reset\n");
|
||||
return PCI_ERS_RESULT_DISCONNECT;
|
||||
}
|
||||
adap->flags |= DEV_ENABLED;
|
||||
adap->flags |= CXGB4_DEV_ENABLED;
|
||||
}
|
||||
|
||||
pci_set_master(pdev);
|
||||
|
@ -4872,7 +4873,7 @@ static pci_ers_result_t eeh_slot_reset(struct pci_dev *pdev)
|
|||
return PCI_ERS_RESULT_DISCONNECT;
|
||||
if (t4_fw_hello(adap, adap->mbox, adap->pf, MASTER_MUST, NULL) < 0)
|
||||
return PCI_ERS_RESULT_DISCONNECT;
|
||||
adap->flags |= FW_OK;
|
||||
adap->flags |= CXGB4_FW_OK;
|
||||
if (adap_init1(adap, &c))
|
||||
return PCI_ERS_RESULT_DISCONNECT;
|
||||
|
||||
|
@ -4984,7 +4985,7 @@ static int cfg_queues(struct adapter *adap)
|
|||
* at all is problematic ...
|
||||
*/
|
||||
niqflint = adap->params.pfres.niqflint - 1;
|
||||
if (!(adap->flags & USING_MSIX))
|
||||
if (!(adap->flags & CXGB4_USING_MSIX))
|
||||
niqflint--;
|
||||
neq = adap->params.pfres.neq / 2;
|
||||
avail_eth_qsets = min(niqflint, neq);
|
||||
|
@ -5266,8 +5267,8 @@ static void print_adapter_info(struct adapter *adapter)
|
|||
/* Software/Hardware configuration */
|
||||
dev_info(adapter->pdev_dev, "Configuration: %sNIC %s, %s capable\n",
|
||||
is_offload(adapter) ? "R" : "",
|
||||
((adapter->flags & USING_MSIX) ? "MSI-X" :
|
||||
(adapter->flags & USING_MSI) ? "MSI" : ""),
|
||||
((adapter->flags & CXGB4_USING_MSIX) ? "MSI-X" :
|
||||
(adapter->flags & CXGB4_USING_MSI) ? "MSI" : ""),
|
||||
is_offload(adapter) ? "Offload" : "non-Offload");
|
||||
}
|
||||
|
||||
|
@ -5342,7 +5343,7 @@ static void free_some_resources(struct adapter *adapter)
|
|||
kfree(adap2pinfo(adapter, i)->rss);
|
||||
free_netdev(adapter->port[i]);
|
||||
}
|
||||
if (adapter->flags & FW_OK)
|
||||
if (adapter->flags & CXGB4_FW_OK)
|
||||
t4_fw_bye(adapter, adapter->pf);
|
||||
}
|
||||
|
||||
|
@ -5646,7 +5647,7 @@ static int init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
|
|||
}
|
||||
|
||||
/* PCI device has been enabled */
|
||||
adapter->flags |= DEV_ENABLED;
|
||||
adapter->flags |= CXGB4_DEV_ENABLED;
|
||||
memset(adapter->chan_map, 0xff, sizeof(adapter->chan_map));
|
||||
|
||||
/* If possible, we use PCIe Relaxed Ordering Attribute to deliver
|
||||
|
@ -5664,7 +5665,7 @@ static int init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
|
|||
* using Relaxed Ordering.
|
||||
*/
|
||||
if (!pcie_relaxed_ordering_enabled(pdev))
|
||||
adapter->flags |= ROOT_NO_RELAXED_ORDERING;
|
||||
adapter->flags |= CXGB4_ROOT_NO_RELAXED_ORDERING;
|
||||
|
||||
spin_lock_init(&adapter->stats_lock);
|
||||
spin_lock_init(&adapter->tid_release_lock);
|
||||
|
@ -5796,7 +5797,7 @@ static int init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
|
|||
|
||||
pci_set_drvdata(pdev, adapter);
|
||||
|
||||
if (adapter->flags & FW_OK) {
|
||||
if (adapter->flags & CXGB4_FW_OK) {
|
||||
err = t4_port_init(adapter, func, func, 0);
|
||||
if (err)
|
||||
goto out_free_dev;
|
||||
|
@ -5818,7 +5819,7 @@ static int init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
|
|||
}
|
||||
}
|
||||
|
||||
if (!(adapter->flags & FW_OK))
|
||||
if (!(adapter->flags & CXGB4_FW_OK))
|
||||
goto fw_attach_fail;
|
||||
|
||||
/* Configure queues and allocate tables now, they can be needed as
|
||||
|
@ -5912,9 +5913,9 @@ static int init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
|
|||
|
||||
/* See what interrupts we'll be using */
|
||||
if (msi > 1 && enable_msix(adapter) == 0)
|
||||
adapter->flags |= USING_MSIX;
|
||||
adapter->flags |= CXGB4_USING_MSIX;
|
||||
else if (msi > 0 && pci_enable_msi(pdev) == 0) {
|
||||
adapter->flags |= USING_MSI;
|
||||
adapter->flags |= CXGB4_USING_MSI;
|
||||
if (msi > 1)
|
||||
free_msix_info(adapter);
|
||||
}
|
||||
|
@ -5982,7 +5983,7 @@ fw_attach_fail:
|
|||
cxgb4_ptp_init(adapter);
|
||||
|
||||
if (IS_REACHABLE(CONFIG_THERMAL) &&
|
||||
!is_t4(adapter->params.chip) && (adapter->flags & FW_OK))
|
||||
!is_t4(adapter->params.chip) && (adapter->flags & CXGB4_FW_OK))
|
||||
cxgb4_thermal_init(adapter);
|
||||
|
||||
print_adapter_info(adapter);
|
||||
|
@ -5991,7 +5992,7 @@ fw_attach_fail:
|
|||
out_free_dev:
|
||||
t4_free_sge_resources(adapter);
|
||||
free_some_resources(adapter);
|
||||
if (adapter->flags & USING_MSIX)
|
||||
if (adapter->flags & CXGB4_USING_MSIX)
|
||||
free_msix_info(adapter);
|
||||
if (adapter->num_uld || adapter->num_ofld_uld)
|
||||
t4_uld_mem_free(adapter);
|
||||
|
@ -6024,7 +6025,7 @@ static void remove_one(struct pci_dev *pdev)
|
|||
return;
|
||||
}
|
||||
|
||||
adapter->flags |= SHUTTING_DOWN;
|
||||
adapter->flags |= CXGB4_SHUTTING_DOWN;
|
||||
|
||||
if (adapter->pf == 4) {
|
||||
int i;
|
||||
|
@ -6059,10 +6060,10 @@ static void remove_one(struct pci_dev *pdev)
|
|||
*/
|
||||
clear_all_filters(adapter);
|
||||
|
||||
if (adapter->flags & FULL_INIT_DONE)
|
||||
if (adapter->flags & CXGB4_FULL_INIT_DONE)
|
||||
cxgb_down(adapter);
|
||||
|
||||
if (adapter->flags & USING_MSIX)
|
||||
if (adapter->flags & CXGB4_USING_MSIX)
|
||||
free_msix_info(adapter);
|
||||
if (adapter->num_uld || adapter->num_ofld_uld)
|
||||
t4_uld_mem_free(adapter);
|
||||
|
@ -6086,9 +6087,9 @@ static void remove_one(struct pci_dev *pdev)
|
|||
#endif
|
||||
iounmap(adapter->regs);
|
||||
pci_disable_pcie_error_reporting(pdev);
|
||||
if ((adapter->flags & DEV_ENABLED)) {
|
||||
if ((adapter->flags & CXGB4_DEV_ENABLED)) {
|
||||
pci_disable_device(pdev);
|
||||
adapter->flags &= ~DEV_ENABLED;
|
||||
adapter->flags &= ~CXGB4_DEV_ENABLED;
|
||||
}
|
||||
pci_release_regions(pdev);
|
||||
kfree(adapter->mbox_log);
|
||||
|
@ -6114,7 +6115,7 @@ static void shutdown_one(struct pci_dev *pdev)
|
|||
return;
|
||||
}
|
||||
|
||||
adapter->flags |= SHUTTING_DOWN;
|
||||
adapter->flags |= CXGB4_SHUTTING_DOWN;
|
||||
|
||||
if (adapter->pf == 4) {
|
||||
int i;
|
||||
|
@ -6132,7 +6133,7 @@ static void shutdown_one(struct pci_dev *pdev)
|
|||
disable_msi(adapter);
|
||||
|
||||
t4_sge_stop(adapter);
|
||||
if (adapter->flags & FW_OK)
|
||||
if (adapter->flags & CXGB4_FW_OK)
|
||||
t4_fw_bye(adapter, adapter->mbox);
|
||||
}
|
||||
}
|
||||
|
|
|
@ -147,7 +147,7 @@ static int alloc_uld_rxqs(struct adapter *adap,
|
|||
|
||||
per_chan = rxq_info->nrxq / adap->params.nports;
|
||||
|
||||
if (adap->flags & USING_MSIX)
|
||||
if (adap->flags & CXGB4_USING_MSIX)
|
||||
msi_idx = 1;
|
||||
else
|
||||
msi_idx = -((int)s->intrq.abs_id + 1);
|
||||
|
@ -195,7 +195,7 @@ setup_sge_queues_uld(struct adapter *adap, unsigned int uld_type, bool lro)
|
|||
struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type];
|
||||
int i, ret = 0;
|
||||
|
||||
if (adap->flags & USING_MSIX) {
|
||||
if (adap->flags & CXGB4_USING_MSIX) {
|
||||
rxq_info->msix_tbl = kcalloc((rxq_info->nrxq + rxq_info->nciq),
|
||||
sizeof(unsigned short),
|
||||
GFP_KERNEL);
|
||||
|
@ -206,7 +206,7 @@ setup_sge_queues_uld(struct adapter *adap, unsigned int uld_type, bool lro)
|
|||
ret = !(!alloc_uld_rxqs(adap, rxq_info, lro));
|
||||
|
||||
/* Tell uP to route control queue completions to rdma rspq */
|
||||
if (adap->flags & FULL_INIT_DONE &&
|
||||
if (adap->flags & CXGB4_FULL_INIT_DONE &&
|
||||
!ret && uld_type == CXGB4_ULD_RDMA) {
|
||||
struct sge *s = &adap->sge;
|
||||
unsigned int cmplqid;
|
||||
|
@ -239,7 +239,7 @@ static void free_sge_queues_uld(struct adapter *adap, unsigned int uld_type)
|
|||
{
|
||||
struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type];
|
||||
|
||||
if (adap->flags & FULL_INIT_DONE && uld_type == CXGB4_ULD_RDMA) {
|
||||
if (adap->flags & CXGB4_FULL_INIT_DONE && uld_type == CXGB4_ULD_RDMA) {
|
||||
struct sge *s = &adap->sge;
|
||||
u32 param, cmdop, cmplqid = 0;
|
||||
int i;
|
||||
|
@ -258,7 +258,7 @@ static void free_sge_queues_uld(struct adapter *adap, unsigned int uld_type)
|
|||
t4_free_uld_rxqs(adap, rxq_info->nciq,
|
||||
rxq_info->uldrxq + rxq_info->nrxq);
|
||||
t4_free_uld_rxqs(adap, rxq_info->nrxq, rxq_info->uldrxq);
|
||||
if (adap->flags & USING_MSIX)
|
||||
if (adap->flags & CXGB4_USING_MSIX)
|
||||
kfree(rxq_info->msix_tbl);
|
||||
}
|
||||
|
||||
|
@ -273,7 +273,7 @@ static int cfg_queues_uld(struct adapter *adap, unsigned int uld_type,
|
|||
if (!rxq_info)
|
||||
return -ENOMEM;
|
||||
|
||||
if (adap->flags & USING_MSIX && uld_info->nrxq > s->nqs_per_uld) {
|
||||
if (adap->flags & CXGB4_USING_MSIX && uld_info->nrxq > s->nqs_per_uld) {
|
||||
i = s->nqs_per_uld;
|
||||
rxq_info->nrxq = roundup(i, adap->params.nports);
|
||||
} else {
|
||||
|
@ -284,7 +284,7 @@ static int cfg_queues_uld(struct adapter *adap, unsigned int uld_type,
|
|||
if (!uld_info->ciq) {
|
||||
rxq_info->nciq = 0;
|
||||
} else {
|
||||
if (adap->flags & USING_MSIX)
|
||||
if (adap->flags & CXGB4_USING_MSIX)
|
||||
rxq_info->nciq = min_t(int, s->nqs_per_uld,
|
||||
num_online_cpus());
|
||||
else
|
||||
|
@ -611,10 +611,10 @@ static void cxgb4_shutdown_uld_adapter(struct adapter *adap, enum cxgb4_uld type
|
|||
adap->uld[type].add = NULL;
|
||||
release_sge_txq_uld(adap, type);
|
||||
|
||||
if (adap->flags & FULL_INIT_DONE)
|
||||
if (adap->flags & CXGB4_FULL_INIT_DONE)
|
||||
quiesce_rx_uld(adap, type);
|
||||
|
||||
if (adap->flags & USING_MSIX)
|
||||
if (adap->flags & CXGB4_USING_MSIX)
|
||||
free_msix_queue_irqs_uld(adap, type);
|
||||
|
||||
free_sge_queues_uld(adap, type);
|
||||
|
@ -673,7 +673,7 @@ static void uld_init(struct adapter *adap, struct cxgb4_lld_info *lld)
|
|||
lld->sge_egrstatuspagesize = adap->sge.stat_len;
|
||||
lld->sge_pktshift = adap->sge.pktshift;
|
||||
lld->ulp_crypto = adap->params.crypto;
|
||||
lld->enable_fw_ofld_conn = adap->flags & FW_OFLD_CONN;
|
||||
lld->enable_fw_ofld_conn = adap->flags & CXGB4_FW_OFLD_CONN;
|
||||
lld->max_ordird_qp = adap->params.max_ordird_qp;
|
||||
lld->max_ird_adapter = adap->params.max_ird_adapter;
|
||||
lld->ulptx_memwrite_dsgl = adap->params.ulptx_memwrite_dsgl;
|
||||
|
@ -702,7 +702,7 @@ static void uld_attach(struct adapter *adap, unsigned int uld)
|
|||
adap->uld[uld].handle = handle;
|
||||
t4_register_netevent_notifier();
|
||||
|
||||
if (adap->flags & FULL_INIT_DONE)
|
||||
if (adap->flags & CXGB4_FULL_INIT_DONE)
|
||||
adap->uld[uld].state_change(handle, CXGB4_STATE_UP);
|
||||
}
|
||||
|
||||
|
@ -737,13 +737,13 @@ void cxgb4_register_uld(enum cxgb4_uld type,
|
|||
ret = setup_sge_queues_uld(adap, type, p->lro);
|
||||
if (ret)
|
||||
goto free_queues;
|
||||
if (adap->flags & USING_MSIX) {
|
||||
if (adap->flags & CXGB4_USING_MSIX) {
|
||||
name_msix_vecs_uld(adap, type);
|
||||
ret = request_msix_queue_irqs_uld(adap, type);
|
||||
if (ret)
|
||||
goto free_rxq;
|
||||
}
|
||||
if (adap->flags & FULL_INIT_DONE)
|
||||
if (adap->flags & CXGB4_FULL_INIT_DONE)
|
||||
enable_rx_uld(adap, type);
|
||||
if (adap->uld[type].add)
|
||||
goto free_irq;
|
||||
|
@ -754,9 +754,9 @@ void cxgb4_register_uld(enum cxgb4_uld type,
|
|||
uld_attach(adap, type);
|
||||
continue;
|
||||
free_irq:
|
||||
if (adap->flags & FULL_INIT_DONE)
|
||||
if (adap->flags & CXGB4_FULL_INIT_DONE)
|
||||
quiesce_rx_uld(adap, type);
|
||||
if (adap->flags & USING_MSIX)
|
||||
if (adap->flags & CXGB4_USING_MSIX)
|
||||
free_msix_queue_irqs_uld(adap, type);
|
||||
free_rxq:
|
||||
free_sge_queues_uld(adap, type);
|
||||
|
|
|
@ -3375,7 +3375,7 @@ static irqreturn_t t4_intr_msi(int irq, void *cookie)
|
|||
{
|
||||
struct adapter *adap = cookie;
|
||||
|
||||
if (adap->flags & MASTER_PF)
|
||||
if (adap->flags & CXGB4_MASTER_PF)
|
||||
t4_slow_intr_handler(adap);
|
||||
process_intrq(adap);
|
||||
return IRQ_HANDLED;
|
||||
|
@ -3391,7 +3391,7 @@ static irqreturn_t t4_intr_intx(int irq, void *cookie)
|
|||
struct adapter *adap = cookie;
|
||||
|
||||
t4_write_reg(adap, MYPF_REG(PCIE_PF_CLI_A), 0);
|
||||
if (((adap->flags & MASTER_PF) && t4_slow_intr_handler(adap)) |
|
||||
if (((adap->flags & CXGB4_MASTER_PF) && t4_slow_intr_handler(adap)) |
|
||||
process_intrq(adap))
|
||||
return IRQ_HANDLED;
|
||||
return IRQ_NONE; /* probably shared interrupt */
|
||||
|
@ -3406,9 +3406,9 @@ static irqreturn_t t4_intr_intx(int irq, void *cookie)
|
|||
*/
|
||||
irq_handler_t t4_intr_handler(struct adapter *adap)
|
||||
{
|
||||
if (adap->flags & USING_MSIX)
|
||||
if (adap->flags & CXGB4_USING_MSIX)
|
||||
return t4_sge_intr_msix;
|
||||
if (adap->flags & USING_MSI)
|
||||
if (adap->flags & CXGB4_USING_MSI)
|
||||
return t4_intr_msi;
|
||||
return t4_intr_intx;
|
||||
}
|
||||
|
@ -3441,7 +3441,7 @@ static void sge_rx_timer_cb(struct timer_list *t)
|
|||
* global Master PF activities like checking for chip ingress stalls,
|
||||
* etc.
|
||||
*/
|
||||
if (!(adap->flags & MASTER_PF))
|
||||
if (!(adap->flags & CXGB4_MASTER_PF))
|
||||
goto done;
|
||||
|
||||
t4_idma_monitor(adap, &s->idma_monitor, HZ, RX_QCHECK_PERIOD);
|
||||
|
@ -3549,7 +3549,7 @@ int t4_sge_alloc_rxq(struct adapter *adap, struct sge_rspq *iq, bool fwevtq,
|
|||
struct fw_iq_cmd c;
|
||||
struct sge *s = &adap->sge;
|
||||
struct port_info *pi = netdev_priv(dev);
|
||||
int relaxed = !(adap->flags & ROOT_NO_RELAXED_ORDERING);
|
||||
int relaxed = !(adap->flags & CXGB4_ROOT_NO_RELAXED_ORDERING);
|
||||
|
||||
/* Size needs to be multiple of 16, including status entry. */
|
||||
iq->size = roundup(iq->size, 16);
|
||||
|
|
|
@ -77,7 +77,7 @@ int cxgb4_get_srq_entry(struct net_device *dev,
|
|||
adap = netdev2adap(dev);
|
||||
s = adap->srq;
|
||||
|
||||
if (!(adap->flags & FULL_INIT_DONE) || !s)
|
||||
if (!(adap->flags & CXGB4_FULL_INIT_DONE) || !s)
|
||||
goto out;
|
||||
|
||||
skb = alloc_skb(sizeof(*req), GFP_KERNEL);
|
||||
|
|
|
@ -198,7 +198,7 @@ static void t4_report_fw_error(struct adapter *adap)
|
|||
if (pcie_fw & PCIE_FW_ERR_F) {
|
||||
dev_err(adap->pdev_dev, "Firmware reports adapter error: %s\n",
|
||||
reason[PCIE_FW_EVAL_G(pcie_fw)]);
|
||||
adap->flags &= ~FW_OK;
|
||||
adap->flags &= ~CXGB4_FW_OK;
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -5243,7 +5243,7 @@ int t4_read_rss(struct adapter *adapter, u16 *map)
|
|||
|
||||
static unsigned int t4_use_ldst(struct adapter *adap)
|
||||
{
|
||||
return (adap->flags & FW_OK) && !adap->use_bd;
|
||||
return (adap->flags & CXGB4_FW_OK) && !adap->use_bd;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -6132,7 +6132,7 @@ unsigned int t4_get_mps_bg_map(struct adapter *adapter, int pidx)
|
|||
* ( MPSBGMAP[Port 1] << 8 ) |
|
||||
* ( MPSBGMAP[Port 0] << 0 ))
|
||||
*/
|
||||
if (adapter->flags & FW_OK) {
|
||||
if (adapter->flags & CXGB4_FW_OK) {
|
||||
u32 param, val;
|
||||
int ret;
|
||||
|
||||
|
@ -7093,10 +7093,10 @@ int t4_fw_upgrade(struct adapter *adap, unsigned int mbox,
|
|||
if (!t4_fw_matches_chip(adap, fw_hdr))
|
||||
return -EINVAL;
|
||||
|
||||
/* Disable FW_OK flag so that mbox commands with FW_OK flag set
|
||||
* wont be sent when we are flashing FW.
|
||||
/* Disable CXGB4_FW_OK flag so that mbox commands with CXGB4_FW_OK flag
|
||||
* set wont be sent when we are flashing FW.
|
||||
*/
|
||||
adap->flags &= ~FW_OK;
|
||||
adap->flags &= ~CXGB4_FW_OK;
|
||||
|
||||
ret = t4_fw_halt(adap, mbox, force);
|
||||
if (ret < 0 && !force)
|
||||
|
@ -7135,7 +7135,7 @@ int t4_fw_upgrade(struct adapter *adap, unsigned int mbox,
|
|||
*/
|
||||
(void)t4_init_devlog_params(adap);
|
||||
out:
|
||||
adap->flags |= FW_OK;
|
||||
adap->flags |= CXGB4_FW_OK;
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
|
Загрузка…
Ссылка в новой задаче