74b793e1ef
Pre-allocate a skb at init time to be used for control messages to the HW if skb allocation fails. Tolerate failures to send messages initializing some memories at the cost of parity error detection for these memories. Retry sending connection id release messages if both alloc_skb(GFP_ATOMIC) and alloc_skb(GFP_KERNEL) fail. Do not bring the interface up if messages binding queue set to port fail to be sent. Signed-off-by: Divy Le Ray <divy@chelsio.com> Signed-off-by: David S. Miller <davem@davemloft.net>
316 lines
11 KiB
C
316 lines
11 KiB
C
/*
|
|
* Copyright (c) 2003-2008 Chelsio, Inc. All rights reserved.
|
|
*
|
|
* This software is available to you under a choice of one of two
|
|
* licenses. You may choose to be licensed under the terms of the GNU
|
|
* General Public License (GPL) Version 2, available from the file
|
|
* COPYING in the main directory of this source tree, or the
|
|
* OpenIB.org BSD license below:
|
|
*
|
|
* Redistribution and use in source and binary forms, with or
|
|
* without modification, are permitted provided that the following
|
|
* conditions are met:
|
|
*
|
|
* - Redistributions of source code must retain the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer.
|
|
*
|
|
* - Redistributions in binary form must reproduce the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer in the documentation and/or other materials
|
|
* provided with the distribution.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
|
|
* BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
|
|
* ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
|
|
* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
* SOFTWARE.
|
|
*/
|
|
|
|
/* This file should not be included directly. Include common.h instead. */
|
|
|
|
#ifndef __T3_ADAPTER_H__
|
|
#define __T3_ADAPTER_H__
|
|
|
|
#include <linux/pci.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/timer.h>
|
|
#include <linux/cache.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/bitops.h>
|
|
#include "t3cdev.h"
|
|
#include <asm/io.h>
|
|
|
|
struct vlan_group;
|
|
struct adapter;
|
|
struct sge_qset;
|
|
|
|
enum { /* rx_offload flags */
|
|
T3_RX_CSUM = 1 << 0,
|
|
T3_LRO = 1 << 1,
|
|
};
|
|
|
|
struct port_info {
|
|
struct adapter *adapter;
|
|
struct vlan_group *vlan_grp;
|
|
struct sge_qset *qs;
|
|
u8 port_id;
|
|
u8 rx_offload;
|
|
u8 nqsets;
|
|
u8 first_qset;
|
|
struct cphy phy;
|
|
struct cmac mac;
|
|
struct link_config link_config;
|
|
struct net_device_stats netstats;
|
|
int activity;
|
|
__be32 iscsi_ipv4addr;
|
|
|
|
int link_fault; /* link fault was detected */
|
|
};
|
|
|
|
enum { /* adapter flags */
|
|
FULL_INIT_DONE = (1 << 0),
|
|
USING_MSI = (1 << 1),
|
|
USING_MSIX = (1 << 2),
|
|
QUEUES_BOUND = (1 << 3),
|
|
TP_PARITY_INIT = (1 << 4),
|
|
NAPI_INIT = (1 << 5),
|
|
};
|
|
|
|
struct fl_pg_chunk {
|
|
struct page *page;
|
|
void *va;
|
|
unsigned int offset;
|
|
unsigned long *p_cnt;
|
|
dma_addr_t mapping;
|
|
};
|
|
|
|
struct rx_desc;
|
|
struct rx_sw_desc;
|
|
|
|
struct sge_fl { /* SGE per free-buffer list state */
|
|
unsigned int buf_size; /* size of each Rx buffer */
|
|
unsigned int credits; /* # of available Rx buffers */
|
|
unsigned int pend_cred; /* new buffers since last FL DB ring */
|
|
unsigned int size; /* capacity of free list */
|
|
unsigned int cidx; /* consumer index */
|
|
unsigned int pidx; /* producer index */
|
|
unsigned int gen; /* free list generation */
|
|
struct fl_pg_chunk pg_chunk;/* page chunk cache */
|
|
unsigned int use_pages; /* whether FL uses pages or sk_buffs */
|
|
unsigned int order; /* order of page allocations */
|
|
unsigned int alloc_size; /* size of allocated buffer */
|
|
struct rx_desc *desc; /* address of HW Rx descriptor ring */
|
|
struct rx_sw_desc *sdesc; /* address of SW Rx descriptor ring */
|
|
dma_addr_t phys_addr; /* physical address of HW ring start */
|
|
unsigned int cntxt_id; /* SGE context id for the free list */
|
|
unsigned long empty; /* # of times queue ran out of buffers */
|
|
unsigned long alloc_failed; /* # of times buffer allocation failed */
|
|
};
|
|
|
|
/*
|
|
* Bundle size for grouping offload RX packets for delivery to the stack.
|
|
* Don't make this too big as we do prefetch on each packet in a bundle.
|
|
*/
|
|
# define RX_BUNDLE_SIZE 8
|
|
|
|
struct rsp_desc;
|
|
|
|
struct sge_rspq { /* state for an SGE response queue */
|
|
unsigned int credits; /* # of pending response credits */
|
|
unsigned int size; /* capacity of response queue */
|
|
unsigned int cidx; /* consumer index */
|
|
unsigned int gen; /* current generation bit */
|
|
unsigned int polling; /* is the queue serviced through NAPI? */
|
|
unsigned int holdoff_tmr; /* interrupt holdoff timer in 100ns */
|
|
unsigned int next_holdoff; /* holdoff time for next interrupt */
|
|
unsigned int rx_recycle_buf; /* whether recycling occurred
|
|
within current sop-eop */
|
|
struct rsp_desc *desc; /* address of HW response ring */
|
|
dma_addr_t phys_addr; /* physical address of the ring */
|
|
unsigned int cntxt_id; /* SGE context id for the response q */
|
|
spinlock_t lock; /* guards response processing */
|
|
struct sk_buff_head rx_queue; /* offload packet receive queue */
|
|
struct sk_buff *pg_skb; /* used to build frag list in napi handler */
|
|
|
|
unsigned long offload_pkts;
|
|
unsigned long offload_bundles;
|
|
unsigned long eth_pkts; /* # of ethernet packets */
|
|
unsigned long pure_rsps; /* # of pure (non-data) responses */
|
|
unsigned long imm_data; /* responses with immediate data */
|
|
unsigned long rx_drops; /* # of packets dropped due to no mem */
|
|
unsigned long async_notif; /* # of asynchronous notification events */
|
|
unsigned long empty; /* # of times queue ran out of credits */
|
|
unsigned long nomem; /* # of responses deferred due to no mem */
|
|
unsigned long unhandled_irqs; /* # of spurious intrs */
|
|
unsigned long starved;
|
|
unsigned long restarted;
|
|
};
|
|
|
|
struct tx_desc;
|
|
struct tx_sw_desc;
|
|
|
|
struct sge_txq { /* state for an SGE Tx queue */
|
|
unsigned long flags; /* HW DMA fetch status */
|
|
unsigned int in_use; /* # of in-use Tx descriptors */
|
|
unsigned int size; /* # of descriptors */
|
|
unsigned int processed; /* total # of descs HW has processed */
|
|
unsigned int cleaned; /* total # of descs SW has reclaimed */
|
|
unsigned int stop_thres; /* SW TX queue suspend threshold */
|
|
unsigned int cidx; /* consumer index */
|
|
unsigned int pidx; /* producer index */
|
|
unsigned int gen; /* current value of generation bit */
|
|
unsigned int unacked; /* Tx descriptors used since last COMPL */
|
|
struct tx_desc *desc; /* address of HW Tx descriptor ring */
|
|
struct tx_sw_desc *sdesc; /* address of SW Tx descriptor ring */
|
|
spinlock_t lock; /* guards enqueueing of new packets */
|
|
unsigned int token; /* WR token */
|
|
dma_addr_t phys_addr; /* physical address of the ring */
|
|
struct sk_buff_head sendq; /* List of backpressured offload packets */
|
|
struct tasklet_struct qresume_tsk; /* restarts the queue */
|
|
unsigned int cntxt_id; /* SGE context id for the Tx q */
|
|
unsigned long stops; /* # of times q has been stopped */
|
|
unsigned long restarts; /* # of queue restarts */
|
|
};
|
|
|
|
enum { /* per port SGE statistics */
|
|
SGE_PSTAT_TSO, /* # of TSO requests */
|
|
SGE_PSTAT_RX_CSUM_GOOD, /* # of successful RX csum offloads */
|
|
SGE_PSTAT_TX_CSUM, /* # of TX checksum offloads */
|
|
SGE_PSTAT_VLANEX, /* # of VLAN tag extractions */
|
|
SGE_PSTAT_VLANINS, /* # of VLAN tag insertions */
|
|
|
|
SGE_PSTAT_MAX /* must be last */
|
|
};
|
|
|
|
struct napi_gro_fraginfo;
|
|
|
|
struct sge_qset { /* an SGE queue set */
|
|
struct adapter *adap;
|
|
struct napi_struct napi;
|
|
struct sge_rspq rspq;
|
|
struct sge_fl fl[SGE_RXQ_PER_SET];
|
|
struct sge_txq txq[SGE_TXQ_PER_SET];
|
|
int nomem;
|
|
int lro_enabled;
|
|
void *lro_va;
|
|
struct net_device *netdev;
|
|
struct netdev_queue *tx_q; /* associated netdev TX queue */
|
|
unsigned long txq_stopped; /* which Tx queues are stopped */
|
|
struct timer_list tx_reclaim_timer; /* reclaims TX buffers */
|
|
struct timer_list rx_reclaim_timer; /* reclaims RX buffers */
|
|
unsigned long port_stats[SGE_PSTAT_MAX];
|
|
} ____cacheline_aligned;
|
|
|
|
struct sge {
|
|
struct sge_qset qs[SGE_QSETS];
|
|
spinlock_t reg_lock; /* guards non-atomic SGE registers (eg context) */
|
|
};
|
|
|
|
struct adapter {
|
|
struct t3cdev tdev;
|
|
struct list_head adapter_list;
|
|
void __iomem *regs;
|
|
struct pci_dev *pdev;
|
|
unsigned long registered_device_map;
|
|
unsigned long open_device_map;
|
|
unsigned long flags;
|
|
|
|
const char *name;
|
|
int msg_enable;
|
|
unsigned int mmio_len;
|
|
|
|
struct adapter_params params;
|
|
unsigned int slow_intr_mask;
|
|
unsigned long irq_stats[IRQ_NUM_STATS];
|
|
|
|
int msix_nvectors;
|
|
struct {
|
|
unsigned short vec;
|
|
char desc[22];
|
|
} msix_info[SGE_QSETS + 1];
|
|
|
|
/* T3 modules */
|
|
struct sge sge;
|
|
struct mc7 pmrx;
|
|
struct mc7 pmtx;
|
|
struct mc7 cm;
|
|
struct mc5 mc5;
|
|
|
|
struct net_device *port[MAX_NPORTS];
|
|
unsigned int check_task_cnt;
|
|
struct delayed_work adap_check_task;
|
|
struct work_struct ext_intr_handler_task;
|
|
struct work_struct fatal_error_handler_task;
|
|
struct work_struct link_fault_handler_task;
|
|
|
|
struct dentry *debugfs_root;
|
|
|
|
struct mutex mdio_lock;
|
|
spinlock_t stats_lock;
|
|
spinlock_t work_lock;
|
|
|
|
struct sk_buff *nofail_skb;
|
|
};
|
|
|
|
static inline u32 t3_read_reg(struct adapter *adapter, u32 reg_addr)
|
|
{
|
|
u32 val = readl(adapter->regs + reg_addr);
|
|
|
|
CH_DBG(adapter, MMIO, "read register 0x%x value 0x%x\n", reg_addr, val);
|
|
return val;
|
|
}
|
|
|
|
static inline void t3_write_reg(struct adapter *adapter, u32 reg_addr, u32 val)
|
|
{
|
|
CH_DBG(adapter, MMIO, "setting register 0x%x to 0x%x\n", reg_addr, val);
|
|
writel(val, adapter->regs + reg_addr);
|
|
}
|
|
|
|
static inline struct port_info *adap2pinfo(struct adapter *adap, int idx)
|
|
{
|
|
return netdev_priv(adap->port[idx]);
|
|
}
|
|
|
|
#define OFFLOAD_DEVMAP_BIT 15
|
|
|
|
#define tdev2adap(d) container_of(d, struct adapter, tdev)
|
|
|
|
static inline int offload_running(struct adapter *adapter)
|
|
{
|
|
return test_bit(OFFLOAD_DEVMAP_BIT, &adapter->open_device_map);
|
|
}
|
|
|
|
int t3_offload_tx(struct t3cdev *tdev, struct sk_buff *skb);
|
|
|
|
void t3_os_ext_intr_handler(struct adapter *adapter);
|
|
void t3_os_link_changed(struct adapter *adapter, int port_id, int link_status,
|
|
int speed, int duplex, int fc);
|
|
void t3_os_phymod_changed(struct adapter *adap, int port_id);
|
|
void t3_os_link_fault(struct adapter *adapter, int port_id, int state);
|
|
void t3_os_link_fault_handler(struct adapter *adapter, int port_id);
|
|
|
|
void t3_sge_start(struct adapter *adap);
|
|
void t3_sge_stop(struct adapter *adap);
|
|
void t3_start_sge_timers(struct adapter *adap);
|
|
void t3_stop_sge_timers(struct adapter *adap);
|
|
void t3_free_sge_resources(struct adapter *adap);
|
|
void t3_sge_err_intr_handler(struct adapter *adapter);
|
|
irq_handler_t t3_intr_handler(struct adapter *adap, int polling);
|
|
int t3_eth_xmit(struct sk_buff *skb, struct net_device *dev);
|
|
int t3_mgmt_tx(struct adapter *adap, struct sk_buff *skb);
|
|
void t3_update_qset_coalesce(struct sge_qset *qs, const struct qset_params *p);
|
|
int t3_sge_alloc_qset(struct adapter *adapter, unsigned int id, int nports,
|
|
int irq_vec_idx, const struct qset_params *p,
|
|
int ntxq, struct net_device *dev,
|
|
struct netdev_queue *netdevq);
|
|
int t3_get_desc(const struct sge_qset *qs, unsigned int qnum, unsigned int idx,
|
|
unsigned char *data);
|
|
irqreturn_t t3_sge_intr_msix(int irq, void *cookie);
|
|
|
|
#endif /* __T3_ADAPTER_H__ */
|