#ifndef _ACENIC_H_ #define _ACENIC_H_ #include <linux/config.h> /* * Generate TX index update each time, when TX ring is closed. * Normally, this is not useful, because results in more dma (and irqs * without TX_COAL_INTS_ONLY). */ #define USE_TX_COAL_NOW 0 /* * Addressing: * * The Tigon uses 64-bit host addresses, regardless of their actual * length, and it expects a big-endian format. For 32 bit systems the * upper 32 bits of the address are simply ignored (zero), however for * little endian 64 bit systems (Alpha) this looks strange with the * two parts of the address word being swapped. * * The addresses are split in two 32 bit words for all architectures * as some of them are in PCI shared memory and it is necessary to use * readl/writel to access them. * * The addressing code is derived from Pete Wyckoff's work, but * modified to deal properly with readl/writel usage. */ struct ace_regs { u32 pad0[16]; /* PCI control registers */ u32 HostCtrl; /* 0x40 */ u32 LocalCtrl; u32 pad1[2]; u32 MiscCfg; /* 0x50 */ u32 pad2[2]; u32 PciState; u32 pad3[2]; /* 0x60 */ u32 WinBase; u32 WinData; u32 pad4[12]; /* 0x70 */ u32 DmaWriteState; /* 0xa0 */ u32 pad5[3]; u32 DmaReadState; /* 0xb0 */ u32 pad6[26]; u32 AssistState; u32 pad7[8]; /* 0x120 */ u32 CpuCtrl; /* 0x140 */ u32 Pc; u32 pad8[3]; u32 SramAddr; /* 0x154 */ u32 SramData; u32 pad9[49]; u32 MacRxState; /* 0x220 */ u32 pad10[7]; u32 CpuBCtrl; /* 0x240 */ u32 PcB; u32 pad11[3]; u32 SramBAddr; /* 0x254 */ u32 SramBData; u32 pad12[105]; u32 pad13[32]; /* 0x400 */ u32 Stats[32]; u32 Mb0Hi; /* 0x500 */ u32 Mb0Lo; u32 Mb1Hi; u32 CmdPrd; u32 Mb2Hi; u32 TxPrd; u32 Mb3Hi; u32 RxStdPrd; u32 Mb4Hi; u32 RxJumboPrd; u32 Mb5Hi; u32 RxMiniPrd; u32 Mb6Hi; u32 Mb6Lo; u32 Mb7Hi; u32 Mb7Lo; u32 Mb8Hi; u32 Mb8Lo; u32 Mb9Hi; u32 Mb9Lo; u32 MbAHi; u32 MbALo; u32 MbBHi; u32 MbBLo; u32 MbCHi; u32 MbCLo; u32 MbDHi; u32 MbDLo; u32 MbEHi; u32 MbELo; u32 MbFHi; u32 MbFLo; u32 pad14[32]; u32 MacAddrHi; /* 0x600 */ u32 MacAddrLo; u32 InfoPtrHi; u32 InfoPtrLo; u32 MultiCastHi; /* 0x610 */ u32 MultiCastLo; u32 ModeStat; u32 DmaReadCfg; u32 DmaWriteCfg; /* 0x620 */ u32 TxBufRat; u32 EvtCsm; u32 CmdCsm; u32 TuneRxCoalTicks;/* 0x630 */ u32 TuneTxCoalTicks; u32 TuneStatTicks; u32 TuneMaxTxDesc; u32 TuneMaxRxDesc; /* 0x640 */ u32 TuneTrace; u32 TuneLink; u32 TuneFastLink; u32 TracePtr; /* 0x650 */ u32 TraceStrt; u32 TraceLen; u32 IfIdx; u32 IfMtu; /* 0x660 */ u32 MaskInt; u32 GigLnkState; u32 FastLnkState; u32 pad16[4]; /* 0x670 */ u32 RxRetCsm; /* 0x680 */ u32 pad17[31]; u32 CmdRng[64]; /* 0x700 */ u32 Window[0x200]; }; typedef struct { u32 addrhi; u32 addrlo; } aceaddr; #define ACE_WINDOW_SIZE 0x800 #define ACE_JUMBO_MTU 9000 #define ACE_STD_MTU 1500 #define ACE_TRACE_SIZE 0x8000 /* * Host control register bits. */ #define IN_INT 0x01 #define CLR_INT 0x02 #define HW_RESET 0x08 #define BYTE_SWAP 0x10 #define WORD_SWAP 0x20 #define MASK_INTS 0x40 /* * Local control register bits. */ #define EEPROM_DATA_IN 0x800000 #define EEPROM_DATA_OUT 0x400000 #define EEPROM_WRITE_ENABLE 0x200000 #define EEPROM_CLK_OUT 0x100000 #define EEPROM_BASE 0xa0000000 #define EEPROM_WRITE_SELECT 0xa0 #define EEPROM_READ_SELECT 0xa1 #define SRAM_BANK_512K 0x200 /* * udelay() values for when clocking the eeprom */ #define ACE_SHORT_DELAY 2 #define ACE_LONG_DELAY 4 /* * Misc Config bits */ #define SYNC_SRAM_TIMING 0x100000 /* * CPU state bits. */ #define CPU_RESET 0x01 #define CPU_TRACE 0x02 #define CPU_PROM_FAILED 0x10 #define CPU_HALT 0x00010000 #define CPU_HALTED 0xffff0000 /* * PCI State bits. */ #define DMA_READ_MAX_4 0x04 #define DMA_READ_MAX_16 0x08 #define DMA_READ_MAX_32 0x0c #define DMA_READ_MAX_64 0x10 #define DMA_READ_MAX_128 0x14 #define DMA_READ_MAX_256 0x18 #define DMA_READ_MAX_1K 0x1c #define DMA_WRITE_MAX_4 0x20 #define DMA_WRITE_MAX_16 0x40 #define DMA_WRITE_MAX_32 0x60 #define DMA_WRITE_MAX_64 0x80 #define DMA_WRITE_MAX_128 0xa0 #define DMA_WRITE_MAX_256 0xc0 #define DMA_WRITE_MAX_1K 0xe0 #define DMA_READ_WRITE_MASK 0xfc #define MEM_READ_MULTIPLE 0x00020000 #define PCI_66MHZ 0x00080000 #define PCI_32BIT 0x00100000 #define DMA_WRITE_ALL_ALIGN 0x00800000 #define READ_CMD_MEM 0x06000000 #define WRITE_CMD_MEM 0x70000000 /* * Mode status */ #define ACE_BYTE_SWAP_BD 0x02 #define ACE_WORD_SWAP_BD 0x04 /* not actually used */ #define ACE_WARN 0x08 #define ACE_BYTE_SWAP_DMA 0x10 #define ACE_NO_JUMBO_FRAG 0x200 #define ACE_FATAL 0x40000000 /* * DMA config */ #define DMA_THRESH_1W 0x10 #define DMA_THRESH_2W 0x20 #define DMA_THRESH_4W 0x40 #define DMA_THRESH_8W 0x80 #define DMA_THRESH_16W 0x100 #define DMA_THRESH_32W 0x0 /* not described in doc, but exists. */ /* * Tuning parameters */ #define TICKS_PER_SEC 1000000 /* * Link bits */ #define LNK_PREF 0x00008000 #define LNK_10MB 0x00010000 #define LNK_100MB 0x00020000 #define LNK_1000MB 0x00040000 #define LNK_FULL_DUPLEX 0x00080000 #define LNK_HALF_DUPLEX 0x00100000 #define LNK_TX_FLOW_CTL_Y 0x00200000 #define LNK_NEG_ADVANCED 0x00400000 #define LNK_RX_FLOW_CTL_Y 0x00800000 #define LNK_NIC 0x01000000 #define LNK_JAM 0x02000000 #define LNK_JUMBO 0x04000000 #define LNK_ALTEON 0x08000000 #define LNK_NEG_FCTL 0x10000000 #define LNK_NEGOTIATE 0x20000000 #define LNK_ENABLE 0x40000000 #define LNK_UP 0x80000000 /* * Event definitions */ #define EVT_RING_ENTRIES 256 #define EVT_RING_SIZE (EVT_RING_ENTRIES * sizeof(struct event)) struct event { #ifdef __LITTLE_ENDIAN_BITFIELD u32 idx:12; u32 code:12; u32 evt:8; #else u32 evt:8; u32 code:12; u32 idx:12; #endif u32 pad; }; /* * Events */ #define E_FW_RUNNING 0x01 #define E_STATS_UPDATED 0x04 #define E_STATS_UPDATE 0x04 #define E_LNK_STATE 0x06 #define E_C_LINK_UP 0x01 #define E_C_LINK_DOWN 0x02 #define E_C_LINK_10_100 0x03 #define E_ERROR 0x07 #define E_C_ERR_INVAL_CMD 0x01 #define E_C_ERR_UNIMP_CMD 0x02 #define E_C_ERR_BAD_CFG 0x03 #define E_MCAST_LIST 0x08 #define E_C_MCAST_ADDR_ADD 0x01 #define E_C_MCAST_ADDR_DEL 0x02 #define E_RESET_JUMBO_RNG 0x09 /* * Commands */ #define CMD_RING_ENTRIES 64 struct cmd { #ifdef __LITTLE_ENDIAN_BITFIELD u32 idx:12; u32 code:12; u32 evt:8; #else u32 evt:8; u32 code:12; u32 idx:12; #endif }; #define C_HOST_STATE 0x01 #define C_C_STACK_UP 0x01 #define C_C_STACK_DOWN 0x02 #define C_FDR_FILTERING 0x02 #define C_C_FDR_FILT_ENABLE 0x01 #define C_C_FDR_FILT_DISABLE 0x02 #define C_SET_RX_PRD_IDX 0x03 #define C_UPDATE_STATS 0x04 #define C_RESET_JUMBO_RNG 0x05 #define C_ADD_MULTICAST_ADDR 0x08 #define C_DEL_MULTICAST_ADDR 0x09 #define C_SET_PROMISC_MODE 0x0a #define C_C_PROMISC_ENABLE 0x01 #define C_C_PROMISC_DISABLE 0x02 #define C_LNK_NEGOTIATION 0x0b #define C_C_NEGOTIATE_BOTH 0x00 #define C_C_NEGOTIATE_GIG 0x01 #define C_C_NEGOTIATE_10_100 0x02 #define C_SET_MAC_ADDR 0x0c #define C_CLEAR_PROFILE 0x0d #define C_SET_MULTICAST_MODE 0x0e #define C_C_MCAST_ENABLE 0x01 #define C_C_MCAST_DISABLE 0x02 #define C_CLEAR_STATS 0x0f #define C_SET_RX_JUMBO_PRD_IDX 0x10 #define C_REFRESH_STATS 0x11 /* * Descriptor flags */ #define BD_FLG_TCP_UDP_SUM 0x01 #define BD_FLG_IP_SUM 0x02 #define BD_FLG_END 0x04 #define BD_FLG_MORE 0x08 #define BD_FLG_JUMBO 0x10 #define BD_FLG_UCAST 0x20 #define BD_FLG_MCAST 0x40 #define BD_FLG_BCAST 0x60 #define BD_FLG_TYP_MASK 0x60 #define BD_FLG_IP_FRAG 0x80 #define BD_FLG_IP_FRAG_END 0x100 #define BD_FLG_VLAN_TAG 0x200 #define BD_FLG_FRAME_ERROR 0x400 #define BD_FLG_COAL_NOW 0x800 #define BD_FLG_MINI 0x1000 /* * Ring Control block flags */ #define RCB_FLG_TCP_UDP_SUM 0x01 #define RCB_FLG_IP_SUM 0x02 #define RCB_FLG_NO_PSEUDO_HDR 0x08 #define RCB_FLG_VLAN_ASSIST 0x10 #define RCB_FLG_COAL_INT_ONLY 0x20 #define RCB_FLG_TX_HOST_RING 0x40 #define RCB_FLG_IEEE_SNAP_SUM 0x80 #define RCB_FLG_EXT_RX_BD 0x100 #define RCB_FLG_RNG_DISABLE 0x200 /* * TX ring - maximum TX ring entries for Tigon I's is 128 */ #define MAX_TX_RING_ENTRIES 256 #define TIGON_I_TX_RING_ENTRIES 128 #define TX_RING_SIZE (MAX_TX_RING_ENTRIES * sizeof(struct tx_desc)) #define TX_RING_BASE 0x3800 struct tx_desc{ aceaddr addr; u32 flagsize; #if 0 /* * This is in PCI shared mem and must be accessed with readl/writel * real layout is: */ #if __LITTLE_ENDIAN u16 flags; u16 size; u16 vlan; u16 reserved; #else u16 size; u16 flags; u16 reserved; u16 vlan; #endif #endif u32 vlanres; }; #define RX_STD_RING_ENTRIES 512 #define RX_STD_RING_SIZE (RX_STD_RING_ENTRIES * sizeof(struct rx_desc)) #define RX_JUMBO_RING_ENTRIES 256 #define RX_JUMBO_RING_SIZE (RX_JUMBO_RING_ENTRIES *sizeof(struct rx_desc)) #define RX_MINI_RING_ENTRIES 1024 #define RX_MINI_RING_SIZE (RX_MINI_RING_ENTRIES *sizeof(struct rx_desc)) #define RX_RETURN_RING_ENTRIES 2048 #define RX_RETURN_RING_SIZE (RX_MAX_RETURN_RING_ENTRIES * \ sizeof(struct rx_desc)) struct rx_desc{ aceaddr addr; #ifdef __LITTLE_ENDIAN u16 size; u16 idx; #else u16 idx; u16 size; #endif #ifdef __LITTLE_ENDIAN u16 flags; u16 type; #else u16 type; u16 flags; #endif #ifdef __LITTLE_ENDIAN u16 tcp_udp_csum; u16 ip_csum; #else u16 ip_csum; u16 tcp_udp_csum; #endif #ifdef __LITTLE_ENDIAN u16 vlan; u16 err_flags; #else u16 err_flags; u16 vlan; #endif u32 reserved; u32 opague; }; /* * This struct is shared with the NIC firmware. */ struct ring_ctrl { aceaddr rngptr; #ifdef __LITTLE_ENDIAN u16 flags; u16 max_len; #else u16 max_len; u16 flags; #endif u32 pad; }; struct ace_mac_stats { u32 excess_colls; u32 coll_1; u32 coll_2; u32 coll_3; u32 coll_4; u32 coll_5; u32 coll_6; u32 coll_7; u32 coll_8; u32 coll_9; u32 coll_10; u32 coll_11; u32 coll_12; u32 coll_13; u32 coll_14; u32 coll_15; u32 late_coll; u32 defers; u32 crc_err; u32 underrun; u32 crs_err; u32 pad[3]; u32 drop_ula; u32 drop_mc; u32 drop_fc; u32 drop_space; u32 coll; u32 kept_bc; u32 kept_mc; u32 kept_uc; }; struct ace_info { union { u32 stats[256]; } s; struct ring_ctrl evt_ctrl; struct ring_ctrl cmd_ctrl; struct ring_ctrl tx_ctrl; struct ring_ctrl rx_std_ctrl; struct ring_ctrl rx_jumbo_ctrl; struct ring_ctrl rx_mini_ctrl; struct ring_ctrl rx_return_ctrl; aceaddr evt_prd_ptr; aceaddr rx_ret_prd_ptr; aceaddr tx_csm_ptr; aceaddr stats2_ptr; }; struct ring_info { struct sk_buff *skb; DECLARE_PCI_UNMAP_ADDR(mapping) }; /* * Funny... As soon as we add maplen on alpha, it starts to work * much slower. Hmm... is it because struct does not fit to one cacheline? * So, split tx_ring_info. */ struct tx_ring_info { struct sk_buff *skb; DECLARE_PCI_UNMAP_ADDR(mapping) DECLARE_PCI_UNMAP_LEN(maplen) }; /* * struct ace_skb holding the rings of skb's. This is an awful lot of * pointers, but I don't see any other smart mode to do this in an * efficient manner ;-( */ struct ace_skb { struct tx_ring_info tx_skbuff[MAX_TX_RING_ENTRIES]; struct ring_info rx_std_skbuff[RX_STD_RING_ENTRIES]; struct ring_info rx_mini_skbuff[RX_MINI_RING_ENTRIES]; struct ring_info rx_jumbo_skbuff[RX_JUMBO_RING_ENTRIES]; }; /* * Struct private for the AceNIC. * * Elements are grouped so variables used by the tx handling goes * together, and will go into the same cache lines etc. in order to * avoid cache line contention between the rx and tx handling on SMP. * * Frequently accessed variables are put at the beginning of the * struct to help the compiler generate better/shorter code. */ struct ace_private { struct ace_info *info; struct ace_regs __iomem *regs; /* register base */ struct ace_skb *skb; dma_addr_t info_dma; /* 32/64 bit */ int version, link; int promisc, mcast_all; /* * TX elements */ struct tx_desc *tx_ring; u32 tx_prd; volatile u32 tx_ret_csm; int tx_ring_entries; /* * RX elements */ unsigned long std_refill_busy __attribute__ ((aligned (SMP_CACHE_BYTES))); unsigned long mini_refill_busy, jumbo_refill_busy; atomic_t cur_rx_bufs; atomic_t cur_mini_bufs; atomic_t cur_jumbo_bufs; u32 rx_std_skbprd, rx_mini_skbprd, rx_jumbo_skbprd; u32 cur_rx; struct rx_desc *rx_std_ring; struct rx_desc *rx_jumbo_ring; struct rx_desc *rx_mini_ring; struct rx_desc *rx_return_ring; #if ACENIC_DO_VLAN struct vlan_group *vlgrp; #endif int tasklet_pending, jumbo; struct tasklet_struct ace_tasklet; struct event *evt_ring; volatile u32 *evt_prd, *rx_ret_prd, *tx_csm; dma_addr_t tx_ring_dma; /* 32/64 bit */ dma_addr_t rx_ring_base_dma; dma_addr_t evt_ring_dma; dma_addr_t evt_prd_dma, rx_ret_prd_dma, tx_csm_dma; unsigned char *trace_buf; struct pci_dev *pdev; struct net_device *next; volatile int fw_running; int board_idx; u16 pci_command; u8 pci_latency; const char *name; #ifdef INDEX_DEBUG spinlock_t debug_lock __attribute__ ((aligned (SMP_CACHE_BYTES))); u32 last_tx, last_std_rx, last_mini_rx; #endif struct net_device_stats stats; int pci_using_dac; }; #define TX_RESERVED MAX_SKB_FRAGS static inline int tx_space (struct ace_private *ap, u32 csm, u32 prd) { return (csm - prd - 1) & (ACE_TX_RING_ENTRIES(ap) - 1); } #define tx_free(ap) tx_space((ap)->tx_ret_csm, (ap)->tx_prd, ap) #define tx_ring_full(ap, csm, prd) (tx_space(ap, csm, prd) <= TX_RESERVED) static inline void set_aceaddr(aceaddr *aa, dma_addr_t addr) { u64 baddr = (u64) addr; aa->addrlo = baddr & 0xffffffff; aa->addrhi = baddr >> 32; wmb(); } static inline void ace_set_txprd(struct ace_regs __iomem *regs, struct ace_private *ap, u32 value) { #ifdef INDEX_DEBUG unsigned long flags; spin_lock_irqsave(&ap->debug_lock, flags); writel(value, ®s->TxPrd); if (value == ap->last_tx) printk(KERN_ERR "AceNIC RACE ALERT! writing identical value " "to tx producer (%i)\n", value); ap->last_tx = value; spin_unlock_irqrestore(&ap->debug_lock, flags); #else writel(value, ®s->TxPrd); #endif wmb(); } static inline void ace_mask_irq(struct net_device *dev) { struct ace_private *ap = netdev_priv(dev); struct ace_regs __iomem *regs = ap->regs; if (ACE_IS_TIGON_I(ap)) writel(1, ®s->MaskInt); else writel(readl(®s->HostCtrl) | MASK_INTS, ®s->HostCtrl); ace_sync_irq(dev->irq); } static inline void ace_unmask_irq(struct net_device *dev) { struct ace_private *ap = netdev_priv(dev); struct ace_regs __iomem *regs = ap->regs; if (ACE_IS_TIGON_I(ap)) writel(0, ®s->MaskInt); else writel(readl(®s->HostCtrl) & ~MASK_INTS, ®s->HostCtrl); } /* * Prototypes */ static int ace_init(struct net_device *dev); static void ace_load_std_rx_ring(struct ace_private *ap, int nr_bufs); static void ace_load_mini_rx_ring(struct ace_private *ap, int nr_bufs); static void ace_load_jumbo_rx_ring(struct ace_private *ap, int nr_bufs); static irqreturn_t ace_interrupt(int irq, void *dev_id, struct pt_regs *regs); static int ace_load_firmware(struct net_device *dev); static int ace_open(struct net_device *dev); static int ace_start_xmit(struct sk_buff *skb, struct net_device *dev); static int ace_close(struct net_device *dev); static void ace_tasklet(unsigned long dev); static void ace_dump_trace(struct ace_private *ap); static void ace_set_multicast_list(struct net_device *dev); static int ace_change_mtu(struct net_device *dev, int new_mtu); static int ace_set_mac_addr(struct net_device *dev, void *p); static void ace_set_rxtx_parms(struct net_device *dev, int jumbo); static int ace_allocate_descriptors(struct net_device *dev); static void ace_free_descriptors(struct net_device *dev); static void ace_init_cleanup(struct net_device *dev); static struct net_device_stats *ace_get_stats(struct net_device *dev); static int read_eeprom_byte(struct net_device *dev, unsigned long offset); #if ACENIC_DO_VLAN static void ace_vlan_rx_register(struct net_device *dev, struct vlan_group *grp); static void ace_vlan_rx_kill_vid(struct net_device *dev, unsigned short vid); #endif #endif /* _ACENIC_H_ */