| /* |
| * libcxgbi.h: Chelsio common library for T3/T4 iSCSI driver. |
| * |
| * Copyright (c) 2010 Chelsio Communications, Inc. |
| * |
| * This program is free software; you can redistribute it and/or modify |
| * it under the terms of the GNU General Public License as published by |
| * the Free Software Foundation. |
| * |
| * Written by: Karen Xie (kxie@chelsio.com) |
| * Written by: Rakesh Ranjan (rranjan@chelsio.com) |
| */ |
| |
| #ifndef __LIBCXGBI_H__ |
| #define __LIBCXGBI_H__ |
| |
| #include <linux/kernel.h> |
| #include <linux/errno.h> |
| #include <linux/types.h> |
| #include <linux/debugfs.h> |
| #include <linux/list.h> |
| #include <linux/netdevice.h> |
| #include <linux/if_vlan.h> |
| #include <linux/scatterlist.h> |
| #include <linux/skbuff.h> |
| #include <linux/vmalloc.h> |
| #include <scsi/scsi_device.h> |
| #include <scsi/libiscsi_tcp.h> |
| |
| enum cxgbi_dbg_flag { |
| CXGBI_DBG_ISCSI, |
| CXGBI_DBG_DDP, |
| CXGBI_DBG_TOE, |
| CXGBI_DBG_SOCK, |
| |
| CXGBI_DBG_PDU_TX, |
| CXGBI_DBG_PDU_RX, |
| CXGBI_DBG_DEV, |
| }; |
| |
| #define log_debug(level, fmt, ...) \ |
| do { \ |
| if (dbg_level & (level)) \ |
| pr_info(fmt, ##__VA_ARGS__); \ |
| } while (0) |
| |
| /* max. connections per adapter */ |
| #define CXGBI_MAX_CONN 16384 |
| |
| /* always allocate rooms for AHS */ |
| #define SKB_TX_ISCSI_PDU_HEADER_MAX \ |
| (sizeof(struct iscsi_hdr) + ISCSI_MAX_AHS_SIZE) |
| |
| #define ISCSI_PDU_NONPAYLOAD_LEN 312 /* bhs(48) + ahs(256) + digest(8)*/ |
| |
| /* |
| * align pdu size to multiple of 512 for better performance |
| */ |
| #define cxgbi_align_pdu_size(n) do { n = (n) & (~511); } while (0) |
| |
| #define ULP2_MODE_ISCSI 2 |
| |
| #define ULP2_MAX_PKT_SIZE 16224 |
| #define ULP2_MAX_PDU_PAYLOAD \ |
| (ULP2_MAX_PKT_SIZE - ISCSI_PDU_NONPAYLOAD_LEN) |
| |
| /* |
| * For iscsi connections HW may inserts digest bytes into the pdu. Those digest |
| * bytes are not sent by the host but are part of the TCP payload and therefore |
| * consume TCP sequence space. |
| */ |
| static const unsigned int ulp2_extra_len[] = { 0, 4, 4, 8 }; |
| static inline unsigned int cxgbi_ulp_extra_len(int submode) |
| { |
| return ulp2_extra_len[submode & 3]; |
| } |
| |
| /* |
| * struct pagepod_hdr, pagepod - pagepod format |
| */ |
| |
| #define CPL_RX_DDP_STATUS_DDP_SHIFT 16 /* ddp'able */ |
| #define CPL_RX_DDP_STATUS_PAD_SHIFT 19 /* pad error */ |
| #define CPL_RX_DDP_STATUS_HCRC_SHIFT 20 /* hcrc error */ |
| #define CPL_RX_DDP_STATUS_DCRC_SHIFT 21 /* dcrc error */ |
| |
| struct cxgbi_pagepod_hdr { |
| u32 vld_tid; |
| u32 pgsz_tag_clr; |
| u32 max_offset; |
| u32 page_offset; |
| u64 rsvd; |
| }; |
| |
| #define PPOD_PAGES_MAX 4 |
| struct cxgbi_pagepod { |
| struct cxgbi_pagepod_hdr hdr; |
| u64 addr[PPOD_PAGES_MAX + 1]; |
| }; |
| |
| struct cxgbi_tag_format { |
| unsigned char sw_bits; |
| unsigned char rsvd_bits; |
| unsigned char rsvd_shift; |
| unsigned char filler[1]; |
| u32 rsvd_mask; |
| }; |
| |
| struct cxgbi_gather_list { |
| unsigned int tag; |
| unsigned int length; |
| unsigned int offset; |
| unsigned int nelem; |
| struct page **pages; |
| dma_addr_t phys_addr[0]; |
| }; |
| |
| struct cxgbi_ddp_info { |
| struct kref refcnt; |
| struct cxgbi_device *cdev; |
| struct pci_dev *pdev; |
| unsigned int max_txsz; |
| unsigned int max_rxsz; |
| unsigned int llimit; |
| unsigned int ulimit; |
| unsigned int nppods; |
| unsigned int idx_last; |
| unsigned char idx_bits; |
| unsigned char filler[3]; |
| unsigned int idx_mask; |
| unsigned int rsvd_tag_mask; |
| spinlock_t map_lock; |
| struct cxgbi_gather_list **gl_map; |
| }; |
| |
| #define DDP_PGIDX_MAX 4 |
| #define DDP_THRESHOLD 2048 |
| |
| #define PPOD_PAGES_SHIFT 2 /* 4 pages per pod */ |
| |
| #define PPOD_SIZE sizeof(struct cxgbi_pagepod) /* 64 */ |
| #define PPOD_SIZE_SHIFT 6 |
| |
| #define ULPMEM_DSGL_MAX_NPPODS 16 /* 1024/PPOD_SIZE */ |
| #define ULPMEM_IDATA_MAX_NPPODS 4 /* 256/PPOD_SIZE */ |
| #define PCIE_MEMWIN_MAX_NPPODS 16 /* 1024/PPOD_SIZE */ |
| |
| #define PPOD_COLOR_SHIFT 0 |
| #define PPOD_COLOR(x) ((x) << PPOD_COLOR_SHIFT) |
| |
| #define PPOD_IDX_SHIFT 6 |
| #define PPOD_IDX_MAX_SIZE 24 |
| |
| #define PPOD_TID_SHIFT 0 |
| #define PPOD_TID(x) ((x) << PPOD_TID_SHIFT) |
| |
| #define PPOD_TAG_SHIFT 6 |
| #define PPOD_TAG(x) ((x) << PPOD_TAG_SHIFT) |
| |
| #define PPOD_VALID_SHIFT 24 |
| #define PPOD_VALID(x) ((x) << PPOD_VALID_SHIFT) |
| #define PPOD_VALID_FLAG PPOD_VALID(1U) |
| |
| /* |
| * sge_opaque_hdr - |
| * Opaque version of structure the SGE stores at skb->head of TX_DATA packets |
| * and for which we must reserve space. |
| */ |
| struct sge_opaque_hdr { |
| void *dev; |
| dma_addr_t addr[MAX_SKB_FRAGS + 1]; |
| }; |
| |
| struct cxgbi_sock { |
| struct cxgbi_device *cdev; |
| |
| int tid; |
| int atid; |
| unsigned long flags; |
| unsigned int mtu; |
| unsigned short rss_qid; |
| unsigned short txq_idx; |
| unsigned short advmss; |
| unsigned int tx_chan; |
| unsigned int rx_chan; |
| unsigned int mss_idx; |
| unsigned int smac_idx; |
| unsigned char port_id; |
| int wr_max_cred; |
| int wr_cred; |
| int wr_una_cred; |
| unsigned char hcrc_len; |
| unsigned char dcrc_len; |
| |
| void *l2t; |
| struct sk_buff *wr_pending_head; |
| struct sk_buff *wr_pending_tail; |
| struct sk_buff *cpl_close; |
| struct sk_buff *cpl_abort_req; |
| struct sk_buff *cpl_abort_rpl; |
| struct sk_buff *skb_ulp_lhdr; |
| spinlock_t lock; |
| struct kref refcnt; |
| unsigned int state; |
| struct sockaddr_in saddr; |
| struct sockaddr_in daddr; |
| struct dst_entry *dst; |
| struct sk_buff_head receive_queue; |
| struct sk_buff_head write_queue; |
| struct timer_list retry_timer; |
| int err; |
| rwlock_t callback_lock; |
| void *user_data; |
| |
| u32 rcv_nxt; |
| u32 copied_seq; |
| u32 rcv_wup; |
| u32 snd_nxt; |
| u32 snd_una; |
| u32 write_seq; |
| }; |
| |
| /* |
| * connection states |
| */ |
| enum cxgbi_sock_states{ |
| CTP_CLOSED, |
| CTP_CONNECTING, |
| CTP_ACTIVE_OPEN, |
| CTP_ESTABLISHED, |
| CTP_ACTIVE_CLOSE, |
| CTP_PASSIVE_CLOSE, |
| CTP_CLOSE_WAIT_1, |
| CTP_CLOSE_WAIT_2, |
| CTP_ABORTING, |
| }; |
| |
| /* |
| * Connection flags -- many to track some close related events. |
| */ |
| enum cxgbi_sock_flags { |
| CTPF_ABORT_RPL_RCVD, /*received one ABORT_RPL_RSS message */ |
| CTPF_ABORT_REQ_RCVD, /*received one ABORT_REQ_RSS message */ |
| CTPF_ABORT_RPL_PENDING, /* expecting an abort reply */ |
| CTPF_TX_DATA_SENT, /* already sent a TX_DATA WR */ |
| CTPF_ACTIVE_CLOSE_NEEDED,/* need to be closed */ |
| CTPF_HAS_ATID, /* reserved atid */ |
| CTPF_HAS_TID, /* reserved hw tid */ |
| CTPF_OFFLOAD_DOWN, /* offload function off */ |
| }; |
| |
| struct cxgbi_skb_rx_cb { |
| __u32 ddigest; |
| __u32 pdulen; |
| }; |
| |
| struct cxgbi_skb_tx_cb { |
| void *l2t; |
| struct sk_buff *wr_next; |
| }; |
| |
| enum cxgbi_skcb_flags { |
| SKCBF_TX_NEED_HDR, /* packet needs a header */ |
| SKCBF_RX_COALESCED, /* received whole pdu */ |
| SKCBF_RX_HDR, /* recieved pdu header */ |
| SKCBF_RX_DATA, /* recieved pdu payload */ |
| SKCBF_RX_STATUS, /* recieved ddp status */ |
| SKCBF_RX_DATA_DDPD, /* pdu payload ddp'd */ |
| SKCBF_RX_HCRC_ERR, /* header digest error */ |
| SKCBF_RX_DCRC_ERR, /* data digest error */ |
| SKCBF_RX_PAD_ERR, /* padding byte error */ |
| }; |
| |
| struct cxgbi_skb_cb { |
| unsigned char ulp_mode; |
| unsigned long flags; |
| unsigned int seq; |
| union { |
| struct cxgbi_skb_rx_cb rx; |
| struct cxgbi_skb_tx_cb tx; |
| }; |
| }; |
| |
| #define CXGBI_SKB_CB(skb) ((struct cxgbi_skb_cb *)&((skb)->cb[0])) |
| #define cxgbi_skcb_flags(skb) (CXGBI_SKB_CB(skb)->flags) |
| #define cxgbi_skcb_ulp_mode(skb) (CXGBI_SKB_CB(skb)->ulp_mode) |
| #define cxgbi_skcb_tcp_seq(skb) (CXGBI_SKB_CB(skb)->seq) |
| #define cxgbi_skcb_rx_ddigest(skb) (CXGBI_SKB_CB(skb)->rx.ddigest) |
| #define cxgbi_skcb_rx_pdulen(skb) (CXGBI_SKB_CB(skb)->rx.pdulen) |
| #define cxgbi_skcb_tx_wr_next(skb) (CXGBI_SKB_CB(skb)->tx.wr_next) |
| |
| static inline void cxgbi_skcb_set_flag(struct sk_buff *skb, |
| enum cxgbi_skcb_flags flag) |
| { |
| __set_bit(flag, &(cxgbi_skcb_flags(skb))); |
| } |
| |
| static inline void cxgbi_skcb_clear_flag(struct sk_buff *skb, |
| enum cxgbi_skcb_flags flag) |
| { |
| __clear_bit(flag, &(cxgbi_skcb_flags(skb))); |
| } |
| |
| static inline int cxgbi_skcb_test_flag(struct sk_buff *skb, |
| enum cxgbi_skcb_flags flag) |
| { |
| return test_bit(flag, &(cxgbi_skcb_flags(skb))); |
| } |
| |
| static inline void cxgbi_sock_set_flag(struct cxgbi_sock *csk, |
| enum cxgbi_sock_flags flag) |
| { |
| __set_bit(flag, &csk->flags); |
| log_debug(1 << CXGBI_DBG_SOCK, |
| "csk 0x%p,%u,0x%lx, bit %d.\n", |
| csk, csk->state, csk->flags, flag); |
| } |
| |
| static inline void cxgbi_sock_clear_flag(struct cxgbi_sock *csk, |
| enum cxgbi_sock_flags flag) |
| { |
| __clear_bit(flag, &csk->flags); |
| log_debug(1 << CXGBI_DBG_SOCK, |
| "csk 0x%p,%u,0x%lx, bit %d.\n", |
| csk, csk->state, csk->flags, flag); |
| } |
| |
| static inline int cxgbi_sock_flag(struct cxgbi_sock *csk, |
| enum cxgbi_sock_flags flag) |
| { |
| if (csk == NULL) |
| return 0; |
| return test_bit(flag, &csk->flags); |
| } |
| |
| static inline void cxgbi_sock_set_state(struct cxgbi_sock *csk, int state) |
| { |
| log_debug(1 << CXGBI_DBG_SOCK, |
| "csk 0x%p,%u,0x%lx, state -> %u.\n", |
| csk, csk->state, csk->flags, state); |
| csk->state = state; |
| } |
| |
| static inline void cxgbi_sock_free(struct kref *kref) |
| { |
| struct cxgbi_sock *csk = container_of(kref, |
| struct cxgbi_sock, |
| refcnt); |
| if (csk) { |
| log_debug(1 << CXGBI_DBG_SOCK, |
| "free csk 0x%p, state %u, flags 0x%lx\n", |
| csk, csk->state, csk->flags); |
| kfree(csk); |
| } |
| } |
| |
| static inline void __cxgbi_sock_put(const char *fn, struct cxgbi_sock *csk) |
| { |
| log_debug(1 << CXGBI_DBG_SOCK, |
| "%s, put csk 0x%p, ref %u-1.\n", |
| fn, csk, atomic_read(&csk->refcnt.refcount)); |
| kref_put(&csk->refcnt, cxgbi_sock_free); |
| } |
| #define cxgbi_sock_put(csk) __cxgbi_sock_put(__func__, csk) |
| |
| static inline void __cxgbi_sock_get(const char *fn, struct cxgbi_sock *csk) |
| { |
| log_debug(1 << CXGBI_DBG_SOCK, |
| "%s, get csk 0x%p, ref %u+1.\n", |
| fn, csk, atomic_read(&csk->refcnt.refcount)); |
| kref_get(&csk->refcnt); |
| } |
| #define cxgbi_sock_get(csk) __cxgbi_sock_get(__func__, csk) |
| |
| static inline int cxgbi_sock_is_closing(struct cxgbi_sock *csk) |
| { |
| return csk->state >= CTP_ACTIVE_CLOSE; |
| } |
| |
| static inline int cxgbi_sock_is_established(struct cxgbi_sock *csk) |
| { |
| return csk->state == CTP_ESTABLISHED; |
| } |
| |
| static inline void cxgbi_sock_purge_write_queue(struct cxgbi_sock *csk) |
| { |
| struct sk_buff *skb; |
| |
| while ((skb = __skb_dequeue(&csk->write_queue))) |
| __kfree_skb(skb); |
| } |
| |
| static inline unsigned int cxgbi_sock_compute_wscale(unsigned int win) |
| { |
| unsigned int wscale = 0; |
| |
| while (wscale < 14 && (65535 << wscale) < win) |
| wscale++; |
| return wscale; |
| } |
| |
| static inline struct sk_buff *alloc_wr(int wrlen, int dlen, gfp_t gfp) |
| { |
| struct sk_buff *skb = alloc_skb(wrlen + dlen, gfp); |
| |
| if (skb) { |
| __skb_put(skb, wrlen); |
| memset(skb->head, 0, wrlen + dlen); |
| } else |
| pr_info("alloc cpl wr skb %u+%u, OOM.\n", wrlen, dlen); |
| return skb; |
| } |
| |
| |
| /* |
| * The number of WRs needed for an skb depends on the number of fragments |
| * in the skb and whether it has any payload in its main body. This maps the |
| * length of the gather list represented by an skb into the # of necessary WRs. |
| * The extra two fragments are for iscsi bhs and payload padding. |
| */ |
| #define SKB_WR_LIST_SIZE (MAX_SKB_FRAGS + 2) |
| |
| static inline void cxgbi_sock_reset_wr_list(struct cxgbi_sock *csk) |
| { |
| csk->wr_pending_head = csk->wr_pending_tail = NULL; |
| } |
| |
| static inline void cxgbi_sock_enqueue_wr(struct cxgbi_sock *csk, |
| struct sk_buff *skb) |
| { |
| cxgbi_skcb_tx_wr_next(skb) = NULL; |
| /* |
| * We want to take an extra reference since both us and the driver |
| * need to free the packet before it's really freed. We know there's |
| * just one user currently so we use atomic_set rather than skb_get |
| * to avoid the atomic op. |
| */ |
| atomic_set(&skb->users, 2); |
| |
| if (!csk->wr_pending_head) |
| csk->wr_pending_head = skb; |
| else |
| cxgbi_skcb_tx_wr_next(csk->wr_pending_tail) = skb; |
| csk->wr_pending_tail = skb; |
| } |
| |
| static inline int cxgbi_sock_count_pending_wrs(const struct cxgbi_sock *csk) |
| { |
| int n = 0; |
| const struct sk_buff *skb = csk->wr_pending_head; |
| |
| while (skb) { |
| n += skb->csum; |
| skb = cxgbi_skcb_tx_wr_next(skb); |
| } |
| return n; |
| } |
| |
| static inline struct sk_buff *cxgbi_sock_peek_wr(const struct cxgbi_sock *csk) |
| { |
| return csk->wr_pending_head; |
| } |
| |
| static inline struct sk_buff *cxgbi_sock_dequeue_wr(struct cxgbi_sock *csk) |
| { |
| struct sk_buff *skb = csk->wr_pending_head; |
| |
| if (likely(skb)) { |
| csk->wr_pending_head = cxgbi_skcb_tx_wr_next(skb); |
| cxgbi_skcb_tx_wr_next(skb) = NULL; |
| } |
| return skb; |
| } |
| |
| void cxgbi_sock_check_wr_invariants(const struct cxgbi_sock *); |
| void cxgbi_sock_purge_wr_queue(struct cxgbi_sock *); |
| void cxgbi_sock_skb_entail(struct cxgbi_sock *, struct sk_buff *); |
| void cxgbi_sock_fail_act_open(struct cxgbi_sock *, int); |
| void cxgbi_sock_act_open_req_arp_failure(void *, struct sk_buff *); |
| void cxgbi_sock_closed(struct cxgbi_sock *); |
| void cxgbi_sock_established(struct cxgbi_sock *, unsigned int, unsigned int); |
| void cxgbi_sock_rcv_abort_rpl(struct cxgbi_sock *); |
| void cxgbi_sock_rcv_peer_close(struct cxgbi_sock *); |
| void cxgbi_sock_rcv_close_conn_rpl(struct cxgbi_sock *, u32); |
| void cxgbi_sock_rcv_wr_ack(struct cxgbi_sock *, unsigned int, unsigned int, |
| int); |
| unsigned int cxgbi_sock_select_mss(struct cxgbi_sock *, unsigned int); |
| void cxgbi_sock_free_cpl_skbs(struct cxgbi_sock *); |
| |
| struct cxgbi_hba { |
| struct net_device *ndev; |
| struct net_device *vdev; /* vlan dev */ |
| struct Scsi_Host *shost; |
| struct cxgbi_device *cdev; |
| __be32 ipv4addr; |
| unsigned char port_id; |
| }; |
| |
| struct cxgbi_ports_map { |
| unsigned int max_connect; |
| unsigned int used; |
| unsigned short sport_base; |
| spinlock_t lock; |
| unsigned int next; |
| struct cxgbi_sock **port_csk; |
| }; |
| |
| #define CXGBI_FLAG_DEV_T3 0x1 |
| #define CXGBI_FLAG_DEV_T4 0x2 |
| #define CXGBI_FLAG_ADAPTER_RESET 0x4 |
| #define CXGBI_FLAG_IPV4_SET 0x10 |
| struct cxgbi_device { |
| struct list_head list_head; |
| unsigned int flags; |
| struct net_device **ports; |
| void *lldev; |
| struct cxgbi_hba **hbas; |
| const unsigned short *mtus; |
| unsigned char nmtus; |
| unsigned char nports; |
| struct pci_dev *pdev; |
| struct dentry *debugfs_root; |
| struct iscsi_transport *itp; |
| |
| unsigned int pfvf; |
| unsigned int snd_win; |
| unsigned int rcv_win; |
| unsigned int rx_credit_thres; |
| unsigned int skb_tx_rsvd; |
| unsigned int skb_rx_extra; /* for msg coalesced mode */ |
| unsigned int tx_max_size; |
| unsigned int rx_max_size; |
| struct cxgbi_ports_map pmap; |
| struct cxgbi_tag_format tag_format; |
| struct cxgbi_ddp_info *ddp; |
| |
| void (*dev_ddp_cleanup)(struct cxgbi_device *); |
| int (*csk_ddp_set)(struct cxgbi_sock *, struct cxgbi_pagepod_hdr *, |
| unsigned int, unsigned int, |
| struct cxgbi_gather_list *); |
| void (*csk_ddp_clear)(struct cxgbi_hba *, |
| unsigned int, unsigned int, unsigned int); |
| int (*csk_ddp_setup_digest)(struct cxgbi_sock *, |
| unsigned int, int, int, int); |
| int (*csk_ddp_setup_pgidx)(struct cxgbi_sock *, |
| unsigned int, int, bool); |
| |
| void (*csk_release_offload_resources)(struct cxgbi_sock *); |
| int (*csk_rx_pdu_ready)(struct cxgbi_sock *, struct sk_buff *); |
| u32 (*csk_send_rx_credits)(struct cxgbi_sock *, u32); |
| int (*csk_push_tx_frames)(struct cxgbi_sock *, int); |
| void (*csk_send_abort_req)(struct cxgbi_sock *); |
| void (*csk_send_close_req)(struct cxgbi_sock *); |
| int (*csk_alloc_cpls)(struct cxgbi_sock *); |
| int (*csk_init_act_open)(struct cxgbi_sock *); |
| |
| void *dd_data; |
| }; |
| #define cxgbi_cdev_priv(cdev) ((cdev)->dd_data) |
| |
| struct cxgbi_conn { |
| struct cxgbi_endpoint *cep; |
| struct iscsi_conn *iconn; |
| struct cxgbi_hba *chba; |
| u32 task_idx_bits; |
| }; |
| |
| struct cxgbi_endpoint { |
| struct cxgbi_conn *cconn; |
| struct cxgbi_hba *chba; |
| struct cxgbi_sock *csk; |
| }; |
| |
| #define MAX_PDU_FRAGS ((ULP2_MAX_PDU_PAYLOAD + 512 - 1) / 512) |
| struct cxgbi_task_data { |
| unsigned short nr_frags; |
| skb_frag_t frags[MAX_PDU_FRAGS]; |
| struct sk_buff *skb; |
| unsigned int offset; |
| unsigned int count; |
| unsigned int sgoffset; |
| }; |
| #define iscsi_task_cxgbi_data(task) \ |
| ((task)->dd_data + sizeof(struct iscsi_tcp_task)) |
| |
| static inline int cxgbi_is_ddp_tag(struct cxgbi_tag_format *tformat, u32 tag) |
| { |
| return !(tag & (1 << (tformat->rsvd_bits + tformat->rsvd_shift - 1))); |
| } |
| |
| static inline int cxgbi_sw_tag_usable(struct cxgbi_tag_format *tformat, |
| u32 sw_tag) |
| { |
| sw_tag >>= (32 - tformat->rsvd_bits); |
| return !sw_tag; |
| } |
| |
| static inline u32 cxgbi_set_non_ddp_tag(struct cxgbi_tag_format *tformat, |
| u32 sw_tag) |
| { |
| unsigned char shift = tformat->rsvd_bits + tformat->rsvd_shift - 1; |
| u32 mask = (1 << shift) - 1; |
| |
| if (sw_tag && (sw_tag & ~mask)) { |
| u32 v1 = sw_tag & ((1 << shift) - 1); |
| u32 v2 = (sw_tag >> (shift - 1)) << shift; |
| |
| return v2 | v1 | 1 << shift; |
| } |
| |
| return sw_tag | 1 << shift; |
| } |
| |
| static inline u32 cxgbi_ddp_tag_base(struct cxgbi_tag_format *tformat, |
| u32 sw_tag) |
| { |
| u32 mask = (1 << tformat->rsvd_shift) - 1; |
| |
| if (sw_tag && (sw_tag & ~mask)) { |
| u32 v1 = sw_tag & mask; |
| u32 v2 = sw_tag >> tformat->rsvd_shift; |
| |
| v2 <<= tformat->rsvd_bits + tformat->rsvd_shift; |
| |
| return v2 | v1; |
| } |
| |
| return sw_tag; |
| } |
| |
| static inline u32 cxgbi_tag_rsvd_bits(struct cxgbi_tag_format *tformat, |
| u32 tag) |
| { |
| if (cxgbi_is_ddp_tag(tformat, tag)) |
| return (tag >> tformat->rsvd_shift) & tformat->rsvd_mask; |
| |
| return 0; |
| } |
| |
| static inline u32 cxgbi_tag_nonrsvd_bits(struct cxgbi_tag_format *tformat, |
| u32 tag) |
| { |
| unsigned char shift = tformat->rsvd_bits + tformat->rsvd_shift - 1; |
| u32 v1, v2; |
| |
| if (cxgbi_is_ddp_tag(tformat, tag)) { |
| v1 = tag & ((1 << tformat->rsvd_shift) - 1); |
| v2 = (tag >> (shift + 1)) << tformat->rsvd_shift; |
| } else { |
| u32 mask = (1 << shift) - 1; |
| tag &= ~(1 << shift); |
| v1 = tag & mask; |
| v2 = (tag >> 1) & ~mask; |
| } |
| return v1 | v2; |
| } |
| |
| static inline void *cxgbi_alloc_big_mem(unsigned int size, |
| gfp_t gfp) |
| { |
| void *p = kmalloc(size, gfp); |
| if (!p) |
| p = vmalloc(size); |
| if (p) |
| memset(p, 0, size); |
| return p; |
| } |
| |
| static inline void cxgbi_free_big_mem(void *addr) |
| { |
| if (is_vmalloc_addr(addr)) |
| vfree(addr); |
| else |
| kfree(addr); |
| } |
| |
| static inline void cxgbi_set_iscsi_ipv4(struct cxgbi_hba *chba, __be32 ipaddr) |
| { |
| if (chba->cdev->flags & CXGBI_FLAG_IPV4_SET) |
| chba->ipv4addr = ipaddr; |
| else |
| pr_info("set iscsi ipv4 NOT supported, using %s ipv4.\n", |
| chba->ndev->name); |
| } |
| |
| static inline __be32 cxgbi_get_iscsi_ipv4(struct cxgbi_hba *chba) |
| { |
| return chba->ipv4addr; |
| } |
| |
| struct cxgbi_device *cxgbi_device_register(unsigned int, unsigned int); |
| void cxgbi_device_unregister(struct cxgbi_device *); |
| void cxgbi_device_unregister_all(unsigned int flag); |
| struct cxgbi_device *cxgbi_device_find_by_lldev(void *); |
| int cxgbi_hbas_add(struct cxgbi_device *, unsigned int, unsigned int, |
| struct scsi_host_template *, |
| struct scsi_transport_template *); |
| void cxgbi_hbas_remove(struct cxgbi_device *); |
| |
| int cxgbi_device_portmap_create(struct cxgbi_device *cdev, unsigned int base, |
| unsigned int max_conn); |
| void cxgbi_device_portmap_cleanup(struct cxgbi_device *cdev); |
| |
| void cxgbi_conn_tx_open(struct cxgbi_sock *); |
| void cxgbi_conn_pdu_ready(struct cxgbi_sock *); |
| int cxgbi_conn_alloc_pdu(struct iscsi_task *, u8); |
| int cxgbi_conn_init_pdu(struct iscsi_task *, unsigned int , unsigned int); |
| int cxgbi_conn_xmit_pdu(struct iscsi_task *); |
| |
| void cxgbi_cleanup_task(struct iscsi_task *task); |
| |
| void cxgbi_get_conn_stats(struct iscsi_cls_conn *, struct iscsi_stats *); |
| int cxgbi_set_conn_param(struct iscsi_cls_conn *, |
| enum iscsi_param, char *, int); |
| int cxgbi_get_ep_param(struct iscsi_endpoint *ep, enum iscsi_param, char *); |
| struct iscsi_cls_conn *cxgbi_create_conn(struct iscsi_cls_session *, u32); |
| int cxgbi_bind_conn(struct iscsi_cls_session *, |
| struct iscsi_cls_conn *, u64, int); |
| void cxgbi_destroy_session(struct iscsi_cls_session *); |
| struct iscsi_cls_session *cxgbi_create_session(struct iscsi_endpoint *, |
| u16, u16, u32); |
| int cxgbi_set_host_param(struct Scsi_Host *, |
| enum iscsi_host_param, char *, int); |
| int cxgbi_get_host_param(struct Scsi_Host *, enum iscsi_host_param, char *); |
| struct iscsi_endpoint *cxgbi_ep_connect(struct Scsi_Host *, |
| struct sockaddr *, int); |
| int cxgbi_ep_poll(struct iscsi_endpoint *, int); |
| void cxgbi_ep_disconnect(struct iscsi_endpoint *); |
| |
| int cxgbi_iscsi_init(struct iscsi_transport *, |
| struct scsi_transport_template **); |
| void cxgbi_iscsi_cleanup(struct iscsi_transport *, |
| struct scsi_transport_template **); |
| void cxgbi_parse_pdu_itt(struct iscsi_conn *, itt_t, int *, int *); |
| int cxgbi_ddp_init(struct cxgbi_device *, unsigned int, unsigned int, |
| unsigned int, unsigned int); |
| int cxgbi_ddp_cleanup(struct cxgbi_device *); |
| void cxgbi_ddp_page_size_factor(int *); |
| void cxgbi_ddp_ppod_clear(struct cxgbi_pagepod *); |
| void cxgbi_ddp_ppod_set(struct cxgbi_pagepod *, struct cxgbi_pagepod_hdr *, |
| struct cxgbi_gather_list *, unsigned int); |
| #endif /*__LIBCXGBI_H__*/ |