| /* |
| * Copyright (c) 2006 Chelsio, Inc. All rights reserved. |
| * |
| * This software is available to you under a choice of one of two |
| * licenses. You may choose to be licensed under the terms of the GNU |
| * General Public License (GPL) Version 2, available from the file |
| * COPYING in the main directory of this source tree, or the |
| * OpenIB.org BSD license below: |
| * |
| * Redistribution and use in source and binary forms, with or |
| * without modification, are permitted provided that the following |
| * conditions are met: |
| * |
| * - Redistributions of source code must retain the above |
| * copyright notice, this list of conditions and the following |
| * disclaimer. |
| * |
| * - Redistributions in binary form must reproduce the above |
| * copyright notice, this list of conditions and the following |
| * disclaimer in the documentation and/or other materials |
| * provided with the distribution. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, |
| * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF |
| * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND |
| * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS |
| * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN |
| * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN |
| * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE |
| * SOFTWARE. |
| */ |
| #ifndef __CXIO_WR_H__ |
| #define __CXIO_WR_H__ |
| |
| #include <asm/io.h> |
| #include <linux/pci.h> |
| #include <linux/timer.h> |
| #include "firmware_exports.h" |
| |
| #define T3_MAX_SGE 4 |
| #define T3_MAX_INLINE 64 |
| #define T3_STAG0_PBL_SIZE (2 * T3_MAX_SGE << 3) |
| #define T3_STAG0_MAX_PBE_LEN (128 * 1024 * 1024) |
| #define T3_STAG0_PAGE_SHIFT 15 |
| |
| #define Q_EMPTY(rptr,wptr) ((rptr)==(wptr)) |
| #define Q_FULL(rptr,wptr,size_log2) ( (((wptr)-(rptr))>>(size_log2)) && \ |
| ((rptr)!=(wptr)) ) |
| #define Q_GENBIT(ptr,size_log2) (!(((ptr)>>size_log2)&0x1)) |
| #define Q_FREECNT(rptr,wptr,size_log2) ((1UL<<size_log2)-((wptr)-(rptr))) |
| #define Q_COUNT(rptr,wptr) ((wptr)-(rptr)) |
| #define Q_PTR2IDX(ptr,size_log2) (ptr & ((1UL<<size_log2)-1)) |
| |
| static inline void ring_doorbell(void __iomem *doorbell, u32 qpid) |
| { |
| writel(((1<<31) | qpid), doorbell); |
| } |
| |
| #define SEQ32_GE(x,y) (!( (((u32) (x)) - ((u32) (y))) & 0x80000000 )) |
| |
| enum t3_wr_flags { |
| T3_COMPLETION_FLAG = 0x01, |
| T3_NOTIFY_FLAG = 0x02, |
| T3_SOLICITED_EVENT_FLAG = 0x04, |
| T3_READ_FENCE_FLAG = 0x08, |
| T3_LOCAL_FENCE_FLAG = 0x10 |
| } __attribute__ ((packed)); |
| |
| enum t3_wr_opcode { |
| T3_WR_BP = FW_WROPCODE_RI_BYPASS, |
| T3_WR_SEND = FW_WROPCODE_RI_SEND, |
| T3_WR_WRITE = FW_WROPCODE_RI_RDMA_WRITE, |
| T3_WR_READ = FW_WROPCODE_RI_RDMA_READ, |
| T3_WR_INV_STAG = FW_WROPCODE_RI_LOCAL_INV, |
| T3_WR_BIND = FW_WROPCODE_RI_BIND_MW, |
| T3_WR_RCV = FW_WROPCODE_RI_RECEIVE, |
| T3_WR_INIT = FW_WROPCODE_RI_RDMA_INIT, |
| T3_WR_QP_MOD = FW_WROPCODE_RI_MODIFY_QP, |
| T3_WR_FASTREG = FW_WROPCODE_RI_FASTREGISTER_MR |
| } __attribute__ ((packed)); |
| |
| enum t3_rdma_opcode { |
| T3_RDMA_WRITE, /* IETF RDMAP v1.0 ... */ |
| T3_READ_REQ, |
| T3_READ_RESP, |
| T3_SEND, |
| T3_SEND_WITH_INV, |
| T3_SEND_WITH_SE, |
| T3_SEND_WITH_SE_INV, |
| T3_TERMINATE, |
| T3_RDMA_INIT, /* CHELSIO RI specific ... */ |
| T3_BIND_MW, |
| T3_FAST_REGISTER, |
| T3_LOCAL_INV, |
| T3_QP_MOD, |
| T3_BYPASS, |
| T3_RDMA_READ_REQ_WITH_INV, |
| } __attribute__ ((packed)); |
| |
| static inline enum t3_rdma_opcode wr2opcode(enum t3_wr_opcode wrop) |
| { |
| switch (wrop) { |
| case T3_WR_BP: return T3_BYPASS; |
| case T3_WR_SEND: return T3_SEND; |
| case T3_WR_WRITE: return T3_RDMA_WRITE; |
| case T3_WR_READ: return T3_READ_REQ; |
| case T3_WR_INV_STAG: return T3_LOCAL_INV; |
| case T3_WR_BIND: return T3_BIND_MW; |
| case T3_WR_INIT: return T3_RDMA_INIT; |
| case T3_WR_QP_MOD: return T3_QP_MOD; |
| case T3_WR_FASTREG: return T3_FAST_REGISTER; |
| default: break; |
| } |
| return -1; |
| } |
| |
| |
| /* Work request id */ |
| union t3_wrid { |
| struct { |
| u32 hi; |
| u32 low; |
| } id0; |
| u64 id1; |
| }; |
| |
| #define WRID(wrid) (wrid.id1) |
| #define WRID_GEN(wrid) (wrid.id0.wr_gen) |
| #define WRID_IDX(wrid) (wrid.id0.wr_idx) |
| #define WRID_LO(wrid) (wrid.id0.wr_lo) |
| |
| struct fw_riwrh { |
| __be32 op_seop_flags; |
| __be32 gen_tid_len; |
| }; |
| |
| #define S_FW_RIWR_OP 24 |
| #define M_FW_RIWR_OP 0xff |
| #define V_FW_RIWR_OP(x) ((x) << S_FW_RIWR_OP) |
| #define G_FW_RIWR_OP(x) ((((x) >> S_FW_RIWR_OP)) & M_FW_RIWR_OP) |
| |
| #define S_FW_RIWR_SOPEOP 22 |
| #define M_FW_RIWR_SOPEOP 0x3 |
| #define V_FW_RIWR_SOPEOP(x) ((x) << S_FW_RIWR_SOPEOP) |
| |
| #define S_FW_RIWR_FLAGS 8 |
| #define M_FW_RIWR_FLAGS 0x3fffff |
| #define V_FW_RIWR_FLAGS(x) ((x) << S_FW_RIWR_FLAGS) |
| #define G_FW_RIWR_FLAGS(x) ((((x) >> S_FW_RIWR_FLAGS)) & M_FW_RIWR_FLAGS) |
| |
| #define S_FW_RIWR_TID 8 |
| #define V_FW_RIWR_TID(x) ((x) << S_FW_RIWR_TID) |
| |
| #define S_FW_RIWR_LEN 0 |
| #define V_FW_RIWR_LEN(x) ((x) << S_FW_RIWR_LEN) |
| |
| #define S_FW_RIWR_GEN 31 |
| #define V_FW_RIWR_GEN(x) ((x) << S_FW_RIWR_GEN) |
| |
| struct t3_sge { |
| __be32 stag; |
| __be32 len; |
| __be64 to; |
| }; |
| |
| /* If num_sgle is zero, flit 5+ contains immediate data.*/ |
| struct t3_send_wr { |
| struct fw_riwrh wrh; /* 0 */ |
| union t3_wrid wrid; /* 1 */ |
| |
| u8 rdmaop; /* 2 */ |
| u8 reserved[3]; |
| __be32 rem_stag; |
| __be32 plen; /* 3 */ |
| __be32 num_sgle; |
| struct t3_sge sgl[T3_MAX_SGE]; /* 4+ */ |
| }; |
| |
| #define T3_MAX_FASTREG_DEPTH 10 |
| #define T3_MAX_FASTREG_FRAG 10 |
| |
| struct t3_fastreg_wr { |
| struct fw_riwrh wrh; /* 0 */ |
| union t3_wrid wrid; /* 1 */ |
| __be32 stag; /* 2 */ |
| __be32 len; |
| __be32 va_base_hi; /* 3 */ |
| __be32 va_base_lo_fbo; |
| __be32 page_type_perms; /* 4 */ |
| __be32 reserved1; |
| __be64 pbl_addrs[0]; /* 5+ */ |
| }; |
| |
| /* |
| * If a fastreg wr spans multiple wqes, then the 2nd fragment look like this. |
| */ |
| struct t3_pbl_frag { |
| struct fw_riwrh wrh; /* 0 */ |
| __be64 pbl_addrs[14]; /* 1..14 */ |
| }; |
| |
| #define S_FR_PAGE_COUNT 24 |
| #define M_FR_PAGE_COUNT 0xff |
| #define V_FR_PAGE_COUNT(x) ((x) << S_FR_PAGE_COUNT) |
| #define G_FR_PAGE_COUNT(x) ((((x) >> S_FR_PAGE_COUNT)) & M_FR_PAGE_COUNT) |
| |
| #define S_FR_PAGE_SIZE 16 |
| #define M_FR_PAGE_SIZE 0x1f |
| #define V_FR_PAGE_SIZE(x) ((x) << S_FR_PAGE_SIZE) |
| #define G_FR_PAGE_SIZE(x) ((((x) >> S_FR_PAGE_SIZE)) & M_FR_PAGE_SIZE) |
| |
| #define S_FR_TYPE 8 |
| #define M_FR_TYPE 0x1 |
| #define V_FR_TYPE(x) ((x) << S_FR_TYPE) |
| #define G_FR_TYPE(x) ((((x) >> S_FR_TYPE)) & M_FR_TYPE) |
| |
| #define S_FR_PERMS 0 |
| #define M_FR_PERMS 0xff |
| #define V_FR_PERMS(x) ((x) << S_FR_PERMS) |
| #define G_FR_PERMS(x) ((((x) >> S_FR_PERMS)) & M_FR_PERMS) |
| |
| struct t3_local_inv_wr { |
| struct fw_riwrh wrh; /* 0 */ |
| union t3_wrid wrid; /* 1 */ |
| __be32 stag; /* 2 */ |
| __be32 reserved; |
| }; |
| |
| struct t3_rdma_write_wr { |
| struct fw_riwrh wrh; /* 0 */ |
| union t3_wrid wrid; /* 1 */ |
| u8 rdmaop; /* 2 */ |
| u8 reserved[3]; |
| __be32 stag_sink; |
| __be64 to_sink; /* 3 */ |
| __be32 plen; /* 4 */ |
| __be32 num_sgle; |
| struct t3_sge sgl[T3_MAX_SGE]; /* 5+ */ |
| }; |
| |
| struct t3_rdma_read_wr { |
| struct fw_riwrh wrh; /* 0 */ |
| union t3_wrid wrid; /* 1 */ |
| u8 rdmaop; /* 2 */ |
| u8 local_inv; |
| u8 reserved[2]; |
| __be32 rem_stag; |
| __be64 rem_to; /* 3 */ |
| __be32 local_stag; /* 4 */ |
| __be32 local_len; |
| __be64 local_to; /* 5 */ |
| }; |
| |
| struct t3_bind_mw_wr { |
| struct fw_riwrh wrh; /* 0 */ |
| union t3_wrid wrid; /* 1 */ |
| u16 reserved; /* 2 */ |
| u8 type; |
| u8 perms; |
| __be32 mr_stag; |
| __be32 mw_stag; /* 3 */ |
| __be32 mw_len; |
| __be64 mw_va; /* 4 */ |
| __be32 mr_pbl_addr; /* 5 */ |
| u8 reserved2[3]; |
| u8 mr_pagesz; |
| }; |
| |
| struct t3_receive_wr { |
| struct fw_riwrh wrh; /* 0 */ |
| union t3_wrid wrid; /* 1 */ |
| u8 pagesz[T3_MAX_SGE]; |
| __be32 num_sgle; /* 2 */ |
| struct t3_sge sgl[T3_MAX_SGE]; /* 3+ */ |
| __be32 pbl_addr[T3_MAX_SGE]; |
| }; |
| |
| struct t3_bypass_wr { |
| struct fw_riwrh wrh; |
| union t3_wrid wrid; /* 1 */ |
| }; |
| |
| struct t3_modify_qp_wr { |
| struct fw_riwrh wrh; /* 0 */ |
| union t3_wrid wrid; /* 1 */ |
| __be32 flags; /* 2 */ |
| __be32 quiesce; /* 2 */ |
| __be32 max_ird; /* 3 */ |
| __be32 max_ord; /* 3 */ |
| __be64 sge_cmd; /* 4 */ |
| __be64 ctx1; /* 5 */ |
| __be64 ctx0; /* 6 */ |
| }; |
| |
| enum t3_modify_qp_flags { |
| MODQP_QUIESCE = 0x01, |
| MODQP_MAX_IRD = 0x02, |
| MODQP_MAX_ORD = 0x04, |
| MODQP_WRITE_EC = 0x08, |
| MODQP_READ_EC = 0x10, |
| }; |
| |
| |
| enum t3_mpa_attrs { |
| uP_RI_MPA_RX_MARKER_ENABLE = 0x1, |
| uP_RI_MPA_TX_MARKER_ENABLE = 0x2, |
| uP_RI_MPA_CRC_ENABLE = 0x4, |
| uP_RI_MPA_IETF_ENABLE = 0x8 |
| } __attribute__ ((packed)); |
| |
| enum t3_qp_caps { |
| uP_RI_QP_RDMA_READ_ENABLE = 0x01, |
| uP_RI_QP_RDMA_WRITE_ENABLE = 0x02, |
| uP_RI_QP_BIND_ENABLE = 0x04, |
| uP_RI_QP_FAST_REGISTER_ENABLE = 0x08, |
| uP_RI_QP_STAG0_ENABLE = 0x10 |
| } __attribute__ ((packed)); |
| |
| enum rdma_init_rtr_types { |
| RTR_READ = 1, |
| RTR_WRITE = 2, |
| RTR_SEND = 3, |
| }; |
| |
| #define S_RTR_TYPE 2 |
| #define M_RTR_TYPE 0x3 |
| #define V_RTR_TYPE(x) ((x) << S_RTR_TYPE) |
| #define G_RTR_TYPE(x) ((((x) >> S_RTR_TYPE)) & M_RTR_TYPE) |
| |
| #define S_CHAN 4 |
| #define M_CHAN 0x3 |
| #define V_CHAN(x) ((x) << S_CHAN) |
| #define G_CHAN(x) ((((x) >> S_CHAN)) & M_CHAN) |
| |
| struct t3_rdma_init_attr { |
| u32 tid; |
| u32 qpid; |
| u32 pdid; |
| u32 scqid; |
| u32 rcqid; |
| u32 rq_addr; |
| u32 rq_size; |
| enum t3_mpa_attrs mpaattrs; |
| enum t3_qp_caps qpcaps; |
| u16 tcp_emss; |
| u32 ord; |
| u32 ird; |
| u64 qp_dma_addr; |
| u32 qp_dma_size; |
| enum rdma_init_rtr_types rtr_type; |
| u16 flags; |
| u16 rqe_count; |
| u32 irs; |
| u32 chan; |
| }; |
| |
| struct t3_rdma_init_wr { |
| struct fw_riwrh wrh; /* 0 */ |
| union t3_wrid wrid; /* 1 */ |
| __be32 qpid; /* 2 */ |
| __be32 pdid; |
| __be32 scqid; /* 3 */ |
| __be32 rcqid; |
| __be32 rq_addr; /* 4 */ |
| __be32 rq_size; |
| u8 mpaattrs; /* 5 */ |
| u8 qpcaps; |
| __be16 ulpdu_size; |
| __be16 flags_rtr_type; |
| __be16 rqe_count; |
| __be32 ord; /* 6 */ |
| __be32 ird; |
| __be64 qp_dma_addr; /* 7 */ |
| __be32 qp_dma_size; /* 8 */ |
| __be32 irs; |
| }; |
| |
| struct t3_genbit { |
| u64 flit[15]; |
| __be64 genbit; |
| }; |
| |
| struct t3_wq_in_err { |
| u64 flit[13]; |
| u64 err; |
| }; |
| |
| enum rdma_init_wr_flags { |
| MPA_INITIATOR = (1<<0), |
| PRIV_QP = (1<<1), |
| }; |
| |
| union t3_wr { |
| struct t3_send_wr send; |
| struct t3_rdma_write_wr write; |
| struct t3_rdma_read_wr read; |
| struct t3_receive_wr recv; |
| struct t3_fastreg_wr fastreg; |
| struct t3_pbl_frag pbl_frag; |
| struct t3_local_inv_wr local_inv; |
| struct t3_bind_mw_wr bind; |
| struct t3_bypass_wr bypass; |
| struct t3_rdma_init_wr init; |
| struct t3_modify_qp_wr qp_mod; |
| struct t3_genbit genbit; |
| struct t3_wq_in_err wq_in_err; |
| __be64 flit[16]; |
| }; |
| |
| #define T3_SQ_CQE_FLIT 13 |
| #define T3_SQ_COOKIE_FLIT 14 |
| |
| #define T3_RQ_COOKIE_FLIT 13 |
| #define T3_RQ_CQE_FLIT 14 |
| |
| static inline enum t3_wr_opcode fw_riwrh_opcode(struct fw_riwrh *wqe) |
| { |
| return G_FW_RIWR_OP(be32_to_cpu(wqe->op_seop_flags)); |
| } |
| |
| enum t3_wr_hdr_bits { |
| T3_EOP = 1, |
| T3_SOP = 2, |
| T3_SOPEOP = T3_EOP|T3_SOP, |
| }; |
| |
| static inline void build_fw_riwrh(struct fw_riwrh *wqe, enum t3_wr_opcode op, |
| enum t3_wr_flags flags, u8 genbit, u32 tid, |
| u8 len, u8 sopeop) |
| { |
| wqe->op_seop_flags = cpu_to_be32(V_FW_RIWR_OP(op) | |
| V_FW_RIWR_SOPEOP(sopeop) | |
| V_FW_RIWR_FLAGS(flags)); |
| wmb(); |
| wqe->gen_tid_len = cpu_to_be32(V_FW_RIWR_GEN(genbit) | |
| V_FW_RIWR_TID(tid) | |
| V_FW_RIWR_LEN(len)); |
| /* 2nd gen bit... */ |
| ((union t3_wr *)wqe)->genbit.genbit = cpu_to_be64(genbit); |
| } |
| |
| /* |
| * T3 ULP2_TX commands |
| */ |
| enum t3_utx_mem_op { |
| T3_UTX_MEM_READ = 2, |
| T3_UTX_MEM_WRITE = 3 |
| }; |
| |
| /* T3 MC7 RDMA TPT entry format */ |
| |
| enum tpt_mem_type { |
| TPT_NON_SHARED_MR = 0x0, |
| TPT_SHARED_MR = 0x1, |
| TPT_MW = 0x2, |
| TPT_MW_RELAXED_PROTECTION = 0x3 |
| }; |
| |
| enum tpt_addr_type { |
| TPT_ZBTO = 0, |
| TPT_VATO = 1 |
| }; |
| |
| enum tpt_mem_perm { |
| TPT_MW_BIND = 0x10, |
| TPT_LOCAL_READ = 0x8, |
| TPT_LOCAL_WRITE = 0x4, |
| TPT_REMOTE_READ = 0x2, |
| TPT_REMOTE_WRITE = 0x1 |
| }; |
| |
| struct tpt_entry { |
| __be32 valid_stag_pdid; |
| __be32 flags_pagesize_qpid; |
| |
| __be32 rsvd_pbl_addr; |
| __be32 len; |
| __be32 va_hi; |
| __be32 va_low_or_fbo; |
| |
| __be32 rsvd_bind_cnt_or_pstag; |
| __be32 rsvd_pbl_size; |
| }; |
| |
| #define S_TPT_VALID 31 |
| #define V_TPT_VALID(x) ((x) << S_TPT_VALID) |
| #define F_TPT_VALID V_TPT_VALID(1U) |
| |
| #define S_TPT_STAG_KEY 23 |
| #define M_TPT_STAG_KEY 0xFF |
| #define V_TPT_STAG_KEY(x) ((x) << S_TPT_STAG_KEY) |
| #define G_TPT_STAG_KEY(x) (((x) >> S_TPT_STAG_KEY) & M_TPT_STAG_KEY) |
| |
| #define S_TPT_STAG_STATE 22 |
| #define V_TPT_STAG_STATE(x) ((x) << S_TPT_STAG_STATE) |
| #define F_TPT_STAG_STATE V_TPT_STAG_STATE(1U) |
| |
| #define S_TPT_STAG_TYPE 20 |
| #define M_TPT_STAG_TYPE 0x3 |
| #define V_TPT_STAG_TYPE(x) ((x) << S_TPT_STAG_TYPE) |
| #define G_TPT_STAG_TYPE(x) (((x) >> S_TPT_STAG_TYPE) & M_TPT_STAG_TYPE) |
| |
| #define S_TPT_PDID 0 |
| #define M_TPT_PDID 0xFFFFF |
| #define V_TPT_PDID(x) ((x) << S_TPT_PDID) |
| #define G_TPT_PDID(x) (((x) >> S_TPT_PDID) & M_TPT_PDID) |
| |
| #define S_TPT_PERM 28 |
| #define M_TPT_PERM 0xF |
| #define V_TPT_PERM(x) ((x) << S_TPT_PERM) |
| #define G_TPT_PERM(x) (((x) >> S_TPT_PERM) & M_TPT_PERM) |
| |
| #define S_TPT_REM_INV_DIS 27 |
| #define V_TPT_REM_INV_DIS(x) ((x) << S_TPT_REM_INV_DIS) |
| #define F_TPT_REM_INV_DIS V_TPT_REM_INV_DIS(1U) |
| |
| #define S_TPT_ADDR_TYPE 26 |
| #define V_TPT_ADDR_TYPE(x) ((x) << S_TPT_ADDR_TYPE) |
| #define F_TPT_ADDR_TYPE V_TPT_ADDR_TYPE(1U) |
| |
| #define S_TPT_MW_BIND_ENABLE 25 |
| #define V_TPT_MW_BIND_ENABLE(x) ((x) << S_TPT_MW_BIND_ENABLE) |
| #define F_TPT_MW_BIND_ENABLE V_TPT_MW_BIND_ENABLE(1U) |
| |
| #define S_TPT_PAGE_SIZE 20 |
| #define M_TPT_PAGE_SIZE 0x1F |
| #define V_TPT_PAGE_SIZE(x) ((x) << S_TPT_PAGE_SIZE) |
| #define G_TPT_PAGE_SIZE(x) (((x) >> S_TPT_PAGE_SIZE) & M_TPT_PAGE_SIZE) |
| |
| #define S_TPT_PBL_ADDR 0 |
| #define M_TPT_PBL_ADDR 0x1FFFFFFF |
| #define V_TPT_PBL_ADDR(x) ((x) << S_TPT_PBL_ADDR) |
| #define G_TPT_PBL_ADDR(x) (((x) >> S_TPT_PBL_ADDR) & M_TPT_PBL_ADDR) |
| |
| #define S_TPT_QPID 0 |
| #define M_TPT_QPID 0xFFFFF |
| #define V_TPT_QPID(x) ((x) << S_TPT_QPID) |
| #define G_TPT_QPID(x) (((x) >> S_TPT_QPID) & M_TPT_QPID) |
| |
| #define S_TPT_PSTAG 0 |
| #define M_TPT_PSTAG 0xFFFFFF |
| #define V_TPT_PSTAG(x) ((x) << S_TPT_PSTAG) |
| #define G_TPT_PSTAG(x) (((x) >> S_TPT_PSTAG) & M_TPT_PSTAG) |
| |
| #define S_TPT_PBL_SIZE 0 |
| #define M_TPT_PBL_SIZE 0xFFFFF |
| #define V_TPT_PBL_SIZE(x) ((x) << S_TPT_PBL_SIZE) |
| #define G_TPT_PBL_SIZE(x) (((x) >> S_TPT_PBL_SIZE) & M_TPT_PBL_SIZE) |
| |
| /* |
| * CQE defs |
| */ |
| struct t3_cqe { |
| __be32 header; |
| __be32 len; |
| union { |
| struct { |
| __be32 stag; |
| __be32 msn; |
| } rcqe; |
| struct { |
| u32 wrid_hi; |
| u32 wrid_low; |
| } scqe; |
| } u; |
| }; |
| |
| #define S_CQE_OOO 31 |
| #define M_CQE_OOO 0x1 |
| #define G_CQE_OOO(x) ((((x) >> S_CQE_OOO)) & M_CQE_OOO) |
| #define V_CEQ_OOO(x) ((x)<<S_CQE_OOO) |
| |
| #define S_CQE_QPID 12 |
| #define M_CQE_QPID 0x7FFFF |
| #define G_CQE_QPID(x) ((((x) >> S_CQE_QPID)) & M_CQE_QPID) |
| #define V_CQE_QPID(x) ((x)<<S_CQE_QPID) |
| |
| #define S_CQE_SWCQE 11 |
| #define M_CQE_SWCQE 0x1 |
| #define G_CQE_SWCQE(x) ((((x) >> S_CQE_SWCQE)) & M_CQE_SWCQE) |
| #define V_CQE_SWCQE(x) ((x)<<S_CQE_SWCQE) |
| |
| #define S_CQE_GENBIT 10 |
| #define M_CQE_GENBIT 0x1 |
| #define G_CQE_GENBIT(x) (((x) >> S_CQE_GENBIT) & M_CQE_GENBIT) |
| #define V_CQE_GENBIT(x) ((x)<<S_CQE_GENBIT) |
| |
| #define S_CQE_STATUS 5 |
| #define M_CQE_STATUS 0x1F |
| #define G_CQE_STATUS(x) ((((x) >> S_CQE_STATUS)) & M_CQE_STATUS) |
| #define V_CQE_STATUS(x) ((x)<<S_CQE_STATUS) |
| |
| #define S_CQE_TYPE 4 |
| #define M_CQE_TYPE 0x1 |
| #define G_CQE_TYPE(x) ((((x) >> S_CQE_TYPE)) & M_CQE_TYPE) |
| #define V_CQE_TYPE(x) ((x)<<S_CQE_TYPE) |
| |
| #define S_CQE_OPCODE 0 |
| #define M_CQE_OPCODE 0xF |
| #define G_CQE_OPCODE(x) ((((x) >> S_CQE_OPCODE)) & M_CQE_OPCODE) |
| #define V_CQE_OPCODE(x) ((x)<<S_CQE_OPCODE) |
| |
| #define SW_CQE(x) (G_CQE_SWCQE(be32_to_cpu((x).header))) |
| #define CQE_OOO(x) (G_CQE_OOO(be32_to_cpu((x).header))) |
| #define CQE_QPID(x) (G_CQE_QPID(be32_to_cpu((x).header))) |
| #define CQE_GENBIT(x) (G_CQE_GENBIT(be32_to_cpu((x).header))) |
| #define CQE_TYPE(x) (G_CQE_TYPE(be32_to_cpu((x).header))) |
| #define SQ_TYPE(x) (CQE_TYPE((x))) |
| #define RQ_TYPE(x) (!CQE_TYPE((x))) |
| #define CQE_STATUS(x) (G_CQE_STATUS(be32_to_cpu((x).header))) |
| #define CQE_OPCODE(x) (G_CQE_OPCODE(be32_to_cpu((x).header))) |
| |
| #define CQE_SEND_OPCODE(x)( \ |
| (G_CQE_OPCODE(be32_to_cpu((x).header)) == T3_SEND) || \ |
| (G_CQE_OPCODE(be32_to_cpu((x).header)) == T3_SEND_WITH_SE) || \ |
| (G_CQE_OPCODE(be32_to_cpu((x).header)) == T3_SEND_WITH_INV) || \ |
| (G_CQE_OPCODE(be32_to_cpu((x).header)) == T3_SEND_WITH_SE_INV)) |
| |
| #define CQE_LEN(x) (be32_to_cpu((x).len)) |
| |
| /* used for RQ completion processing */ |
| #define CQE_WRID_STAG(x) (be32_to_cpu((x).u.rcqe.stag)) |
| #define CQE_WRID_MSN(x) (be32_to_cpu((x).u.rcqe.msn)) |
| |
| /* used for SQ completion processing */ |
| #define CQE_WRID_SQ_WPTR(x) ((x).u.scqe.wrid_hi) |
| #define CQE_WRID_WPTR(x) ((x).u.scqe.wrid_low) |
| |
| /* generic accessor macros */ |
| #define CQE_WRID_HI(x) ((x).u.scqe.wrid_hi) |
| #define CQE_WRID_LOW(x) ((x).u.scqe.wrid_low) |
| |
| #define TPT_ERR_SUCCESS 0x0 |
| #define TPT_ERR_STAG 0x1 /* STAG invalid: either the */ |
| /* STAG is offlimt, being 0, */ |
| /* or STAG_key mismatch */ |
| #define TPT_ERR_PDID 0x2 /* PDID mismatch */ |
| #define TPT_ERR_QPID 0x3 /* QPID mismatch */ |
| #define TPT_ERR_ACCESS 0x4 /* Invalid access right */ |
| #define TPT_ERR_WRAP 0x5 /* Wrap error */ |
| #define TPT_ERR_BOUND 0x6 /* base and bounds voilation */ |
| #define TPT_ERR_INVALIDATE_SHARED_MR 0x7 /* attempt to invalidate a */ |
| /* shared memory region */ |
| #define TPT_ERR_INVALIDATE_MR_WITH_MW_BOUND 0x8 /* attempt to invalidate a */ |
| /* shared memory region */ |
| #define TPT_ERR_ECC 0x9 /* ECC error detected */ |
| #define TPT_ERR_ECC_PSTAG 0xA /* ECC error detected when */ |
| /* reading PSTAG for a MW */ |
| /* Invalidate */ |
| #define TPT_ERR_PBL_ADDR_BOUND 0xB /* pbl addr out of bounds: */ |
| /* software error */ |
| #define TPT_ERR_SWFLUSH 0xC /* SW FLUSHED */ |
| #define TPT_ERR_CRC 0x10 /* CRC error */ |
| #define TPT_ERR_MARKER 0x11 /* Marker error */ |
| #define TPT_ERR_PDU_LEN_ERR 0x12 /* invalid PDU length */ |
| #define TPT_ERR_OUT_OF_RQE 0x13 /* out of RQE */ |
| #define TPT_ERR_DDP_VERSION 0x14 /* wrong DDP version */ |
| #define TPT_ERR_RDMA_VERSION 0x15 /* wrong RDMA version */ |
| #define TPT_ERR_OPCODE 0x16 /* invalid rdma opcode */ |
| #define TPT_ERR_DDP_QUEUE_NUM 0x17 /* invalid ddp queue number */ |
| #define TPT_ERR_MSN 0x18 /* MSN error */ |
| #define TPT_ERR_TBIT 0x19 /* tag bit not set correctly */ |
| #define TPT_ERR_MO 0x1A /* MO not 0 for TERMINATE */ |
| /* or READ_REQ */ |
| #define TPT_ERR_MSN_GAP 0x1B |
| #define TPT_ERR_MSN_RANGE 0x1C |
| #define TPT_ERR_IRD_OVERFLOW 0x1D |
| #define TPT_ERR_RQE_ADDR_BOUND 0x1E /* RQE addr out of bounds: */ |
| /* software error */ |
| #define TPT_ERR_INTERNAL_ERR 0x1F /* internal error (opcode */ |
| /* mismatch) */ |
| |
| struct t3_swsq { |
| __u64 wr_id; |
| struct t3_cqe cqe; |
| __u32 sq_wptr; |
| __be32 read_len; |
| int opcode; |
| int complete; |
| int signaled; |
| }; |
| |
| struct t3_swrq { |
| __u64 wr_id; |
| __u32 pbl_addr; |
| }; |
| |
| /* |
| * A T3 WQ implements both the SQ and RQ. |
| */ |
| struct t3_wq { |
| union t3_wr *queue; /* DMA accessable memory */ |
| dma_addr_t dma_addr; /* DMA address for HW */ |
| DEFINE_DMA_UNMAP_ADDR(mapping); /* unmap kruft */ |
| u32 error; /* 1 once we go to ERROR */ |
| u32 qpid; |
| u32 wptr; /* idx to next available WR slot */ |
| u32 size_log2; /* total wq size */ |
| struct t3_swsq *sq; /* SW SQ */ |
| struct t3_swsq *oldest_read; /* tracks oldest pending read */ |
| u32 sq_wptr; /* sq_wptr - sq_rptr == count of */ |
| u32 sq_rptr; /* pending wrs */ |
| u32 sq_size_log2; /* sq size */ |
| struct t3_swrq *rq; /* SW RQ (holds consumer wr_ids */ |
| u32 rq_wptr; /* rq_wptr - rq_rptr == count of */ |
| u32 rq_rptr; /* pending wrs */ |
| struct t3_swrq *rq_oldest_wr; /* oldest wr on the SW RQ */ |
| u32 rq_size_log2; /* rq size */ |
| u32 rq_addr; /* rq adapter address */ |
| void __iomem *doorbell; /* kernel db */ |
| u64 udb; /* user db if any */ |
| struct cxio_rdev *rdev; |
| }; |
| |
| struct t3_cq { |
| u32 cqid; |
| u32 rptr; |
| u32 wptr; |
| u32 size_log2; |
| dma_addr_t dma_addr; |
| DEFINE_DMA_UNMAP_ADDR(mapping); |
| struct t3_cqe *queue; |
| struct t3_cqe *sw_queue; |
| u32 sw_rptr; |
| u32 sw_wptr; |
| }; |
| |
| #define CQ_VLD_ENTRY(ptr,size_log2,cqe) (Q_GENBIT(ptr,size_log2) == \ |
| CQE_GENBIT(*cqe)) |
| |
| struct t3_cq_status_page { |
| u32 cq_err; |
| }; |
| |
| static inline int cxio_cq_in_error(struct t3_cq *cq) |
| { |
| return ((struct t3_cq_status_page *) |
| &cq->queue[1 << cq->size_log2])->cq_err; |
| } |
| |
| static inline void cxio_set_cq_in_error(struct t3_cq *cq) |
| { |
| ((struct t3_cq_status_page *) |
| &cq->queue[1 << cq->size_log2])->cq_err = 1; |
| } |
| |
| static inline void cxio_set_wq_in_error(struct t3_wq *wq) |
| { |
| wq->queue->wq_in_err.err |= 1; |
| } |
| |
| static inline void cxio_disable_wq_db(struct t3_wq *wq) |
| { |
| wq->queue->wq_in_err.err |= 2; |
| } |
| |
| static inline void cxio_enable_wq_db(struct t3_wq *wq) |
| { |
| wq->queue->wq_in_err.err &= ~2; |
| } |
| |
| static inline int cxio_wq_db_enabled(struct t3_wq *wq) |
| { |
| return !(wq->queue->wq_in_err.err & 2); |
| } |
| |
| static inline struct t3_cqe *cxio_next_hw_cqe(struct t3_cq *cq) |
| { |
| struct t3_cqe *cqe; |
| |
| cqe = cq->queue + (Q_PTR2IDX(cq->rptr, cq->size_log2)); |
| if (CQ_VLD_ENTRY(cq->rptr, cq->size_log2, cqe)) |
| return cqe; |
| return NULL; |
| } |
| |
| static inline struct t3_cqe *cxio_next_sw_cqe(struct t3_cq *cq) |
| { |
| struct t3_cqe *cqe; |
| |
| if (!Q_EMPTY(cq->sw_rptr, cq->sw_wptr)) { |
| cqe = cq->sw_queue + (Q_PTR2IDX(cq->sw_rptr, cq->size_log2)); |
| return cqe; |
| } |
| return NULL; |
| } |
| |
| static inline struct t3_cqe *cxio_next_cqe(struct t3_cq *cq) |
| { |
| struct t3_cqe *cqe; |
| |
| if (!Q_EMPTY(cq->sw_rptr, cq->sw_wptr)) { |
| cqe = cq->sw_queue + (Q_PTR2IDX(cq->sw_rptr, cq->size_log2)); |
| return cqe; |
| } |
| cqe = cq->queue + (Q_PTR2IDX(cq->rptr, cq->size_log2)); |
| if (CQ_VLD_ENTRY(cq->rptr, cq->size_log2, cqe)) |
| return cqe; |
| return NULL; |
| } |
| |
| #endif |