/drivers/infiniband/hw/ipath/ |
H A D | ipath_keys.c | 113 * @sge: SGE to check 122 struct ib_sge *sge, int acc) 134 if (sge->lkey == 0) { 143 isge->vaddr = (void *) sge->addr; 144 isge->length = sge->length; 145 isge->sge_length = sge->length; 149 mr = rkt->table[(sge->lkey >> (32 - ib_ipath_lkey_table_size))]; 150 if (unlikely(mr == NULL || mr->lkey != sge->lkey || 156 off = sge->addr - mr->user_base; 157 if (unlikely(sge 121 ipath_lkey_ok(struct ipath_qp *qp, struct ipath_sge *isge, struct ib_sge *sge, int acc) argument 204 struct ipath_sge *sge = &ss->sge; local [all...] |
H A D | ipath_ruc.c | 134 if (!ipath_lkey_ok(qp, j ? &ss->sg_list[j - 1] : &ss->sge, 265 struct ipath_sge *sge; local 328 sqp->s_sge.sge = wqe->sg_list[0]; 370 qp->r_sge.sge = wqe->sg_list[0]; 385 maddr = (atomic64_t *) qp->r_sge.sge.vaddr; 387 *(u64 *) sqp->s_sge.sge.vaddr = 390 (u64) cmpxchg((u64 *) qp->r_sge.sge.vaddr, 399 sge = &sqp->s_sge.sge; 403 if (len > sge [all...] |
H A D | ipath_sdma.c | 670 struct ipath_sge *sge; local 729 sge = &ss->sge; 735 if (len > sge->length) 736 len = sge->length; 737 if (len > sge->sge_length) 738 len = sge->sge_length; 741 addr = dma_map_single(&dd->pcidev->dev, sge->vaddr, dw << 2, 759 sge->vaddr += len; 760 sge [all...] |
H A D | ipath_ud.c | 59 struct ipath_sge *sge; local 184 sge = swqe->sg_list; 186 u32 len = sge->length; 190 if (len > sge->sge_length) 191 len = sge->sge_length; 193 ipath_copy_sge(&rsge, sge->vaddr, len); 194 sge->vaddr += len; 195 sge->length -= len; 196 sge->sge_length -= len; 197 if (sge [all...] |
H A D | ipath_verbs.c | 173 struct ipath_sge *sge = &ss->sge; local 176 u32 len = sge->length; 180 if (len > sge->sge_length) 181 len = sge->sge_length; 183 memcpy(sge->vaddr, data, len); 184 sge->vaddr += len; 185 sge->length -= len; 186 sge->sge_length -= len; 187 if (sge 213 struct ipath_sge *sge = &ss->sge; local 252 struct ipath_sge sge = ss->sge; local 298 struct ipath_sge *sge = &ss->sge; local 782 struct ipath_sge *sge = &ss->sge; local [all...] |
/drivers/infiniband/hw/qib/ |
H A D | qib_keys.c | 125 * @sge: SGE to check 134 struct qib_sge *isge, struct ib_sge *sge, int acc) 146 if (sge->lkey == 0) { 157 isge->vaddr = (void *) sge->addr; 158 isge->length = sge->length; 159 isge->sge_length = sge->length; 164 mr = rkt->table[(sge->lkey >> (32 - ib_qib_lkey_table_size))]; 165 if (unlikely(mr == NULL || mr->lkey != sge->lkey || 169 off = sge->addr - mr->user_base; 170 if (unlikely(sge 133 qib_lkey_ok(struct qib_lkey_table *rkt, struct qib_pd *pd, struct qib_sge *isge, struct ib_sge *sge, int acc) argument 226 qib_rkey_ok(struct qib_qp *qp, struct qib_sge *sge, u32 len, u64 vaddr, u32 rkey, int acc) argument [all...] |
H A D | qib_ruc.c | 98 if (!qib_lkey_ok(rkt, pd, j ? &ss->sg_list[j - 1] : &ss->sge, 111 struct qib_sge *sge = --j ? &ss->sg_list[j - 1] : &ss->sge; local 113 atomic_dec(&sge->mr->refcount); 360 struct qib_sge *sge; local 427 sqp->s_sge.sge = wqe->sg_list[0]; 460 if (unlikely(!qib_rkey_ok(qp, &qp->r_sge.sge, wqe->length, 473 if (unlikely(!qib_rkey_ok(qp, &sqp->s_sge.sge, wqe->length, 481 qp->r_sge.sge = wqe->sg_list[0]; 491 if (unlikely(!qib_rkey_ok(qp, &qp->r_sge.sge, sizeo 783 struct qib_sge *sge = &wqe->sg_list[i]; local [all...] |
H A D | qib_ud.c | 57 struct qib_sge *sge; local 166 ssge.sge = *swqe->sg_list; 168 sge = &ssge.sge; 170 u32 len = sge->length; 174 if (len > sge->sge_length) 175 len = sge->sge_length; 177 qib_copy_sge(&qp->r_sge, sge->vaddr, len, 1); 178 sge->vaddr += len; 179 sge [all...] |
H A D | qib_qp.c | 410 atomic_dec(&qp->s_rdma_read_sge.sge.mr->refcount); 412 qp->s_rdma_read_sge.sge = 417 atomic_dec(&qp->r_sge.sge.mr->refcount); 419 qp->r_sge.sge = *qp->r_sge.sg_list++; 428 struct qib_sge *sge = &wqe->sg_list[i]; local 430 atomic_dec(&sge->mr->refcount);
|
H A D | qib_rc.c | 49 ss->sge = wqe->sg_list[0]; 137 qp->s_ack_rdma_sge.sge = e->rdma_sge; 173 qp->s_rdma_mr = qp->s_ack_rdma_sge.sge.mr; 176 len = qp->s_ack_rdma_sge.sge.sge_length; 487 qp->s_sge.sge = wqe->sg_list[0]; 1013 struct qib_sge *sge = &wqe->sg_list[i]; local 1015 atomic_dec(&sge->mr->refcount); 1069 struct qib_sge *sge = &wqe->sg_list[i]; local 1071 atomic_dec(&sge->mr->refcount); 2028 atomic_dec(&qp->r_sge.sge [all...] |
H A D | qib_sdma.c | 531 struct qib_sge *sge; local 579 sge = &ss->sge; 585 if (len > sge->length) 586 len = sge->length; 587 if (len > sge->sge_length) 588 len = sge->sge_length; 591 addr = dma_map_single(&ppd->dd->pcidev->dev, sge->vaddr, 610 sge->vaddr += len; 611 sge [all...] |
H A D | qib_verbs.c | 170 struct qib_sge *sge = &ss->sge; local 173 u32 len = sge->length; 177 if (len > sge->sge_length) 178 len = sge->sge_length; 180 memcpy(sge->vaddr, data, len); 181 sge->vaddr += len; 182 sge->length -= len; 183 sge->sge_length -= len; 184 if (sge 212 struct qib_sge *sge = &ss->sge; local 253 struct qib_sge sge = ss->sge; local 298 struct qib_sge *sge = &ss->sge; local 436 struct qib_sge *sge = &wqe->sg_list[--j]; local 725 struct qib_sge *sge = &ss->sge; local [all...] |
/drivers/scsi/be2iscsi/ |
H A D | be_mgmt.c | 61 struct be_sge *sge; local 75 sge = nonembedded_sgl(wrb); 85 sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd->dma)); 86 sge->pa_lo = cpu_to_le32(nonemb_cmd->dma & 0xFFFFFFFF); 87 sge->len = cpu_to_le32(nonemb_cmd->size); 144 struct be_sge *sge = nonembedded_sgl(wrb); local 164 sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd.dma)); 165 sge->pa_lo = cpu_to_le32(nonemb_cmd.dma & 0xFFFFFFFF); 166 sge->len = cpu_to_le32(nonemb_cmd.size); 224 struct be_sge *sge; local 347 struct be_sge *sge; local [all...] |
/drivers/infiniband/hw/ehca/ |
H A D | ehca_reqs.c | 119 struct ib_sge *sge = send_wr->sg_list; local 138 u8 *data = (u8 *)abs_to_virt(sge->addr); 139 ehca_gen_dbg("send_wr#%x sge#%x addr=%p length=%x " 141 idx, j, data, sge->length, sge->lkey); 143 ehca_dmp(data, sge->length, "send_wr#%x sge#%x", 145 sge++;
|
/drivers/infiniband/ulp/iser/ |
H A D | iser_verbs.c | 689 struct ib_sge sge; local 692 sge.addr = ib_conn->login_resp_dma; 693 sge.length = ISER_RX_LOGIN_SIZE; 694 sge.lkey = ib_conn->device->mr->lkey; 697 rx_wr.sg_list = &sge;
|
/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_cm.c | 117 struct ib_sge *sge, int id) 126 sge[i].addr = rx->rx_ring[id].mapping[i]; 328 struct ib_sge *sge) 334 sge[i].lkey = priv->mr->lkey; 336 sge[0].length = IPOIB_CM_HEAD_SIZE; 338 sge[i].length = PAGE_SIZE; 341 wr->sg_list = sge; 351 struct ib_sge sge[IPOIB_CM_RX_SG]; member in struct:__anon1066 369 ipoib_cm_init_rx_wr(dev, &t->wr, t->sge); 390 ret = ipoib_cm_post_receive_nonsrq(dev, rx, &t->wr, t->sge, 114 ipoib_cm_post_receive_nonsrq(struct net_device *dev, struct ipoib_cm_rx *rx, struct ib_recv_wr *wr, struct ib_sge *sge, int id) argument 326 ipoib_cm_init_rx_wr(struct net_device *dev, struct ib_recv_wr *wr, struct ib_sge *sge) argument [all...] |
/drivers/net/ethernet/chelsio/cxgb4vf/ |
H A D | t4vf_common.h | 183 struct sge_params sge; /* Scatter Gather Engine */ member in struct:adapter_params
|
H A D | adapter.h | 260 struct sge { struct 331 #define for_each_ethrxq(sge, iter) \ 332 for (iter = 0; iter < (sge)->ethqsets; iter++) 354 struct sge sge; member in struct:adapter
|
/drivers/infiniband/ulp/srpt/ |
H A D | ib_srpt.h | 156 struct ib_sge *sge; member in struct:rdma_iu
|
/drivers/net/ethernet/broadcom/bnx2x/ |
H A D | bnx2x_cmn.h | 809 struct eth_rx_sge *sge = &fp->rx_sge_ring[index]; local 820 sge->addr_hi = 0; 821 sge->addr_lo = 0; 891 struct eth_rx_sge *sge = &fp->rx_sge_ring[index]; local 895 BNX2X_ERR("Can't alloc sge\n"); 903 BNX2X_ERR("Can't map sge\n"); 910 sge->addr_hi = cpu_to_le32(U64_HI(mapping)); 911 sge->addr_lo = cpu_to_le32(U64_LO(mapping)); 1112 struct eth_rx_sge *sge; local 1114 sge [all...] |
/drivers/net/ethernet/chelsio/cxgb3/ |
H A D | adapter.h | 216 struct sge { struct 245 struct sge sge; member in struct:adapter
|
/drivers/scsi/lpfc/ |
H A D | lpfc_mbox.c | 1648 * lpfc_sli4_mbx_sge_set - Set a sge entry in non-embedded mailbox command 1650 * @sgentry: sge entry index. 1651 * @phyaddr: physical address for the sge 1652 * @length: Length of the sge. 1654 * This routine sets up an entry in the non-embedded mailbox command at the sge 1665 nembed_sge->sge[sgentry].pa_lo = putPaddrLow(phyaddr); 1666 nembed_sge->sge[sgentry].pa_hi = putPaddrHigh(phyaddr); 1667 nembed_sge->sge[sgentry].length = length; 1671 * lpfc_sli4_mbx_sge_get - Get a sge entry from non-embedded mailbox command 1673 * @sgentry: sge entr 1679 lpfc_sli4_mbx_sge_get(struct lpfcMboxq *mbox, uint32_t sgentry, struct lpfc_mbx_sge *sge) argument 1702 struct lpfc_mbx_sge sge; local 2002 struct lpfc_mbx_sge sge; local [all...] |
/drivers/scsi/ |
H A D | vmw_pvscsi.c | 51 struct PVSCSISGElement sge[PVSCSI_MAX_NUM_SG_ENTRIES_PER_SEGMENT]; member in struct:pvscsi_sg_list 319 struct PVSCSISGElement *sge; local 323 sge = &ctx->sgl->sge[0]; 325 sge[i].addr = sg_dma_address(sg); 326 sge[i].length = sg_dma_len(sg); 327 sge[i].flags = 0;
|
/drivers/infiniband/core/ |
H A D | mad.c | 1010 struct ib_sge *sge; local 1020 sge = mad_send_wr->sg_list; 1021 sge[0].addr = ib_dma_map_single(mad_agent->device, 1023 sge[0].length, 1025 mad_send_wr->header_mapping = sge[0].addr; 1027 sge[1].addr = ib_dma_map_single(mad_agent->device, 1029 sge[1].length, 1031 mad_send_wr->payload_mapping = sge[1].addr; 1051 sge[0].length, DMA_TO_DEVICE); 1054 sge[ [all...] |
/drivers/net/ethernet/chelsio/cxgb/ |
H A D | common.h | 187 struct sge_params sge; member in struct:adapter_params 226 struct sge; 244 struct sge *sge; member in struct:adapter
|