/net/9p/ |
H A D | trans_rdma.c | 49 #include <rdma/ib_verbs.h> 50 #include <rdma/rdma_cm.h> 161 * parse_opts - parse mount options into rdma options structure 163 * @opts: rdma transport-specific structure to parse options into 231 struct p9_trans_rdma *rdma = c->trans; local 234 BUG_ON(rdma->state != P9_RDMA_INIT); 235 rdma->state = P9_RDMA_ADDR_RESOLVED; 239 BUG_ON(rdma->state != P9_RDMA_ADDR_RESOLVED); 240 rdma->state = P9_RDMA_ROUTE_RESOLVED; 244 BUG_ON(rdma 280 handle_recv(struct p9_client *client, struct p9_trans_rdma *rdma, struct p9_rdma_context *c, enum ib_wc_status status, u32 byte_len) argument 321 handle_send(struct p9_client *client, struct p9_trans_rdma *rdma, struct p9_rdma_context *c, enum ib_wc_status status, u32 byte_len) argument 338 struct p9_trans_rdma *rdma = client->trans; local 371 rdma_destroy_trans(struct p9_trans_rdma *rdma) argument 397 struct p9_trans_rdma *rdma = client->trans; local 425 struct p9_trans_rdma *rdma = client->trans; local 554 struct p9_trans_rdma *rdma; local 574 struct p9_trans_rdma *rdma; local 605 struct p9_trans_rdma *rdma = client->trans; local 621 struct p9_trans_rdma *rdma; local [all...] |
/net/rds/ |
H A D | rdma_transport.h | 4 #include <rdma/ib_verbs.h> 5 #include <rdma/rdma_cm.h>
|
H A D | Makefile | 4 loop.o page.o rdma.o
|
H A D | send.c | 232 if (rm->rdma.op_active && 258 /* The transport either sends the whole rdma or none of it */ 259 if (rm->rdma.op_active && !conn->c_xmit_rdma_sent) { 260 rm->m_final_op = &rm->rdma; 261 ret = conn->c_trans->xmit_rdma(conn, &rm->rdma); 294 ops_present = (rm->atomic.op_active || rm->rdma.op_active); 297 if (rm->rdma.op_active && !rm->rdma.op_silent) 431 ro = &rm->rdma; 502 ro = &rm->rdma; [all...] |
H A D | message.c | 71 if (rm->rdma.op_active) 72 rds_rdma_free_op(&rm->rdma); 73 if (rm->rdma.op_rdma_mr) 74 rds_mr_put(rm->rdma.op_rdma_mr);
|
H A D | iw_send.c | 89 if (rm->rdma.op_active) { 90 rds_iw_send_unmap_rdma(ic, &rm->rdma); 114 if (rm->rdma.op_write) 115 rds_stats_add(s_send_rdma_bytes, rm->rdma.op_bytes); 117 rds_stats_add(s_recv_rdma_bytes, rm->rdma.op_bytes); 594 if (rm->rdma.op_active) { 597 ext_hdr.h_rdma_rkey = cpu_to_be32(rm->rdma.op_rkey); 635 if (rm->rdma.op_active && rm->rdma.op_fence) 836 * Instead of knowing how to return a partial rdma rea [all...] |
H A D | ib_send.c | 137 * we would need to take an event for the rdma WR. To implement #1, 146 rds_ib_send_complete(container_of(op, struct rds_message, rdma), 199 rm = container_of(send->s_op, struct rds_message, rdma); 618 if (rm->rdma.op_active) { 621 ext_hdr.h_rdma_rkey = cpu_to_be32(rm->rdma.op_rkey); 654 if (rm->rdma.op_active && rm->rdma.op_fence) 783 * A simplified version of the rdma case, we always map 1 SG, and 911 * Instead of knowing how to return a partial rdma read/write we insist that there 939 send->s_wr.wr.rdma [all...] |
H A D | rdma_transport.c | 34 #include <rdma/rdma_cm.h>
|
H A D | ib.h | 4 #include <rdma/ib_verbs.h> 5 #include <rdma/rdma_cm.h>
|
H A D | iw.h | 5 #include <rdma/ib_verbs.h> 6 #include <rdma/rdma_cm.h>
|
H A D | recv.c | 114 struct rds_ext_header_rdma rdma; member in union:__anon135 126 rds_rdma_unuse(rs, be32_to_cpu(buffer.rdma.h_rdma_rkey), 0);
|
H A D | rdma.c | 45 * - an rdma is an mlock, apply rlimit? 550 struct rm_rdma_op *op = &rm->rdma; 560 || rm->rdma.op_active) 642 rdsdebug("RDS: rdma prepare nr_local %llu rva %llx rkey %x\n", 743 rm->rdma.op_rdma_mr = mr; 761 return __rds_rdma_map(rs, CMSG_DATA(cmsg), &rm->m_rdma_cookie, &rm->rdma.op_rdma_mr);
|
H A D | rds.h | 7 #include <rdma/rdma_cm.h> 360 } rdma; member in struct:rds_message::__anon131 730 /* rdma.c */
|
H A D | ib_recv.c | 37 #include <rdma/rdma_cm.h> 731 * long-lived bitmaps. We could have posted the bitmaps and rdma written into
|
H A D | iw_recv.c | 37 #include <rdma/rdma_cm.h> 562 * long-lived bitmaps. We could have posted the bitmaps and rdma written into
|
/net/sunrpc/xprtrdma/ |
H A D | svc_rdma_sendto.c | 47 #include <rdma/ib_verbs.h> 48 #include <rdma/rdma_cm.h> 207 write_wr.wr.rdma.rkey = rmr; 208 write_wr.wr.rdma.remote_addr = to; 366 static int send_reply(struct svcxprt_rdma *rdma, argument 382 ret = svc_rdma_post_recv(rdma); 386 "Closing transport %p.\n", ret, rdma); 387 set_bit(XPT_CLOSE, &rdma->sc_xprt.xpt_flags); 397 ctxt->sge[0].lkey = rdma->sc_dma_lkey; 400 ib_dma_map_page(rdma 485 struct svcxprt_rdma *rdma = local [all...] |
H A D | svc_rdma_transport.c | 51 #include <rdma/ib_verbs.h> 52 #include <rdma/rdma_cm.h> 91 .xcl_name = "rdma", 644 struct svcxprt_rdma *rdma = local 652 clear_bit(RDMAXPRT_CONN_PENDING, &rdma->sc_flags); 786 struct svc_rdma_fastreg_mr *svc_rdma_get_frmr(struct svcxprt_rdma *rdma) argument 790 spin_lock_bh(&rdma->sc_frmr_q_lock); 791 if (!list_empty(&rdma->sc_frmr_q)) { 792 frmr = list_entry(rdma->sc_frmr_q.next, 798 spin_unlock_bh(&rdma 819 svc_rdma_put_frmr(struct svcxprt_rdma *rdma, struct svc_rdma_fastreg_mr *frmr) argument 1113 struct svcxprt_rdma *rdma = local 1123 struct svcxprt_rdma *rdma = local 1186 struct svcxprt_rdma *rdma = local 1194 struct svcxprt_rdma *rdma = local [all...] |
H A D | svc_rdma_recvfrom.c | 48 #include <rdma/ib_verbs.h> 49 #include <rdma/rdma_cm.h> 207 read_wr.wr.rdma.rkey = rs_handle; 208 read_wr.wr.rdma.remote_addr = rs_offset; 327 read_wr.wr.rdma.rkey = rs_handle; 328 read_wr.wr.rdma.remote_addr = rs_offset;
|
H A D | xprt_rdma.h | 48 #include <rdma/rdma_cm.h> /* RDMA connection api */ 49 #include <rdma/ib_verbs.h> /* RDMA verbs api */ 272 unsigned int inline_rsize; /* max non-rdma read data payload */ 273 unsigned int inline_wsize; /* max non-rdma write data payload */ 274 unsigned int padding; /* non-rdma write header padding */
|