/drivers/staging/lustre/lnet/selftest/ |
H A D | brw_test.c | 54 srpc_bulk_t *bulk; local 60 bulk = tsu->tsu_private; 61 if (bulk == NULL) 64 srpc_free_bulk(bulk); 77 srpc_bulk_t *bulk; local 117 bulk = srpc_alloc_bulk(lnet_cpt_of_nid(tsu->tsu_dest.nid), 119 if (bulk == NULL) { 124 tsu->tsu_private = bulk; 258 srpc_bulk_t *bulk = tsu->tsu_private; local 270 LASSERT(bulk ! [all...] |
H A D | conrpc.c | 153 srpc_bulk_t *bulk = &crpc->crp_rpc->crpc_bulk; local 158 for (i = 0; i < bulk->bk_niov; i++) { 159 if (bulk->bk_iovs[i].kiov_page == NULL) 162 __free_page(bulk->bk_iovs[i].kiov_page); 818 srpc_bulk_t *bulk; local 846 bulk = &(*crpc)->crp_rpc->crpc_bulk; 857 bulk->bk_iovs[i].kiov_offset = 0; 858 bulk->bk_iovs[i].kiov_len = len; 859 bulk->bk_iovs[i].kiov_page = 862 if (bulk [all...] |
H A D | framework.c | 700 test_bulk_req_t *bulk = &req->tsr_u.bulk_v0; local 702 __swab32s(&bulk->blk_opc); 703 __swab32s(&bulk->blk_npg); 704 __swab32s(&bulk->blk_flags); 707 test_bulk_req_v1_t *bulk = &req->tsr_u.bulk_v1; local 709 __swab16s(&bulk->blk_opc); 710 __swab16s(&bulk->blk_flags); 711 __swab32s(&bulk->blk_offset); 712 __swab32s(&bulk->blk_len);
|
/drivers/scsi/ |
H A D | imm.c | 611 * The bulk flag enables some optimisations in the data transfer loops, 630 int fast, bulk, status; local 633 bulk = ((v == READ_6) || 666 fast = (bulk 671 fast = (bulk
|
H A D | ppa.c | 506 * The bulk flag enables some optimisations in the data transfer loops, 525 int fast, bulk, status; local 528 bulk = ((v == READ_6) || 574 fast = (bulk && (cmd->SCp.this_residual >= PPA_BURST_SIZE))
|
/drivers/staging/lustre/lustre/ptlrpc/ |
H A D | sec_config.c | 98 char *bulk, *alg; local 110 bulk = strchr(buf, '-'); 111 if (bulk) 112 *bulk++ = '\0'; 119 * currently only base flavor "plain" can have bulk specification. 123 if (bulk) { 127 alg = strchr(bulk, ':'); 132 if (strcmp(bulk, "hash")) 145 if (bulk)
|
/drivers/usb/gadget/udc/ |
H A D | r8a66597-udc.h | 112 unsigned char bulk; member in struct:r8a66597
|
H A D | m66592-udc.h | 490 int bulk; member in struct:m66592
|
/drivers/media/usb/dvb-usb/ |
H A D | dvb-usb.h | 112 } bulk; member in union:usb_data_stream_properties::__anon2391 198 * @bulk_mode: device supports bulk mode for RC (disable polling mode) 209 bool bulk_mode; /* uses bulk mode */ 246 * endpoint which received control messages with bulk transfers. When this 251 * endpoint for responses to control messages sent with bulk transfers via
|
/drivers/media/usb/dvb-usb-v2/ |
H A D | dvb_usb.h | 82 .bulk = { \ 135 * @bulk_mode: device supports bulk mode for rc (disable polling mode) 163 } bulk; member in union:usb_data_stream_properties::__anon2397 201 * @generic_bulk_ctrl_endpoint: bulk control endpoint number for sent 202 * @generic_bulk_ctrl_endpoint_response: bulk control endpoint number for 204 * @generic_bulk_ctrl_delay: delay between bulk control sent and receive message
|
/drivers/media/usb/gspca/ |
H A D | gspca.h | 50 u32 bulk_size; /* buffer size when image transfer by bulk */ 54 u8 bulk_nurbs; /* number of URBs in bulk mode 58 u8 bulk; /* image transfer by 0:isoc / 1:bulk */ member in struct:cam
|
/drivers/regulator/ |
H A D | core.c | 1855 * this for bulk operations so that the regulators can ramp 3222 struct regulator_bulk_data *bulk = data; local 3224 bulk->ret = regulator_enable(bulk->consumer);
|
/drivers/usb/core/ |
H A D | devio.c | 362 static const char *types[] = {"isoc", "int", "ctrl", "bulk"}; 1058 struct usbdevfs_bulktransfer bulk; local 1064 if (copy_from_user(&bulk, arg, sizeof(bulk))) 1066 ret = findintfep(ps->dev, bulk.ep); 1072 if (bulk.ep & USB_DIR_IN) 1073 pipe = usb_rcvbulkpipe(dev, bulk.ep & 0x7f); 1075 pipe = usb_sndbulkpipe(dev, bulk.ep & 0x7f); 1076 if (!usb_maxpacket(dev, pipe, !(bulk.ep & USB_DIR_IN))) 1078 len1 = bulk [all...] |
/drivers/media/usb/uvc/ |
H A D | uvcvideo.h | 472 /* Context data used by the bulk completion handler. */ 479 } bulk; member in struct:uvc_streaming
|
/drivers/staging/lustre/lustre/include/ |
H A D | lustre_net.h | 76 * Max # of bulk operations in one request. 79 * value. The client is free to limit the actual RPC size for any bulk 88 * use the negotiated per-client ocd_brw_size to determine the bulk 93 * Define maxima for bulk I/O. 299 /* XXX only ones left are those used by the bulk descs as well! */ 1384 rq_bulk_read:1, /* request bulk read */ 1385 rq_bulk_write:1, /* request bulk write */ 1473 /** For bulk requests on client only: bulk descriptor */ 1490 /** when req bulk unlin 2415 ptlrpc_free_bulk_pin(struct ptlrpc_bulk_desc *bulk) argument 2419 ptlrpc_free_bulk_nopin(struct ptlrpc_bulk_desc *bulk) argument [all...] |