/drivers/crypto/ccp/ |
H A D | ccp-crypto-main.c | 48 struct list_head *backlog; member in struct:ccp_crypto_queue 90 struct ccp_crypto_cmd *crypto_cmd, struct ccp_crypto_cmd **backlog) 95 *backlog = NULL; 110 /* Process the backlog: 112 * special precautions have to be taken when handling the backlog. 114 if (req_queue.backlog != &req_queue.cmds) { 115 /* Skip over this cmd if it is the next backlog cmd */ 116 if (req_queue.backlog == &crypto_cmd->entry) 117 req_queue.backlog = crypto_cmd->entry.next; 119 *backlog 89 ccp_crypto_cmd_complete( struct ccp_crypto_cmd *crypto_cmd, struct ccp_crypto_cmd **backlog) argument 140 struct ccp_crypto_cmd *held, *next, *backlog; local [all...] |
H A D | ccp-dev.c | 83 * backlogged) or advancement out of the backlog. If the cmd has 84 * advanced out of the backlog the "err" value of the callback 115 list_add_tail(&cmd->entry, &ccp->backlog); 175 struct ccp_cmd *backlog = NULL; local 200 if (!list_empty(&ccp->backlog)) { 201 backlog = list_first_entry(&ccp->backlog, struct ccp_cmd, 203 list_del(&backlog->entry); 208 if (backlog) { 209 INIT_WORK(&backlog [all...] |
H A D | ccp-dev.h | 214 * backlog list is neeeded so that the backlog completion call 220 struct list_head backlog; member in struct:ccp_device
|
/drivers/crypto/qce/ |
H A D | core.c | 82 struct crypto_async_request *async_req, *backlog; local 97 backlog = crypto_get_backlog(&qce->queue); 107 if (backlog) { 109 backlog->complete(backlog, -EINPROGRESS);
|
/drivers/atm/ |
H A D | zatm.h | 54 struct sk_buff_head backlog; /* list of buffers waiting for ring */ member in struct:zatm_vcc
|
H A D | eni.h | 49 struct sk_buff_head backlog; /* queue of waiting TX buffers */ member in struct:eni_tx
|
/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_cm.h | 172 int backlog; member in struct:iwch_listen_ep 218 int iwch_create_listen(struct iw_cm_id *cm_id, int backlog);
|
/drivers/crypto/ |
H A D | mxs-dcp.c | 347 struct crypto_async_request *backlog; local 356 backlog = crypto_get_backlog(&sdcp->queue[chan]); 360 if (backlog) 361 backlog->complete(backlog, -EINPROGRESS); 640 struct crypto_async_request *backlog; local 652 backlog = crypto_get_backlog(&sdcp->queue[chan]); 656 if (backlog) 657 backlog->complete(backlog, [all...] |
H A D | s5p-sss.c | 466 struct crypto_async_request *async_req, *backlog; local 471 backlog = crypto_get_backlog(&dev->queue); 481 if (backlog) 482 backlog->complete(backlog, -EINPROGRESS);
|
H A D | bfin_crc.c | 302 struct crypto_async_request *async_req, *backlog; local 318 backlog = crypto_get_backlog(&crc->queue); 327 if (backlog) 328 backlog->complete(backlog, -EINPROGRESS);
|
H A D | sahara.c | 535 struct crypto_async_request *async_req, *backlog; local 542 backlog = crypto_get_backlog(&dev->queue); 551 if (backlog) 552 backlog->complete(backlog, -EINPROGRESS);
|
H A D | mv_cesa.c | 598 struct crypto_async_request *backlog; local 607 backlog = crypto_get_backlog(&cpg->queue); 616 if (backlog) { 617 backlog->complete(backlog, -EINPROGRESS); 618 backlog = NULL;
|
H A D | atmel-aes.c | 553 struct crypto_async_request *async_req, *backlog; local 566 backlog = crypto_get_backlog(&dd->queue); 575 if (backlog) 576 backlog->complete(backlog, -EINPROGRESS);
|
H A D | atmel-tdes.c | 587 struct crypto_async_request *async_req, *backlog; local 600 backlog = crypto_get_backlog(&dd->queue); 609 if (backlog) 610 backlog->complete(backlog, -EINPROGRESS);
|
H A D | omap-aes.c | 602 struct crypto_async_request *async_req, *backlog; local 615 backlog = crypto_get_backlog(&dd->queue); 624 if (backlog) 625 backlog->complete(backlog, -EINPROGRESS);
|
/drivers/staging/ozwpan/ |
H A D | ozpd.c | 522 void oz_send_queued_frames(struct oz_pd *pd, int backlog) argument 525 backlog++; 530 backlog += pd->nb_queued_isoc_frames; 531 if (backlog <= 0) 533 if (backlog > OZ_MAX_SUBMITTED_ISOC) 534 backlog = OZ_MAX_SUBMITTED_ISOC; 538 if ((backlog <= 0) && (pd->isoc_sent == 0)) 543 if (backlog <= 0) 548 while (backlog--) { 549 if (oz_send_next_queued_frame(pd, backlog) < [all...] |
H A D | ozpd.h | 122 void oz_send_queued_frames(struct oz_pd *pd, int backlog);
|
/drivers/infiniband/hw/amso1100/ |
H A D | c2_cm.c | 135 int c2_llp_service_create(struct iw_cm_id *cm_id, int backlog) argument 166 wr.backlog = cpu_to_be32(backlog);
|
/drivers/staging/lustre/include/linux/libcfs/ |
H A D | libcfs.h | 92 int libcfs_sock_listen(struct socket **sockp, __u32 ip, int port, int backlog);
|
/drivers/staging/lustre/lustre/libcfs/linux/ |
H A D | linux-tcpip.c | 504 __u32 local_ip, int local_port, int backlog) 517 rc = (*sockp)->ops->listen(*sockp, backlog); 521 CERROR("Can't set listen backlog %d: %d\n", backlog, rc); 503 libcfs_sock_listen(struct socket **sockp, __u32 local_ip, int local_port, int backlog) argument
|
/drivers/staging/octeon/ |
H A D | ethernet-rx.c | 296 int backlog; local 299 backlog = counts.s.iq_cnt + counts.s.ds_cnt; 300 if (backlog > budget * cores_in_use && napi != NULL)
|
/drivers/infiniband/hw/nes/ |
H A D | nes_cm.h | 303 int backlog; member in struct:nes_cm_listener 375 int backlog; member in struct:nes_cm_info
|
/drivers/infiniband/core/ |
H A D | ucma.c | 84 int backlog; member in struct:ucma_context 268 if (!ctx->backlog) { 273 ctx->backlog--; 328 uevent->ctx->backlog++; 963 ctx->backlog = cmd.backlog > 0 && cmd.backlog < max_backlog ? 964 cmd.backlog : max_backlog; 965 ret = rdma_listen(ctx->cm_id, ctx->backlog);
|
H A D | iwcm.c | 87 * listen backlog. 98 * the backlog is exceeded, then no more connection request events will 435 int iw_cm_listen(struct iw_cm_id *cm_id, int backlog) argument 443 if (!backlog) 444 backlog = default_backlog; 446 ret = alloc_work_entries(cm_id_priv, backlog); 455 ret = cm_id->device->iwcm->create_listen(cm_id, backlog);
|
/drivers/net/wireless/p54/ |
H A D | lmac.h | 258 u8 backlog; member in struct:p54_tx_data 452 u8 backlog; member in struct:p54_burst
|