/drivers/scsi/aic7xxx/aicasm/ |
H A D | aicasm_scan.l | 55 #include "../queue.h" 57 #include <sys/queue.h>
|
/drivers/staging/lustre/lustre/llite/ |
H A D | rw.c | 453 struct cl_page_list *queue, struct cl_page *page, 468 cl_page_list_add(queue, page); 486 * \retval +ve: page was added to \a queue. 491 * \retval -ve, 0: page wasn't added to \a queue for other reason. 494 struct cl_page_list *queue, 511 rc = cl_read_ahead_page(env, io, queue, 638 struct cl_io *io, struct cl_page_list *queue, 655 rc = ll_read_ahead_page(env, io, queue, 690 struct cl_page_list *queue, int flags) 792 ret = ll_read_ahead_pages(env, io, queue, 452 cl_read_ahead_page(const struct lu_env *env, struct cl_io *io, struct cl_page_list *queue, struct cl_page *page, struct page *vmpage) argument 493 ll_read_ahead_page(const struct lu_env *env, struct cl_io *io, struct cl_page_list *queue, pgoff_t index, struct address_space *mapping) argument 637 ll_read_ahead_pages(const struct lu_env *env, struct cl_io *io, struct cl_page_list *queue, struct ra_io_arg *ria, unsigned long *reserved_pages, struct address_space *mapping, unsigned long *ra_end) argument 688 ll_readahead(const struct lu_env *env, struct cl_io *io, struct ll_readahead_state *ras, struct address_space *mapping, struct cl_page_list *queue, int flags) argument [all...] |
/drivers/staging/unisys/include/ |
H A D | uisqueue.h | 62 * to the specified queue, at the tail if the queue is full but 64 * wait for the queue to become non-full. If command is queued, return 86 unsigned int queue, 93 /* uisqueue_get_cmdrsp gets the cmdrsp entry at the head of the queue 95 * returns 0 if queue is empty, 1 otherwise 100 unsigned int queue); 238 * guest linux/windows IO drivers when data channel queue for
|
/drivers/usb/gadget/udc/ |
H A D | atmel_usba_udc.h | 227 #define DBG_QUEUE 0x0020 /* debug request queue processing */ 276 struct list_head queue; member in struct:usba_ep 297 struct list_head queue; member in struct:usba_request
|
H A D | goku_udc.h | 218 struct list_head queue; member in struct:goku_ep 227 struct list_head queue; member in struct:goku_request
|
H A D | mv_u3d.h | 293 struct list_head queue; /* ep request queued hardware */ member in struct:mv_u3d_ep 311 struct list_head queue; /* ep requst queued on hardware */ member in struct:mv_u3d_req
|
H A D | mv_udc.h | 232 struct list_head queue; member in struct:mv_ep 247 struct list_head queue; member in struct:mv_req
|
H A D | r8a66597-udc.h | 55 struct list_head queue; member in struct:r8a66597_request 63 struct list_head queue; member in struct:r8a66597_ep
|
H A D | r8a66597-udc.c | 101 INIT_LIST_HEAD(&r8a66597->ep[0].queue); 435 INIT_LIST_HEAD(&ep->queue); 897 return list_entry(ep->queue.next, struct r8a66597_request, queue); 915 list_del_init(&req->queue); 921 if (!list_empty(&ep->queue)) 1154 if (!list_empty(&ep->queue)) 1229 if (list_empty(&ep->queue)) 1232 } else if (!list_empty(&ep->queue)) 1574 while (!list_empty(&ep->queue)) { [all...] |
/drivers/usb/host/ |
H A D | imx21-hcd.h | 344 struct list_head queue; member in struct:ep_priv 366 struct list_head queue; member in struct:etd_priv
|
/drivers/char/tpm/ |
H A D | tpm_i2c_nuvoton.c | 168 * set queue to NULL to avoid waiting for interrupt 177 u32 timeout, wait_queue_head_t *queue) 179 if (chip->vendor.irq && queue) { 185 rc = wait_event_interruptible_timeout(*queue, 222 wait_queue_head_t *queue) 227 timeout, queue); 176 i2c_nuvoton_wait_for_stat(struct tpm_chip *chip, u8 mask, u8 value, u32 timeout, wait_queue_head_t *queue) argument 221 i2c_nuvoton_wait_for_data_avail(struct tpm_chip *chip, u32 timeout, wait_queue_head_t *queue) argument
|
/drivers/crypto/ |
H A D | s5p-sss.c | 190 struct crypto_queue queue; member in struct:s5p_aes_dev 471 backlog = crypto_get_backlog(&dev->queue); 472 async_req = crypto_dequeue_request(&dev->queue); 498 err = ablkcipher_enqueue_request(&dev->queue, req); 696 crypto_init_queue(&pdata->queue, CRYPTO_QUEUE_LEN);
|
/drivers/ide/ |
H A D | ide-atapi.c | 86 * Add a special packet command request to the tail of the request queue, 95 rq = blk_get_request(drive->queue, READ, __GFP_WAIT); 100 error = blk_rq_map_kern(drive->queue, rq, buf, bufflen, 109 error = blk_execute_rq(drive->queue, disk, rq, 0); 201 err = blk_rq_map_kern(drive->queue, sense_rq, sense, sense_len, 237 elv_add_request(drive->queue, &drive->sense_rq, ELEVATOR_INSERT_FRONT); 244 * We queue a request sense packet command at the head of the request 245 * queue.
|
/drivers/infiniband/hw/ehca/ |
H A D | ehca_qp.c | 263 * init userspace queue info from ipz_queue data 266 struct ipz_queue *queue) 268 resp->qe_size = queue->qe_size; 269 resp->act_nr_of_sg = queue->act_nr_of_sg; 270 resp->queue_length = queue->queue_length; 271 resp->pagesize = queue->pagesize; 272 resp->toggle_state = queue->toggle_state; 273 resp->offset = queue->offset; 277 * init_qp_queue initializes/constructs r/squeue and registers queue pages. 282 struct ipz_queue *queue, 265 queue2resp(struct ipzu_queue_resp *resp, struct ipz_queue *queue) argument 279 init_qp_queue(struct ehca_shca *shca, struct ehca_pd *pd, struct ehca_qp *my_qp, struct ipz_queue *queue, int q_type, u64 expected_hret, struct ehca_alloc_queue_parms *parms, int wqe_size) argument 373 ehca_determine_small_queue(struct ehca_alloc_queue_parms *queue, int req_nr_sge, int is_llqp) argument [all...] |
/drivers/net/wireless/ath/carl9170/ |
H A D | rx.c | 577 unsigned int queue; local 585 queue = TID_TO_WME_AC(((le16_to_cpu(bar->control) & 590 list_for_each_entry_rcu(entry, &ar->bar_list[queue], list) { 608 spin_lock_bh(&ar->bar_list_lock[queue]); 610 spin_unlock_bh(&ar->bar_list_lock[queue]);
|
/drivers/dma/ |
H A D | txx9dmac.c | 180 return list_entry(dc->queue.next, struct txx9dmac_desc, desc_node); 347 /* The tasklet will hopefully advance the queue... */ 452 } while (!list_empty(&dc->queue)); 465 if (!list_empty(&dc->queue)) { 528 if (list_empty(&dc->active_list) && !list_empty(&dc->queue)) 605 if (!list_empty(&dc->queue)) { 709 list_add_tail(&desc->desc_node, &dc->queue); 777 * the dc->queue list or dc->active_list after a 921 list_splice_init(&dc->queue, &list); 981 if (!list_empty(&dc->queue)) { [all...] |
H A D | ep93xx_dma.c | 120 * @node: link used for putting this into a channel queue 144 * @queue: pending descriptors which are handled next 155 * @queue holds pending transactions. These are linked through the first 156 * descriptor in the chain. When a descriptor is moved to the @active queue, 178 struct list_head queue; member in struct:ep93xx_dma_chan 711 * function takes the next queued transaction from the @edmac->queue and 720 if (!list_empty(&edmac->active) || list_empty(&edmac->queue)) { 725 /* Take the next descriptor from the pending queue */ 726 new = list_first_entry(&edmac->queue, struct ep93xx_dma_desc, node); 763 /* Pick up the next descriptor from the queue */ [all...] |
/drivers/media/pci/cx23885/ |
H A D | cx23885-video.c | 105 struct cx23885_buffer, queue); 111 list_del(&buf->queue); 464 list_add_tail(&buf->queue, &q->active); 470 queue); 471 list_add_tail(&buf->queue, &q->active); 484 struct cx23885_buffer, queue); 500 struct cx23885_buffer, queue); 502 list_del(&buf->queue); 1260 dev->video_dev->queue = &dev->vb2_vidq; 1274 dev->vbi_dev->queue [all...] |
/drivers/media/platform/soc_camera/ |
H A D | mx3_camera.c | 67 struct list_head queue; member in struct:mx3_camera_buffer 157 list_del_init(&buf->queue); 176 struct mx3_camera_buffer, queue); 331 list_add_tail(&buf->queue, &mx3_cam->capture); 348 list_del_init(&buf->queue); 368 "Release%s DMA 0x%08x, queue %sempty\n", 370 list_empty(&buf->queue) ? "" : "not "); 378 list_del_init(&buf->queue); 400 INIT_LIST_HEAD(&buf->queue); 425 list_for_each_entry_safe(buf, tmp, &mx3_cam->capture, queue) { [all...] |
/drivers/scsi/ibmvscsi/ |
H A D | ibmvfc.c | 560 list_for_each_entry(tgt, &vhost->targets, queue) 593 list_for_each_entry(tgt, &vhost->targets, queue) 722 /* Clean out the queue */ 771 list_add_tail(&evt->queue, &vhost->free); 812 list_del(&evt->queue); 831 list_for_each_entry_safe(evt, pos, &vhost->sent, queue) 917 list_for_each_entry(tgt, &vhost->targets, queue) 1232 list_add_tail(&evt->queue, &vhost->free); 1251 list_del(&pool->events[i].queue); 1277 evt = list_entry(vhost->free.next, struct ibmvfc_event, queue); 3178 struct ibmvfc_crq_queue *queue = &vhost->crq; local [all...] |
/drivers/spi/ |
H A D | spi-bfin5xx.c | 70 /* Driver message queue */ 74 struct list_head queue; member in struct:bfin_spi_master_data 875 /* pop a msg from queue and kick off real transfer */ 883 /* Lock queue and check for queue work */ 885 if (list_empty(&drv_data->queue) || !drv_data->running) { 898 /* Extract head of queue */ 899 drv_data->cur_msg = list_entry(drv_data->queue.next, 900 struct spi_message, queue); 906 list_del_init(&drv_data->cur_msg->queue); [all...] |
/drivers/memstick/core/ |
H A D | ms_block.c | 1891 msb->req = blk_fetch_request(msb->queue); 1907 blk_rq_map_sg(msb->queue, msb->req, sg); 2053 blk_stop_queue(msb->queue); 2062 blk_requeue_request(msb->queue, msb->req); 2090 blk_start_queue(msb->queue); 2129 msb->queue = blk_init_queue(msb_submit_req, &msb->q_lock); 2130 if (!msb->queue) { 2135 msb->queue->queuedata = card; 2136 blk_queue_prep_rq(msb->queue, msb_prepare_req); 2138 blk_queue_bounce_limit(msb->queue, limi [all...] |
/drivers/net/wireless/rtlwifi/ |
H A D | pci.c | 540 (ring->entries - skb_queue_len(&ring->queue) > 569 while (skb_queue_len(&ring->queue)) { 585 skb = __skb_dequeue(&ring->queue); 599 skb_queue_len(&ring->queue), 644 if ((ring->entries - skb_queue_len(&ring->queue)) 650 skb_queue_len(&ring->queue)); 1111 pskb = __skb_dequeue(&ring->queue); 1129 __skb_queue_tail(&ring->queue, pskb); 1159 *we just alloc 2 desc for beacon queue, 1164 /*BE queue nee [all...] |
/drivers/infiniband/hw/cxgb4/ |
H A D | cq.c | 72 cq->memsize, cq->queue, 102 cq->queue = dma_alloc_coherent(&rdev->lldi.pdev->dev, cq->memsize, 104 if (!cq->queue) { 109 memset(cq->queue, 0, cq->memsize); 168 dma_free_coherent(&rdev->lldi.pdev->dev, cq->memsize, cq->queue, 900 * Make actual HW queue 2x to avoid cdix_inc overflows. 905 * Make HW queue at least 64 entries so GTS updates aren't too 911 memsize = hwentries * sizeof *chp->cq.queue; 962 mm->addr = virt_to_phys(chp->cq.queue);
|
/drivers/net/wireless/iwlwifi/mvm/ |
H A D | mac-ctxt.c | 373 u8 queue = find_first_zero_bit(&used_hw_queues, local 376 if (queue >= mvm->first_agg_queue) { 377 IWL_ERR(mvm, "Failed to allocate queue\n"); 382 __set_bit(queue, &used_hw_queues); 383 vif->hw_queue[ac] = queue; 386 /* Allocate the CAB queue for softAP and GO interfaces */ 388 u8 queue = find_first_zero_bit(&used_hw_queues, local 391 if (queue >= mvm->first_agg_queue) { 392 IWL_ERR(mvm, "Failed to allocate cab queue\n"); 397 vif->cab_queue = queue; [all...] |