Searched refs:queue (Results 276 - 300 of 742) sorted by relevance

<<11121314151617181920>>

/drivers/scsi/aic7xxx/aicasm/
H A Daicasm_scan.l55 #include "../queue.h"
57 #include <sys/queue.h>
/drivers/staging/lustre/lustre/llite/
H A Drw.c453 struct cl_page_list *queue, struct cl_page *page,
468 cl_page_list_add(queue, page);
486 * \retval +ve: page was added to \a queue.
491 * \retval -ve, 0: page wasn't added to \a queue for other reason.
494 struct cl_page_list *queue,
511 rc = cl_read_ahead_page(env, io, queue,
638 struct cl_io *io, struct cl_page_list *queue,
655 rc = ll_read_ahead_page(env, io, queue,
690 struct cl_page_list *queue, int flags)
792 ret = ll_read_ahead_pages(env, io, queue,
452 cl_read_ahead_page(const struct lu_env *env, struct cl_io *io, struct cl_page_list *queue, struct cl_page *page, struct page *vmpage) argument
493 ll_read_ahead_page(const struct lu_env *env, struct cl_io *io, struct cl_page_list *queue, pgoff_t index, struct address_space *mapping) argument
637 ll_read_ahead_pages(const struct lu_env *env, struct cl_io *io, struct cl_page_list *queue, struct ra_io_arg *ria, unsigned long *reserved_pages, struct address_space *mapping, unsigned long *ra_end) argument
688 ll_readahead(const struct lu_env *env, struct cl_io *io, struct ll_readahead_state *ras, struct address_space *mapping, struct cl_page_list *queue, int flags) argument
[all...]
/drivers/staging/unisys/include/
H A Duisqueue.h62 * to the specified queue, at the tail if the queue is full but
64 * wait for the queue to become non-full. If command is queued, return
86 unsigned int queue,
93 /* uisqueue_get_cmdrsp gets the cmdrsp entry at the head of the queue
95 * returns 0 if queue is empty, 1 otherwise
100 unsigned int queue);
238 * guest linux/windows IO drivers when data channel queue for
/drivers/usb/gadget/udc/
H A Datmel_usba_udc.h227 #define DBG_QUEUE 0x0020 /* debug request queue processing */
276 struct list_head queue; member in struct:usba_ep
297 struct list_head queue; member in struct:usba_request
H A Dgoku_udc.h218 struct list_head queue; member in struct:goku_ep
227 struct list_head queue; member in struct:goku_request
H A Dmv_u3d.h293 struct list_head queue; /* ep request queued hardware */ member in struct:mv_u3d_ep
311 struct list_head queue; /* ep requst queued on hardware */ member in struct:mv_u3d_req
H A Dmv_udc.h232 struct list_head queue; member in struct:mv_ep
247 struct list_head queue; member in struct:mv_req
H A Dr8a66597-udc.h55 struct list_head queue; member in struct:r8a66597_request
63 struct list_head queue; member in struct:r8a66597_ep
H A Dr8a66597-udc.c101 INIT_LIST_HEAD(&r8a66597->ep[0].queue);
435 INIT_LIST_HEAD(&ep->queue);
897 return list_entry(ep->queue.next, struct r8a66597_request, queue);
915 list_del_init(&req->queue);
921 if (!list_empty(&ep->queue))
1154 if (!list_empty(&ep->queue))
1229 if (list_empty(&ep->queue))
1232 } else if (!list_empty(&ep->queue))
1574 while (!list_empty(&ep->queue)) {
[all...]
/drivers/usb/host/
H A Dimx21-hcd.h344 struct list_head queue; member in struct:ep_priv
366 struct list_head queue; member in struct:etd_priv
/drivers/char/tpm/
H A Dtpm_i2c_nuvoton.c168 * set queue to NULL to avoid waiting for interrupt
177 u32 timeout, wait_queue_head_t *queue)
179 if (chip->vendor.irq && queue) {
185 rc = wait_event_interruptible_timeout(*queue,
222 wait_queue_head_t *queue)
227 timeout, queue);
176 i2c_nuvoton_wait_for_stat(struct tpm_chip *chip, u8 mask, u8 value, u32 timeout, wait_queue_head_t *queue) argument
221 i2c_nuvoton_wait_for_data_avail(struct tpm_chip *chip, u32 timeout, wait_queue_head_t *queue) argument
/drivers/crypto/
H A Ds5p-sss.c190 struct crypto_queue queue; member in struct:s5p_aes_dev
471 backlog = crypto_get_backlog(&dev->queue);
472 async_req = crypto_dequeue_request(&dev->queue);
498 err = ablkcipher_enqueue_request(&dev->queue, req);
696 crypto_init_queue(&pdata->queue, CRYPTO_QUEUE_LEN);
/drivers/ide/
H A Dide-atapi.c86 * Add a special packet command request to the tail of the request queue,
95 rq = blk_get_request(drive->queue, READ, __GFP_WAIT);
100 error = blk_rq_map_kern(drive->queue, rq, buf, bufflen,
109 error = blk_execute_rq(drive->queue, disk, rq, 0);
201 err = blk_rq_map_kern(drive->queue, sense_rq, sense, sense_len,
237 elv_add_request(drive->queue, &drive->sense_rq, ELEVATOR_INSERT_FRONT);
244 * We queue a request sense packet command at the head of the request
245 * queue.
/drivers/infiniband/hw/ehca/
H A Dehca_qp.c263 * init userspace queue info from ipz_queue data
266 struct ipz_queue *queue)
268 resp->qe_size = queue->qe_size;
269 resp->act_nr_of_sg = queue->act_nr_of_sg;
270 resp->queue_length = queue->queue_length;
271 resp->pagesize = queue->pagesize;
272 resp->toggle_state = queue->toggle_state;
273 resp->offset = queue->offset;
277 * init_qp_queue initializes/constructs r/squeue and registers queue pages.
282 struct ipz_queue *queue,
265 queue2resp(struct ipzu_queue_resp *resp, struct ipz_queue *queue) argument
279 init_qp_queue(struct ehca_shca *shca, struct ehca_pd *pd, struct ehca_qp *my_qp, struct ipz_queue *queue, int q_type, u64 expected_hret, struct ehca_alloc_queue_parms *parms, int wqe_size) argument
373 ehca_determine_small_queue(struct ehca_alloc_queue_parms *queue, int req_nr_sge, int is_llqp) argument
[all...]
/drivers/net/wireless/ath/carl9170/
H A Drx.c577 unsigned int queue; local
585 queue = TID_TO_WME_AC(((le16_to_cpu(bar->control) &
590 list_for_each_entry_rcu(entry, &ar->bar_list[queue], list) {
608 spin_lock_bh(&ar->bar_list_lock[queue]);
610 spin_unlock_bh(&ar->bar_list_lock[queue]);
/drivers/dma/
H A Dtxx9dmac.c180 return list_entry(dc->queue.next, struct txx9dmac_desc, desc_node);
347 /* The tasklet will hopefully advance the queue... */
452 } while (!list_empty(&dc->queue));
465 if (!list_empty(&dc->queue)) {
528 if (list_empty(&dc->active_list) && !list_empty(&dc->queue))
605 if (!list_empty(&dc->queue)) {
709 list_add_tail(&desc->desc_node, &dc->queue);
777 * the dc->queue list or dc->active_list after a
921 list_splice_init(&dc->queue, &list);
981 if (!list_empty(&dc->queue)) {
[all...]
H A Dep93xx_dma.c120 * @node: link used for putting this into a channel queue
144 * @queue: pending descriptors which are handled next
155 * @queue holds pending transactions. These are linked through the first
156 * descriptor in the chain. When a descriptor is moved to the @active queue,
178 struct list_head queue; member in struct:ep93xx_dma_chan
711 * function takes the next queued transaction from the @edmac->queue and
720 if (!list_empty(&edmac->active) || list_empty(&edmac->queue)) {
725 /* Take the next descriptor from the pending queue */
726 new = list_first_entry(&edmac->queue, struct ep93xx_dma_desc, node);
763 /* Pick up the next descriptor from the queue */
[all...]
/drivers/media/pci/cx23885/
H A Dcx23885-video.c105 struct cx23885_buffer, queue);
111 list_del(&buf->queue);
464 list_add_tail(&buf->queue, &q->active);
470 queue);
471 list_add_tail(&buf->queue, &q->active);
484 struct cx23885_buffer, queue);
500 struct cx23885_buffer, queue);
502 list_del(&buf->queue);
1260 dev->video_dev->queue = &dev->vb2_vidq;
1274 dev->vbi_dev->queue
[all...]
/drivers/media/platform/soc_camera/
H A Dmx3_camera.c67 struct list_head queue; member in struct:mx3_camera_buffer
157 list_del_init(&buf->queue);
176 struct mx3_camera_buffer, queue);
331 list_add_tail(&buf->queue, &mx3_cam->capture);
348 list_del_init(&buf->queue);
368 "Release%s DMA 0x%08x, queue %sempty\n",
370 list_empty(&buf->queue) ? "" : "not ");
378 list_del_init(&buf->queue);
400 INIT_LIST_HEAD(&buf->queue);
425 list_for_each_entry_safe(buf, tmp, &mx3_cam->capture, queue) {
[all...]
/drivers/scsi/ibmvscsi/
H A Dibmvfc.c560 list_for_each_entry(tgt, &vhost->targets, queue)
593 list_for_each_entry(tgt, &vhost->targets, queue)
722 /* Clean out the queue */
771 list_add_tail(&evt->queue, &vhost->free);
812 list_del(&evt->queue);
831 list_for_each_entry_safe(evt, pos, &vhost->sent, queue)
917 list_for_each_entry(tgt, &vhost->targets, queue)
1232 list_add_tail(&evt->queue, &vhost->free);
1251 list_del(&pool->events[i].queue);
1277 evt = list_entry(vhost->free.next, struct ibmvfc_event, queue);
3178 struct ibmvfc_crq_queue *queue = &vhost->crq; local
[all...]
/drivers/spi/
H A Dspi-bfin5xx.c70 /* Driver message queue */
74 struct list_head queue; member in struct:bfin_spi_master_data
875 /* pop a msg from queue and kick off real transfer */
883 /* Lock queue and check for queue work */
885 if (list_empty(&drv_data->queue) || !drv_data->running) {
898 /* Extract head of queue */
899 drv_data->cur_msg = list_entry(drv_data->queue.next,
900 struct spi_message, queue);
906 list_del_init(&drv_data->cur_msg->queue);
[all...]
/drivers/memstick/core/
H A Dms_block.c1891 msb->req = blk_fetch_request(msb->queue);
1907 blk_rq_map_sg(msb->queue, msb->req, sg);
2053 blk_stop_queue(msb->queue);
2062 blk_requeue_request(msb->queue, msb->req);
2090 blk_start_queue(msb->queue);
2129 msb->queue = blk_init_queue(msb_submit_req, &msb->q_lock);
2130 if (!msb->queue) {
2135 msb->queue->queuedata = card;
2136 blk_queue_prep_rq(msb->queue, msb_prepare_req);
2138 blk_queue_bounce_limit(msb->queue, limi
[all...]
/drivers/net/wireless/rtlwifi/
H A Dpci.c540 (ring->entries - skb_queue_len(&ring->queue) >
569 while (skb_queue_len(&ring->queue)) {
585 skb = __skb_dequeue(&ring->queue);
599 skb_queue_len(&ring->queue),
644 if ((ring->entries - skb_queue_len(&ring->queue))
650 skb_queue_len(&ring->queue));
1111 pskb = __skb_dequeue(&ring->queue);
1129 __skb_queue_tail(&ring->queue, pskb);
1159 *we just alloc 2 desc for beacon queue,
1164 /*BE queue nee
[all...]
/drivers/infiniband/hw/cxgb4/
H A Dcq.c72 cq->memsize, cq->queue,
102 cq->queue = dma_alloc_coherent(&rdev->lldi.pdev->dev, cq->memsize,
104 if (!cq->queue) {
109 memset(cq->queue, 0, cq->memsize);
168 dma_free_coherent(&rdev->lldi.pdev->dev, cq->memsize, cq->queue,
900 * Make actual HW queue 2x to avoid cdix_inc overflows.
905 * Make HW queue at least 64 entries so GTS updates aren't too
911 memsize = hwentries * sizeof *chp->cq.queue;
962 mm->addr = virt_to_phys(chp->cq.queue);
/drivers/net/wireless/iwlwifi/mvm/
H A Dmac-ctxt.c373 u8 queue = find_first_zero_bit(&used_hw_queues, local
376 if (queue >= mvm->first_agg_queue) {
377 IWL_ERR(mvm, "Failed to allocate queue\n");
382 __set_bit(queue, &used_hw_queues);
383 vif->hw_queue[ac] = queue;
386 /* Allocate the CAB queue for softAP and GO interfaces */
388 u8 queue = find_first_zero_bit(&used_hw_queues, local
391 if (queue >= mvm->first_agg_queue) {
392 IWL_ERR(mvm, "Failed to allocate cab queue\n");
397 vif->cab_queue = queue;
[all...]

Completed in 585 milliseconds

<<11121314151617181920>>