/drivers/spi/ |
H A D | spi-bfin-sport.c | 67 /* Driver message queue */ 71 struct list_head queue; member in struct:bfin_sport_spi_master_data 480 /* pop a msg from queue and kick off real transfer */ 490 /* Lock queue and check for queue work */ 492 if (list_empty(&drv_data->queue) || !drv_data->run) { 505 /* Extract head of queue */ 506 next_msg = list_entry(drv_data->queue.next, 507 struct spi_message, queue); 514 list_del_init(&drv_data->cur_msg->queue); [all...] |
/drivers/staging/lustre/lustre/include/ |
H A D | cl_object.h | 772 /** Link to a queue, for debugging. */ 1687 * becomes possible, cl_lock::cll_wq wait-queue 1876 * A 2-queue of pages. A convenience data-type for common use case, 2-queue 2109 * Submit pages from \a queue->c2_qin for IO, and move 2110 * successfully submitted pages into \a queue->c2_qout. Return 2113 * queue->c2_qout, completion callback with non-zero ioret is 2119 struct cl_2queue *queue); 2999 int cl_queue_match (const struct list_head *queue, 3065 enum cl_req_type iot, struct cl_2queue *queue); [all...] |
/drivers/staging/rtl8188eu/core/ |
H A D | rtw_cmd.c | 52 static int _rtw_enqueue_cmd(struct __queue *queue, struct cmd_obj *obj) argument 60 spin_lock_irqsave(&queue->lock, irqL); 62 list_add_tail(&obj->list, &queue->queue); 64 spin_unlock_irqrestore(&queue->lock, irqL); 72 struct cmd_obj *rtw_dequeue_cmd(struct __queue *queue) argument 78 spin_lock_irqsave(&queue->lock, irqL); 79 if (list_empty(&(queue->queue))) { 82 obj = container_of((&queue [all...] |
/drivers/staging/rtl8723au/core/ |
H A D | rtw_xmit.c | 89 &pxmitpriv->free_xmit_queue.queue); 118 &pxmitpriv->free_xmitbuf_queue.queue); 125 /* init xframe_ext queue, the same count as extbuf */ 145 &pxmitpriv->free_xframe_ext_queue.queue); 170 &pxmitpriv->free_xmit_extbuf_queue.queue); 214 list_for_each_safe(plist, ptmp, &pxmitpriv->free_xmit_queue.queue) { 228 /* free xframe_ext queue, the same count as extbuf */ 230 &pxmitpriv->free_xframe_ext_queue.queue) { 1001 return (!list_empty(&pxmitpriv->be_pending.queue)) || 1002 (!list_empty(&pxmitpriv->bk_pending.queue)) || 1528 struct rtw_queue *queue = &pxmitpriv->free_xframe_ext_queue; local 1554 struct rtw_queue *queue = NULL; local [all...] |
/drivers/usb/gadget/udc/ |
H A D | fsl_qe_udc.c | 85 /* the req->queue pointer is used by ep_queue() func, in which 86 * the request will be added into a udc_ep->queue 'd tail 87 * so here the req will be dropped from the ep->queue 89 list_del_init(&req->queue); 117 /* don't modify queue heads during completion callback */ 134 while (!list_empty(&ep->queue)) { 136 req = list_entry(ep->queue.next, struct qe_req, queue); 783 if (ep->has_data <= 0 && (!list_empty(&ep->queue))) 910 if (list_empty(&ep->queue)) { [all...] |
H A D | mv_u3d_core.c | 193 dev_dbg(u3d->dev, "mv_u3d_done: remove req->queue\n"); 194 /* Removed the req from ep queue */ 195 list_del_init(&req->queue); 248 if (!list_empty(&ep->queue)) { 249 dev_err(u3d->dev, "add trb to non-empty queue!\n"); 505 /* build trbs and push them to device queue */ 518 /* irq handler advances the queue */ 520 list_add_tail(&req->queue, &ep->queue); 687 INIT_LIST_HEAD(&req->queue); [all...] |
/drivers/block/mtip32xx/ |
H A D | mtip32xx.c | 161 if (dd->queue) 162 set_bit(QUEUE_FLAG_DEAD, &dd->queue->queue_flags); 165 "%s: dd->queue is NULL\n", __func__); 182 rq = blk_mq_alloc_request(dd->queue, 0, __GFP_WAIT, true); 197 struct blk_mq_hw_ctx *hctx = dd->queue->queue_hw_ctx[0]; 1038 blk_mq_stop_hw_queues(port->dd->queue); 1066 blk_mq_start_stopped_hw_queues(port->dd->queue, true); 1069 blk_mq_start_stopped_hw_queues(port->dd->queue, true); 2814 if (dd->queue) { 2815 dd->queue [all...] |
/drivers/gpu/drm/via/ |
H A D | via_dmablit.c | 398 via_dmablit_active(drm_via_blitq_t *blitq, int engine, uint32_t handle, wait_queue_head_t **queue) argument 413 if (queue && active) { 417 *queue = blitq->blit_queue + slot; 435 wait_queue_head_t *queue; local 438 if (via_dmablit_active(blitq, engine, handle, &queue)) { 439 DRM_WAIT_ON(ret, *queue, 3 * HZ, 676 * Reserve one free slot in the blit queue. Will wait for one second for one 720 * Grab a free slot. Build blit info and queue a blit. 795 * while waiting for a free slot in the blit queue. In that case it returns with -EAGAIN and should
|
/drivers/media/v4l2-core/ |
H A D | v4l2-mem2mem.c | 48 /* Offset base for buffers on the destination queue - used to distinguish 184 struct v4l2_m2m_ctx, queue); 193 * the pending job queue and add it if so. 201 * If a queue is buffered (for example a decoder hardware ringbuffer that has 203 * on that queue. 236 dprintk("On job queue already\n"); 270 list_add_tail(&m2m_ctx->queue, &m2m_dev->job_queue); 303 list_del(&m2m_ctx->queue); 306 dprintk("m2m_ctx: %p had been on queue and was removed\n", 309 /* Do nothing, was not on queue/runnin [all...] |
/drivers/net/wireless/rtlwifi/rtl8192cu/ |
H A D | phy.c | 461 if (skb_queue_len(&ring->queue) == 0 || 470 skb_queue_len(&ring->queue)); 479 skb_queue_len(&ring->queue)); 504 if (skb_queue_len(&ring->queue) == 0) { 511 skb_queue_len(&ring->queue)); 520 skb_queue_len(&ring->queue));
|
/drivers/block/rsxx/ |
H A D | rsxx_priv.h | 98 u32 sw_q_depth; /* Number of DMAs on the SW queue. */ 110 struct list_head queue; member in struct:rsxx_dma_ctrl 141 struct list_head queue; member in struct:rsxx_cardinfo::__anon366 172 struct request_queue *queue; member in struct:rsxx_cardinfo
|
/drivers/gpu/host1x/ |
H A D | intr.c | 44 * add a waiter to a waiter queue, sorted by threshold 45 * returns true if it was added at the head of the queue 48 struct list_head *queue) 53 list_for_each_entry_reverse(pos, queue, list) 59 list_add(&waiter->list, queue); 64 * run through a waiter queue for a single sync point ID 47 add_waiter_to_queue(struct host1x_waitlist *waiter, struct list_head *queue) argument
|
/drivers/net/wireless/iwlwifi/pcie/ |
H A D | internal.h | 77 * struct iwl_rxq - Rx queue 81 * @queue: 99 struct iwl_rx_mem_buffer *queue[RX_QUEUE_SIZE]; member in struct:iwl_rxq 119 * iwl_queue_inc_wrap - increment queue index, wrap back to beginning 128 * iwl_queue_dec_wrap - decrement queue index, wrap back to end 143 * Generic queue structure 150 * queues, n_window, which is the size of the software queue data 151 * is also 256; however, for the command queue, n_window is only 161 * data is a window overlayed over the HW queue. 166 /* use for monitoring and recovering the stuck queue */ [all...] |
/drivers/net/wireless/ti/wlcore/ |
H A D | tx.c | 301 /* queue */ 554 /* we found the first non-empty low priority queue */ 1154 /* only one flush should be in progress, for consistent queue state */ 1211 u8 queue, enum wlcore_queue_stop_reason reason) 1213 int hwq = wlcore_tx_get_mac80211_queue(wlvif, queue); 1216 /* queue should not be stopped for this reason */ 1225 void wlcore_stop_queue(struct wl1271 *wl, struct wl12xx_vif *wlvif, u8 queue, argument 1231 wlcore_stop_queue_locked(wl, wlvif, queue, reason); 1235 void wlcore_wake_queue(struct wl1271 *wl, struct wl12xx_vif *wlvif, u8 queue, argument 1239 int hwq = wlcore_tx_get_mac80211_queue(wlvif, queue); 1210 wlcore_stop_queue_locked(struct wl1271 *wl, struct wl12xx_vif *wlvif, u8 queue, enum wlcore_queue_stop_reason reason) argument 1297 wlcore_is_queue_stopped_by_reason(struct wl1271 *wl, struct wl12xx_vif *wlvif, u8 queue, enum wlcore_queue_stop_reason reason) argument 1312 wlcore_is_queue_stopped_by_reason_locked(struct wl1271 *wl, struct wl12xx_vif *wlvif, u8 queue, enum wlcore_queue_stop_reason reason) argument 1322 wlcore_is_queue_stopped_locked(struct wl1271 *wl, struct wl12xx_vif *wlvif, u8 queue) argument [all...] |
/drivers/pps/ |
H A D | pps.c | 56 poll_wait(file, &pps->queue, wait); 159 err = wait_event_interruptible(pps->queue, 172 pps->queue,
|
/drivers/scsi/aacraid/ |
H A D | dpcsup.c | 49 * know there is a response on our normal priority queue. We will pull off 51 * take a spinlock out on the queue before operating on it. 65 * Keep pulling response QEs off the response queue and waking 80 * Remove this fib from the Outstanding I/O queue. 87 dev->queues->queue[AdapNormCmdQueue].numpending--; 165 * @q: queue to process 168 * let us know there is a command on our normal priority queue. We will 170 * We will take a spinlock out on the queue before operating on it. 182 * Keep pulling response QEs off the response queue and waking 282 * know there is a response on our normal priority queue [all...] |
/drivers/block/ |
H A D | umem.c | 25 * - Never block on make_request. queue 113 struct request_queue *queue; member in struct:cardinfo 893 card->queue = blk_alloc_queue(GFP_KERNEL); 894 if (!card->queue) 897 blk_queue_make_request(card->queue, mm_make_request); 898 card->queue->queue_lock = &card->lock; 899 card->queue->queuedata = card; 1039 blk_cleanup_queue(card->queue); 1097 disk->queue = cards[i].queue; [all...] |
/drivers/dma/ |
H A D | pch_dma.c | 112 struct list_head queue; member in struct:pch_dma_chan 185 return list_first_entry(&pd_chan->queue, 381 if (!list_empty(&pd_chan->queue)) 385 list_splice_init(&pd_chan->queue, &pd_chan->active_list); 398 list_splice_init(&pd_chan->queue, pd_chan->active_list.prev); 434 list_add_tail(&desc->desc_node, &pd_chan->queue); 552 BUG_ON(!list_empty(&pd_chan->queue)); 683 list_splice_init(&pd_chan->queue, &list); 918 INIT_LIST_HEAD(&pd_chan->queue);
|
H A D | timb_dma.c | 91 struct list_head queue; member in struct:timb_dma_chan 296 BUG_ON(list_empty(&td_chan->queue)); 299 td_desc = list_entry(td_chan->queue.next, struct timb_dma_desc, 329 list_add_tail(&td_desc->desc_node, &td_chan->queue); 462 BUG_ON(!list_empty(&td_chan->queue)); 502 if (list_empty(&td_chan->active_list) && !list_empty(&td_chan->queue)) 576 /* first the easy part, put the queue into the free list */ 578 list_for_each_entry_safe(td_desc, _td_desc, &td_chan->queue, 608 if (!list_empty(&td_chan->queue)) 722 INIT_LIST_HEAD(&td_chan->queue); [all...] |
/drivers/hsi/controllers/ |
H A D | omap_ssi_port.c | 310 static int ssi_start_transfer(struct list_head *queue) argument 315 if (list_empty(queue)) 317 msg = list_first_entry(queue, struct hsi_msg, link); 370 struct list_head *queue; local 383 queue = &omap_port->txqueue[msg->channel]; 386 queue = &omap_port->rxqueue[msg->channel]; 390 list_add_tail(&msg->link, queue); 391 err = ssi_start_transfer(queue); 419 static void ssi_flush_queue(struct list_head *queue, struct hsi_client *cl) argument 424 list_for_each_safe(node, tmp, queue) { 602 ssi_transfer(struct omap_ssi_port *omap_port, struct list_head *queue) argument 850 ssi_pio_complete(struct hsi_port *port, struct list_head *queue) argument [all...] |
/drivers/i2c/busses/ |
H A D | i2c-octeon.c | 64 wait_queue_head_t queue; member in struct:octeon_i2c 184 wake_up(&i2c->queue); 207 result = wait_event_timeout(i2c->queue, 564 init_waitqueue_head(&i2c->queue);
|
/drivers/media/usb/au0828/ |
H A D | au0828-vbi.c | 124 list_add_tail(&buf->vb.queue, &vbiq->active);
|
/drivers/net/wireless/prism54/ |
H A D | isl_38xx.h | 59 /* High end mobos queue up pci writes, the following 158 /* determine number of entries currently in queue */ 159 int isl38xx_in_queue(isl38xx_control_block *cb, int queue);
|
/drivers/net/wireless/rt2x00/ |
H A D | rt2x00dev.c | 278 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; 323 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; 481 * If the data queue was below the threshold before the txdone 482 * handler we must make sure the packet queue in the mac80211 stack 484 * serialized with rt2x00mac_tx(), otherwise we can wake up queue 487 spin_lock_bh(&entry->queue->tx_lock); 488 if (!rt2x00queue_threshold(entry->queue)) 489 rt2x00queue_unpause_queue(entry->queue); 490 spin_unlock_bh(&entry->queue->tx_lock); 685 struct rt2x00_dev *rt2x00dev = entry->queue [all...] |
/drivers/s390/block/ |
H A D | dasd_genhd.c | 76 gdp->queue = block->request_queue;
|