Searched refs:queue (Results 1 - 25 of 575) sorted by relevance

1234567891011>>

/drivers/media/video/uvc/
H A Duvc_queue.c28 * Video buffers queue management.
34 * the videobuf2 queue operations by serializing calls to videobuf2 and a
35 * spinlock to protect the IRQ queue that holds the buffers to be processed by
40 * videobuf2 queue operations
47 struct uvc_video_queue *queue = vb2_get_drv_priv(vq); local
49 container_of(queue, struct uvc_streaming, queue);
63 struct uvc_video_queue *queue = vb2_get_drv_priv(vb->vb2_queue); local
72 if (unlikely(queue->flags & UVC_QUEUE_DISCONNECTED))
89 struct uvc_video_queue *queue local
109 struct uvc_video_queue *queue = vb2_get_drv_priv(vb->vb2_queue); local
125 uvc_queue_init(struct uvc_video_queue *queue, enum v4l2_buf_type type, int drop_corrupted) argument
146 uvc_alloc_buffers(struct uvc_video_queue *queue, struct v4l2_requestbuffers *rb) argument
158 uvc_free_buffers(struct uvc_video_queue *queue) argument
165 uvc_query_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf) argument
176 uvc_queue_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf) argument
187 uvc_dequeue_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf, int nonblocking) argument
199 uvc_queue_mmap(struct uvc_video_queue *queue, struct vm_area_struct *vma) argument
210 uvc_queue_poll(struct uvc_video_queue *queue, struct file *file, poll_table *wait) argument
229 uvc_queue_allocated(struct uvc_video_queue *queue) argument
246 uvc_queue_get_unmapped_area(struct uvc_video_queue *queue, unsigned long pgoff) argument
286 uvc_queue_enable(struct uvc_video_queue *queue, int enable) argument
325 uvc_queue_cancel(struct uvc_video_queue *queue, int disconnect) argument
349 uvc_queue_next_buffer(struct uvc_video_queue *queue, struct uvc_buffer *buf) argument
[all...]
/drivers/scsi/arm/
H A Dqueue.h2 * linux/drivers/acorn/scsi/queue.h: queue handling
21 * Function: void queue_initialise (Queue_t *queue)
22 * Purpose : initialise a queue
23 * Params : queue - queue to initialise
25 extern int queue_initialise (Queue_t *queue);
28 * Function: void queue_free (Queue_t *queue)
29 * Purpose : free a queue
30 * Params : queue
[all...]
H A Dqueue.c2 * linux/drivers/acorn/scsi/queue.c: queue handling primitives
49 #include "queue.h"
54 * Function: void queue_initialise (Queue_t *queue)
55 * Purpose : initialise a queue
56 * Params : queue - queue to initialise
58 int queue_initialise (Queue_t *queue) argument
63 spin_lock_init(&queue->queue_lock);
64 INIT_LIST_HEAD(&queue
90 queue_free(Queue_t *queue) argument
106 __queue_add(Queue_t *queue, struct scsi_cmnd *SCpnt, int head) argument
137 __queue_remove(Queue_t *queue, struct list_head *ent) argument
161 queue_remove_exclude(Queue_t *queue, unsigned long *exclude) argument
186 queue_remove(Queue_t *queue) argument
208 queue_remove_tgtluntag(Queue_t *queue, int target, int lun, int tag) argument
236 queue_remove_all_target(Queue_t *queue, int target) argument
259 queue_probetgtlun(Queue_t *queue, int target, int lun) argument
285 queue_remove_cmd(Queue_t *queue, struct scsi_cmnd *SCpnt) argument
[all...]
H A DMakefile7 obj-$(CONFIG_SCSI_ACORNSCSI_3) += acornscsi_mod.o queue.o msgqueue.o
8 obj-$(CONFIG_SCSI_ARXESCSI) += arxescsi.o fas216.o queue.o msgqueue.o
10 obj-$(CONFIG_SCSI_CUMANA_2) += cumana_2.o fas216.o queue.o msgqueue.o
12 obj-$(CONFIG_SCSI_POWERTECSCSI) += powertec.o fas216.o queue.o msgqueue.o
13 obj-$(CONFIG_SCSI_EESOXSCSI) += eesox.o fas216.o queue.o msgqueue.o
/drivers/infiniband/hw/ehca/
H A Dipz_pt_fn.c4 * internal queue handling
51 void *ipz_qpageit_get_inc(struct ipz_queue *queue) argument
53 void *ret = ipz_qeit_get(queue);
54 queue->current_q_offset += queue->pagesize;
55 if (queue->current_q_offset > queue->queue_length) {
56 queue->current_q_offset -= queue->pagesize;
59 if (((u64)ret) % queue
66 ipz_qeit_eq_get_inc(struct ipz_queue *queue) argument
80 ipz_queue_abs_to_offset(struct ipz_queue *queue, u64 addr, u64 *q_offset) argument
102 alloc_queue_pages(struct ipz_queue *queue, const u32 nr_of_pages) argument
127 alloc_small_queue_page(struct ipz_queue *queue, struct ehca_pd *pd) argument
172 free_small_queue_page(struct ipz_queue *queue, struct ehca_pd *pd) argument
204 ipz_queue_ctor(struct ehca_pd *pd, struct ipz_queue *queue, const u32 nr_of_pages, const u32 pagesize, const u32 qe_size, const u32 nr_of_sg, int is_small) argument
255 ipz_queue_dtor(struct ehca_pd *pd, struct ipz_queue *queue) argument
[all...]
H A Dipz_pt_fn.h4 * internal queue handling
75 /* struct generic queue in linux kernel virtual memory (kv) */
77 u64 current_q_offset; /* current queue entry */
79 struct ipz_page **queue_pages; /* array of pages belonging to queue */
80 u32 qe_size; /* queue entry size */
82 u32 queue_length; /* queue length allocated in bytes */
93 static inline void *ipz_qeit_calc(struct ipz_queue *queue, u64 q_offset) argument
96 if (q_offset >= queue->queue_length)
98 current_page = (queue->queue_pages)[q_offset >> EHCA_PAGESHIFT];
106 static inline void *ipz_qeit_get(struct ipz_queue *queue) argument
126 ipz_qeit_get_inc(struct ipz_queue *queue) argument
142 ipz_qeit_is_valid(struct ipz_queue *queue) argument
155 ipz_qeit_get_inc_valid(struct ipz_queue *queue) argument
164 ipz_qeit_reset(struct ipz_queue *queue) argument
178 ipz_queue_advance_offset(struct ipz_queue *queue, u64 offset) argument
258 ipz_eqit_eq_get_inc_valid(struct ipz_queue *queue) argument
268 ipz_eqit_eq_peek_valid(struct ipz_queue *queue) argument
[all...]
/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_marker.c37 void vmw_marker_queue_init(struct vmw_marker_queue *queue) argument
39 INIT_LIST_HEAD(&queue->head);
40 queue->lag = ns_to_timespec(0);
41 getrawmonotonic(&queue->lag_time);
42 spin_lock_init(&queue->lock);
45 void vmw_marker_queue_takedown(struct vmw_marker_queue *queue) argument
49 spin_lock(&queue->lock);
50 list_for_each_entry_safe(marker, next, &queue->head, head) {
53 spin_unlock(&queue->lock);
56 int vmw_marker_push(struct vmw_marker_queue *queue, argument
73 vmw_marker_pull(struct vmw_marker_queue *queue, uint32_t signaled_seqno) argument
120 vmw_fifo_lag(struct vmw_marker_queue *queue) argument
134 vmw_lag_lt(struct vmw_marker_queue *queue, uint32_t us) argument
144 vmw_wait_lag(struct vmw_private *dev_priv, struct vmw_marker_queue *queue, uint32_t us) argument
[all...]
/drivers/net/wireless/b43legacy/
H A Dpio.c35 static void tx_start(struct b43legacy_pioqueue *queue) argument
37 b43legacy_pio_write(queue, B43legacy_PIO_TXCTL,
41 static void tx_octet(struct b43legacy_pioqueue *queue, argument
44 if (queue->need_workarounds) {
45 b43legacy_pio_write(queue, B43legacy_PIO_TXDATA, octet);
46 b43legacy_pio_write(queue, B43legacy_PIO_TXCTL,
49 b43legacy_pio_write(queue, B43legacy_PIO_TXCTL,
51 b43legacy_pio_write(queue, B43legacy_PIO_TXDATA, octet);
76 static void tx_data(struct b43legacy_pioqueue *queue, argument
84 if (queue
102 tx_complete(struct b43legacy_pioqueue *queue, struct sk_buff *skb) argument
116 generate_cookie(struct b43legacy_pioqueue *queue, struct b43legacy_pio_txpacket *packet) argument
154 struct b43legacy_pioqueue *queue = NULL; local
185 pio_tx_write_fragment(struct b43legacy_pioqueue *queue, struct sk_buff *skb, struct b43legacy_pio_txpacket *packet, size_t txhdr_size) argument
218 struct b43legacy_pioqueue *queue = packet->queue; local
232 struct b43legacy_pioqueue *queue = packet->queue; local
282 struct b43legacy_pioqueue *queue = (struct b43legacy_pioqueue *)d; local
312 setup_txqueues(struct b43legacy_pioqueue *queue) argument
332 struct b43legacy_pioqueue *queue; local
381 cancel_transfers(struct b43legacy_pioqueue *queue) argument
393 b43legacy_destroy_pioqueue(struct b43legacy_pioqueue *queue) argument
423 struct b43legacy_pioqueue *queue; local
469 struct b43legacy_pioqueue *queue = dev->pio.queue1; local
491 struct b43legacy_pioqueue *queue; local
549 pio_rx_error(struct b43legacy_pioqueue *queue, int clear_buffers, const char *error) argument
567 b43legacy_pio_rx(struct b43legacy_pioqueue *queue) argument
648 b43legacy_pio_tx_suspend(struct b43legacy_pioqueue *queue) argument
656 b43legacy_pio_tx_resume(struct b43legacy_pioqueue *queue) argument
[all...]
H A Dpio.h42 struct b43legacy_pioqueue *queue; member in struct:b43legacy_pio_txpacket
48 (packet)->queue->tx_packets_cache))
73 /* Packets on the txrunning queue are completely
83 u16 b43legacy_pio_read(struct b43legacy_pioqueue *queue, argument
86 return b43legacy_read16(queue->dev, queue->mmio_base + offset);
90 void b43legacy_pio_write(struct b43legacy_pioqueue *queue, argument
93 b43legacy_write16(queue->dev, queue->mmio_base + offset, value);
105 void b43legacy_pio_rx(struct b43legacy_pioqueue *queue);
137 b43legacy_pio_rx(struct b43legacy_pioqueue *queue) argument
141 b43legacy_pio_tx_suspend(struct b43legacy_pioqueue *queue) argument
145 b43legacy_pio_tx_resume(struct b43legacy_pioqueue *queue) argument
[all...]
/drivers/usb/gadget/
H A Duvc_queue.c26 * Video buffers queue management.
32 * uvc_free_buffers respectively. The former acquires the video queue lock,
38 * drivers that use an in queue and an out queue, we use a main queue to hold
39 * all queued buffers (both 'empty' and 'done' buffers), and an irq queue to
41 * in interrupt, as only one queue is shared between interrupt and user
47 * Unless stated otherwise, all operations that modify the irq buffers queue
53 * protected by the queue lock, and the later is protected by the irq
56 * The completion handler fetches a buffer from the irq queue an
81 uvc_queue_init(struct uvc_video_queue *queue, enum v4l2_buf_type type) argument
95 uvc_free_buffers(struct uvc_video_queue *queue) argument
121 uvc_alloc_buffers(struct uvc_video_queue *queue, unsigned int nbuffers, unsigned int buflength) argument
200 uvc_query_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *v4l2_buf) argument
222 uvc_queue_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *v4l2_buf) argument
303 uvc_dequeue_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *v4l2_buf, int nonblocking) argument
365 uvc_queue_poll(struct uvc_video_queue *queue, struct file *file, poll_table *wait) argument
414 uvc_queue_mmap(struct uvc_video_queue *queue, struct vm_area_struct *vma) argument
476 uvc_queue_cancel(struct uvc_video_queue *queue, int disconnect) argument
517 uvc_queue_enable(struct uvc_video_queue *queue, int enable) argument
548 uvc_queue_next_buffer(struct uvc_video_queue *queue, struct uvc_buffer *buf) argument
573 uvc_queue_head(struct uvc_video_queue *queue) argument
[all...]
H A Duvc_video.c35 if (buf->buf.bytesused - video->queue.buf_used <= len - 2)
45 struct uvc_video_queue *queue = &video->queue; local
50 mem = queue->mem + buf->buf.m.offset + queue->buf_used;
51 nbytes = min((unsigned int)len, buf->buf.bytesused - queue->buf_used);
54 queue->buf_used += nbytes;
85 if (buf->buf.bytesused == video->queue.buf_used) {
86 video->queue.buf_used = 0;
88 uvc_queue_next_buffer(&video->queue, bu
[all...]
H A Duvc_queue.h33 struct list_head queue; member in struct:uvc_buffer
61 static inline int uvc_queue_streaming(struct uvc_video_queue *queue) argument
63 return queue->flags & UVC_QUEUE_STREAMING;
/drivers/net/wireless/rt2x00/
H A Drt2x00queue.c25 Abstract: rt2x00 queue specific routines.
38 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
49 frame_size = entry->queue->data_size + entry->queue->desc_size;
102 struct device *dev = entry->queue->rt2x00dev->dev;
113 struct device *dev = entry->queue->rt2x00dev->dev;
494 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
499 * a queue corruption!
504 "Corrupt queue %d, accessing entry which is not ours.\n"
506 entry->queue
534 struct data_queue *queue = entry->queue; local
545 rt2x00queue_kick_tx_queue(struct data_queue *queue, struct txentry_desc *txdesc) argument
562 rt2x00queue_write_tx_frame(struct data_queue *queue, struct sk_buff *skb, bool local) argument
753 rt2x00queue_for_each_entry(struct data_queue *queue, enum queue_index start, enum queue_index end, void *data, bool (*fn)(struct queue_entry *entry, void *data)) argument
808 rt2x00queue_get_entry(struct data_queue *queue, enum queue_index index) argument
832 struct data_queue *queue = entry->queue; local
859 rt2x00queue_pause_queue(struct data_queue *queue) argument
883 rt2x00queue_unpause_queue(struct data_queue *queue) argument
913 rt2x00queue_start_queue(struct data_queue *queue) argument
933 rt2x00queue_stop_queue(struct data_queue *queue) argument
950 rt2x00queue_flush_queue(struct data_queue *queue, bool drop) argument
1011 struct data_queue *queue; local
1026 struct data_queue *queue; local
1045 struct data_queue *queue; local
1054 rt2x00queue_reset(struct data_queue *queue) argument
1072 struct data_queue *queue; local
1083 rt2x00queue_alloc_entries(struct data_queue *queue, const struct data_queue_desc *qdesc) argument
1126 rt2x00queue_free_skbs(struct data_queue *queue) argument
1138 rt2x00queue_alloc_rxskbs(struct data_queue *queue) argument
1155 struct data_queue *queue; local
1195 struct data_queue *queue; local
1205 rt2x00queue_init(struct rt2x00_dev *rt2x00dev, struct data_queue *queue, enum data_queue_qid qid) argument
1222 struct data_queue *queue; local
[all...]
H A Drt2x00usb.c244 struct data_queue *queue; local
247 tx_queue_for_each(rt2x00dev, queue) {
248 while (!rt2x00queue_empty(queue)) {
249 entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE);
263 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
290 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
318 usb_sndbulkpipe(usb_dev, entry->queue->usb_endpoint),
356 skbdesc->desc_len = entry->queue->desc_size;
368 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
383 if (urb->actual_length < entry->queue
422 rt2x00usb_kick_queue(struct data_queue *queue) argument
471 rt2x00usb_flush_queue(struct data_queue *queue, bool drop) argument
521 rt2x00usb_watchdog_tx_dma(struct data_queue *queue) argument
529 rt2x00usb_dma_timeout(struct data_queue *queue) argument
539 struct data_queue *queue; local
572 rt2x00usb_assign_endpoint(struct data_queue *queue, struct usb_endpoint_descriptor *ep_desc) argument
597 struct data_queue *queue = rt2x00dev->tx; local
642 rt2x00usb_alloc_entries(struct data_queue *queue) argument
675 rt2x00usb_free_entries(struct data_queue *queue) argument
709 struct data_queue *queue; local
739 struct data_queue *queue; local
[all...]
H A Drt2x00pci.c65 struct data_queue *queue = rt2x00dev->rx; local
72 entry = rt2x00queue_get_entry(queue, Q_INDEX);
83 skbdesc->desc_len = entry->queue->desc_size;
102 void rt2x00pci_flush_queue(struct data_queue *queue, bool drop) argument
106 for (i = 0; !rt2x00queue_empty(queue) && i < 10; i++)
115 struct data_queue *queue)
126 queue->limit * queue->desc_size,
131 memset(addr, 0, queue->limit * queue
114 rt2x00pci_alloc_queue_dma(struct rt2x00_dev *rt2x00dev, struct data_queue *queue) argument
145 rt2x00pci_free_queue_dma(struct rt2x00_dev *rt2x00dev, struct data_queue *queue) argument
160 struct data_queue *queue; local
196 struct data_queue *queue; local
[all...]
/drivers/media/video/omap3isp/
H A Dispqueue.h4 * TI OMAP3 ISP - Video buffers queue handling
65 * @stream: List head for insertion into main queue
66 * @queue: ISP buffers queue this buffer belongs to
78 * @irqlist: List head for insertion into IRQ queue
80 * @wait: Wait queue to signal buffer completion
85 struct isp_video_queue *queue; member in struct:isp_video_buffer
118 * the userspace memory address for a USERPTR buffer, with the queue lock
121 * @buffer_queue: Called when a buffer is being added to the queue with the
122 * queue irqloc
162 struct list_head queue; member in struct:isp_video_queue
[all...]
H A Dispqueue.c4 * TI OMAP3 ISP - Video buffers queue handling
266 if (buf->queue->ops->buffer_cleanup)
267 buf->queue->ops->buffer_cleanup(buf);
272 dma_unmap_sg(buf->queue->dev, buf->sglist, buf->sglen,
520 ret = dma_map_sg(buf->queue->dev, buf->sglist, buf->sglen,
528 if (buf->queue->ops->buffer_prepare)
529 ret = buf->queue->ops->buffer_prepare(buf);
543 * Locking: must be called with the queue lock held.
575 * queue using the same condition.
597 * Buffers can only be freed if the queue is
602 isp_video_queue_free(struct isp_video_queue *queue) argument
636 isp_video_queue_alloc(struct isp_video_queue *queue, unsigned int nbuffers, unsigned int size, enum v4l2_memory memory) argument
704 omap3isp_video_queue_cleanup(struct isp_video_queue *queue) argument
730 omap3isp_video_queue_init(struct isp_video_queue *queue, enum v4l2_buf_type type, const struct isp_video_queue_operations *ops, struct device *dev, unsigned int bufsize) argument
772 omap3isp_video_queue_reqbufs(struct isp_video_queue *queue, struct v4l2_requestbuffers *rb) argument
810 omap3isp_video_queue_querybuf(struct isp_video_queue *queue, struct v4l2_buffer *vbuf) argument
847 omap3isp_video_queue_qbuf(struct isp_video_queue *queue, struct v4l2_buffer *vbuf) argument
919 omap3isp_video_queue_dqbuf(struct isp_video_queue *queue, struct v4l2_buffer *vbuf, int nonblocking) argument
960 omap3isp_video_queue_streamon(struct isp_video_queue *queue) argument
992 omap3isp_video_queue_streamoff(struct isp_video_queue *queue) argument
1033 omap3isp_video_queue_discard_done(struct isp_video_queue *queue) argument
1081 omap3isp_video_queue_mmap(struct isp_video_queue *queue, struct vm_area_struct *vma) argument
1132 omap3isp_video_queue_poll(struct isp_video_queue *queue, struct file *file, poll_table *wait) argument
[all...]
/drivers/net/ethernet/ibm/ehea/
H A Dehea_qmr.h210 static inline void *hw_qeit_calc(struct hw_queue *queue, u64 q_offset) argument
214 if (q_offset >= queue->queue_length)
215 q_offset -= queue->queue_length;
216 current_page = (queue->queue_pages)[q_offset >> EHEA_PAGESHIFT];
220 static inline void *hw_qeit_get(struct hw_queue *queue) argument
222 return hw_qeit_calc(queue, queue->current_q_offset);
225 static inline void hw_qeit_inc(struct hw_queue *queue) argument
227 queue->current_q_offset += queue
235 hw_qeit_get_inc(struct hw_queue *queue) argument
242 hw_qeit_get_inc_valid(struct hw_queue *queue) argument
259 hw_qeit_get_valid(struct hw_queue *queue) argument
275 hw_qeit_reset(struct hw_queue *queue) argument
281 hw_qeit_eq_get_inc(struct hw_queue *queue) argument
295 hw_eqit_eq_get_inc_valid(struct hw_queue *queue) argument
309 struct hw_queue *queue; local
324 struct hw_queue *queue = &my_qp->hw_squeue; local
341 struct hw_queue *queue = &qp->hw_rqueue1; local
[all...]
/drivers/scsi/ibmvscsi/
H A Drpa_vscsi.c44 * Routines for managing the command/response queue
65 * @queue: crq_queue to initialize and register
71 static void rpavscsi_release_crq_queue(struct crq_queue *queue, argument
85 queue->msg_token,
86 queue->size * sizeof(*queue->msgs), DMA_BIDIRECTIONAL);
87 free_page((unsigned long)queue->msgs);
91 * crq_queue_next_crq: - Returns the next entry in message queue
92 * @queue: crq_queue to use
94 * Returns pointer to next entry in queue, o
97 crq_queue_next_crq(struct crq_queue *queue) argument
203 rpavscsi_reset_crq_queue(struct crq_queue *queue, struct ibmvscsi_host_data *hostdata) argument
244 rpavscsi_init_crq_queue(struct crq_queue *queue, struct ibmvscsi_host_data *hostdata, int max_requests) argument
331 rpavscsi_reenable_crq_queue(struct crq_queue *queue, struct ibmvscsi_host_data *hostdata) argument
[all...]
/drivers/net/wireless/ath/ath5k/
H A Dqcu.c33 * basically we have 10 queues to play with. Each queue has a matching
34 * QCU that controls when the queue will get triggered and multiple QCUs
37 * and DCUs allowing us to have different DFS settings for each queue.
39 * When a frame goes into a TX queue, QCU decides when it'll trigger a
41 * it's buffer or -if it's a beacon queue- if it's time to fire up the queue
56 * ath5k_hw_num_tx_pending() - Get number of pending frames for a given queue
58 * @queue: One of enum ath5k_tx_queue_id
61 ath5k_hw_num_tx_pending(struct ath5k_hw *ah, unsigned int queue) argument
64 AR5K_ASSERT_ENTRY(queue, a
92 ath5k_hw_release_tx_queue(struct ath5k_hw *ah, unsigned int queue) argument
136 ath5k_hw_get_tx_queueprops(struct ath5k_hw *ah, int queue, struct ath5k_txq_info *queue_info) argument
152 ath5k_hw_set_tx_queueprops(struct ath5k_hw *ah, int queue, const struct ath5k_txq_info *qinfo) argument
204 unsigned int queue; local
285 ath5k_hw_set_tx_retry_limits(struct ath5k_hw *ah, unsigned int queue) argument
328 ath5k_hw_reset_tx_queue(struct ath5k_hw *ah, unsigned int queue) argument
[all...]
H A Ddma.c27 * handle queue setup for 5210 chipset (rest are handled on qcu.c).
114 * ath5k_hw_start_tx_dma() - Start DMA transmit for a specific queue
116 * @queue: The hw queue number
118 * Start DMA transmit for a specific queue and since 5210 doesn't have
119 * QCU/DCU, set up queue parameters for 5210 here based on queue type (one
120 * queue for normal data and one queue for beacons). For queue setu
128 ath5k_hw_start_tx_dma(struct ath5k_hw *ah, unsigned int queue) argument
186 ath5k_hw_stop_tx_dma(struct ath5k_hw *ah, unsigned int queue) argument
326 ath5k_hw_stop_beacon_queue(struct ath5k_hw *ah, unsigned int queue) argument
351 ath5k_hw_get_txdp(struct ath5k_hw *ah, unsigned int queue) argument
394 ath5k_hw_set_txdp(struct ath5k_hw *ah, unsigned int queue, u32 phys_addr) argument
[all...]
/drivers/net/
H A Deql.c140 static void eql_kill_one_slave(slave_queue_t *queue, slave_t *slave);
147 spin_lock(&eql->queue.lock);
148 head = &eql->queue.all_slaves;
157 eql_kill_one_slave(&eql->queue, slave);
161 spin_unlock(&eql->queue.lock);
186 spin_lock_init(&eql->queue.lock);
187 INIT_LIST_HEAD(&eql->queue.all_slaves);
188 eql->queue.master_dev = dev;
213 BUG_ON(!list_empty(&eql->queue.all_slaves));
223 static void eql_kill_one_slave(slave_queue_t *queue, slave_ argument
232 eql_kill_slave_queue(slave_queue_t *queue) argument
298 __eql_schedule_slaves(slave_queue_t *queue) argument
364 __eql_find_slave_dev(slave_queue_t *queue, struct net_device *dev) argument
379 eql_is_full(slave_queue_t *queue) argument
389 __eql_insert_slave(slave_queue_t *queue, slave_t *slave) argument
[all...]
/drivers/mmc/card/
H A DMakefile6 mmc_block-objs := block.o queue.o
/drivers/isdn/i4l/
H A Disdn_net.h85 lp = nd->queue; /* get lp on top of queue */
86 while (isdn_net_lp_busy(nd->queue)) {
87 nd->queue = nd->queue->next;
88 if (nd->queue == lp) { /* not found -- should never happen */
93 lp = nd->queue;
94 nd->queue = nd->queue->next;
114 lp = nd->queue;
[all...]
/drivers/message/i2o/
H A Di2o_block.h26 * Added a queue depth.
29 * Removed queue walk, fixed for 64bitness.
34 * Heavily chop down the queue depths
59 /* request queue sizes */
75 spinlock_t lock; /* queue lock */
78 unsigned int open_queue_depth; /* number of requests in the queue */
89 struct list_head queue; member in struct:i2o_block_request
100 struct request_queue *queue; member in struct:i2o_block_delayed_request

Completed in 427 milliseconds

1234567891011>>