Searched refs:queue (Results 1 - 25 of 742) sorted by relevance

1234567891011>>

/drivers/media/usb/uvc/
H A Duvc_queue.c28 * Video buffers queue management.
34 * the videobuf2 queue operations by serializing calls to videobuf2 and a
35 * spinlock to protect the IRQ queue that holds the buffers to be processed by
40 * videobuf2 queue operations
47 struct uvc_video_queue *queue = vb2_get_drv_priv(vq); local
49 container_of(queue, struct uvc_streaming, queue);
65 struct uvc_video_queue *queue = vb2_get_drv_priv(vb->vb2_queue); local
74 if (unlikely(queue->flags & UVC_QUEUE_DISCONNECTED))
91 struct uvc_video_queue *queue local
111 struct uvc_video_queue *queue = vb2_get_drv_priv(vb->vb2_queue); local
122 struct uvc_video_queue *queue = vb2_get_drv_priv(vq); local
129 struct uvc_video_queue *queue = vb2_get_drv_priv(vq); local
143 uvc_queue_init(struct uvc_video_queue *queue, enum v4l2_buf_type type, int drop_corrupted) argument
172 uvc_alloc_buffers(struct uvc_video_queue *queue, struct v4l2_requestbuffers *rb) argument
184 uvc_free_buffers(struct uvc_video_queue *queue) argument
191 uvc_query_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf) argument
202 uvc_create_buffers(struct uvc_video_queue *queue, struct v4l2_create_buffers *cb) argument
214 uvc_queue_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf) argument
225 uvc_dequeue_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf, int nonblocking) argument
237 uvc_queue_mmap(struct uvc_video_queue *queue, struct vm_area_struct *vma) argument
249 uvc_queue_get_unmapped_area(struct uvc_video_queue *queue, unsigned long pgoff) argument
261 uvc_queue_poll(struct uvc_video_queue *queue, struct file *file, poll_table *wait) argument
280 uvc_queue_allocated(struct uvc_video_queue *queue) argument
307 uvc_queue_enable(struct uvc_video_queue *queue, int enable) argument
346 uvc_queue_cancel(struct uvc_video_queue *queue, int disconnect) argument
370 uvc_queue_next_buffer(struct uvc_video_queue *queue, struct uvc_buffer *buf) argument
[all...]
/drivers/usb/gadget/function/
H A Duvc_queue.c29 * Video buffers queue management.
35 * the videobuf2 queue operations by serializing calls to videobuf2 and a
36 * spinlock to protect the IRQ queue that holds the buffers to be processed by
41 * videobuf2 queue operations
48 struct uvc_video_queue *queue = vb2_get_drv_priv(vq); local
49 struct uvc_video *video = container_of(queue, struct uvc_video, queue);
63 struct uvc_video_queue *queue = vb2_get_drv_priv(vb->vb2_queue); local
72 if (unlikely(queue->flags & UVC_QUEUE_DISCONNECTED))
88 struct uvc_video_queue *queue local
109 struct uvc_video_queue *queue = vb2_get_drv_priv(vq); local
116 struct uvc_video_queue *queue = vb2_get_drv_priv(vq); local
129 uvcg_queue_init(struct uvc_video_queue *queue, enum v4l2_buf_type type) argument
156 uvcg_free_buffers(struct uvc_video_queue *queue) argument
166 uvcg_alloc_buffers(struct uvc_video_queue *queue, struct v4l2_requestbuffers *rb) argument
178 uvcg_query_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf) argument
189 uvcg_queue_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf) argument
213 uvcg_dequeue_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf, int nonblocking) argument
231 uvcg_queue_poll(struct uvc_video_queue *queue, struct file *file, poll_table *wait) argument
243 uvcg_queue_mmap(struct uvc_video_queue *queue, struct vm_area_struct *vma) argument
260 uvcg_queue_get_unmapped_area(struct uvc_video_queue *queue, unsigned long pgoff) argument
284 uvcg_queue_cancel(struct uvc_video_queue *queue, int disconnect) argument
325 uvcg_queue_enable(struct uvc_video_queue *queue, int enable) argument
362 uvcg_queue_next_buffer(struct uvc_video_queue *queue, struct uvc_buffer *buf) argument
391 uvcg_queue_head(struct uvc_video_queue *queue) argument
[all...]
H A Duvc_queue.h30 struct list_head queue; member in struct:uvc_buffer
43 struct vb2_queue queue; member in struct:uvc_video_queue
44 struct mutex mutex; /* Protects queue */
55 static inline int uvc_queue_streaming(struct uvc_video_queue *queue) argument
57 return vb2_is_streaming(&queue->queue);
60 int uvcg_queue_init(struct uvc_video_queue *queue, enum v4l2_buf_type type);
62 void uvcg_free_buffers(struct uvc_video_queue *queue);
64 int uvcg_alloc_buffers(struct uvc_video_queue *queue,
67 int uvcg_query_buffer(struct uvc_video_queue *queue, struc
[all...]
H A Duvc_video.c36 if (buf->bytesused - video->queue.buf_used <= len - 2)
46 struct uvc_video_queue *queue = &video->queue; local
51 mem = buf->mem + queue->buf_used;
52 nbytes = min((unsigned int)len, buf->bytesused - queue->buf_used);
55 queue->buf_used += nbytes;
86 if (buf->bytesused == video->queue.buf_used) {
87 video->queue.buf_used = 0;
89 uvcg_queue_next_buffer(&video->queue, buf);
96 buf->bytesused == video->queue
165 struct uvc_video_queue *queue = &video->queue; local
285 struct uvc_video_queue *queue = &video->queue; local
[all...]
/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_marker.c37 void vmw_marker_queue_init(struct vmw_marker_queue *queue) argument
39 INIT_LIST_HEAD(&queue->head);
40 queue->lag = 0;
41 queue->lag_time = ktime_get_raw_ns();
42 spin_lock_init(&queue->lock);
45 void vmw_marker_queue_takedown(struct vmw_marker_queue *queue) argument
49 spin_lock(&queue->lock);
50 list_for_each_entry_safe(marker, next, &queue->head, head) {
53 spin_unlock(&queue->lock);
56 int vmw_marker_push(struct vmw_marker_queue *queue, argument
73 vmw_marker_pull(struct vmw_marker_queue *queue, uint32_t signaled_seqno) argument
107 vmw_fifo_lag(struct vmw_marker_queue *queue) argument
120 vmw_lag_lt(struct vmw_marker_queue *queue, uint32_t us) argument
128 vmw_wait_lag(struct vmw_private *dev_priv, struct vmw_marker_queue *queue, uint32_t us) argument
[all...]
/drivers/scsi/arm/
H A Dqueue.h2 * linux/drivers/acorn/scsi/queue.h: queue handling
21 * Function: void queue_initialise (Queue_t *queue)
22 * Purpose : initialise a queue
23 * Params : queue - queue to initialise
25 extern int queue_initialise (Queue_t *queue);
28 * Function: void queue_free (Queue_t *queue)
29 * Purpose : free a queue
30 * Params : queue
[all...]
H A Dqueue.c2 * linux/drivers/acorn/scsi/queue.c: queue handling primitives
49 #include "queue.h"
54 * Function: void queue_initialise (Queue_t *queue)
55 * Purpose : initialise a queue
56 * Params : queue - queue to initialise
58 int queue_initialise (Queue_t *queue) argument
63 spin_lock_init(&queue->queue_lock);
64 INIT_LIST_HEAD(&queue
90 queue_free(Queue_t *queue) argument
106 __queue_add(Queue_t *queue, struct scsi_cmnd *SCpnt, int head) argument
137 __queue_remove(Queue_t *queue, struct list_head *ent) argument
161 queue_remove_exclude(Queue_t *queue, unsigned long *exclude) argument
187 queue_remove(Queue_t *queue) argument
209 queue_remove_tgtluntag(Queue_t *queue, int target, int lun, int tag) argument
237 queue_remove_all_target(Queue_t *queue, int target) argument
260 queue_probetgtlun(Queue_t *queue, int target, int lun) argument
286 queue_remove_cmd(Queue_t *queue, struct scsi_cmnd *SCpnt) argument
[all...]
H A DMakefile7 obj-$(CONFIG_SCSI_ACORNSCSI_3) += acornscsi_mod.o queue.o msgqueue.o
8 obj-$(CONFIG_SCSI_ARXESCSI) += arxescsi.o fas216.o queue.o msgqueue.o
10 obj-$(CONFIG_SCSI_CUMANA_2) += cumana_2.o fas216.o queue.o msgqueue.o
12 obj-$(CONFIG_SCSI_POWERTECSCSI) += powertec.o fas216.o queue.o msgqueue.o
13 obj-$(CONFIG_SCSI_EESOXSCSI) += eesox.o fas216.o queue.o msgqueue.o
/drivers/net/wireless/cw1200/
H A Dqueue.c2 * O(1) TX queue with built-in allocator for ST-Ericsson CW1200 drivers
14 #include "queue.h"
29 static inline void __cw1200_queue_lock(struct cw1200_queue *queue) argument
31 struct cw1200_queue_stats *stats = queue->stats;
32 if (queue->tx_locked_cnt++ == 0) {
34 queue->queue_id);
35 ieee80211_stop_queue(stats->priv->hw, queue->queue_id);
39 static inline void __cw1200_queue_unlock(struct cw1200_queue *queue) argument
41 struct cw1200_queue_stats *stats = queue->stats;
42 BUG_ON(!queue
92 __cw1200_queue_gc(struct cw1200_queue *queue, struct list_head *head, bool unlock) argument
136 struct cw1200_queue *queue = local
165 cw1200_queue_init(struct cw1200_queue *queue, struct cw1200_queue_stats *stats, u8 queue_id, size_t capacity, unsigned long ttl) argument
205 cw1200_queue_clear(struct cw1200_queue *queue) argument
247 cw1200_queue_deinit(struct cw1200_queue *queue) argument
259 cw1200_queue_get_num_queued(struct cw1200_queue *queue, u32 link_id_map) argument
283 cw1200_queue_put(struct cw1200_queue *queue, struct sk_buff *skb, struct cw1200_txpriv *txpriv) argument
335 cw1200_queue_get(struct cw1200_queue *queue, u32 link_id_map, struct wsm_tx **tx, struct ieee80211_tx_info **tx_info, const struct cw1200_txpriv **txpriv) argument
376 cw1200_queue_requeue(struct cw1200_queue *queue, u32 packet_id) argument
418 cw1200_queue_requeue_all(struct cw1200_queue *queue) argument
445 cw1200_queue_remove(struct cw1200_queue *queue, u32 packet_id) argument
496 cw1200_queue_get_skb(struct cw1200_queue *queue, u32 packet_id, struct sk_buff **skb, const struct cw1200_txpriv **txpriv) argument
526 cw1200_queue_lock(struct cw1200_queue *queue) argument
533 cw1200_queue_unlock(struct cw1200_queue *queue) argument
540 cw1200_queue_get_xmit_timestamp(struct cw1200_queue *queue, unsigned long *timestamp, u32 pending_frame_id) argument
[all...]
H A Dqueue.h2 * O(1) TX queue with built-in allocator for ST-Ericsson CW1200 drivers
36 struct list_head queue; member in struct:cw1200_queue
42 spinlock_t lock; /* Protect queue entry */
71 int cw1200_queue_init(struct cw1200_queue *queue,
76 int cw1200_queue_clear(struct cw1200_queue *queue);
78 void cw1200_queue_deinit(struct cw1200_queue *queue);
80 size_t cw1200_queue_get_num_queued(struct cw1200_queue *queue,
82 int cw1200_queue_put(struct cw1200_queue *queue,
85 int cw1200_queue_get(struct cw1200_queue *queue,
90 int cw1200_queue_requeue(struct cw1200_queue *queue, u3
[all...]
/drivers/infiniband/hw/ehca/
H A Dipz_pt_fn.c4 * internal queue handling
51 void *ipz_qpageit_get_inc(struct ipz_queue *queue) argument
53 void *ret = ipz_qeit_get(queue);
54 queue->current_q_offset += queue->pagesize;
55 if (queue->current_q_offset > queue->queue_length) {
56 queue->current_q_offset -= queue->pagesize;
59 if (((u64)ret) % queue
66 ipz_qeit_eq_get_inc(struct ipz_queue *queue) argument
80 ipz_queue_abs_to_offset(struct ipz_queue *queue, u64 addr, u64 *q_offset) argument
102 alloc_queue_pages(struct ipz_queue *queue, const u32 nr_of_pages) argument
127 alloc_small_queue_page(struct ipz_queue *queue, struct ehca_pd *pd) argument
172 free_small_queue_page(struct ipz_queue *queue, struct ehca_pd *pd) argument
204 ipz_queue_ctor(struct ehca_pd *pd, struct ipz_queue *queue, const u32 nr_of_pages, const u32 pagesize, const u32 qe_size, const u32 nr_of_sg, int is_small) argument
256 ipz_queue_dtor(struct ehca_pd *pd, struct ipz_queue *queue) argument
[all...]
H A Dipz_pt_fn.h4 * internal queue handling
75 /* struct generic queue in linux kernel virtual memory (kv) */
77 u64 current_q_offset; /* current queue entry */
79 struct ipz_page **queue_pages; /* array of pages belonging to queue */
80 u32 qe_size; /* queue entry size */
82 u32 queue_length; /* queue length allocated in bytes */
93 static inline void *ipz_qeit_calc(struct ipz_queue *queue, u64 q_offset) argument
96 if (q_offset >= queue->queue_length)
98 current_page = (queue->queue_pages)[q_offset >> EHCA_PAGESHIFT];
106 static inline void *ipz_qeit_get(struct ipz_queue *queue) argument
126 ipz_qeit_get_inc(struct ipz_queue *queue) argument
142 ipz_qeit_is_valid(struct ipz_queue *queue) argument
155 ipz_qeit_get_inc_valid(struct ipz_queue *queue) argument
164 ipz_qeit_reset(struct ipz_queue *queue) argument
178 ipz_queue_advance_offset(struct ipz_queue *queue, u64 offset) argument
258 ipz_eqit_eq_get_inc_valid(struct ipz_queue *queue) argument
268 ipz_eqit_eq_peek_valid(struct ipz_queue *queue) argument
[all...]
/drivers/net/wireless/b43legacy/
H A Dpio.c35 static void tx_start(struct b43legacy_pioqueue *queue) argument
37 b43legacy_pio_write(queue, B43legacy_PIO_TXCTL,
41 static void tx_octet(struct b43legacy_pioqueue *queue, argument
44 if (queue->need_workarounds) {
45 b43legacy_pio_write(queue, B43legacy_PIO_TXDATA, octet);
46 b43legacy_pio_write(queue, B43legacy_PIO_TXCTL,
49 b43legacy_pio_write(queue, B43legacy_PIO_TXCTL,
51 b43legacy_pio_write(queue, B43legacy_PIO_TXDATA, octet);
76 static void tx_data(struct b43legacy_pioqueue *queue, argument
84 if (queue
102 tx_complete(struct b43legacy_pioqueue *queue, struct sk_buff *skb) argument
116 generate_cookie(struct b43legacy_pioqueue *queue, struct b43legacy_pio_txpacket *packet) argument
154 struct b43legacy_pioqueue *queue = NULL; local
185 pio_tx_write_fragment(struct b43legacy_pioqueue *queue, struct sk_buff *skb, struct b43legacy_pio_txpacket *packet, size_t txhdr_size) argument
218 struct b43legacy_pioqueue *queue = packet->queue; local
232 struct b43legacy_pioqueue *queue = packet->queue; local
282 struct b43legacy_pioqueue *queue = (struct b43legacy_pioqueue *)d; local
312 setup_txqueues(struct b43legacy_pioqueue *queue) argument
332 struct b43legacy_pioqueue *queue; local
381 cancel_transfers(struct b43legacy_pioqueue *queue) argument
393 b43legacy_destroy_pioqueue(struct b43legacy_pioqueue *queue) argument
423 struct b43legacy_pioqueue *queue; local
469 struct b43legacy_pioqueue *queue = dev->pio.queue1; local
491 struct b43legacy_pioqueue *queue; local
549 pio_rx_error(struct b43legacy_pioqueue *queue, int clear_buffers, const char *error) argument
567 b43legacy_pio_rx(struct b43legacy_pioqueue *queue) argument
648 b43legacy_pio_tx_suspend(struct b43legacy_pioqueue *queue) argument
656 b43legacy_pio_tx_resume(struct b43legacy_pioqueue *queue) argument
[all...]
H A Dpio.h42 struct b43legacy_pioqueue *queue; member in struct:b43legacy_pio_txpacket
48 (packet)->queue->tx_packets_cache))
73 /* Packets on the txrunning queue are completely
83 u16 b43legacy_pio_read(struct b43legacy_pioqueue *queue, argument
86 return b43legacy_read16(queue->dev, queue->mmio_base + offset);
90 void b43legacy_pio_write(struct b43legacy_pioqueue *queue, argument
93 b43legacy_write16(queue->dev, queue->mmio_base + offset, value);
105 void b43legacy_pio_rx(struct b43legacy_pioqueue *queue);
137 b43legacy_pio_rx(struct b43legacy_pioqueue *queue) argument
141 b43legacy_pio_tx_suspend(struct b43legacy_pioqueue *queue) argument
145 b43legacy_pio_tx_resume(struct b43legacy_pioqueue *queue) argument
[all...]
/drivers/net/xen-netback/
H A Dnetback.c58 /* The time that packets can stay on the guest Rx internal queue
85 static void xenvif_idx_release(struct xenvif_queue *queue, u16 pending_idx,
88 static void make_tx_response(struct xenvif_queue *queue,
92 static inline int tx_work_todo(struct xenvif_queue *queue);
94 static struct xen_netif_rx_response *make_rx_response(struct xenvif_queue *queue,
101 static inline unsigned long idx_to_pfn(struct xenvif_queue *queue, argument
104 return page_to_pfn(queue->mmap_pages[idx]);
107 static inline unsigned long idx_to_kaddr(struct xenvif_queue *queue, argument
110 return (unsigned long)pfn_to_kaddr(idx_to_pfn(queue, idx));
150 bool xenvif_rx_ring_slots_available(struct xenvif_queue *queue, in argument
172 xenvif_rx_queue_tail(struct xenvif_queue *queue, struct sk_buff *skb) argument
187 xenvif_rx_dequeue(struct xenvif_queue *queue) argument
202 xenvif_rx_queue_maybe_wake(struct xenvif_queue *queue) argument
213 xenvif_rx_queue_purge(struct xenvif_queue *queue) argument
220 xenvif_rx_queue_drop_expired(struct xenvif_queue *queue) argument
289 get_next_rx_buffer(struct xenvif_queue *queue, struct netrx_pending_operations *npo) argument
313 xenvif_gop_frag_copy(struct xenvif_queue *queue, struct sk_buff *skb, struct netrx_pending_operations *npo, struct page *page, unsigned long size, unsigned long offset, int *head, struct xenvif_queue *foreign_queue, grant_ref_t foreign_gref) argument
448 xenvif_gop_skb(struct sk_buff *skb, struct netrx_pending_operations *npo, struct xenvif_queue *queue) argument
602 xenvif_add_frag_responses(struct xenvif_queue *queue, int status, struct xenvif_rx_meta *meta, int nr_meta_slots) argument
628 xenvif_kick_thread(struct xenvif_queue *queue) argument
633 xenvif_rx_action(struct xenvif_queue *queue) argument
796 xenvif_napi_schedule_or_enable_events(struct xenvif_queue *queue) argument
806 tx_add_credit(struct xenvif_queue *queue) argument
828 struct xenvif_queue *queue = (struct xenvif_queue *)data; local
833 xenvif_tx_err(struct xenvif_queue *queue, struct xen_netif_tx_request *txp, RING_IDX end) argument
859 xenvif_count_requests(struct xenvif_queue *queue, struct xen_netif_tx_request *first, struct xen_netif_tx_request *txp, int work_to_do) argument
964 xenvif_tx_create_map_op(struct xenvif_queue *queue, u16 pending_idx, struct xen_netif_tx_request *txp, struct gnttab_map_grant_ref *mop) argument
995 xenvif_get_requests(struct xenvif_queue *queue, struct sk_buff *skb, struct xen_netif_tx_request *txp, struct gnttab_map_grant_ref *gop) argument
1055 xenvif_grant_handle_set(struct xenvif_queue *queue, u16 pending_idx, grant_handle_t handle) argument
1069 xenvif_grant_handle_reset(struct xenvif_queue *queue, u16 pending_idx) argument
1082 xenvif_tx_check_gop(struct xenvif_queue *queue, struct sk_buff *skb, struct gnttab_map_grant_ref **gopp_map, struct gnttab_copy **gopp_copy) argument
1207 xenvif_fill_frags(struct xenvif_queue *queue, struct sk_buff *skb) argument
1251 xenvif_get_extras(struct xenvif_queue *queue, struct xen_netif_extra_info *extras, int work_to_do) argument
1312 checksum_setup(struct xenvif_queue *queue, struct sk_buff *skb) argument
1334 tx_credit_exceeded(struct xenvif_queue *queue, unsigned size) argument
1366 xenvif_tx_build_gops(struct xenvif_queue *queue, int budget, unsigned *copy_ops, unsigned *map_ops) argument
1530 xenvif_handle_frag_list(struct xenvif_queue *queue, struct sk_buff *skb) argument
1595 xenvif_tx_submit(struct xenvif_queue *queue) argument
1716 struct xenvif_queue *queue = ubuf_to_queue(ubuf); local
1745 xenvif_tx_dealloc_action(struct xenvif_queue *queue) argument
1813 xenvif_tx_action(struct xenvif_queue *queue, int budget) argument
1840 xenvif_idx_release(struct xenvif_queue *queue, u16 pending_idx, u8 status) argument
1859 make_tx_response(struct xenvif_queue *queue, struct xen_netif_tx_request *txp, s8 st) argument
1880 make_rx_response(struct xenvif_queue *queue, u16 id, s8 st, u16 offset, u16 size, u16 flags) argument
1903 xenvif_idx_unmap(struct xenvif_queue *queue, u16 pending_idx) argument
1928 tx_work_todo(struct xenvif_queue *queue) argument
1936 tx_dealloc_work_todo(struct xenvif_queue *queue) argument
1941 xenvif_unmap_frontend_rings(struct xenvif_queue *queue) argument
1951 xenvif_map_frontend_rings(struct xenvif_queue *queue, grant_ref_t tx_ring_ref, grant_ref_t rx_ring_ref) argument
1984 xenvif_queue_carrier_off(struct xenvif_queue *queue) argument
1999 xenvif_queue_carrier_on(struct xenvif_queue *queue) argument
2015 xenvif_rx_queue_stalled(struct xenvif_queue *queue) argument
2028 xenvif_rx_queue_ready(struct xenvif_queue *queue) argument
2039 xenvif_have_rx_work(struct xenvif_queue *queue) argument
2049 xenvif_rx_queue_timeout(struct xenvif_queue *queue) argument
2072 xenvif_wait_for_rx_work(struct xenvif_queue *queue) argument
2094 struct xenvif_queue *queue = data; local
2146 xenvif_dealloc_kthread_should_stop(struct xenvif_queue *queue) argument
2157 struct xenvif_queue *queue = data; local
[all...]
H A Dinterface.c46 /* Number of bytes allowed on the internal guest Rx queue. */
54 void xenvif_skb_zerocopy_prepare(struct xenvif_queue *queue, argument
58 atomic_inc(&queue->inflight_packets);
61 void xenvif_skb_zerocopy_complete(struct xenvif_queue *queue) argument
63 atomic_dec(&queue->inflight_packets);
75 struct xenvif_queue *queue = dev_id; local
77 if (RING_HAS_UNCONSUMED_REQUESTS(&queue->tx))
78 napi_schedule(&queue->napi);
85 struct xenvif_queue *queue = local
93 if (unlikely(queue
110 struct xenvif_queue *queue = dev_id; local
125 xenvif_queue_stopped(struct xenvif_queue *queue) argument
132 xenvif_wake_queue(struct xenvif_queue *queue) argument
142 struct xenvif_queue *queue = NULL; local
185 struct xenvif_queue *queue = NULL; local
216 struct xenvif_queue *queue = NULL; local
232 struct xenvif_queue *queue = NULL; local
458 xenvif_init_queue(struct xenvif_queue *queue) argument
515 xenvif_connect(struct xenvif_queue *queue, unsigned long tx_ring_ref, unsigned long rx_ring_ref, unsigned int tx_evtchn, unsigned int rx_evtchn) argument
622 struct xenvif_queue *queue = NULL; local
661 xenvif_deinit_queue(struct xenvif_queue *queue) argument
668 struct xenvif_queue *queue = NULL; local
[all...]
/drivers/misc/genwqe/
H A Dcard_ddcb.c22 * Device Driver Control Block (DDCB) queue support. Definition of
23 * interrupt handlers for queue support as well as triggering the
49 * Situation (1): Empty queue
91 static int queue_empty(struct ddcb_queue *queue) argument
93 return queue->ddcb_next == queue->ddcb_act;
96 static int queue_enqueued_ddcbs(struct ddcb_queue *queue) argument
98 if (queue->ddcb_next >= queue->ddcb_act)
99 return queue
104 queue_free_ddcbs(struct ddcb_queue *queue) argument
172 print_ddcb_info(struct genwqe_dev *cd, struct ddcb_queue *queue) argument
274 enqueue_ddcb(struct genwqe_dev *cd, struct ddcb_queue *queue, struct ddcb *pddcb, int ddcb_no) argument
339 struct ddcb_queue *queue = req->queue; local
371 genwqe_check_ddcb_queue(struct genwqe_dev *cd, struct ddcb_queue *queue) argument
487 struct ddcb_queue *queue; local
512 struct ddcb_queue *queue = req->queue; local
571 get_next_ddcb(struct genwqe_dev *cd, struct ddcb_queue *queue, int *num) argument
629 struct ddcb_queue *queue = req->queue; local
763 struct ddcb_queue *queue; local
984 struct ddcb_queue *queue = &cd->queue; local
1014 struct ddcb_queue *queue = &cd->queue; local
1023 setup_ddcb_queue(struct genwqe_dev *cd, struct ddcb_queue *queue) argument
1101 ddcb_queue_initialized(struct ddcb_queue *queue) argument
1106 free_ddcb_queue(struct genwqe_dev *cd, struct ddcb_queue *queue) argument
1233 struct ddcb_queue *queue; local
1321 struct ddcb_queue *queue = &cd->queue; local
1347 struct ddcb_queue *queue = &cd->queue; local
[all...]
/drivers/net/wireless/rt2x00/
H A Drt2x00queue.c23 Abstract: rt2x00 queue specific routines.
36 struct data_queue *queue = entry->queue; local
37 struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
48 frame_size = queue->data_size + queue->desc_size + queue->winfo_size;
107 struct device *dev = entry->queue->rt2x00dev->dev;
123 struct device *dev = entry->queue->rt2x00dev->dev;
498 struct rt2x00_dev *rt2x00dev = entry->queue
539 struct data_queue *queue = entry->queue; local
550 rt2x00queue_kick_tx_queue(struct data_queue *queue, struct txentry_desc *txdesc) argument
609 rt2x00queue_write_tx_frame(struct data_queue *queue, struct sk_buff *skb, struct ieee80211_sta *sta, bool local) argument
788 rt2x00queue_for_each_entry(struct data_queue *queue, enum queue_index start, enum queue_index end, void *data, bool (*fn)(struct queue_entry *entry, void *data)) argument
843 rt2x00queue_get_entry(struct data_queue *queue, enum queue_index index) argument
867 struct data_queue *queue = entry->queue; local
894 rt2x00queue_pause_queue_nocheck(struct data_queue *queue) argument
911 rt2x00queue_pause_queue(struct data_queue *queue) argument
922 rt2x00queue_unpause_queue(struct data_queue *queue) argument
952 rt2x00queue_start_queue(struct data_queue *queue) argument
972 rt2x00queue_stop_queue(struct data_queue *queue) argument
989 rt2x00queue_flush_queue(struct data_queue *queue, bool drop) argument
1026 struct data_queue *queue; local
1041 struct data_queue *queue; local
1060 struct data_queue *queue; local
1069 rt2x00queue_reset(struct data_queue *queue) argument
1087 struct data_queue *queue; local
1098 rt2x00queue_alloc_entries(struct data_queue *queue) argument
1135 rt2x00queue_free_skbs(struct data_queue *queue) argument
1147 rt2x00queue_alloc_rxskbs(struct data_queue *queue) argument
1164 struct data_queue *queue; local
1203 struct data_queue *queue; local
1213 rt2x00queue_init(struct rt2x00_dev *rt2x00dev, struct data_queue *queue, enum data_queue_qid qid) argument
1234 struct data_queue *queue; local
[all...]
H A Drt2x00mmio.c62 struct data_queue *queue = rt2x00dev->rx; local
69 entry = rt2x00queue_get_entry(queue, Q_INDEX);
80 skbdesc->desc_len = entry->queue->desc_size;
99 void rt2x00mmio_flush_queue(struct data_queue *queue, bool drop) argument
103 for (i = 0; !rt2x00queue_empty(queue) && i < 10; i++)
112 struct data_queue *queue)
123 queue->limit * queue->desc_size, &dma,
129 * Initialize all queue entries to contain valid addresses.
131 for (i = 0; i < queue
111 rt2x00mmio_alloc_queue_dma(struct rt2x00_dev *rt2x00dev, struct data_queue *queue) argument
140 rt2x00mmio_free_queue_dma(struct rt2x00_dev *rt2x00dev, struct data_queue *queue) argument
155 struct data_queue *queue; local
191 struct data_queue *queue; local
[all...]
H A Drt2x00usb.c248 struct data_queue *queue; local
251 tx_queue_for_each(rt2x00dev, queue) {
252 while (!rt2x00queue_empty(queue)) {
253 entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE);
267 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
294 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
322 usb_sndbulkpipe(usb_dev, entry->queue->usb_endpoint),
360 skbdesc->desc_len = entry->queue->desc_size;
372 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
387 if (urb->actual_length < entry->queue
426 rt2x00usb_kick_queue(struct data_queue *queue) argument
475 rt2x00usb_flush_queue(struct data_queue *queue, bool drop) argument
525 rt2x00usb_watchdog_tx_dma(struct data_queue *queue) argument
535 rt2x00usb_dma_timeout(struct data_queue *queue) argument
545 struct data_queue *queue; local
578 rt2x00usb_assign_endpoint(struct data_queue *queue, struct usb_endpoint_descriptor *ep_desc) argument
603 struct data_queue *queue = rt2x00dev->tx; local
648 rt2x00usb_alloc_entries(struct data_queue *queue) argument
681 rt2x00usb_free_entries(struct data_queue *queue) argument
715 struct data_queue *queue; local
745 struct data_queue *queue; local
[all...]
/drivers/net/ethernet/ibm/ehea/
H A Dehea_qmr.h210 static inline void *hw_qeit_calc(struct hw_queue *queue, u64 q_offset) argument
214 if (q_offset >= queue->queue_length)
215 q_offset -= queue->queue_length;
216 current_page = (queue->queue_pages)[q_offset >> EHEA_PAGESHIFT];
220 static inline void *hw_qeit_get(struct hw_queue *queue) argument
222 return hw_qeit_calc(queue, queue->current_q_offset);
225 static inline void hw_qeit_inc(struct hw_queue *queue) argument
227 queue->current_q_offset += queue
235 hw_qeit_get_inc(struct hw_queue *queue) argument
242 hw_qeit_get_inc_valid(struct hw_queue *queue) argument
259 hw_qeit_get_valid(struct hw_queue *queue) argument
275 hw_qeit_reset(struct hw_queue *queue) argument
281 hw_qeit_eq_get_inc(struct hw_queue *queue) argument
295 hw_eqit_eq_get_inc_valid(struct hw_queue *queue) argument
309 struct hw_queue *queue; local
324 struct hw_queue *queue = &my_qp->hw_squeue; local
341 struct hw_queue *queue = &qp->hw_rqueue1; local
[all...]
/drivers/net/wireless/ath/ath5k/
H A Dqcu.c35 * basically we have 10 queues to play with. Each queue has a matching
36 * QCU that controls when the queue will get triggered and multiple QCUs
39 * and DCUs allowing us to have different DFS settings for each queue.
41 * When a frame goes into a TX queue, QCU decides when it'll trigger a
43 * it's buffer or -if it's a beacon queue- if it's time to fire up the queue
58 * ath5k_hw_num_tx_pending() - Get number of pending frames for a given queue
60 * @queue: One of enum ath5k_tx_queue_id
63 ath5k_hw_num_tx_pending(struct ath5k_hw *ah, unsigned int queue) argument
66 AR5K_ASSERT_ENTRY(queue, a
94 ath5k_hw_release_tx_queue(struct ath5k_hw *ah, unsigned int queue) argument
138 ath5k_hw_get_tx_queueprops(struct ath5k_hw *ah, int queue, struct ath5k_txq_info *queue_info) argument
154 ath5k_hw_set_tx_queueprops(struct ath5k_hw *ah, int queue, const struct ath5k_txq_info *qinfo) argument
206 unsigned int queue; local
287 ath5k_hw_set_tx_retry_limits(struct ath5k_hw *ah, unsigned int queue) argument
330 ath5k_hw_reset_tx_queue(struct ath5k_hw *ah, unsigned int queue) argument
[all...]
/drivers/net/
H A Dxen-netfront.c85 /* IRQ name is queue name with "-tx" or "-rx" appended */
164 /* Multi-queue support */
213 static struct sk_buff *xennet_get_rx_skb(struct netfront_queue *queue, argument
217 struct sk_buff *skb = queue->rx_skbs[i];
218 queue->rx_skbs[i] = NULL;
222 static grant_ref_t xennet_get_rx_ref(struct netfront_queue *queue, argument
226 grant_ref_t ref = queue->grant_rx_ref[i];
227 queue->grant_rx_ref[i] = GRANT_INVALID_REF;
247 struct netfront_queue *queue = (struct netfront_queue *)data; local
248 napi_schedule(&queue
251 netfront_tx_slot_available(struct netfront_queue *queue) argument
257 xennet_maybe_wake_tx(struct netfront_queue *queue) argument
268 xennet_alloc_rx_buffers(struct netfront_queue *queue) argument
375 struct netfront_queue *queue = NULL; local
396 xennet_tx_buf_gc(struct netfront_queue *queue) argument
451 xennet_make_frags(struct sk_buff *skb, struct netfront_queue *queue, struct xen_netif_tx_request *tx) argument
602 struct netfront_queue *queue = NULL; local
721 struct netfront_queue *queue; local
730 xennet_move_rx_slot(struct netfront_queue *queue, struct sk_buff *skb, grant_ref_t ref) argument
743 xennet_get_extras(struct netfront_queue *queue, struct xen_netif_extra_info *extras, RING_IDX rp) argument
787 xennet_get_responses(struct netfront_queue *queue, struct netfront_rx_info *rinfo, RING_IDX rp, struct sk_buff_head *list) argument
896 xennet_fill_frags(struct netfront_queue *queue, struct sk_buff *skb, struct sk_buff_head *list) argument
951 handle_incoming_queue(struct netfront_queue *queue, struct sk_buff_head *rxq) argument
989 struct netfront_queue *queue = container_of(napi, struct netfront_queue, napi); local
1139 xennet_release_tx_bufs(struct netfront_queue *queue) argument
1161 xennet_release_rx_bufs(struct netfront_queue *queue) argument
1253 struct netfront_queue *queue = dev_id; local
1265 struct netfront_queue *queue = dev_id; local
1424 struct netfront_queue *queue = &info->queues[i]; local
1491 setup_netfront_single(struct netfront_queue *queue) argument
1516 setup_netfront_split(struct netfront_queue *queue) argument
1560 setup_netfront(struct xenbus_device *dev, struct netfront_queue *queue, unsigned int feature_split_evtchn) argument
1633 xennet_init_queue(struct netfront_queue *queue) argument
1691 write_queue_xenstore_keys(struct netfront_queue *queue, struct xenbus_transaction *xbt, int write_hierarchical) argument
1780 struct netfront_queue *queue = &info->queues[i]; local
1807 struct netfront_queue *queue = &info->queues[i]; local
1847 struct netfront_queue *queue = NULL; local
2000 struct netfront_queue *queue = NULL; local
2167 struct netfront_queue *queue; local
2220 struct netfront_queue *queue = NULL; local
2302 struct netfront_queue *queue = NULL; local
[all...]
H A Deql.c140 static void eql_kill_one_slave(slave_queue_t *queue, slave_t *slave);
147 spin_lock(&eql->queue.lock);
148 head = &eql->queue.all_slaves;
157 eql_kill_one_slave(&eql->queue, slave);
161 spin_unlock(&eql->queue.lock);
186 spin_lock_init(&eql->queue.lock);
187 INIT_LIST_HEAD(&eql->queue.all_slaves);
188 eql->queue.master_dev = dev;
213 BUG_ON(!list_empty(&eql->queue.all_slaves));
223 static void eql_kill_one_slave(slave_queue_t *queue, slave_ argument
232 eql_kill_slave_queue(slave_queue_t *queue) argument
298 __eql_schedule_slaves(slave_queue_t *queue) argument
364 __eql_find_slave_dev(slave_queue_t *queue, struct net_device *dev) argument
379 eql_is_full(slave_queue_t *queue) argument
389 __eql_insert_slave(slave_queue_t *queue, slave_t *slave) argument
[all...]
/drivers/mmc/card/
H A DMakefile6 mmc_block-objs := block.o queue.o

Completed in 491 milliseconds

1234567891011>>