/drivers/media/usb/uvc/ |
H A D | uvc_queue.c | 28 * Video buffers queue management. 34 * the videobuf2 queue operations by serializing calls to videobuf2 and a 35 * spinlock to protect the IRQ queue that holds the buffers to be processed by 40 * videobuf2 queue operations 47 struct uvc_video_queue *queue = vb2_get_drv_priv(vq); local 49 container_of(queue, struct uvc_streaming, queue); 65 struct uvc_video_queue *queue = vb2_get_drv_priv(vb->vb2_queue); local 74 if (unlikely(queue->flags & UVC_QUEUE_DISCONNECTED)) 91 struct uvc_video_queue *queue local 111 struct uvc_video_queue *queue = vb2_get_drv_priv(vb->vb2_queue); local 122 struct uvc_video_queue *queue = vb2_get_drv_priv(vq); local 129 struct uvc_video_queue *queue = vb2_get_drv_priv(vq); local 143 uvc_queue_init(struct uvc_video_queue *queue, enum v4l2_buf_type type, int drop_corrupted) argument 172 uvc_alloc_buffers(struct uvc_video_queue *queue, struct v4l2_requestbuffers *rb) argument 184 uvc_free_buffers(struct uvc_video_queue *queue) argument 191 uvc_query_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf) argument 202 uvc_create_buffers(struct uvc_video_queue *queue, struct v4l2_create_buffers *cb) argument 214 uvc_queue_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf) argument 225 uvc_dequeue_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf, int nonblocking) argument 237 uvc_queue_mmap(struct uvc_video_queue *queue, struct vm_area_struct *vma) argument 249 uvc_queue_get_unmapped_area(struct uvc_video_queue *queue, unsigned long pgoff) argument 261 uvc_queue_poll(struct uvc_video_queue *queue, struct file *file, poll_table *wait) argument 280 uvc_queue_allocated(struct uvc_video_queue *queue) argument 307 uvc_queue_enable(struct uvc_video_queue *queue, int enable) argument 346 uvc_queue_cancel(struct uvc_video_queue *queue, int disconnect) argument 370 uvc_queue_next_buffer(struct uvc_video_queue *queue, struct uvc_buffer *buf) argument [all...] |
/drivers/usb/gadget/function/ |
H A D | uvc_queue.c | 29 * Video buffers queue management. 35 * the videobuf2 queue operations by serializing calls to videobuf2 and a 36 * spinlock to protect the IRQ queue that holds the buffers to be processed by 41 * videobuf2 queue operations 48 struct uvc_video_queue *queue = vb2_get_drv_priv(vq); local 49 struct uvc_video *video = container_of(queue, struct uvc_video, queue); 63 struct uvc_video_queue *queue = vb2_get_drv_priv(vb->vb2_queue); local 72 if (unlikely(queue->flags & UVC_QUEUE_DISCONNECTED)) 88 struct uvc_video_queue *queue local 109 struct uvc_video_queue *queue = vb2_get_drv_priv(vq); local 116 struct uvc_video_queue *queue = vb2_get_drv_priv(vq); local 129 uvcg_queue_init(struct uvc_video_queue *queue, enum v4l2_buf_type type) argument 156 uvcg_free_buffers(struct uvc_video_queue *queue) argument 166 uvcg_alloc_buffers(struct uvc_video_queue *queue, struct v4l2_requestbuffers *rb) argument 178 uvcg_query_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf) argument 189 uvcg_queue_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf) argument 213 uvcg_dequeue_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf, int nonblocking) argument 231 uvcg_queue_poll(struct uvc_video_queue *queue, struct file *file, poll_table *wait) argument 243 uvcg_queue_mmap(struct uvc_video_queue *queue, struct vm_area_struct *vma) argument 260 uvcg_queue_get_unmapped_area(struct uvc_video_queue *queue, unsigned long pgoff) argument 284 uvcg_queue_cancel(struct uvc_video_queue *queue, int disconnect) argument 325 uvcg_queue_enable(struct uvc_video_queue *queue, int enable) argument 362 uvcg_queue_next_buffer(struct uvc_video_queue *queue, struct uvc_buffer *buf) argument 391 uvcg_queue_head(struct uvc_video_queue *queue) argument [all...] |
H A D | uvc_queue.h | 30 struct list_head queue; member in struct:uvc_buffer 43 struct vb2_queue queue; member in struct:uvc_video_queue 44 struct mutex mutex; /* Protects queue */ 55 static inline int uvc_queue_streaming(struct uvc_video_queue *queue) argument 57 return vb2_is_streaming(&queue->queue); 60 int uvcg_queue_init(struct uvc_video_queue *queue, enum v4l2_buf_type type); 62 void uvcg_free_buffers(struct uvc_video_queue *queue); 64 int uvcg_alloc_buffers(struct uvc_video_queue *queue, 67 int uvcg_query_buffer(struct uvc_video_queue *queue, struc [all...] |
H A D | uvc_video.c | 36 if (buf->bytesused - video->queue.buf_used <= len - 2) 46 struct uvc_video_queue *queue = &video->queue; local 51 mem = buf->mem + queue->buf_used; 52 nbytes = min((unsigned int)len, buf->bytesused - queue->buf_used); 55 queue->buf_used += nbytes; 86 if (buf->bytesused == video->queue.buf_used) { 87 video->queue.buf_used = 0; 89 uvcg_queue_next_buffer(&video->queue, buf); 96 buf->bytesused == video->queue 165 struct uvc_video_queue *queue = &video->queue; local 285 struct uvc_video_queue *queue = &video->queue; local [all...] |
/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_marker.c | 37 void vmw_marker_queue_init(struct vmw_marker_queue *queue) argument 39 INIT_LIST_HEAD(&queue->head); 40 queue->lag = 0; 41 queue->lag_time = ktime_get_raw_ns(); 42 spin_lock_init(&queue->lock); 45 void vmw_marker_queue_takedown(struct vmw_marker_queue *queue) argument 49 spin_lock(&queue->lock); 50 list_for_each_entry_safe(marker, next, &queue->head, head) { 53 spin_unlock(&queue->lock); 56 int vmw_marker_push(struct vmw_marker_queue *queue, argument 73 vmw_marker_pull(struct vmw_marker_queue *queue, uint32_t signaled_seqno) argument 107 vmw_fifo_lag(struct vmw_marker_queue *queue) argument 120 vmw_lag_lt(struct vmw_marker_queue *queue, uint32_t us) argument 128 vmw_wait_lag(struct vmw_private *dev_priv, struct vmw_marker_queue *queue, uint32_t us) argument [all...] |
/drivers/scsi/arm/ |
H A D | queue.h | 2 * linux/drivers/acorn/scsi/queue.h: queue handling 21 * Function: void queue_initialise (Queue_t *queue) 22 * Purpose : initialise a queue 23 * Params : queue - queue to initialise 25 extern int queue_initialise (Queue_t *queue); 28 * Function: void queue_free (Queue_t *queue) 29 * Purpose : free a queue 30 * Params : queue [all...] |
H A D | queue.c | 2 * linux/drivers/acorn/scsi/queue.c: queue handling primitives 49 #include "queue.h" 54 * Function: void queue_initialise (Queue_t *queue) 55 * Purpose : initialise a queue 56 * Params : queue - queue to initialise 58 int queue_initialise (Queue_t *queue) argument 63 spin_lock_init(&queue->queue_lock); 64 INIT_LIST_HEAD(&queue 90 queue_free(Queue_t *queue) argument 106 __queue_add(Queue_t *queue, struct scsi_cmnd *SCpnt, int head) argument 137 __queue_remove(Queue_t *queue, struct list_head *ent) argument 161 queue_remove_exclude(Queue_t *queue, unsigned long *exclude) argument 187 queue_remove(Queue_t *queue) argument 209 queue_remove_tgtluntag(Queue_t *queue, int target, int lun, int tag) argument 237 queue_remove_all_target(Queue_t *queue, int target) argument 260 queue_probetgtlun(Queue_t *queue, int target, int lun) argument 286 queue_remove_cmd(Queue_t *queue, struct scsi_cmnd *SCpnt) argument [all...] |
H A D | Makefile | 7 obj-$(CONFIG_SCSI_ACORNSCSI_3) += acornscsi_mod.o queue.o msgqueue.o 8 obj-$(CONFIG_SCSI_ARXESCSI) += arxescsi.o fas216.o queue.o msgqueue.o 10 obj-$(CONFIG_SCSI_CUMANA_2) += cumana_2.o fas216.o queue.o msgqueue.o 12 obj-$(CONFIG_SCSI_POWERTECSCSI) += powertec.o fas216.o queue.o msgqueue.o 13 obj-$(CONFIG_SCSI_EESOXSCSI) += eesox.o fas216.o queue.o msgqueue.o
|
/drivers/net/wireless/cw1200/ |
H A D | queue.c | 2 * O(1) TX queue with built-in allocator for ST-Ericsson CW1200 drivers 14 #include "queue.h" 29 static inline void __cw1200_queue_lock(struct cw1200_queue *queue) argument 31 struct cw1200_queue_stats *stats = queue->stats; 32 if (queue->tx_locked_cnt++ == 0) { 34 queue->queue_id); 35 ieee80211_stop_queue(stats->priv->hw, queue->queue_id); 39 static inline void __cw1200_queue_unlock(struct cw1200_queue *queue) argument 41 struct cw1200_queue_stats *stats = queue->stats; 42 BUG_ON(!queue 92 __cw1200_queue_gc(struct cw1200_queue *queue, struct list_head *head, bool unlock) argument 136 struct cw1200_queue *queue = local 165 cw1200_queue_init(struct cw1200_queue *queue, struct cw1200_queue_stats *stats, u8 queue_id, size_t capacity, unsigned long ttl) argument 205 cw1200_queue_clear(struct cw1200_queue *queue) argument 247 cw1200_queue_deinit(struct cw1200_queue *queue) argument 259 cw1200_queue_get_num_queued(struct cw1200_queue *queue, u32 link_id_map) argument 283 cw1200_queue_put(struct cw1200_queue *queue, struct sk_buff *skb, struct cw1200_txpriv *txpriv) argument 335 cw1200_queue_get(struct cw1200_queue *queue, u32 link_id_map, struct wsm_tx **tx, struct ieee80211_tx_info **tx_info, const struct cw1200_txpriv **txpriv) argument 376 cw1200_queue_requeue(struct cw1200_queue *queue, u32 packet_id) argument 418 cw1200_queue_requeue_all(struct cw1200_queue *queue) argument 445 cw1200_queue_remove(struct cw1200_queue *queue, u32 packet_id) argument 496 cw1200_queue_get_skb(struct cw1200_queue *queue, u32 packet_id, struct sk_buff **skb, const struct cw1200_txpriv **txpriv) argument 526 cw1200_queue_lock(struct cw1200_queue *queue) argument 533 cw1200_queue_unlock(struct cw1200_queue *queue) argument 540 cw1200_queue_get_xmit_timestamp(struct cw1200_queue *queue, unsigned long *timestamp, u32 pending_frame_id) argument [all...] |
H A D | queue.h | 2 * O(1) TX queue with built-in allocator for ST-Ericsson CW1200 drivers 36 struct list_head queue; member in struct:cw1200_queue 42 spinlock_t lock; /* Protect queue entry */ 71 int cw1200_queue_init(struct cw1200_queue *queue, 76 int cw1200_queue_clear(struct cw1200_queue *queue); 78 void cw1200_queue_deinit(struct cw1200_queue *queue); 80 size_t cw1200_queue_get_num_queued(struct cw1200_queue *queue, 82 int cw1200_queue_put(struct cw1200_queue *queue, 85 int cw1200_queue_get(struct cw1200_queue *queue, 90 int cw1200_queue_requeue(struct cw1200_queue *queue, u3 [all...] |
/drivers/infiniband/hw/ehca/ |
H A D | ipz_pt_fn.c | 4 * internal queue handling 51 void *ipz_qpageit_get_inc(struct ipz_queue *queue) argument 53 void *ret = ipz_qeit_get(queue); 54 queue->current_q_offset += queue->pagesize; 55 if (queue->current_q_offset > queue->queue_length) { 56 queue->current_q_offset -= queue->pagesize; 59 if (((u64)ret) % queue 66 ipz_qeit_eq_get_inc(struct ipz_queue *queue) argument 80 ipz_queue_abs_to_offset(struct ipz_queue *queue, u64 addr, u64 *q_offset) argument 102 alloc_queue_pages(struct ipz_queue *queue, const u32 nr_of_pages) argument 127 alloc_small_queue_page(struct ipz_queue *queue, struct ehca_pd *pd) argument 172 free_small_queue_page(struct ipz_queue *queue, struct ehca_pd *pd) argument 204 ipz_queue_ctor(struct ehca_pd *pd, struct ipz_queue *queue, const u32 nr_of_pages, const u32 pagesize, const u32 qe_size, const u32 nr_of_sg, int is_small) argument 256 ipz_queue_dtor(struct ehca_pd *pd, struct ipz_queue *queue) argument [all...] |
H A D | ipz_pt_fn.h | 4 * internal queue handling 75 /* struct generic queue in linux kernel virtual memory (kv) */ 77 u64 current_q_offset; /* current queue entry */ 79 struct ipz_page **queue_pages; /* array of pages belonging to queue */ 80 u32 qe_size; /* queue entry size */ 82 u32 queue_length; /* queue length allocated in bytes */ 93 static inline void *ipz_qeit_calc(struct ipz_queue *queue, u64 q_offset) argument 96 if (q_offset >= queue->queue_length) 98 current_page = (queue->queue_pages)[q_offset >> EHCA_PAGESHIFT]; 106 static inline void *ipz_qeit_get(struct ipz_queue *queue) argument 126 ipz_qeit_get_inc(struct ipz_queue *queue) argument 142 ipz_qeit_is_valid(struct ipz_queue *queue) argument 155 ipz_qeit_get_inc_valid(struct ipz_queue *queue) argument 164 ipz_qeit_reset(struct ipz_queue *queue) argument 178 ipz_queue_advance_offset(struct ipz_queue *queue, u64 offset) argument 258 ipz_eqit_eq_get_inc_valid(struct ipz_queue *queue) argument 268 ipz_eqit_eq_peek_valid(struct ipz_queue *queue) argument [all...] |
/drivers/net/wireless/b43legacy/ |
H A D | pio.c | 35 static void tx_start(struct b43legacy_pioqueue *queue) argument 37 b43legacy_pio_write(queue, B43legacy_PIO_TXCTL, 41 static void tx_octet(struct b43legacy_pioqueue *queue, argument 44 if (queue->need_workarounds) { 45 b43legacy_pio_write(queue, B43legacy_PIO_TXDATA, octet); 46 b43legacy_pio_write(queue, B43legacy_PIO_TXCTL, 49 b43legacy_pio_write(queue, B43legacy_PIO_TXCTL, 51 b43legacy_pio_write(queue, B43legacy_PIO_TXDATA, octet); 76 static void tx_data(struct b43legacy_pioqueue *queue, argument 84 if (queue 102 tx_complete(struct b43legacy_pioqueue *queue, struct sk_buff *skb) argument 116 generate_cookie(struct b43legacy_pioqueue *queue, struct b43legacy_pio_txpacket *packet) argument 154 struct b43legacy_pioqueue *queue = NULL; local 185 pio_tx_write_fragment(struct b43legacy_pioqueue *queue, struct sk_buff *skb, struct b43legacy_pio_txpacket *packet, size_t txhdr_size) argument 218 struct b43legacy_pioqueue *queue = packet->queue; local 232 struct b43legacy_pioqueue *queue = packet->queue; local 282 struct b43legacy_pioqueue *queue = (struct b43legacy_pioqueue *)d; local 312 setup_txqueues(struct b43legacy_pioqueue *queue) argument 332 struct b43legacy_pioqueue *queue; local 381 cancel_transfers(struct b43legacy_pioqueue *queue) argument 393 b43legacy_destroy_pioqueue(struct b43legacy_pioqueue *queue) argument 423 struct b43legacy_pioqueue *queue; local 469 struct b43legacy_pioqueue *queue = dev->pio.queue1; local 491 struct b43legacy_pioqueue *queue; local 549 pio_rx_error(struct b43legacy_pioqueue *queue, int clear_buffers, const char *error) argument 567 b43legacy_pio_rx(struct b43legacy_pioqueue *queue) argument 648 b43legacy_pio_tx_suspend(struct b43legacy_pioqueue *queue) argument 656 b43legacy_pio_tx_resume(struct b43legacy_pioqueue *queue) argument [all...] |
H A D | pio.h | 42 struct b43legacy_pioqueue *queue; member in struct:b43legacy_pio_txpacket 48 (packet)->queue->tx_packets_cache)) 73 /* Packets on the txrunning queue are completely 83 u16 b43legacy_pio_read(struct b43legacy_pioqueue *queue, argument 86 return b43legacy_read16(queue->dev, queue->mmio_base + offset); 90 void b43legacy_pio_write(struct b43legacy_pioqueue *queue, argument 93 b43legacy_write16(queue->dev, queue->mmio_base + offset, value); 105 void b43legacy_pio_rx(struct b43legacy_pioqueue *queue); 137 b43legacy_pio_rx(struct b43legacy_pioqueue *queue) argument 141 b43legacy_pio_tx_suspend(struct b43legacy_pioqueue *queue) argument 145 b43legacy_pio_tx_resume(struct b43legacy_pioqueue *queue) argument [all...] |
/drivers/net/xen-netback/ |
H A D | netback.c | 58 /* The time that packets can stay on the guest Rx internal queue 85 static void xenvif_idx_release(struct xenvif_queue *queue, u16 pending_idx, 88 static void make_tx_response(struct xenvif_queue *queue, 92 static inline int tx_work_todo(struct xenvif_queue *queue); 94 static struct xen_netif_rx_response *make_rx_response(struct xenvif_queue *queue, 101 static inline unsigned long idx_to_pfn(struct xenvif_queue *queue, argument 104 return page_to_pfn(queue->mmap_pages[idx]); 107 static inline unsigned long idx_to_kaddr(struct xenvif_queue *queue, argument 110 return (unsigned long)pfn_to_kaddr(idx_to_pfn(queue, idx)); 150 bool xenvif_rx_ring_slots_available(struct xenvif_queue *queue, in argument 172 xenvif_rx_queue_tail(struct xenvif_queue *queue, struct sk_buff *skb) argument 187 xenvif_rx_dequeue(struct xenvif_queue *queue) argument 202 xenvif_rx_queue_maybe_wake(struct xenvif_queue *queue) argument 213 xenvif_rx_queue_purge(struct xenvif_queue *queue) argument 220 xenvif_rx_queue_drop_expired(struct xenvif_queue *queue) argument 289 get_next_rx_buffer(struct xenvif_queue *queue, struct netrx_pending_operations *npo) argument 313 xenvif_gop_frag_copy(struct xenvif_queue *queue, struct sk_buff *skb, struct netrx_pending_operations *npo, struct page *page, unsigned long size, unsigned long offset, int *head, struct xenvif_queue *foreign_queue, grant_ref_t foreign_gref) argument 448 xenvif_gop_skb(struct sk_buff *skb, struct netrx_pending_operations *npo, struct xenvif_queue *queue) argument 602 xenvif_add_frag_responses(struct xenvif_queue *queue, int status, struct xenvif_rx_meta *meta, int nr_meta_slots) argument 628 xenvif_kick_thread(struct xenvif_queue *queue) argument 633 xenvif_rx_action(struct xenvif_queue *queue) argument 796 xenvif_napi_schedule_or_enable_events(struct xenvif_queue *queue) argument 806 tx_add_credit(struct xenvif_queue *queue) argument 828 struct xenvif_queue *queue = (struct xenvif_queue *)data; local 833 xenvif_tx_err(struct xenvif_queue *queue, struct xen_netif_tx_request *txp, RING_IDX end) argument 859 xenvif_count_requests(struct xenvif_queue *queue, struct xen_netif_tx_request *first, struct xen_netif_tx_request *txp, int work_to_do) argument 964 xenvif_tx_create_map_op(struct xenvif_queue *queue, u16 pending_idx, struct xen_netif_tx_request *txp, struct gnttab_map_grant_ref *mop) argument 995 xenvif_get_requests(struct xenvif_queue *queue, struct sk_buff *skb, struct xen_netif_tx_request *txp, struct gnttab_map_grant_ref *gop) argument 1055 xenvif_grant_handle_set(struct xenvif_queue *queue, u16 pending_idx, grant_handle_t handle) argument 1069 xenvif_grant_handle_reset(struct xenvif_queue *queue, u16 pending_idx) argument 1082 xenvif_tx_check_gop(struct xenvif_queue *queue, struct sk_buff *skb, struct gnttab_map_grant_ref **gopp_map, struct gnttab_copy **gopp_copy) argument 1207 xenvif_fill_frags(struct xenvif_queue *queue, struct sk_buff *skb) argument 1251 xenvif_get_extras(struct xenvif_queue *queue, struct xen_netif_extra_info *extras, int work_to_do) argument 1312 checksum_setup(struct xenvif_queue *queue, struct sk_buff *skb) argument 1334 tx_credit_exceeded(struct xenvif_queue *queue, unsigned size) argument 1366 xenvif_tx_build_gops(struct xenvif_queue *queue, int budget, unsigned *copy_ops, unsigned *map_ops) argument 1530 xenvif_handle_frag_list(struct xenvif_queue *queue, struct sk_buff *skb) argument 1595 xenvif_tx_submit(struct xenvif_queue *queue) argument 1716 struct xenvif_queue *queue = ubuf_to_queue(ubuf); local 1745 xenvif_tx_dealloc_action(struct xenvif_queue *queue) argument 1813 xenvif_tx_action(struct xenvif_queue *queue, int budget) argument 1840 xenvif_idx_release(struct xenvif_queue *queue, u16 pending_idx, u8 status) argument 1859 make_tx_response(struct xenvif_queue *queue, struct xen_netif_tx_request *txp, s8 st) argument 1880 make_rx_response(struct xenvif_queue *queue, u16 id, s8 st, u16 offset, u16 size, u16 flags) argument 1903 xenvif_idx_unmap(struct xenvif_queue *queue, u16 pending_idx) argument 1928 tx_work_todo(struct xenvif_queue *queue) argument 1936 tx_dealloc_work_todo(struct xenvif_queue *queue) argument 1941 xenvif_unmap_frontend_rings(struct xenvif_queue *queue) argument 1951 xenvif_map_frontend_rings(struct xenvif_queue *queue, grant_ref_t tx_ring_ref, grant_ref_t rx_ring_ref) argument 1984 xenvif_queue_carrier_off(struct xenvif_queue *queue) argument 1999 xenvif_queue_carrier_on(struct xenvif_queue *queue) argument 2015 xenvif_rx_queue_stalled(struct xenvif_queue *queue) argument 2028 xenvif_rx_queue_ready(struct xenvif_queue *queue) argument 2039 xenvif_have_rx_work(struct xenvif_queue *queue) argument 2049 xenvif_rx_queue_timeout(struct xenvif_queue *queue) argument 2072 xenvif_wait_for_rx_work(struct xenvif_queue *queue) argument 2094 struct xenvif_queue *queue = data; local 2146 xenvif_dealloc_kthread_should_stop(struct xenvif_queue *queue) argument 2157 struct xenvif_queue *queue = data; local [all...] |
H A D | interface.c | 46 /* Number of bytes allowed on the internal guest Rx queue. */ 54 void xenvif_skb_zerocopy_prepare(struct xenvif_queue *queue, argument 58 atomic_inc(&queue->inflight_packets); 61 void xenvif_skb_zerocopy_complete(struct xenvif_queue *queue) argument 63 atomic_dec(&queue->inflight_packets); 75 struct xenvif_queue *queue = dev_id; local 77 if (RING_HAS_UNCONSUMED_REQUESTS(&queue->tx)) 78 napi_schedule(&queue->napi); 85 struct xenvif_queue *queue = local 93 if (unlikely(queue 110 struct xenvif_queue *queue = dev_id; local 125 xenvif_queue_stopped(struct xenvif_queue *queue) argument 132 xenvif_wake_queue(struct xenvif_queue *queue) argument 142 struct xenvif_queue *queue = NULL; local 185 struct xenvif_queue *queue = NULL; local 216 struct xenvif_queue *queue = NULL; local 232 struct xenvif_queue *queue = NULL; local 458 xenvif_init_queue(struct xenvif_queue *queue) argument 515 xenvif_connect(struct xenvif_queue *queue, unsigned long tx_ring_ref, unsigned long rx_ring_ref, unsigned int tx_evtchn, unsigned int rx_evtchn) argument 622 struct xenvif_queue *queue = NULL; local 661 xenvif_deinit_queue(struct xenvif_queue *queue) argument 668 struct xenvif_queue *queue = NULL; local [all...] |
/drivers/misc/genwqe/ |
H A D | card_ddcb.c | 22 * Device Driver Control Block (DDCB) queue support. Definition of 23 * interrupt handlers for queue support as well as triggering the 49 * Situation (1): Empty queue 91 static int queue_empty(struct ddcb_queue *queue) argument 93 return queue->ddcb_next == queue->ddcb_act; 96 static int queue_enqueued_ddcbs(struct ddcb_queue *queue) argument 98 if (queue->ddcb_next >= queue->ddcb_act) 99 return queue 104 queue_free_ddcbs(struct ddcb_queue *queue) argument 172 print_ddcb_info(struct genwqe_dev *cd, struct ddcb_queue *queue) argument 274 enqueue_ddcb(struct genwqe_dev *cd, struct ddcb_queue *queue, struct ddcb *pddcb, int ddcb_no) argument 339 struct ddcb_queue *queue = req->queue; local 371 genwqe_check_ddcb_queue(struct genwqe_dev *cd, struct ddcb_queue *queue) argument 487 struct ddcb_queue *queue; local 512 struct ddcb_queue *queue = req->queue; local 571 get_next_ddcb(struct genwqe_dev *cd, struct ddcb_queue *queue, int *num) argument 629 struct ddcb_queue *queue = req->queue; local 763 struct ddcb_queue *queue; local 984 struct ddcb_queue *queue = &cd->queue; local 1014 struct ddcb_queue *queue = &cd->queue; local 1023 setup_ddcb_queue(struct genwqe_dev *cd, struct ddcb_queue *queue) argument 1101 ddcb_queue_initialized(struct ddcb_queue *queue) argument 1106 free_ddcb_queue(struct genwqe_dev *cd, struct ddcb_queue *queue) argument 1233 struct ddcb_queue *queue; local 1321 struct ddcb_queue *queue = &cd->queue; local 1347 struct ddcb_queue *queue = &cd->queue; local [all...] |
/drivers/net/wireless/rt2x00/ |
H A D | rt2x00queue.c | 23 Abstract: rt2x00 queue specific routines. 36 struct data_queue *queue = entry->queue; local 37 struct rt2x00_dev *rt2x00dev = queue->rt2x00dev; 48 frame_size = queue->data_size + queue->desc_size + queue->winfo_size; 107 struct device *dev = entry->queue->rt2x00dev->dev; 123 struct device *dev = entry->queue->rt2x00dev->dev; 498 struct rt2x00_dev *rt2x00dev = entry->queue 539 struct data_queue *queue = entry->queue; local 550 rt2x00queue_kick_tx_queue(struct data_queue *queue, struct txentry_desc *txdesc) argument 609 rt2x00queue_write_tx_frame(struct data_queue *queue, struct sk_buff *skb, struct ieee80211_sta *sta, bool local) argument 788 rt2x00queue_for_each_entry(struct data_queue *queue, enum queue_index start, enum queue_index end, void *data, bool (*fn)(struct queue_entry *entry, void *data)) argument 843 rt2x00queue_get_entry(struct data_queue *queue, enum queue_index index) argument 867 struct data_queue *queue = entry->queue; local 894 rt2x00queue_pause_queue_nocheck(struct data_queue *queue) argument 911 rt2x00queue_pause_queue(struct data_queue *queue) argument 922 rt2x00queue_unpause_queue(struct data_queue *queue) argument 952 rt2x00queue_start_queue(struct data_queue *queue) argument 972 rt2x00queue_stop_queue(struct data_queue *queue) argument 989 rt2x00queue_flush_queue(struct data_queue *queue, bool drop) argument 1026 struct data_queue *queue; local 1041 struct data_queue *queue; local 1060 struct data_queue *queue; local 1069 rt2x00queue_reset(struct data_queue *queue) argument 1087 struct data_queue *queue; local 1098 rt2x00queue_alloc_entries(struct data_queue *queue) argument 1135 rt2x00queue_free_skbs(struct data_queue *queue) argument 1147 rt2x00queue_alloc_rxskbs(struct data_queue *queue) argument 1164 struct data_queue *queue; local 1203 struct data_queue *queue; local 1213 rt2x00queue_init(struct rt2x00_dev *rt2x00dev, struct data_queue *queue, enum data_queue_qid qid) argument 1234 struct data_queue *queue; local [all...] |
H A D | rt2x00mmio.c | 62 struct data_queue *queue = rt2x00dev->rx; local 69 entry = rt2x00queue_get_entry(queue, Q_INDEX); 80 skbdesc->desc_len = entry->queue->desc_size; 99 void rt2x00mmio_flush_queue(struct data_queue *queue, bool drop) argument 103 for (i = 0; !rt2x00queue_empty(queue) && i < 10; i++) 112 struct data_queue *queue) 123 queue->limit * queue->desc_size, &dma, 129 * Initialize all queue entries to contain valid addresses. 131 for (i = 0; i < queue 111 rt2x00mmio_alloc_queue_dma(struct rt2x00_dev *rt2x00dev, struct data_queue *queue) argument 140 rt2x00mmio_free_queue_dma(struct rt2x00_dev *rt2x00dev, struct data_queue *queue) argument 155 struct data_queue *queue; local 191 struct data_queue *queue; local [all...] |
H A D | rt2x00usb.c | 248 struct data_queue *queue; local 251 tx_queue_for_each(rt2x00dev, queue) { 252 while (!rt2x00queue_empty(queue)) { 253 entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE); 267 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; 294 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; 322 usb_sndbulkpipe(usb_dev, entry->queue->usb_endpoint), 360 skbdesc->desc_len = entry->queue->desc_size; 372 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; 387 if (urb->actual_length < entry->queue 426 rt2x00usb_kick_queue(struct data_queue *queue) argument 475 rt2x00usb_flush_queue(struct data_queue *queue, bool drop) argument 525 rt2x00usb_watchdog_tx_dma(struct data_queue *queue) argument 535 rt2x00usb_dma_timeout(struct data_queue *queue) argument 545 struct data_queue *queue; local 578 rt2x00usb_assign_endpoint(struct data_queue *queue, struct usb_endpoint_descriptor *ep_desc) argument 603 struct data_queue *queue = rt2x00dev->tx; local 648 rt2x00usb_alloc_entries(struct data_queue *queue) argument 681 rt2x00usb_free_entries(struct data_queue *queue) argument 715 struct data_queue *queue; local 745 struct data_queue *queue; local [all...] |
/drivers/net/ethernet/ibm/ehea/ |
H A D | ehea_qmr.h | 210 static inline void *hw_qeit_calc(struct hw_queue *queue, u64 q_offset) argument 214 if (q_offset >= queue->queue_length) 215 q_offset -= queue->queue_length; 216 current_page = (queue->queue_pages)[q_offset >> EHEA_PAGESHIFT]; 220 static inline void *hw_qeit_get(struct hw_queue *queue) argument 222 return hw_qeit_calc(queue, queue->current_q_offset); 225 static inline void hw_qeit_inc(struct hw_queue *queue) argument 227 queue->current_q_offset += queue 235 hw_qeit_get_inc(struct hw_queue *queue) argument 242 hw_qeit_get_inc_valid(struct hw_queue *queue) argument 259 hw_qeit_get_valid(struct hw_queue *queue) argument 275 hw_qeit_reset(struct hw_queue *queue) argument 281 hw_qeit_eq_get_inc(struct hw_queue *queue) argument 295 hw_eqit_eq_get_inc_valid(struct hw_queue *queue) argument 309 struct hw_queue *queue; local 324 struct hw_queue *queue = &my_qp->hw_squeue; local 341 struct hw_queue *queue = &qp->hw_rqueue1; local [all...] |
/drivers/net/wireless/ath/ath5k/ |
H A D | qcu.c | 35 * basically we have 10 queues to play with. Each queue has a matching 36 * QCU that controls when the queue will get triggered and multiple QCUs 39 * and DCUs allowing us to have different DFS settings for each queue. 41 * When a frame goes into a TX queue, QCU decides when it'll trigger a 43 * it's buffer or -if it's a beacon queue- if it's time to fire up the queue 58 * ath5k_hw_num_tx_pending() - Get number of pending frames for a given queue 60 * @queue: One of enum ath5k_tx_queue_id 63 ath5k_hw_num_tx_pending(struct ath5k_hw *ah, unsigned int queue) argument 66 AR5K_ASSERT_ENTRY(queue, a 94 ath5k_hw_release_tx_queue(struct ath5k_hw *ah, unsigned int queue) argument 138 ath5k_hw_get_tx_queueprops(struct ath5k_hw *ah, int queue, struct ath5k_txq_info *queue_info) argument 154 ath5k_hw_set_tx_queueprops(struct ath5k_hw *ah, int queue, const struct ath5k_txq_info *qinfo) argument 206 unsigned int queue; local 287 ath5k_hw_set_tx_retry_limits(struct ath5k_hw *ah, unsigned int queue) argument 330 ath5k_hw_reset_tx_queue(struct ath5k_hw *ah, unsigned int queue) argument [all...] |
/drivers/net/ |
H A D | xen-netfront.c | 85 /* IRQ name is queue name with "-tx" or "-rx" appended */ 164 /* Multi-queue support */ 213 static struct sk_buff *xennet_get_rx_skb(struct netfront_queue *queue, argument 217 struct sk_buff *skb = queue->rx_skbs[i]; 218 queue->rx_skbs[i] = NULL; 222 static grant_ref_t xennet_get_rx_ref(struct netfront_queue *queue, argument 226 grant_ref_t ref = queue->grant_rx_ref[i]; 227 queue->grant_rx_ref[i] = GRANT_INVALID_REF; 247 struct netfront_queue *queue = (struct netfront_queue *)data; local 248 napi_schedule(&queue 251 netfront_tx_slot_available(struct netfront_queue *queue) argument 257 xennet_maybe_wake_tx(struct netfront_queue *queue) argument 268 xennet_alloc_rx_buffers(struct netfront_queue *queue) argument 375 struct netfront_queue *queue = NULL; local 396 xennet_tx_buf_gc(struct netfront_queue *queue) argument 451 xennet_make_frags(struct sk_buff *skb, struct netfront_queue *queue, struct xen_netif_tx_request *tx) argument 602 struct netfront_queue *queue = NULL; local 721 struct netfront_queue *queue; local 730 xennet_move_rx_slot(struct netfront_queue *queue, struct sk_buff *skb, grant_ref_t ref) argument 743 xennet_get_extras(struct netfront_queue *queue, struct xen_netif_extra_info *extras, RING_IDX rp) argument 787 xennet_get_responses(struct netfront_queue *queue, struct netfront_rx_info *rinfo, RING_IDX rp, struct sk_buff_head *list) argument 896 xennet_fill_frags(struct netfront_queue *queue, struct sk_buff *skb, struct sk_buff_head *list) argument 951 handle_incoming_queue(struct netfront_queue *queue, struct sk_buff_head *rxq) argument 989 struct netfront_queue *queue = container_of(napi, struct netfront_queue, napi); local 1139 xennet_release_tx_bufs(struct netfront_queue *queue) argument 1161 xennet_release_rx_bufs(struct netfront_queue *queue) argument 1253 struct netfront_queue *queue = dev_id; local 1265 struct netfront_queue *queue = dev_id; local 1424 struct netfront_queue *queue = &info->queues[i]; local 1491 setup_netfront_single(struct netfront_queue *queue) argument 1516 setup_netfront_split(struct netfront_queue *queue) argument 1560 setup_netfront(struct xenbus_device *dev, struct netfront_queue *queue, unsigned int feature_split_evtchn) argument 1633 xennet_init_queue(struct netfront_queue *queue) argument 1691 write_queue_xenstore_keys(struct netfront_queue *queue, struct xenbus_transaction *xbt, int write_hierarchical) argument 1780 struct netfront_queue *queue = &info->queues[i]; local 1807 struct netfront_queue *queue = &info->queues[i]; local 1847 struct netfront_queue *queue = NULL; local 2000 struct netfront_queue *queue = NULL; local 2167 struct netfront_queue *queue; local 2220 struct netfront_queue *queue = NULL; local 2302 struct netfront_queue *queue = NULL; local [all...] |
H A D | eql.c | 140 static void eql_kill_one_slave(slave_queue_t *queue, slave_t *slave); 147 spin_lock(&eql->queue.lock); 148 head = &eql->queue.all_slaves; 157 eql_kill_one_slave(&eql->queue, slave); 161 spin_unlock(&eql->queue.lock); 186 spin_lock_init(&eql->queue.lock); 187 INIT_LIST_HEAD(&eql->queue.all_slaves); 188 eql->queue.master_dev = dev; 213 BUG_ON(!list_empty(&eql->queue.all_slaves)); 223 static void eql_kill_one_slave(slave_queue_t *queue, slave_ argument 232 eql_kill_slave_queue(slave_queue_t *queue) argument 298 __eql_schedule_slaves(slave_queue_t *queue) argument 364 __eql_find_slave_dev(slave_queue_t *queue, struct net_device *dev) argument 379 eql_is_full(slave_queue_t *queue) argument 389 __eql_insert_slave(slave_queue_t *queue, slave_t *slave) argument [all...] |
/drivers/mmc/card/ |
H A D | Makefile | 6 mmc_block-objs := block.o queue.o
|