/drivers/media/video/uvc/ |
H A D | uvc_queue.c | 28 * Video buffers queue management. 34 * the videobuf2 queue operations by serializing calls to videobuf2 and a 35 * spinlock to protect the IRQ queue that holds the buffers to be processed by 40 * videobuf2 queue operations 47 struct uvc_video_queue *queue = vb2_get_drv_priv(vq); local 49 container_of(queue, struct uvc_streaming, queue); 63 struct uvc_video_queue *queue = vb2_get_drv_priv(vb->vb2_queue); local 72 if (unlikely(queue->flags & UVC_QUEUE_DISCONNECTED)) 89 struct uvc_video_queue *queue local 109 struct uvc_video_queue *queue = vb2_get_drv_priv(vb->vb2_queue); local 125 uvc_queue_init(struct uvc_video_queue *queue, enum v4l2_buf_type type, int drop_corrupted) argument 146 uvc_alloc_buffers(struct uvc_video_queue *queue, struct v4l2_requestbuffers *rb) argument 158 uvc_free_buffers(struct uvc_video_queue *queue) argument 165 uvc_query_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf) argument 176 uvc_queue_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf) argument 187 uvc_dequeue_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf, int nonblocking) argument 199 uvc_queue_mmap(struct uvc_video_queue *queue, struct vm_area_struct *vma) argument 210 uvc_queue_poll(struct uvc_video_queue *queue, struct file *file, poll_table *wait) argument 229 uvc_queue_allocated(struct uvc_video_queue *queue) argument 246 uvc_queue_get_unmapped_area(struct uvc_video_queue *queue, unsigned long pgoff) argument 286 uvc_queue_enable(struct uvc_video_queue *queue, int enable) argument 325 uvc_queue_cancel(struct uvc_video_queue *queue, int disconnect) argument 349 uvc_queue_next_buffer(struct uvc_video_queue *queue, struct uvc_buffer *buf) argument [all...] |
/drivers/scsi/arm/ |
H A D | queue.h | 2 * linux/drivers/acorn/scsi/queue.h: queue handling 21 * Function: void queue_initialise (Queue_t *queue) 22 * Purpose : initialise a queue 23 * Params : queue - queue to initialise 25 extern int queue_initialise (Queue_t *queue); 28 * Function: void queue_free (Queue_t *queue) 29 * Purpose : free a queue 30 * Params : queue [all...] |
H A D | queue.c | 2 * linux/drivers/acorn/scsi/queue.c: queue handling primitives 49 #include "queue.h" 54 * Function: void queue_initialise (Queue_t *queue) 55 * Purpose : initialise a queue 56 * Params : queue - queue to initialise 58 int queue_initialise (Queue_t *queue) argument 63 spin_lock_init(&queue->queue_lock); 64 INIT_LIST_HEAD(&queue 90 queue_free(Queue_t *queue) argument 106 __queue_add(Queue_t *queue, struct scsi_cmnd *SCpnt, int head) argument 137 __queue_remove(Queue_t *queue, struct list_head *ent) argument 161 queue_remove_exclude(Queue_t *queue, unsigned long *exclude) argument 186 queue_remove(Queue_t *queue) argument 208 queue_remove_tgtluntag(Queue_t *queue, int target, int lun, int tag) argument 236 queue_remove_all_target(Queue_t *queue, int target) argument 259 queue_probetgtlun(Queue_t *queue, int target, int lun) argument 285 queue_remove_cmd(Queue_t *queue, struct scsi_cmnd *SCpnt) argument [all...] |
H A D | Makefile | 7 obj-$(CONFIG_SCSI_ACORNSCSI_3) += acornscsi_mod.o queue.o msgqueue.o 8 obj-$(CONFIG_SCSI_ARXESCSI) += arxescsi.o fas216.o queue.o msgqueue.o 10 obj-$(CONFIG_SCSI_CUMANA_2) += cumana_2.o fas216.o queue.o msgqueue.o 12 obj-$(CONFIG_SCSI_POWERTECSCSI) += powertec.o fas216.o queue.o msgqueue.o 13 obj-$(CONFIG_SCSI_EESOXSCSI) += eesox.o fas216.o queue.o msgqueue.o
|
/drivers/infiniband/hw/ehca/ |
H A D | ipz_pt_fn.c | 4 * internal queue handling 51 void *ipz_qpageit_get_inc(struct ipz_queue *queue) argument 53 void *ret = ipz_qeit_get(queue); 54 queue->current_q_offset += queue->pagesize; 55 if (queue->current_q_offset > queue->queue_length) { 56 queue->current_q_offset -= queue->pagesize; 59 if (((u64)ret) % queue 66 ipz_qeit_eq_get_inc(struct ipz_queue *queue) argument 80 ipz_queue_abs_to_offset(struct ipz_queue *queue, u64 addr, u64 *q_offset) argument 102 alloc_queue_pages(struct ipz_queue *queue, const u32 nr_of_pages) argument 127 alloc_small_queue_page(struct ipz_queue *queue, struct ehca_pd *pd) argument 172 free_small_queue_page(struct ipz_queue *queue, struct ehca_pd *pd) argument 204 ipz_queue_ctor(struct ehca_pd *pd, struct ipz_queue *queue, const u32 nr_of_pages, const u32 pagesize, const u32 qe_size, const u32 nr_of_sg, int is_small) argument 255 ipz_queue_dtor(struct ehca_pd *pd, struct ipz_queue *queue) argument [all...] |
H A D | ipz_pt_fn.h | 4 * internal queue handling 75 /* struct generic queue in linux kernel virtual memory (kv) */ 77 u64 current_q_offset; /* current queue entry */ 79 struct ipz_page **queue_pages; /* array of pages belonging to queue */ 80 u32 qe_size; /* queue entry size */ 82 u32 queue_length; /* queue length allocated in bytes */ 93 static inline void *ipz_qeit_calc(struct ipz_queue *queue, u64 q_offset) argument 96 if (q_offset >= queue->queue_length) 98 current_page = (queue->queue_pages)[q_offset >> EHCA_PAGESHIFT]; 106 static inline void *ipz_qeit_get(struct ipz_queue *queue) argument 126 ipz_qeit_get_inc(struct ipz_queue *queue) argument 142 ipz_qeit_is_valid(struct ipz_queue *queue) argument 155 ipz_qeit_get_inc_valid(struct ipz_queue *queue) argument 164 ipz_qeit_reset(struct ipz_queue *queue) argument 178 ipz_queue_advance_offset(struct ipz_queue *queue, u64 offset) argument 258 ipz_eqit_eq_get_inc_valid(struct ipz_queue *queue) argument 268 ipz_eqit_eq_peek_valid(struct ipz_queue *queue) argument [all...] |
/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_marker.c | 37 void vmw_marker_queue_init(struct vmw_marker_queue *queue) argument 39 INIT_LIST_HEAD(&queue->head); 40 queue->lag = ns_to_timespec(0); 41 getrawmonotonic(&queue->lag_time); 42 spin_lock_init(&queue->lock); 45 void vmw_marker_queue_takedown(struct vmw_marker_queue *queue) argument 49 spin_lock(&queue->lock); 50 list_for_each_entry_safe(marker, next, &queue->head, head) { 53 spin_unlock(&queue->lock); 56 int vmw_marker_push(struct vmw_marker_queue *queue, argument 73 vmw_marker_pull(struct vmw_marker_queue *queue, uint32_t signaled_seqno) argument 120 vmw_fifo_lag(struct vmw_marker_queue *queue) argument 134 vmw_lag_lt(struct vmw_marker_queue *queue, uint32_t us) argument 144 vmw_wait_lag(struct vmw_private *dev_priv, struct vmw_marker_queue *queue, uint32_t us) argument [all...] |
/drivers/net/wireless/b43legacy/ |
H A D | pio.c | 35 static void tx_start(struct b43legacy_pioqueue *queue) argument 37 b43legacy_pio_write(queue, B43legacy_PIO_TXCTL, 41 static void tx_octet(struct b43legacy_pioqueue *queue, argument 44 if (queue->need_workarounds) { 45 b43legacy_pio_write(queue, B43legacy_PIO_TXDATA, octet); 46 b43legacy_pio_write(queue, B43legacy_PIO_TXCTL, 49 b43legacy_pio_write(queue, B43legacy_PIO_TXCTL, 51 b43legacy_pio_write(queue, B43legacy_PIO_TXDATA, octet); 76 static void tx_data(struct b43legacy_pioqueue *queue, argument 84 if (queue 102 tx_complete(struct b43legacy_pioqueue *queue, struct sk_buff *skb) argument 116 generate_cookie(struct b43legacy_pioqueue *queue, struct b43legacy_pio_txpacket *packet) argument 154 struct b43legacy_pioqueue *queue = NULL; local 185 pio_tx_write_fragment(struct b43legacy_pioqueue *queue, struct sk_buff *skb, struct b43legacy_pio_txpacket *packet, size_t txhdr_size) argument 218 struct b43legacy_pioqueue *queue = packet->queue; local 232 struct b43legacy_pioqueue *queue = packet->queue; local 282 struct b43legacy_pioqueue *queue = (struct b43legacy_pioqueue *)d; local 312 setup_txqueues(struct b43legacy_pioqueue *queue) argument 332 struct b43legacy_pioqueue *queue; local 381 cancel_transfers(struct b43legacy_pioqueue *queue) argument 393 b43legacy_destroy_pioqueue(struct b43legacy_pioqueue *queue) argument 423 struct b43legacy_pioqueue *queue; local 469 struct b43legacy_pioqueue *queue = dev->pio.queue1; local 491 struct b43legacy_pioqueue *queue; local 549 pio_rx_error(struct b43legacy_pioqueue *queue, int clear_buffers, const char *error) argument 567 b43legacy_pio_rx(struct b43legacy_pioqueue *queue) argument 648 b43legacy_pio_tx_suspend(struct b43legacy_pioqueue *queue) argument 656 b43legacy_pio_tx_resume(struct b43legacy_pioqueue *queue) argument [all...] |
H A D | pio.h | 42 struct b43legacy_pioqueue *queue; member in struct:b43legacy_pio_txpacket 48 (packet)->queue->tx_packets_cache)) 73 /* Packets on the txrunning queue are completely 83 u16 b43legacy_pio_read(struct b43legacy_pioqueue *queue, argument 86 return b43legacy_read16(queue->dev, queue->mmio_base + offset); 90 void b43legacy_pio_write(struct b43legacy_pioqueue *queue, argument 93 b43legacy_write16(queue->dev, queue->mmio_base + offset, value); 105 void b43legacy_pio_rx(struct b43legacy_pioqueue *queue); 137 b43legacy_pio_rx(struct b43legacy_pioqueue *queue) argument 141 b43legacy_pio_tx_suspend(struct b43legacy_pioqueue *queue) argument 145 b43legacy_pio_tx_resume(struct b43legacy_pioqueue *queue) argument [all...] |
/drivers/usb/gadget/ |
H A D | uvc_queue.c | 26 * Video buffers queue management. 32 * uvc_free_buffers respectively. The former acquires the video queue lock, 38 * drivers that use an in queue and an out queue, we use a main queue to hold 39 * all queued buffers (both 'empty' and 'done' buffers), and an irq queue to 41 * in interrupt, as only one queue is shared between interrupt and user 47 * Unless stated otherwise, all operations that modify the irq buffers queue 53 * protected by the queue lock, and the later is protected by the irq 56 * The completion handler fetches a buffer from the irq queue an 81 uvc_queue_init(struct uvc_video_queue *queue, enum v4l2_buf_type type) argument 95 uvc_free_buffers(struct uvc_video_queue *queue) argument 121 uvc_alloc_buffers(struct uvc_video_queue *queue, unsigned int nbuffers, unsigned int buflength) argument 200 uvc_query_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *v4l2_buf) argument 222 uvc_queue_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *v4l2_buf) argument 303 uvc_dequeue_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *v4l2_buf, int nonblocking) argument 365 uvc_queue_poll(struct uvc_video_queue *queue, struct file *file, poll_table *wait) argument 414 uvc_queue_mmap(struct uvc_video_queue *queue, struct vm_area_struct *vma) argument 476 uvc_queue_cancel(struct uvc_video_queue *queue, int disconnect) argument 517 uvc_queue_enable(struct uvc_video_queue *queue, int enable) argument 548 uvc_queue_next_buffer(struct uvc_video_queue *queue, struct uvc_buffer *buf) argument 573 uvc_queue_head(struct uvc_video_queue *queue) argument [all...] |
H A D | uvc_video.c | 35 if (buf->buf.bytesused - video->queue.buf_used <= len - 2) 45 struct uvc_video_queue *queue = &video->queue; local 50 mem = queue->mem + buf->buf.m.offset + queue->buf_used; 51 nbytes = min((unsigned int)len, buf->buf.bytesused - queue->buf_used); 54 queue->buf_used += nbytes; 85 if (buf->buf.bytesused == video->queue.buf_used) { 86 video->queue.buf_used = 0; 88 uvc_queue_next_buffer(&video->queue, bu [all...] |
H A D | uvc_queue.h | 33 struct list_head queue; member in struct:uvc_buffer 61 static inline int uvc_queue_streaming(struct uvc_video_queue *queue) argument 63 return queue->flags & UVC_QUEUE_STREAMING;
|
/drivers/net/wireless/rt2x00/ |
H A D | rt2x00queue.c | 25 Abstract: rt2x00 queue specific routines. 38 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; 49 frame_size = entry->queue->data_size + entry->queue->desc_size; 102 struct device *dev = entry->queue->rt2x00dev->dev; 113 struct device *dev = entry->queue->rt2x00dev->dev; 494 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; 499 * a queue corruption! 504 "Corrupt queue %d, accessing entry which is not ours.\n" 506 entry->queue 534 struct data_queue *queue = entry->queue; local 545 rt2x00queue_kick_tx_queue(struct data_queue *queue, struct txentry_desc *txdesc) argument 562 rt2x00queue_write_tx_frame(struct data_queue *queue, struct sk_buff *skb, bool local) argument 753 rt2x00queue_for_each_entry(struct data_queue *queue, enum queue_index start, enum queue_index end, void *data, bool (*fn)(struct queue_entry *entry, void *data)) argument 808 rt2x00queue_get_entry(struct data_queue *queue, enum queue_index index) argument 832 struct data_queue *queue = entry->queue; local 859 rt2x00queue_pause_queue(struct data_queue *queue) argument 883 rt2x00queue_unpause_queue(struct data_queue *queue) argument 913 rt2x00queue_start_queue(struct data_queue *queue) argument 933 rt2x00queue_stop_queue(struct data_queue *queue) argument 950 rt2x00queue_flush_queue(struct data_queue *queue, bool drop) argument 1011 struct data_queue *queue; local 1026 struct data_queue *queue; local 1045 struct data_queue *queue; local 1054 rt2x00queue_reset(struct data_queue *queue) argument 1072 struct data_queue *queue; local 1083 rt2x00queue_alloc_entries(struct data_queue *queue, const struct data_queue_desc *qdesc) argument 1126 rt2x00queue_free_skbs(struct data_queue *queue) argument 1138 rt2x00queue_alloc_rxskbs(struct data_queue *queue) argument 1155 struct data_queue *queue; local 1195 struct data_queue *queue; local 1205 rt2x00queue_init(struct rt2x00_dev *rt2x00dev, struct data_queue *queue, enum data_queue_qid qid) argument 1222 struct data_queue *queue; local [all...] |
H A D | rt2x00usb.c | 244 struct data_queue *queue; local 247 tx_queue_for_each(rt2x00dev, queue) { 248 while (!rt2x00queue_empty(queue)) { 249 entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE); 263 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; 290 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; 318 usb_sndbulkpipe(usb_dev, entry->queue->usb_endpoint), 356 skbdesc->desc_len = entry->queue->desc_size; 368 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; 383 if (urb->actual_length < entry->queue 422 rt2x00usb_kick_queue(struct data_queue *queue) argument 471 rt2x00usb_flush_queue(struct data_queue *queue, bool drop) argument 521 rt2x00usb_watchdog_tx_dma(struct data_queue *queue) argument 529 rt2x00usb_dma_timeout(struct data_queue *queue) argument 539 struct data_queue *queue; local 572 rt2x00usb_assign_endpoint(struct data_queue *queue, struct usb_endpoint_descriptor *ep_desc) argument 597 struct data_queue *queue = rt2x00dev->tx; local 642 rt2x00usb_alloc_entries(struct data_queue *queue) argument 675 rt2x00usb_free_entries(struct data_queue *queue) argument 709 struct data_queue *queue; local 739 struct data_queue *queue; local [all...] |
H A D | rt2x00pci.c | 65 struct data_queue *queue = rt2x00dev->rx; local 72 entry = rt2x00queue_get_entry(queue, Q_INDEX); 83 skbdesc->desc_len = entry->queue->desc_size; 102 void rt2x00pci_flush_queue(struct data_queue *queue, bool drop) argument 106 for (i = 0; !rt2x00queue_empty(queue) && i < 10; i++) 115 struct data_queue *queue) 126 queue->limit * queue->desc_size, 131 memset(addr, 0, queue->limit * queue 114 rt2x00pci_alloc_queue_dma(struct rt2x00_dev *rt2x00dev, struct data_queue *queue) argument 145 rt2x00pci_free_queue_dma(struct rt2x00_dev *rt2x00dev, struct data_queue *queue) argument 160 struct data_queue *queue; local 196 struct data_queue *queue; local [all...] |
/drivers/media/video/omap3isp/ |
H A D | ispqueue.h | 4 * TI OMAP3 ISP - Video buffers queue handling 65 * @stream: List head for insertion into main queue 66 * @queue: ISP buffers queue this buffer belongs to 78 * @irqlist: List head for insertion into IRQ queue 80 * @wait: Wait queue to signal buffer completion 85 struct isp_video_queue *queue; member in struct:isp_video_buffer 118 * the userspace memory address for a USERPTR buffer, with the queue lock 121 * @buffer_queue: Called when a buffer is being added to the queue with the 122 * queue irqloc 162 struct list_head queue; member in struct:isp_video_queue [all...] |
H A D | ispqueue.c | 4 * TI OMAP3 ISP - Video buffers queue handling 266 if (buf->queue->ops->buffer_cleanup) 267 buf->queue->ops->buffer_cleanup(buf); 272 dma_unmap_sg(buf->queue->dev, buf->sglist, buf->sglen, 520 ret = dma_map_sg(buf->queue->dev, buf->sglist, buf->sglen, 528 if (buf->queue->ops->buffer_prepare) 529 ret = buf->queue->ops->buffer_prepare(buf); 543 * Locking: must be called with the queue lock held. 575 * queue using the same condition. 597 * Buffers can only be freed if the queue is 602 isp_video_queue_free(struct isp_video_queue *queue) argument 636 isp_video_queue_alloc(struct isp_video_queue *queue, unsigned int nbuffers, unsigned int size, enum v4l2_memory memory) argument 704 omap3isp_video_queue_cleanup(struct isp_video_queue *queue) argument 730 omap3isp_video_queue_init(struct isp_video_queue *queue, enum v4l2_buf_type type, const struct isp_video_queue_operations *ops, struct device *dev, unsigned int bufsize) argument 772 omap3isp_video_queue_reqbufs(struct isp_video_queue *queue, struct v4l2_requestbuffers *rb) argument 810 omap3isp_video_queue_querybuf(struct isp_video_queue *queue, struct v4l2_buffer *vbuf) argument 847 omap3isp_video_queue_qbuf(struct isp_video_queue *queue, struct v4l2_buffer *vbuf) argument 919 omap3isp_video_queue_dqbuf(struct isp_video_queue *queue, struct v4l2_buffer *vbuf, int nonblocking) argument 960 omap3isp_video_queue_streamon(struct isp_video_queue *queue) argument 992 omap3isp_video_queue_streamoff(struct isp_video_queue *queue) argument 1033 omap3isp_video_queue_discard_done(struct isp_video_queue *queue) argument 1081 omap3isp_video_queue_mmap(struct isp_video_queue *queue, struct vm_area_struct *vma) argument 1132 omap3isp_video_queue_poll(struct isp_video_queue *queue, struct file *file, poll_table *wait) argument [all...] |
/drivers/net/ethernet/ibm/ehea/ |
H A D | ehea_qmr.h | 210 static inline void *hw_qeit_calc(struct hw_queue *queue, u64 q_offset) argument 214 if (q_offset >= queue->queue_length) 215 q_offset -= queue->queue_length; 216 current_page = (queue->queue_pages)[q_offset >> EHEA_PAGESHIFT]; 220 static inline void *hw_qeit_get(struct hw_queue *queue) argument 222 return hw_qeit_calc(queue, queue->current_q_offset); 225 static inline void hw_qeit_inc(struct hw_queue *queue) argument 227 queue->current_q_offset += queue 235 hw_qeit_get_inc(struct hw_queue *queue) argument 242 hw_qeit_get_inc_valid(struct hw_queue *queue) argument 259 hw_qeit_get_valid(struct hw_queue *queue) argument 275 hw_qeit_reset(struct hw_queue *queue) argument 281 hw_qeit_eq_get_inc(struct hw_queue *queue) argument 295 hw_eqit_eq_get_inc_valid(struct hw_queue *queue) argument 309 struct hw_queue *queue; local 324 struct hw_queue *queue = &my_qp->hw_squeue; local 341 struct hw_queue *queue = &qp->hw_rqueue1; local [all...] |
/drivers/scsi/ibmvscsi/ |
H A D | rpa_vscsi.c | 44 * Routines for managing the command/response queue 65 * @queue: crq_queue to initialize and register 71 static void rpavscsi_release_crq_queue(struct crq_queue *queue, argument 85 queue->msg_token, 86 queue->size * sizeof(*queue->msgs), DMA_BIDIRECTIONAL); 87 free_page((unsigned long)queue->msgs); 91 * crq_queue_next_crq: - Returns the next entry in message queue 92 * @queue: crq_queue to use 94 * Returns pointer to next entry in queue, o 97 crq_queue_next_crq(struct crq_queue *queue) argument 203 rpavscsi_reset_crq_queue(struct crq_queue *queue, struct ibmvscsi_host_data *hostdata) argument 244 rpavscsi_init_crq_queue(struct crq_queue *queue, struct ibmvscsi_host_data *hostdata, int max_requests) argument 331 rpavscsi_reenable_crq_queue(struct crq_queue *queue, struct ibmvscsi_host_data *hostdata) argument [all...] |
/drivers/net/wireless/ath/ath5k/ |
H A D | qcu.c | 33 * basically we have 10 queues to play with. Each queue has a matching 34 * QCU that controls when the queue will get triggered and multiple QCUs 37 * and DCUs allowing us to have different DFS settings for each queue. 39 * When a frame goes into a TX queue, QCU decides when it'll trigger a 41 * it's buffer or -if it's a beacon queue- if it's time to fire up the queue 56 * ath5k_hw_num_tx_pending() - Get number of pending frames for a given queue 58 * @queue: One of enum ath5k_tx_queue_id 61 ath5k_hw_num_tx_pending(struct ath5k_hw *ah, unsigned int queue) argument 64 AR5K_ASSERT_ENTRY(queue, a 92 ath5k_hw_release_tx_queue(struct ath5k_hw *ah, unsigned int queue) argument 136 ath5k_hw_get_tx_queueprops(struct ath5k_hw *ah, int queue, struct ath5k_txq_info *queue_info) argument 152 ath5k_hw_set_tx_queueprops(struct ath5k_hw *ah, int queue, const struct ath5k_txq_info *qinfo) argument 204 unsigned int queue; local 285 ath5k_hw_set_tx_retry_limits(struct ath5k_hw *ah, unsigned int queue) argument 328 ath5k_hw_reset_tx_queue(struct ath5k_hw *ah, unsigned int queue) argument [all...] |
H A D | dma.c | 27 * handle queue setup for 5210 chipset (rest are handled on qcu.c). 114 * ath5k_hw_start_tx_dma() - Start DMA transmit for a specific queue 116 * @queue: The hw queue number 118 * Start DMA transmit for a specific queue and since 5210 doesn't have 119 * QCU/DCU, set up queue parameters for 5210 here based on queue type (one 120 * queue for normal data and one queue for beacons). For queue setu 128 ath5k_hw_start_tx_dma(struct ath5k_hw *ah, unsigned int queue) argument 186 ath5k_hw_stop_tx_dma(struct ath5k_hw *ah, unsigned int queue) argument 326 ath5k_hw_stop_beacon_queue(struct ath5k_hw *ah, unsigned int queue) argument 351 ath5k_hw_get_txdp(struct ath5k_hw *ah, unsigned int queue) argument 394 ath5k_hw_set_txdp(struct ath5k_hw *ah, unsigned int queue, u32 phys_addr) argument [all...] |
/drivers/net/ |
H A D | eql.c | 140 static void eql_kill_one_slave(slave_queue_t *queue, slave_t *slave); 147 spin_lock(&eql->queue.lock); 148 head = &eql->queue.all_slaves; 157 eql_kill_one_slave(&eql->queue, slave); 161 spin_unlock(&eql->queue.lock); 186 spin_lock_init(&eql->queue.lock); 187 INIT_LIST_HEAD(&eql->queue.all_slaves); 188 eql->queue.master_dev = dev; 213 BUG_ON(!list_empty(&eql->queue.all_slaves)); 223 static void eql_kill_one_slave(slave_queue_t *queue, slave_ argument 232 eql_kill_slave_queue(slave_queue_t *queue) argument 298 __eql_schedule_slaves(slave_queue_t *queue) argument 364 __eql_find_slave_dev(slave_queue_t *queue, struct net_device *dev) argument 379 eql_is_full(slave_queue_t *queue) argument 389 __eql_insert_slave(slave_queue_t *queue, slave_t *slave) argument [all...] |
/drivers/mmc/card/ |
H A D | Makefile | 6 mmc_block-objs := block.o queue.o
|
/drivers/isdn/i4l/ |
H A D | isdn_net.h | 85 lp = nd->queue; /* get lp on top of queue */ 86 while (isdn_net_lp_busy(nd->queue)) { 87 nd->queue = nd->queue->next; 88 if (nd->queue == lp) { /* not found -- should never happen */ 93 lp = nd->queue; 94 nd->queue = nd->queue->next; 114 lp = nd->queue; [all...] |
/drivers/message/i2o/ |
H A D | i2o_block.h | 26 * Added a queue depth. 29 * Removed queue walk, fixed for 64bitness. 34 * Heavily chop down the queue depths 59 /* request queue sizes */ 75 spinlock_t lock; /* queue lock */ 78 unsigned int open_queue_depth; /* number of requests in the queue */ 89 struct list_head queue; member in struct:i2o_block_request 100 struct request_queue *queue; member in struct:i2o_block_delayed_request
|