/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_cq.c | 49 struct t3_wq *wq; local 62 wq = NULL; 65 wq = &(qhp->wq); 67 ret = cxio_poll_cq(wq, &(chp->cq), &cqe, &cqe_flushed, &cookie, 190 if (wq)
|
/drivers/scsi/fnic/ |
H A D | vnic_wq_copy.c | 25 void vnic_wq_copy_enable(struct vnic_wq_copy *wq) argument 27 iowrite32(1, &wq->ctrl->enable); 30 int vnic_wq_copy_disable(struct vnic_wq_copy *wq) argument 34 iowrite32(0, &wq->ctrl->enable); 38 if (!(ioread32(&wq->ctrl->running))) 45 wq->index, ioread32(&wq->ctrl->fetch_index), 46 ioread32(&wq->ctrl->posted_index)); 51 void vnic_wq_copy_clean(struct vnic_wq_copy *wq, argument 52 void (*q_clean)(struct vnic_wq_copy *wq, 69 vnic_wq_copy_free(struct vnic_wq_copy *wq) argument 78 vnic_wq_copy_alloc(struct vnic_dev *vdev, struct vnic_wq_copy *wq, unsigned int index, unsigned int desc_count, unsigned int desc_size) argument 102 vnic_wq_copy_init(struct vnic_wq_copy *wq, unsigned int cq_index, unsigned int error_interrupt_enable, unsigned int error_interrupt_offset) argument [all...] |
H A D | vnic_wq.c | 27 static int vnic_wq_alloc_bufs(struct vnic_wq *wq) argument 31 unsigned int i, j, count = wq->ring.desc_count; 34 vdev = wq->vdev; 37 wq->bufs[i] = kzalloc(VNIC_WQ_BUF_BLK_SZ, GFP_ATOMIC); 38 if (!wq->bufs[i]) { 45 buf = wq->bufs[i]; 48 buf->desc = (u8 *)wq->ring.descs + 49 wq->ring.desc_size * buf->index; 51 buf->next = wq->bufs[0]; 54 buf->next = wq 67 vnic_wq_free(struct vnic_wq *wq) argument 85 vnic_wq_alloc(struct vnic_dev *vdev, struct vnic_wq *wq, unsigned int index, unsigned int desc_count, unsigned int desc_size) argument 114 vnic_wq_init(struct vnic_wq *wq, unsigned int cq_index, unsigned int error_interrupt_enable, unsigned int error_interrupt_offset) argument 131 vnic_wq_error_status(struct vnic_wq *wq) argument 136 vnic_wq_enable(struct vnic_wq *wq) argument 141 vnic_wq_disable(struct vnic_wq *wq) argument 159 vnic_wq_clean(struct vnic_wq *wq, void (*buf_clean)(struct vnic_wq *wq, struct vnic_wq_buf *buf)) argument [all...] |
H A D | vnic_wq_copy.h | 36 static inline unsigned int vnic_wq_copy_desc_avail(struct vnic_wq_copy *wq) argument 38 return wq->ring.desc_avail; 41 static inline unsigned int vnic_wq_copy_desc_in_use(struct vnic_wq_copy *wq) argument 43 return wq->ring.desc_count - 1 - wq->ring.desc_avail; 46 static inline void *vnic_wq_copy_next_desc(struct vnic_wq_copy *wq) argument 48 struct fcpio_host_req *desc = wq->ring.descs; 49 return &desc[wq->to_use_index]; 52 static inline void vnic_wq_copy_post(struct vnic_wq_copy *wq) argument 55 ((wq 69 vnic_wq_copy_desc_process(struct vnic_wq_copy *wq, u16 index) argument 83 vnic_wq_copy_service(struct vnic_wq_copy *wq, u16 completed_index, void (*q_service)(struct vnic_wq_copy *wq, struct fcpio_host_req *wq_desc)) argument [all...] |
H A D | fnic_res.h | 30 static inline void fnic_queue_wq_desc(struct vnic_wq *wq, argument 37 struct wq_enet_desc *desc = vnic_wq_next_desc(wq); 51 vnic_wq_post(wq, os_buf, dma_addr, len, sop, eop); 54 static inline void fnic_queue_wq_eth_desc(struct vnic_wq *wq, argument 61 struct wq_enet_desc *desc = vnic_wq_next_desc(wq); 76 vnic_wq_post(wq, os_buf, dma_addr, len, 1, 1); 79 static inline void fnic_queue_wq_copy_desc_icmnd_16(struct vnic_wq_copy *wq, argument 91 struct fcpio_host_req *desc = vnic_wq_copy_next_desc(wq); 121 vnic_wq_copy_post(wq); 124 static inline void fnic_queue_wq_copy_desc_itmf(struct vnic_wq_copy *wq, argument 150 fnic_queue_wq_copy_desc_flogi_reg(struct vnic_wq_copy *wq, u32 req_id, u8 format, u32 s_id, u8 *gw_mac) argument 169 fnic_queue_wq_copy_desc_fip_reg(struct vnic_wq_copy *wq, u32 req_id, u32 s_id, u8 *fcf_mac, u8 *ha_mac, u32 r_a_tov, u32 e_d_tov) argument 193 fnic_queue_wq_copy_desc_fw_reset(struct vnic_wq_copy *wq, u32 req_id) argument 206 fnic_queue_wq_copy_desc_lunmap(struct vnic_wq_copy *wq, u32 req_id, u64 lunmap_addr, u32 lunmap_len) argument [all...] |
H A D | vnic_wq.h | 96 static inline unsigned int vnic_wq_desc_avail(struct vnic_wq *wq) argument 99 return wq->ring.desc_avail; 102 static inline unsigned int vnic_wq_desc_used(struct vnic_wq *wq) argument 105 return wq->ring.desc_count - wq->ring.desc_avail - 1; 108 static inline void *vnic_wq_next_desc(struct vnic_wq *wq) argument 110 return wq->to_use->desc; 113 static inline void vnic_wq_post(struct vnic_wq *wq, argument 117 struct vnic_wq_buf *buf = wq->to_use; 132 iowrite32(buf->index, &wq 139 vnic_wq_service(struct vnic_wq *wq, struct cq_desc *cq_desc, u16 completed_index, void (*buf_service)(struct vnic_wq *wq, struct cq_desc *cq_desc, struct vnic_wq_buf *buf, void *opaque), void *opaque) argument [all...] |
/drivers/net/ethernet/cisco/enic/ |
H A D | vnic_wq.c | 30 static int vnic_wq_alloc_bufs(struct vnic_wq *wq) argument 34 unsigned int i, j, count = wq->ring.desc_count; 37 vdev = wq->vdev; 40 wq->bufs[i] = kzalloc(VNIC_WQ_BUF_BLK_SZ(count), GFP_ATOMIC); 41 if (!wq->bufs[i]) 46 buf = wq->bufs[i]; 49 buf->desc = (u8 *)wq->ring.descs + 50 wq->ring.desc_size * buf->index; 52 buf->next = wq->bufs[0]; 55 buf->next = wq 68 vnic_wq_free(struct vnic_wq *wq) argument 87 vnic_wq_alloc(struct vnic_dev *vdev, struct vnic_wq *wq, unsigned int index, unsigned int desc_count, unsigned int desc_size) argument 116 vnic_wq_init_start(struct vnic_wq *wq, unsigned int cq_index, unsigned int fetch_index, unsigned int posted_index, unsigned int error_interrupt_enable, unsigned int error_interrupt_offset) argument 139 vnic_wq_init(struct vnic_wq *wq, unsigned int cq_index, unsigned int error_interrupt_enable, unsigned int error_interrupt_offset) argument 148 vnic_wq_error_status(struct vnic_wq *wq) argument 153 vnic_wq_enable(struct vnic_wq *wq) argument 158 vnic_wq_disable(struct vnic_wq *wq) argument 176 vnic_wq_clean(struct vnic_wq *wq, void (*buf_clean)(struct vnic_wq *wq, struct vnic_wq_buf *buf)) argument [all...] |
H A D | enic_res.h | 43 static inline void enic_queue_wq_desc_ex(struct vnic_wq *wq, argument 49 struct wq_enet_desc *desc = vnic_wq_next_desc(wq); 62 vnic_wq_post(wq, os_buf, dma_addr, len, sop, eop); 65 static inline void enic_queue_wq_desc_cont(struct vnic_wq *wq, argument 69 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, 74 static inline void enic_queue_wq_desc(struct vnic_wq *wq, void *os_buf, argument 78 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, 84 static inline void enic_queue_wq_desc_csum(struct vnic_wq *wq, argument 89 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, 96 static inline void enic_queue_wq_desc_csum_l4(struct vnic_wq *wq, argument 107 enic_queue_wq_desc_tso(struct vnic_wq *wq, void *os_buf, dma_addr_t dma_addr, unsigned int len, unsigned int mss, unsigned int hdr_len, int vlan_tag_insert, unsigned int vlan_tag, int eop, int loopback) argument [all...] |
H A D | vnic_wq.h | 86 static inline unsigned int vnic_wq_desc_avail(struct vnic_wq *wq) argument 89 return wq->ring.desc_avail; 92 static inline unsigned int vnic_wq_desc_used(struct vnic_wq *wq) argument 95 return wq->ring.desc_count - wq->ring.desc_avail - 1; 98 static inline void *vnic_wq_next_desc(struct vnic_wq *wq) argument 100 return wq->to_use->desc; 103 static inline void vnic_wq_post(struct vnic_wq *wq, argument 107 struct vnic_wq_buf *buf = wq->to_use; 122 iowrite32(buf->index, &wq 129 vnic_wq_service(struct vnic_wq *wq, struct cq_desc *cq_desc, u16 completed_index, void (*buf_service)(struct vnic_wq *wq, struct cq_desc *cq_desc, struct vnic_wq_buf *buf, void *opaque), void *opaque) argument [all...] |
/drivers/scsi/libsas/ |
H A D | sas_event.c | 59 struct workqueue_struct *wq = ha->core.shost->work_q; local 67 drain_workqueue(wq);
|
/drivers/infiniband/hw/ipath/ |
H A D | ipath_srq.c | 52 struct ipath_rwq *wq; local 68 wq = srq->rq.wq; 69 next = wq->head + 1; 72 if (next == wq->tail) { 79 wqe = get_rwqe_ptr(&srq->rq, wq->head); 86 wq->head = next; 139 srq->rq.wq = vmalloc_user(sizeof(struct ipath_rwq) + srq->rq.size * sz); 140 if (!srq->rq.wq) { 156 srq->rq.wq); 220 struct ipath_rwq *wq; local [all...] |
H A D | ipath_ruc.c | 170 struct ipath_rwq *wq; local 193 wq = rq->wq; 194 tail = wq->tail; 199 if (unlikely(tail == wq->head)) { 213 wq->tail = tail; 224 n = wq->head;
|
H A D | ipath_ud.c | 60 struct ipath_rwq *wq; local 123 wq = rq->wq; 124 tail = wq->tail; 128 if (unlikely(tail == wq->head)) { 148 wq->tail = tail; 157 n = wq->head;
|
/drivers/infiniband/hw/qib/ |
H A D | qib_srq.c | 52 struct qib_rwq *wq; local 68 wq = srq->rq.wq; 69 next = wq->head + 1; 72 if (next == wq->tail) { 79 wqe = get_rwqe_ptr(&srq->rq, wq->head); 86 wq->head = next; 136 srq->rq.wq = vmalloc_user(sizeof(struct qib_rwq) + srq->rq.size * sz); 137 if (!srq->rq.wq) { 152 srq->rq.wq); 216 struct qib_rwq *wq; local [all...] |
H A D | qib_ruc.c | 142 struct qib_rwq *wq; local 165 wq = rq->wq; 166 tail = wq->tail; 170 if (unlikely(tail == wq->head)) { 184 wq->tail = tail; 200 n = wq->head;
|
/drivers/net/ethernet/sfc/ |
H A D | mcdi.h | 37 * @wq: Wait queue for threads waiting for state != STATE_RUNNING 50 wait_queue_head_t wq; member in struct:efx_mcdi_iface
|
/drivers/staging/omapdrm/ |
H A D | omap_drv.h | 57 struct workqueue_struct *wq; member in struct:omap_drm_private
|
/drivers/media/video/ |
H A D | saa7110.c | 64 wait_queue_head_t wq; member in struct:saa7110 200 prepare_to_wait(&decoder->wq, &wait, TASK_UNINTERRUPTIBLE); 202 finish_wait(&decoder->wq, &wait); 235 prepare_to_wait(&decoder->wq, &wait, TASK_UNINTERRUPTIBLE); 237 finish_wait(&decoder->wq, &wait); 436 init_waitqueue_head(&decoder->wq);
|
/drivers/message/i2o/ |
H A D | exec-osm.c | 49 wait_queue_head_t *wq; /* Pointer to Wait queue */ member in struct:i2o_exec_wait 125 DECLARE_WAIT_QUEUE_HEAD_ONSTACK(wq); 152 wait->wq = &wq; 165 wait_event_interruptible_timeout(wq, wait->complete, timeout * HZ); 169 wait->wq = NULL; 240 if (wait->wq) 257 wake_up_interruptible(wait->wq);
|
/drivers/sbus/char/ |
H A D | bbc_i2c.h | 61 wait_queue_head_t wq; member in struct:bbc_i2c_bus
|
/drivers/staging/nvec/ |
H A D | nvec.h | 142 * @wq: The work queue in which @rx_work and @tx_work are executed 168 struct workqueue_struct *wq; member in struct:nvec_chip
|
/drivers/usb/misc/ |
H A D | appledisplay.c | 86 static struct workqueue_struct *wq; variable in typeref:struct:workqueue_struct 122 queue_delayed_work(wq, &pdata->work, 0); 360 wq = create_singlethread_workqueue("appledisplay"); 361 if (!wq) { 371 flush_workqueue(wq); 372 destroy_workqueue(wq);
|
/drivers/block/xen-blkback/ |
H A D | common.h | 185 wait_queue_head_t wq; member in struct:xen_blkif
|
/drivers/gpu/drm/via/ |
H A D | via_dmablit.h | 77 struct work_struct wq; member in struct:_drm_via_blitq
|
/drivers/i2c/busses/ |
H A D | i2c-ibm_iic.h | 48 wait_queue_head_t wq; member in struct:ibm_iic_private
|