/drivers/message/i2o/ |
H A D | exec-osm.c | 49 wait_queue_head_t *wq; /* Pointer to Wait queue */ member in struct:i2o_exec_wait 125 DECLARE_WAIT_QUEUE_HEAD_ONSTACK(wq); 152 wait->wq = &wq; 165 wait_event_interruptible_timeout(wq, wait->complete, timeout * HZ); 169 wait->wq = NULL; 240 if (wait->wq) 257 wake_up_interruptible(wait->wq);
|
/drivers/net/ethernet/cisco/enic/ |
H A D | enic_res.h | 43 static inline void enic_queue_wq_desc_ex(struct vnic_wq *wq, argument 49 struct wq_enet_desc *desc = vnic_wq_next_desc(wq); 65 vnic_wq_post(wq, os_buf, dma_addr, len, sop, eop, desc_skip_cnt, 69 static inline void enic_queue_wq_desc_cont(struct vnic_wq *wq, argument 73 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, 78 static inline void enic_queue_wq_desc(struct vnic_wq *wq, void *os_buf, argument 82 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, 88 static inline void enic_queue_wq_desc_csum(struct vnic_wq *wq, argument 93 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, 100 static inline void enic_queue_wq_desc_csum_l4(struct vnic_wq *wq, argument 111 enic_queue_wq_desc_tso(struct vnic_wq *wq, void *os_buf, dma_addr_t dma_addr, unsigned int len, unsigned int mss, unsigned int hdr_len, int vlan_tag_insert, unsigned int vlan_tag, int eop, int loopback) argument [all...] |
H A D | vnic_wq.h | 90 static inline unsigned int vnic_wq_desc_avail(struct vnic_wq *wq) argument 93 return wq->ring.desc_avail; 96 static inline unsigned int vnic_wq_desc_used(struct vnic_wq *wq) argument 99 return wq->ring.desc_count - wq->ring.desc_avail - 1; 102 static inline void *vnic_wq_next_desc(struct vnic_wq *wq) argument 104 return wq->to_use->desc; 107 static inline void vnic_wq_post(struct vnic_wq *wq, argument 113 struct vnic_wq_buf *buf = wq->to_use; 132 iowrite32(buf->index, &wq 139 vnic_wq_service(struct vnic_wq *wq, struct cq_desc *cq_desc, u16 completed_index, void (*buf_service)(struct vnic_wq *wq, struct cq_desc *cq_desc, struct vnic_wq_buf *buf, void *opaque), void *opaque) argument [all...] |
/drivers/net/wireless/cw1200/ |
H A D | cw1200_spi.c | 43 wait_queue_head_t wq; member in struct:hwbus_priv 205 add_wait_queue(&self->wq, &wait); 218 remove_wait_queue(&self->wq, &wait); 230 wake_up(&self->wq); 413 init_waitqueue_head(&self->wq);
|
/drivers/nfc/ |
H A D | nfcsim.c | 63 static struct workqueue_struct *wq; variable in typeref:struct:workqueue_struct 211 queue_delayed_work(wq, &dev->poll_work, 0); 326 queue_delayed_work(wq, &peer->recv_work, msecs_to_jiffies(5)); 427 * Because the wq is ordered and only 1 work item is executed at a time, 431 queue_delayed_work(wq, &dev->poll_work, msecs_to_jiffies(200)); 488 /* We need an ordered wq to ensure that poll_work items are executed 491 wq = alloc_ordered_workqueue("nfcsim", 0); 492 if (!wq) { 533 destroy_workqueue(wq);
|
/drivers/pci/hotplug/ |
H A D | pciehp.h | 80 struct workqueue_struct *wq; member in struct:slot
|
/drivers/power/ |
H A D | ipaq_micro_battery.c | 43 struct workqueue_struct *wq; member in struct:micro_battery 91 queue_delayed_work(mb->wq, &mb->update, msecs_to_jiffies(BATT_PERIOD)); 235 mb->wq = create_singlethread_workqueue("ipaq-battery-wq"); 238 queue_delayed_work(mb->wq, &mb->update, 1); 270 queue_delayed_work(mb->wq, &mb->update, msecs_to_jiffies(BATT_PERIOD));
|
/drivers/sbus/char/ |
H A D | bbc_i2c.h | 61 wait_queue_head_t wq; member in struct:bbc_i2c_bus
|
/drivers/scsi/fnic/ |
H A D | fnic_res.h | 30 static inline void fnic_queue_wq_desc(struct vnic_wq *wq, argument 37 struct wq_enet_desc *desc = vnic_wq_next_desc(wq); 51 vnic_wq_post(wq, os_buf, dma_addr, len, sop, eop); 54 static inline void fnic_queue_wq_eth_desc(struct vnic_wq *wq, argument 61 struct wq_enet_desc *desc = vnic_wq_next_desc(wq); 76 vnic_wq_post(wq, os_buf, dma_addr, len, 1, 1); 79 static inline void fnic_queue_wq_copy_desc_icmnd_16(struct vnic_wq_copy *wq, argument 91 struct fcpio_host_req *desc = vnic_wq_copy_next_desc(wq); 121 vnic_wq_copy_post(wq); 124 static inline void fnic_queue_wq_copy_desc_itmf(struct vnic_wq_copy *wq, argument 150 fnic_queue_wq_copy_desc_flogi_reg(struct vnic_wq_copy *wq, u32 req_id, u8 format, u32 s_id, u8 *gw_mac) argument 169 fnic_queue_wq_copy_desc_fip_reg(struct vnic_wq_copy *wq, u32 req_id, u32 s_id, u8 *fcf_mac, u8 *ha_mac, u32 r_a_tov, u32 e_d_tov) argument 193 fnic_queue_wq_copy_desc_fw_reset(struct vnic_wq_copy *wq, u32 req_id) argument 206 fnic_queue_wq_copy_desc_lunmap(struct vnic_wq_copy *wq, u32 req_id, u64 lunmap_addr, u32 lunmap_len) argument [all...] |
H A D | vnic_wq.h | 96 static inline unsigned int vnic_wq_desc_avail(struct vnic_wq *wq) argument 99 return wq->ring.desc_avail; 102 static inline unsigned int vnic_wq_desc_used(struct vnic_wq *wq) argument 105 return wq->ring.desc_count - wq->ring.desc_avail - 1; 108 static inline void *vnic_wq_next_desc(struct vnic_wq *wq) argument 110 return wq->to_use->desc; 113 static inline void vnic_wq_post(struct vnic_wq *wq, argument 117 struct vnic_wq_buf *buf = wq->to_use; 132 iowrite32(buf->index, &wq 139 vnic_wq_service(struct vnic_wq *wq, struct cq_desc *cq_desc, u16 completed_index, void (*buf_service)(struct vnic_wq *wq, struct cq_desc *cq_desc, struct vnic_wq_buf *buf, void *opaque), void *opaque) argument [all...] |
/drivers/staging/android/ |
H A D | sync.h | 151 * @wq: wait queue for fence signaling 163 wait_queue_head_t wq; member in struct:sync_fence
|
/drivers/staging/lustre/lustre/ptlrpc/ |
H A D | niobuf.c | 246 wait_queue_head_t *wq; local 279 wq = &req->rq_set->set_waitq; 281 wq = &req->rq_reply_waitq; 288 rc = l_wait_event(*wq, !ptlrpc_client_bulk_active(req), &lwi);
|
/drivers/staging/nvec/ |
H A D | nvec.h | 122 * @wq: The work queue in which @rx_work and @tx_work are executed 149 struct workqueue_struct *wq; member in struct:nvec_chip
|
/drivers/usb/misc/ |
H A D | appledisplay.c | 88 static struct workqueue_struct *wq; variable in typeref:struct:workqueue_struct 125 queue_delayed_work(wq, &pdata->work, 0); 368 wq = create_singlethread_workqueue("appledisplay"); 369 if (!wq) { 379 flush_workqueue(wq); 380 destroy_workqueue(wq);
|
/drivers/gpu/drm/via/ |
H A D | via_dmablit.h | 77 struct work_struct wq; member in struct:_drm_via_blitq
|
/drivers/i2c/busses/ |
H A D | i2c-ibm_iic.h | 48 wait_queue_head_t wq; member in struct:ibm_iic_private
|
/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_qp.c | 150 u8 *flit_cnt, int *wr_cnt, struct t3_wq *wq) 174 wqe = (union t3_wr *)(wq->queue + 175 Q_PTR2IDX((wq->wptr+1), wq->size_log2)); 177 Q_GENBIT(wq->wptr + 1, wq->size_log2), 280 qhp->wq.rq[Q_PTR2IDX(qhp->wq.rq_wptr, 281 qhp->wq.rq_size_log2)].wr_id = wr->wr_id; 282 qhp->wq 149 build_fastreg(union t3_wr *wqe, struct ib_send_wr *wr, u8 *flit_cnt, int *wr_cnt, struct t3_wq *wq) argument [all...] |
/drivers/infiniband/hw/cxgb4/ |
H A D | cq.c | 178 static void insert_recv_cqe(struct t4_wq *wq, struct t4_cq *cq) argument 182 PDBG("%s wq %p cq %p sw_cidx %u sw_pidx %u\n", __func__, 183 wq, cq, cq->sw_cidx, cq->sw_pidx); 189 V_CQE_QPID(wq->sq.qid)); 195 int c4iw_flush_rq(struct t4_wq *wq, struct t4_cq *cq, int count) argument 198 int in_use = wq->rq.in_use - count; 201 PDBG("%s wq %p cq %p rq.in_use %u skip count %u\n", __func__, 202 wq, cq, wq->rq.in_use, count); 204 insert_recv_cqe(wq, c 210 insert_sq_cqe(struct t4_wq *wq, struct t4_cq *cq, struct t4_swsqe *swcqe) argument 234 struct t4_wq *wq = &qhp->wq; local 263 flush_completed_wrs(struct t4_wq *wq, struct t4_cq *cq) argument 299 create_read_req_cqe(struct t4_wq *wq, struct t4_cqe *hw_cqe, struct t4_cqe *read_cqe) argument 311 advance_oldest_read(struct t4_wq *wq) argument 412 cqe_completes_wr(struct t4_cqe *cqe, struct t4_wq *wq) argument 428 c4iw_count_rcqes(struct t4_cq *cq, struct t4_wq *wq, int *count) argument 463 poll_cq(struct t4_wq *wq, struct t4_cq *cq, struct t4_cqe *cqe, u8 *cqe_flushed, u64 *cookie, u32 *credit) argument 681 struct t4_wq *wq; local [all...] |
/drivers/infiniband/hw/ipath/ |
H A D | ipath_qp.c | 360 if (qp->r_rq.wq) { 361 qp->r_rq.wq->head = 0; 362 qp->r_rq.wq->tail = 0; 410 if (qp->r_rq.wq) { 411 struct ipath_rwq *wq; local 418 wq = qp->r_rq.wq; 419 head = wq->head; 422 tail = wq->tail; 431 wq 690 struct ipath_rwq *wq = qp->r_rq.wq; local [all...] |
/drivers/infiniband/hw/mlx4/ |
H A D | cq.c | 600 struct mlx4_ib_wq *wq; local 687 wq = &(*cur_qp)->sq; 690 wq->tail += (u16) (wqe_ctr - (u16) wq->tail); 692 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; 693 ++wq->tail; 705 wq = &(*cur_qp)->rq; 706 tail = wq [all...] |
/drivers/media/i2c/ |
H A D | msp3400-driver.h | 102 wait_queue_head_t wq; member in struct:msp_state
|
/drivers/media/pci/ddbridge/ |
H A D | ddbridge.h | 85 wait_queue_head_t wq; member in struct:ddb_input 113 wait_queue_head_t wq; member in struct:ddb_output 132 wait_queue_head_t wq; member in struct:ddb_i2c
|
/drivers/mtd/ubi/ |
H A D | block.c | 86 struct workqueue_struct *wq; member in struct:ubiblock 302 queue_work(dev->wq, &dev->work); 436 dev->wq = alloc_workqueue("%s", 0, 0, gd->disk_name); 437 if (!dev->wq) { 495 destroy_workqueue(dev->wq); 625 destroy_workqueue(dev->wq);
|
H A D | io.c | 335 wait_queue_head_t wq; local 346 init_waitqueue_head(&wq); 353 ei.priv = (unsigned long)&wq; 368 err = wait_event_interruptible(wq, ei.state == MTD_ERASE_DONE ||
|
/drivers/net/ethernet/sfc/ |
H A D | mcdi.h | 46 * @state: Request handling state. Waited for by @wq. 48 * @wq: Wait queue for threads waiting for @state != %MCDI_STATE_RUNNING 66 wait_queue_head_t wq; member in struct:efx_mcdi_iface
|