Searched refs:wq (Results 26 - 50 of 265) sorted by relevance

1234567891011

/drivers/i2c/busses/
H A Di2c-taos-evm.c38 static DECLARE_WAIT_QUEUE_HEAD(wq);
112 wait_event_interruptible_timeout(wq, taos->state == TAOS_STATE_IDLE,
163 wake_up_interruptible(&wq);
168 wake_up_interruptible(&wq);
175 wake_up_interruptible(&wq);
228 wait_event_interruptible_timeout(wq, taos->state == TAOS_STATE_IDLE,
250 wait_event_interruptible_timeout(wq, taos->state == TAOS_STATE_IDLE,
/drivers/infiniband/hw/cxgb3/
H A Dcxio_hal.h165 int cxio_create_qp(struct cxio_rdev *rdev, u32 kernel_domain, struct t3_wq *wq,
167 int cxio_destroy_qp(struct cxio_rdev *rdev, struct t3_wq *wq,
190 int cxio_flush_rq(struct t3_wq *wq, struct t3_cq *cq, int count);
191 int cxio_flush_sq(struct t3_wq *wq, struct t3_cq *cq, int count);
192 void cxio_count_rcqes(struct t3_cq *cq, struct t3_wq *wq, int *count);
193 void cxio_count_scqes(struct t3_cq *cq, struct t3_wq *wq, int *count);
195 int cxio_poll_cq(struct t3_wq *wq, struct t3_cq *cq, struct t3_cqe *cqe,
H A Diwch_ev.c66 qhp->attr.state, qhp->wq.qpid, CQE_STATUS(rsp_msg->cqe));
141 __func__, qhp->wq.qpid, qhp->ep);
145 qhp->wq.qpid);
222 CQE_STATUS(rsp_msg->cqe), qhp->wq.qpid);
/drivers/infiniband/hw/cxgb4/
H A Ddevice.c116 void c4iw_log_wr_stats(struct t4_wq *wq, struct t4_cqe *cqe) argument
121 if (!wq->rdev->wr_log)
124 idx = (atomic_inc_return(&wq->rdev->wr_log_idx) - 1) &
125 (wq->rdev->wr_log_size - 1);
126 le.poll_sge_ts = cxgb4_read_sge_timestamp(wq->rdev->lldi.ports[0]);
131 le.qid = wq->sq.qid;
133 le.post_host_ts = wq->sq.sw_sq[wq->sq.cidx].host_ts;
134 le.post_sge_ts = wq->sq.sw_sq[wq
[all...]
/drivers/input/
H A Dkeycombo.c30 struct workqueue_struct *wq; member in struct:keycombo_state
97 if (queue_delayed_work(state->wq, &state->key_down_work,
103 queue_work(state->wq, &state->key_up_work);
202 state->wq = alloc_ordered_workqueue("keycombo", 0);
203 if (!state->wq)
238 destroy_workqueue(state->wq);
/drivers/infiniband/hw/ipath/
H A Dipath_qp.c360 if (qp->r_rq.wq) {
361 qp->r_rq.wq->head = 0;
362 qp->r_rq.wq->tail = 0;
410 if (qp->r_rq.wq) {
411 struct ipath_rwq *wq; local
418 wq = qp->r_rq.wq;
419 head = wq->head;
422 tail = wq->tail;
431 wq
690 struct ipath_rwq *wq = qp->r_rq.wq; local
[all...]
/drivers/infiniband/hw/mlx5/
H A Dcq.c102 static enum ib_wc_opcode get_umr_comp(struct mlx5_ib_wq *wq, int idx) argument
104 switch (wq->wr_data[idx]) {
121 struct mlx5_ib_wq *wq, int idx)
160 wc->opcode = get_umr_comp(wq, idx);
175 struct mlx5_ib_wq *wq; local
197 wq = &qp->rq;
198 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)];
199 ++wq
120 handle_good_req(struct ib_wc *wc, struct mlx5_cqe64 *cqe, struct mlx5_ib_wq *wq, int idx) argument
410 struct mlx5_ib_wq *wq; local
[all...]
/drivers/thunderbolt/
H A Dtb.c214 * Executes on tb->wq.
296 queue_work(tb->wq, &ev->work);
332 if (tb->wq) {
333 flush_workqueue(tb->wq);
334 destroy_workqueue(tb->wq);
335 tb->wq = NULL;
366 tb->wq = alloc_ordered_workqueue("thunderbolt", 0);
367 if (!tb->wq)
/drivers/mtd/chips/
H A Dcfi_cmdset_0020.c159 init_waitqueue_head(&(cfi->chips[i].wq));
297 wake_up(&chip->wq);
352 add_wait_queue(&chip->wq, &wait);
355 remove_wait_queue(&chip->wq, &wait);
377 wake_up(&chip->wq);
485 add_wait_queue(&chip->wq, &wait);
488 remove_wait_queue(&chip->wq, &wait);
542 add_wait_queue(&chip->wq, &wait);
545 remove_wait_queue(&chip->wq, &wait);
595 wake_up(&chip->wq);
[all...]
/drivers/power/
H A Dipaq_micro_battery.c43 struct workqueue_struct *wq; member in struct:micro_battery
91 queue_delayed_work(mb->wq, &mb->update, msecs_to_jiffies(BATT_PERIOD));
235 mb->wq = create_singlethread_workqueue("ipaq-battery-wq");
238 queue_delayed_work(mb->wq, &mb->update, 1);
270 queue_delayed_work(mb->wq, &mb->update, msecs_to_jiffies(BATT_PERIOD));
/drivers/scsi/fnic/
H A Dfnic_scsi.c143 static int free_wq_copy_descs(struct fnic *fnic, struct vnic_wq_copy *wq) argument
153 if (wq->to_clean_index <= fnic->fw_ack_index[0])
154 wq->ring.desc_avail += (fnic->fw_ack_index[0]
155 - wq->to_clean_index + 1);
157 wq->ring.desc_avail += (wq->ring.desc_count
158 - wq->to_clean_index
166 wq->to_clean_index =
167 (fnic->fw_ack_index[0] + 1) % wq->ring.desc_count;
208 struct vnic_wq_copy *wq local
261 struct vnic_wq_copy *wq = &fnic->wq_copy[0]; local
317 fnic_queue_wq_copy_desc(struct fnic *fnic, struct vnic_wq_copy *wq, struct fnic_io_req *io_req, struct scsi_cmnd *sc, int sg_count) argument
427 struct vnic_wq_copy *wq; local
707 is_ack_index_in_range(struct vnic_wq_copy *wq, u16 request_out) argument
736 struct vnic_wq_copy *wq; local
1340 fnic_wq_copy_cleanup_handler(struct vnic_wq_copy *wq, struct fcpio_host_req *desc) argument
1405 struct vnic_wq_copy *wq = &fnic->wq_copy[0]; local
1913 struct vnic_wq_copy *wq = &fnic->wq_copy[0]; local
[all...]
/drivers/media/pci/ddbridge/
H A Dddbridge.h85 wait_queue_head_t wq; member in struct:ddb_input
113 wait_queue_head_t wq; member in struct:ddb_output
132 wait_queue_head_t wq; member in struct:ddb_i2c
/drivers/usb/misc/
H A Dappledisplay.c88 static struct workqueue_struct *wq; variable in typeref:struct:workqueue_struct
125 queue_delayed_work(wq, &pdata->work, 0);
368 wq = create_singlethread_workqueue("appledisplay");
369 if (!wq) {
379 flush_workqueue(wq);
380 destroy_workqueue(wq);
/drivers/message/i2o/
H A Dexec-osm.c49 wait_queue_head_t *wq; /* Pointer to Wait queue */ member in struct:i2o_exec_wait
125 DECLARE_WAIT_QUEUE_HEAD_ONSTACK(wq);
152 wait->wq = &wq;
165 wait_event_interruptible_timeout(wq, wait->complete, timeout * HZ);
169 wait->wq = NULL;
240 if (wait->wq)
257 wake_up_interruptible(wait->wq);
/drivers/infiniband/hw/qib/
H A Dqib_qp.c410 if (qp->r_rq.wq) {
411 qp->r_rq.wq->head = 0;
412 qp->r_rq.wq->tail = 0;
528 if (qp->r_rq.wq) {
529 struct qib_rwq *wq; local
536 wq = qp->r_rq.wq;
537 head = wq->head;
540 tail = wq->tail;
549 wq
923 struct qib_rwq *wq = qp->r_rq.wq; local
[all...]
/drivers/char/tpm/
H A Dtpm_ibmvtpm.h45 wait_queue_head_t wq; member in struct:ibmvtpm_dev
/drivers/gpu/drm/via/
H A Dvia_dmablit.h77 struct work_struct wq; member in struct:_drm_via_blitq
/drivers/gpu/host1x/
H A Dcdma.h53 struct delayed_work wq; /* work queue */ member in struct:buffer_timeout
78 struct buffer_timeout timeout; /* channel's timeout state/wq */
/drivers/sbus/char/
H A Dbbc_i2c.h61 wait_queue_head_t wq; member in struct:bbc_i2c_bus
/drivers/nfc/
H A Dnfcsim.c63 static struct workqueue_struct *wq; variable in typeref:struct:workqueue_struct
211 queue_delayed_work(wq, &dev->poll_work, 0);
326 queue_delayed_work(wq, &peer->recv_work, msecs_to_jiffies(5));
427 * Because the wq is ordered and only 1 work item is executed at a time,
431 queue_delayed_work(wq, &dev->poll_work, msecs_to_jiffies(200));
488 /* We need an ordered wq to ensure that poll_work items are executed
491 wq = alloc_ordered_workqueue("nfcsim", 0);
492 if (!wq) {
533 destroy_workqueue(wq);
/drivers/pci/hotplug/
H A Dpciehp_ctrl.c52 queue_work(p_slot->wq, &info->work);
361 queue_work(p_slot->wq, &info->work);
389 queue_delayed_work(p_slot->wq, &p_slot->work, 5*HZ);
450 queue_work(p_slot->wq, &info->work);
479 queue_work(p_slot->wq, &info->work);
492 queue_work(p_slot->wq, &info->work);
501 queue_work(p_slot->wq, &info->work);
/drivers/net/wireless/cw1200/
H A Dcw1200_spi.c43 wait_queue_head_t wq; member in struct:hwbus_priv
205 add_wait_queue(&self->wq, &wait);
218 remove_wait_queue(&self->wq, &wait);
230 wake_up(&self->wq);
413 init_waitqueue_head(&self->wq);
/drivers/infiniband/hw/mthca/
H A Dmthca_cq.c489 struct mthca_wq *wq; local
539 wq = &(*cur_qp)->sq;
541 >> wq->wqe_shift);
547 wq = NULL;
553 wq = &(*cur_qp)->rq;
555 wqe_index = wqe >> wq->wqe_shift;
562 wqe_index = wq->max - 1;
566 if (wq) {
567 if (wq->last_comp < wqe_index)
568 wq
[all...]
/drivers/net/ethernet/cisco/enic/
H A Denic_main.c138 static void enic_free_wq_buf(struct vnic_wq *wq, struct vnic_wq_buf *buf) argument
140 struct enic *enic = vnic_dev_priv(wq->vdev);
153 static void enic_wq_free_buf(struct vnic_wq *wq, argument
156 enic_free_wq_buf(wq, buf);
166 vnic_wq_service(&enic->wq[q_number], cq_desc,
171 vnic_wq_desc_avail(&enic->wq[q_number]) >=
186 error_status = vnic_wq_error_status(&enic->wq[i]);
357 struct vnic_wq *wq, struct sk_buff *skb,
365 enic_queue_wq_desc_cont(wq, skb,
376 struct vnic_wq *wq, struc
356 enic_queue_wq_skb_cont(struct enic *enic, struct vnic_wq *wq, struct sk_buff *skb, unsigned int len_left, int loopback) argument
375 enic_queue_wq_skb_vlan(struct enic *enic, struct vnic_wq *wq, struct sk_buff *skb, int vlan_tag_insert, unsigned int vlan_tag, int loopback) argument
399 enic_queue_wq_skb_csum_l4(struct enic *enic, struct vnic_wq *wq, struct sk_buff *skb, int vlan_tag_insert, unsigned int vlan_tag, int loopback) argument
427 enic_queue_wq_skb_tso(struct enic *enic, struct vnic_wq *wq, struct sk_buff *skb, unsigned int mss, int vlan_tag_insert, unsigned int vlan_tag, int loopback) argument
500 enic_queue_wq_skb(struct enic *enic, struct vnic_wq *wq, struct sk_buff *skb) argument
533 struct vnic_wq *wq; local
1296 struct vnic_wq *wq = &enic->wq[wq_index]; local
1450 int wq = enic_cq_wq(enic, i); local
[all...]
/drivers/staging/android/
H A Dsync.c169 init_waitqueue_head(&fence->wq);
187 wake_up_all(&fence->wq);
342 spin_lock_irqsave(&fence->wq.lock, flags);
345 __add_wait_queue_tail(&fence->wq, &waiter->work);
346 spin_unlock_irqrestore(&fence->wq.lock, flags);
361 spin_lock_irqsave(&fence->wq.lock, flags);
366 spin_unlock_irqrestore(&fence->wq.lock, flags);
384 ret = wait_event_interruptible_timeout(fence->wq,
549 poll_wait(file, &fence->wq, wait);

Completed in 734 milliseconds

1234567891011