/drivers/i2c/busses/ |
H A D | i2c-taos-evm.c | 38 static DECLARE_WAIT_QUEUE_HEAD(wq); 112 wait_event_interruptible_timeout(wq, taos->state == TAOS_STATE_IDLE, 163 wake_up_interruptible(&wq); 168 wake_up_interruptible(&wq); 175 wake_up_interruptible(&wq); 228 wait_event_interruptible_timeout(wq, taos->state == TAOS_STATE_IDLE, 250 wait_event_interruptible_timeout(wq, taos->state == TAOS_STATE_IDLE,
|
/drivers/infiniband/hw/cxgb3/ |
H A D | cxio_hal.h | 165 int cxio_create_qp(struct cxio_rdev *rdev, u32 kernel_domain, struct t3_wq *wq, 167 int cxio_destroy_qp(struct cxio_rdev *rdev, struct t3_wq *wq, 190 int cxio_flush_rq(struct t3_wq *wq, struct t3_cq *cq, int count); 191 int cxio_flush_sq(struct t3_wq *wq, struct t3_cq *cq, int count); 192 void cxio_count_rcqes(struct t3_cq *cq, struct t3_wq *wq, int *count); 193 void cxio_count_scqes(struct t3_cq *cq, struct t3_wq *wq, int *count); 195 int cxio_poll_cq(struct t3_wq *wq, struct t3_cq *cq, struct t3_cqe *cqe,
|
H A D | iwch_ev.c | 66 qhp->attr.state, qhp->wq.qpid, CQE_STATUS(rsp_msg->cqe)); 141 __func__, qhp->wq.qpid, qhp->ep); 145 qhp->wq.qpid); 222 CQE_STATUS(rsp_msg->cqe), qhp->wq.qpid);
|
/drivers/infiniband/hw/cxgb4/ |
H A D | device.c | 116 void c4iw_log_wr_stats(struct t4_wq *wq, struct t4_cqe *cqe) argument 121 if (!wq->rdev->wr_log) 124 idx = (atomic_inc_return(&wq->rdev->wr_log_idx) - 1) & 125 (wq->rdev->wr_log_size - 1); 126 le.poll_sge_ts = cxgb4_read_sge_timestamp(wq->rdev->lldi.ports[0]); 131 le.qid = wq->sq.qid; 133 le.post_host_ts = wq->sq.sw_sq[wq->sq.cidx].host_ts; 134 le.post_sge_ts = wq->sq.sw_sq[wq [all...] |
/drivers/input/ |
H A D | keycombo.c | 30 struct workqueue_struct *wq; member in struct:keycombo_state 97 if (queue_delayed_work(state->wq, &state->key_down_work, 103 queue_work(state->wq, &state->key_up_work); 202 state->wq = alloc_ordered_workqueue("keycombo", 0); 203 if (!state->wq) 238 destroy_workqueue(state->wq);
|
/drivers/infiniband/hw/ipath/ |
H A D | ipath_qp.c | 360 if (qp->r_rq.wq) { 361 qp->r_rq.wq->head = 0; 362 qp->r_rq.wq->tail = 0; 410 if (qp->r_rq.wq) { 411 struct ipath_rwq *wq; local 418 wq = qp->r_rq.wq; 419 head = wq->head; 422 tail = wq->tail; 431 wq 690 struct ipath_rwq *wq = qp->r_rq.wq; local [all...] |
/drivers/infiniband/hw/mlx5/ |
H A D | cq.c | 102 static enum ib_wc_opcode get_umr_comp(struct mlx5_ib_wq *wq, int idx) argument 104 switch (wq->wr_data[idx]) { 121 struct mlx5_ib_wq *wq, int idx) 160 wc->opcode = get_umr_comp(wq, idx); 175 struct mlx5_ib_wq *wq; local 197 wq = &qp->rq; 198 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; 199 ++wq 120 handle_good_req(struct ib_wc *wc, struct mlx5_cqe64 *cqe, struct mlx5_ib_wq *wq, int idx) argument 410 struct mlx5_ib_wq *wq; local [all...] |
/drivers/thunderbolt/ |
H A D | tb.c | 214 * Executes on tb->wq. 296 queue_work(tb->wq, &ev->work); 332 if (tb->wq) { 333 flush_workqueue(tb->wq); 334 destroy_workqueue(tb->wq); 335 tb->wq = NULL; 366 tb->wq = alloc_ordered_workqueue("thunderbolt", 0); 367 if (!tb->wq)
|
/drivers/mtd/chips/ |
H A D | cfi_cmdset_0020.c | 159 init_waitqueue_head(&(cfi->chips[i].wq)); 297 wake_up(&chip->wq); 352 add_wait_queue(&chip->wq, &wait); 355 remove_wait_queue(&chip->wq, &wait); 377 wake_up(&chip->wq); 485 add_wait_queue(&chip->wq, &wait); 488 remove_wait_queue(&chip->wq, &wait); 542 add_wait_queue(&chip->wq, &wait); 545 remove_wait_queue(&chip->wq, &wait); 595 wake_up(&chip->wq); [all...] |
/drivers/power/ |
H A D | ipaq_micro_battery.c | 43 struct workqueue_struct *wq; member in struct:micro_battery 91 queue_delayed_work(mb->wq, &mb->update, msecs_to_jiffies(BATT_PERIOD)); 235 mb->wq = create_singlethread_workqueue("ipaq-battery-wq"); 238 queue_delayed_work(mb->wq, &mb->update, 1); 270 queue_delayed_work(mb->wq, &mb->update, msecs_to_jiffies(BATT_PERIOD));
|
/drivers/scsi/fnic/ |
H A D | fnic_scsi.c | 143 static int free_wq_copy_descs(struct fnic *fnic, struct vnic_wq_copy *wq) argument 153 if (wq->to_clean_index <= fnic->fw_ack_index[0]) 154 wq->ring.desc_avail += (fnic->fw_ack_index[0] 155 - wq->to_clean_index + 1); 157 wq->ring.desc_avail += (wq->ring.desc_count 158 - wq->to_clean_index 166 wq->to_clean_index = 167 (fnic->fw_ack_index[0] + 1) % wq->ring.desc_count; 208 struct vnic_wq_copy *wq local 261 struct vnic_wq_copy *wq = &fnic->wq_copy[0]; local 317 fnic_queue_wq_copy_desc(struct fnic *fnic, struct vnic_wq_copy *wq, struct fnic_io_req *io_req, struct scsi_cmnd *sc, int sg_count) argument 427 struct vnic_wq_copy *wq; local 707 is_ack_index_in_range(struct vnic_wq_copy *wq, u16 request_out) argument 736 struct vnic_wq_copy *wq; local 1340 fnic_wq_copy_cleanup_handler(struct vnic_wq_copy *wq, struct fcpio_host_req *desc) argument 1405 struct vnic_wq_copy *wq = &fnic->wq_copy[0]; local 1913 struct vnic_wq_copy *wq = &fnic->wq_copy[0]; local [all...] |
/drivers/media/pci/ddbridge/ |
H A D | ddbridge.h | 85 wait_queue_head_t wq; member in struct:ddb_input 113 wait_queue_head_t wq; member in struct:ddb_output 132 wait_queue_head_t wq; member in struct:ddb_i2c
|
/drivers/usb/misc/ |
H A D | appledisplay.c | 88 static struct workqueue_struct *wq; variable in typeref:struct:workqueue_struct 125 queue_delayed_work(wq, &pdata->work, 0); 368 wq = create_singlethread_workqueue("appledisplay"); 369 if (!wq) { 379 flush_workqueue(wq); 380 destroy_workqueue(wq);
|
/drivers/message/i2o/ |
H A D | exec-osm.c | 49 wait_queue_head_t *wq; /* Pointer to Wait queue */ member in struct:i2o_exec_wait 125 DECLARE_WAIT_QUEUE_HEAD_ONSTACK(wq); 152 wait->wq = &wq; 165 wait_event_interruptible_timeout(wq, wait->complete, timeout * HZ); 169 wait->wq = NULL; 240 if (wait->wq) 257 wake_up_interruptible(wait->wq);
|
/drivers/infiniband/hw/qib/ |
H A D | qib_qp.c | 410 if (qp->r_rq.wq) { 411 qp->r_rq.wq->head = 0; 412 qp->r_rq.wq->tail = 0; 528 if (qp->r_rq.wq) { 529 struct qib_rwq *wq; local 536 wq = qp->r_rq.wq; 537 head = wq->head; 540 tail = wq->tail; 549 wq 923 struct qib_rwq *wq = qp->r_rq.wq; local [all...] |
/drivers/char/tpm/ |
H A D | tpm_ibmvtpm.h | 45 wait_queue_head_t wq; member in struct:ibmvtpm_dev
|
/drivers/gpu/drm/via/ |
H A D | via_dmablit.h | 77 struct work_struct wq; member in struct:_drm_via_blitq
|
/drivers/gpu/host1x/ |
H A D | cdma.h | 53 struct delayed_work wq; /* work queue */ member in struct:buffer_timeout 78 struct buffer_timeout timeout; /* channel's timeout state/wq */
|
/drivers/sbus/char/ |
H A D | bbc_i2c.h | 61 wait_queue_head_t wq; member in struct:bbc_i2c_bus
|
/drivers/nfc/ |
H A D | nfcsim.c | 63 static struct workqueue_struct *wq; variable in typeref:struct:workqueue_struct 211 queue_delayed_work(wq, &dev->poll_work, 0); 326 queue_delayed_work(wq, &peer->recv_work, msecs_to_jiffies(5)); 427 * Because the wq is ordered and only 1 work item is executed at a time, 431 queue_delayed_work(wq, &dev->poll_work, msecs_to_jiffies(200)); 488 /* We need an ordered wq to ensure that poll_work items are executed 491 wq = alloc_ordered_workqueue("nfcsim", 0); 492 if (!wq) { 533 destroy_workqueue(wq);
|
/drivers/pci/hotplug/ |
H A D | pciehp_ctrl.c | 52 queue_work(p_slot->wq, &info->work); 361 queue_work(p_slot->wq, &info->work); 389 queue_delayed_work(p_slot->wq, &p_slot->work, 5*HZ); 450 queue_work(p_slot->wq, &info->work); 479 queue_work(p_slot->wq, &info->work); 492 queue_work(p_slot->wq, &info->work); 501 queue_work(p_slot->wq, &info->work);
|
/drivers/net/wireless/cw1200/ |
H A D | cw1200_spi.c | 43 wait_queue_head_t wq; member in struct:hwbus_priv 205 add_wait_queue(&self->wq, &wait); 218 remove_wait_queue(&self->wq, &wait); 230 wake_up(&self->wq); 413 init_waitqueue_head(&self->wq);
|
/drivers/infiniband/hw/mthca/ |
H A D | mthca_cq.c | 489 struct mthca_wq *wq; local 539 wq = &(*cur_qp)->sq; 541 >> wq->wqe_shift); 547 wq = NULL; 553 wq = &(*cur_qp)->rq; 555 wqe_index = wqe >> wq->wqe_shift; 562 wqe_index = wq->max - 1; 566 if (wq) { 567 if (wq->last_comp < wqe_index) 568 wq [all...] |
/drivers/net/ethernet/cisco/enic/ |
H A D | enic_main.c | 138 static void enic_free_wq_buf(struct vnic_wq *wq, struct vnic_wq_buf *buf) argument 140 struct enic *enic = vnic_dev_priv(wq->vdev); 153 static void enic_wq_free_buf(struct vnic_wq *wq, argument 156 enic_free_wq_buf(wq, buf); 166 vnic_wq_service(&enic->wq[q_number], cq_desc, 171 vnic_wq_desc_avail(&enic->wq[q_number]) >= 186 error_status = vnic_wq_error_status(&enic->wq[i]); 357 struct vnic_wq *wq, struct sk_buff *skb, 365 enic_queue_wq_desc_cont(wq, skb, 376 struct vnic_wq *wq, struc 356 enic_queue_wq_skb_cont(struct enic *enic, struct vnic_wq *wq, struct sk_buff *skb, unsigned int len_left, int loopback) argument 375 enic_queue_wq_skb_vlan(struct enic *enic, struct vnic_wq *wq, struct sk_buff *skb, int vlan_tag_insert, unsigned int vlan_tag, int loopback) argument 399 enic_queue_wq_skb_csum_l4(struct enic *enic, struct vnic_wq *wq, struct sk_buff *skb, int vlan_tag_insert, unsigned int vlan_tag, int loopback) argument 427 enic_queue_wq_skb_tso(struct enic *enic, struct vnic_wq *wq, struct sk_buff *skb, unsigned int mss, int vlan_tag_insert, unsigned int vlan_tag, int loopback) argument 500 enic_queue_wq_skb(struct enic *enic, struct vnic_wq *wq, struct sk_buff *skb) argument 533 struct vnic_wq *wq; local 1296 struct vnic_wq *wq = &enic->wq[wq_index]; local 1450 int wq = enic_cq_wq(enic, i); local [all...] |
/drivers/staging/android/ |
H A D | sync.c | 169 init_waitqueue_head(&fence->wq); 187 wake_up_all(&fence->wq); 342 spin_lock_irqsave(&fence->wq.lock, flags); 345 __add_wait_queue_tail(&fence->wq, &waiter->work); 346 spin_unlock_irqrestore(&fence->wq.lock, flags); 361 spin_lock_irqsave(&fence->wq.lock, flags); 366 spin_unlock_irqrestore(&fence->wq.lock, flags); 384 ret = wait_event_interruptible_timeout(fence->wq, 549 poll_wait(file, &fence->wq, wait);
|