Searched defs:wq (Results 26 - 50 of 99) sorted by relevance

1234

/drivers/infiniband/hw/cxgb3/
H A Diwch_qp.c150 u8 *flit_cnt, int *wr_cnt, struct t3_wq *wq)
174 wqe = (union t3_wr *)(wq->queue +
175 Q_PTR2IDX((wq->wptr+1), wq->size_log2));
177 Q_GENBIT(wq->wptr + 1, wq->size_log2),
280 qhp->wq.rq[Q_PTR2IDX(qhp->wq.rq_wptr,
281 qhp->wq.rq_size_log2)].wr_id = wr->wr_id;
282 qhp->wq
149 build_fastreg(union t3_wr *wqe, struct ib_send_wr *wr, u8 *flit_cnt, int *wr_cnt, struct t3_wq *wq) argument
[all...]
H A Dcxio_hal.c275 struct t3_wq *wq, struct cxio_ucontext *uctx)
277 int depth = 1UL << wq->size_log2;
278 int rqsize = 1UL << wq->rq_size_log2;
280 wq->qpid = get_qpid(rdev_p, uctx);
281 if (!wq->qpid)
284 wq->rq = kzalloc(depth * sizeof(struct t3_swrq), GFP_KERNEL);
285 if (!wq->rq)
288 wq->rq_addr = cxio_hal_rqtpool_alloc(rdev_p, rqsize);
289 if (!wq->rq_addr)
292 wq
274 cxio_create_qp(struct cxio_rdev *rdev_p, u32 kernel_domain, struct t3_wq *wq, struct cxio_ucontext *uctx) argument
336 cxio_destroy_qp(struct cxio_rdev *rdev_p, struct t3_wq *wq, struct cxio_ucontext *uctx) argument
350 insert_recv_cqe(struct t3_wq *wq, struct t3_cq *cq) argument
368 cxio_flush_rq(struct t3_wq *wq, struct t3_cq *cq, int count) argument
386 insert_sq_cqe(struct t3_wq *wq, struct t3_cq *cq, struct t3_swsq *sqp) argument
407 cxio_flush_sq(struct t3_wq *wq, struct t3_cq *cq, int count) argument
446 cqe_completes_wr(struct t3_cqe *cqe, struct t3_wq *wq) argument
464 cxio_count_scqes(struct t3_cq *cq, struct t3_wq *wq, int *count) argument
482 cxio_count_rcqes(struct t3_cq *cq, struct t3_wq *wq, int *count) argument
1075 flush_completed_wrs(struct t3_wq *wq, struct t3_cq *cq) argument
1104 create_read_req_cqe(struct t3_wq *wq, struct t3_cqe *hw_cqe, struct t3_cqe *read_cqe) argument
1118 advance_oldest_read(struct t3_wq *wq) argument
1149 cxio_poll_cq(struct t3_wq *wq, struct t3_cq *cq, struct t3_cqe *cqe, u8 *cqe_flushed, u64 *cookie, u32 *credit) argument
[all...]
H A Diwch_provider.h164 struct t3_wq wq; member in struct:iwch_qp
/drivers/infiniband/hw/cxgb4/
H A Dcq.c177 static void insert_recv_cqe(struct t4_wq *wq, struct t4_cq *cq) argument
181 PDBG("%s wq %p cq %p sw_cidx %u sw_pidx %u\n", __func__,
182 wq, cq, cq->sw_cidx, cq->sw_pidx);
188 V_CQE_QPID(wq->sq.qid));
194 int c4iw_flush_rq(struct t4_wq *wq, struct t4_cq *cq, int count) argument
197 int in_use = wq->rq.in_use - count;
200 PDBG("%s wq %p cq %p rq.in_use %u skip count %u\n", __func__,
201 wq, cq, wq->rq.in_use, count);
203 insert_recv_cqe(wq, c
209 insert_sq_cqe(struct t4_wq *wq, struct t4_cq *cq, struct t4_swsqe *swcqe) argument
228 c4iw_flush_sq(struct t4_wq *wq, struct t4_cq *cq, int count) argument
268 cqe_completes_wr(struct t4_cqe *cqe, struct t4_wq *wq) argument
284 c4iw_count_scqes(struct t4_cq *cq, struct t4_wq *wq, int *count) argument
303 c4iw_count_rcqes(struct t4_cq *cq, struct t4_wq *wq, int *count) argument
322 flush_completed_wrs(struct t4_wq *wq, struct t4_cq *cq) argument
353 create_read_req_cqe(struct t4_wq *wq, struct t4_cqe *hw_cqe, struct t4_cqe *read_cqe) argument
368 advance_oldest_read(struct t4_wq *wq) argument
402 poll_cq(struct t4_wq *wq, struct t4_cq *cq, struct t4_cqe *cqe, u8 *cqe_flushed, u64 *cookie, u32 *credit) argument
569 struct t4_wq *wq; local
[all...]
H A Dqp.c94 static int destroy_qp(struct c4iw_rdev *rdev, struct t4_wq *wq, argument
102 wq->rq.memsize, wq->rq.queue,
103 dma_unmap_addr(&wq->rq, mapping));
104 dealloc_sq(rdev, &wq->sq);
105 c4iw_rqtpool_free(rdev, wq->rq.rqt_hwaddr, wq->rq.rqt_size);
106 kfree(wq->rq.sw_rq);
107 kfree(wq->sq.sw_sq);
108 c4iw_put_qpid(rdev, wq
113 create_qp(struct c4iw_rdev *rdev, struct t4_wq *wq, struct t4_cq *rcq, struct t4_cq *scq, struct c4iw_dev_ucontext *uctx) argument
[all...]
/drivers/infiniband/hw/ipath/
H A Dipath_qp.c360 if (qp->r_rq.wq) {
361 qp->r_rq.wq->head = 0;
362 qp->r_rq.wq->tail = 0;
410 if (qp->r_rq.wq) {
411 struct ipath_rwq *wq; local
418 wq = qp->r_rq.wq;
419 head = wq->head;
422 tail = wq->tail;
431 wq
690 struct ipath_rwq *wq = qp->r_rq.wq; local
[all...]
/drivers/infiniband/hw/mlx4/
H A Dcq.c553 struct mlx4_ib_wq *wq; local
620 wq = &(*cur_qp)->sq;
623 wq->tail += (u16) (wqe_ctr - (u16) wq->tail);
625 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)];
626 ++wq->tail;
633 wq = &(*cur_qp)->rq;
634 wc->wr_id = wq
[all...]
H A Dmain.c73 static struct workqueue_struct *wq; variable in typeref:struct:workqueue_struct
1003 queue_work(wq, &work->work);
1257 flush_workqueue(wq);
1350 wq = create_singlethread_workqueue("mlx4_ib");
1351 if (!wq)
1356 destroy_workqueue(wq);
1366 destroy_workqueue(wq);
/drivers/infiniband/hw/qib/
H A Dqib_qp.c397 if (qp->r_rq.wq) {
398 qp->r_rq.wq->head = 0;
399 qp->r_rq.wq->tail = 0;
524 if (qp->r_rq.wq) {
525 struct qib_rwq *wq; local
532 wq = qp->r_rq.wq;
533 head = wq->head;
536 tail = wq->tail;
545 wq
919 struct qib_rwq *wq = qp->r_rq.wq; local
[all...]
/drivers/media/dvb/ddbridge/
H A Dddbridge.h85 wait_queue_head_t wq; member in struct:ddb_input
113 wait_queue_head_t wq; member in struct:ddb_output
132 wait_queue_head_t wq; member in struct:ddb_i2c
/drivers/media/video/
H A Dmsp3400-driver.h102 wait_queue_head_t wq; member in struct:msp_state
/drivers/mtd/ubi/
H A Dio.c344 wait_queue_head_t wq; local
355 init_waitqueue_head(&wq);
362 ei.priv = (unsigned long)&wq;
377 err = wait_event_interruptible(wq, ei.state == MTD_ERASE_DONE ||
/drivers/net/ethernet/cisco/enic/
H A Denic.h102 ____cacheline_aligned struct vnic_wq wq[ENIC_WQ_MAX]; member in struct:enic
/drivers/net/wireless/brcm80211/brcmfmac/
H A Dbcmsdh_sdmmc.c69 brcmf_pm_resume_wait(struct brcmf_sdio_dev *sdiodev, wait_queue_head_t *wq) argument
74 wait_event_timeout(*wq, false, HZ/100);
/drivers/staging/android/
H A Dlogger.c42 wait_queue_head_t wq; /* wait queue for readers */ member in struct:logger_log
266 prepare_to_wait(&log->wq, &wait, TASK_INTERRUPTIBLE);
286 finish_wait(&log->wq, &wait);
503 wake_up_interruptible(&log->wq);
596 poll_wait(file, &log->wq, wait);
723 .wq = __WAIT_QUEUE_HEAD_INITIALIZER(VAR .wq), \
/drivers/target/tcm_fc/
H A Dtfc_conf.c303 struct workqueue_struct *wq; local
325 wq = alloc_workqueue("tcm_fc", 0, 1);
326 if (!wq) {
334 destroy_workqueue(wq);
338 tpg->workqueue = wq;
/drivers/video/msm/
H A Dmdp.c153 static int mdp_wait(struct mdp_info *mdp, uint32_t mask, wait_queue_head_t *wq) argument
158 wait_event_timeout(*wq, !mdp_check_mask(mask), HZ);
/drivers/ata/
H A Dpata_octeon_cf.c41 struct workqueue_struct *wq; member in struct:octeon_cf_port
692 queue_delayed_work(cf_port->wq,
729 queue_delayed_work(cf_port->wq,
884 cf_port->wq = create_singlethread_workqueue(DRV_NAME);
885 if (!cf_port->wq)
/drivers/infiniband/core/
H A Dmad_priv.h211 struct workqueue_struct *wq; member in struct:ib_mad_port_private
/drivers/iommu/
H A Damd_iommu_v2.c55 wait_queue_head_t wq; /* To wait for count == 0 */ member in struct:pasid_state
68 wait_queue_head_t wq; member in struct:device_state
146 wake_up(&dev_state->wq);
153 prepare_to_wait(&dev_state->wq, &wait, TASK_UNINTERRUPTIBLE);
156 finish_wait(&dev_state->wq, &wait);
295 wake_up(&pasid_state->wq);
303 prepare_to_wait(&pasid_state->wq, &wait, TASK_UNINTERRUPTIBLE);
310 finish_wait(&pasid_state->wq, &wait);
683 init_waitqueue_head(&pasid_state->wq);
771 init_waitqueue_head(&dev_state->wq);
[all...]
/drivers/scsi/bfa/
H A Dbfad_im.c156 wait_queue_head_t *wq; local
160 wq = (wait_queue_head_t *) cmnd->SCp.ptr;
163 if (wq)
164 wake_up(wq);
293 DECLARE_WAIT_QUEUE_HEAD_ONSTACK(wq);
321 cmnd->SCp.ptr = (char *)&wq;
329 wait_event(wq, test_bit(IO_DONE_BIT,
356 DECLARE_WAIT_QUEUE_HEAD_ONSTACK(wq);
363 cmnd->SCp.ptr = (char *)&wq;
372 wait_event(wq, test_bi
[all...]
/drivers/scsi/fnic/
H A Dfnic_fcs.c471 struct vnic_wq *wq = &fnic->wq[0]; local
490 if (!vnic_wq_desc_avail(wq)) {
497 fnic_queue_wq_eth_desc(wq, skb, pa, skb->len,
507 struct vnic_wq *wq = &fnic->wq[0]; local
558 if (!vnic_wq_desc_avail(wq)) {
565 fnic_queue_wq_desc(wq, skb, pa, tot_len, fr_eof(fp),
664 static void fnic_wq_complete_frame_send(struct vnic_wq *wq, argument
670 struct fnic *fnic = vnic_dev_priv(wq
710 fnic_free_wq_buf(struct vnic_wq *wq, struct vnic_wq_buf *buf) argument
[all...]
H A Dfnic_scsi.c136 static int free_wq_copy_descs(struct fnic *fnic, struct vnic_wq_copy *wq) argument
146 if (wq->to_clean_index <= fnic->fw_ack_index[0])
147 wq->ring.desc_avail += (fnic->fw_ack_index[0]
148 - wq->to_clean_index + 1);
150 wq->ring.desc_avail += (wq->ring.desc_count
151 - wq->to_clean_index
159 wq->to_clean_index =
160 (fnic->fw_ack_index[0] + 1) % wq->ring.desc_count;
174 struct vnic_wq_copy *wq local
209 struct vnic_wq_copy *wq = &fnic->wq_copy[0]; local
259 fnic_queue_wq_copy_desc(struct fnic *fnic, struct vnic_wq_copy *wq, struct fnic_io_req *io_req, struct scsi_cmnd *sc, int sg_count) argument
358 struct vnic_wq_copy *wq; local
593 is_ack_index_in_range(struct vnic_wq_copy *wq, u16 request_out) argument
622 struct vnic_wq_copy *wq; local
992 fnic_wq_copy_cleanup_handler(struct vnic_wq_copy *wq, struct fcpio_host_req *desc) argument
1046 struct vnic_wq_copy *wq = &fnic->wq_copy[0]; local
1377 struct vnic_wq_copy *wq = &fnic->wq_copy[0]; local
[all...]
/drivers/infiniband/hw/mthca/
H A Dmthca_cq.c489 struct mthca_wq *wq; local
539 wq = &(*cur_qp)->sq;
541 >> wq->wqe_shift);
547 wq = NULL;
553 wq = &(*cur_qp)->rq;
555 wqe_index = wqe >> wq->wqe_shift;
562 wqe_index = wq->max - 1;
566 if (wq) {
567 if (wq->last_comp < wqe_index)
568 wq
[all...]
/drivers/media/video/au0828/
H A Dau0828.h179 wait_queue_head_t wq; member in struct:au0828_dmaqueue

Completed in 1034 milliseconds

1234