/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_qp.c | 150 u8 *flit_cnt, int *wr_cnt, struct t3_wq *wq) 174 wqe = (union t3_wr *)(wq->queue + 175 Q_PTR2IDX((wq->wptr+1), wq->size_log2)); 177 Q_GENBIT(wq->wptr + 1, wq->size_log2), 280 qhp->wq.rq[Q_PTR2IDX(qhp->wq.rq_wptr, 281 qhp->wq.rq_size_log2)].wr_id = wr->wr_id; 282 qhp->wq 149 build_fastreg(union t3_wr *wqe, struct ib_send_wr *wr, u8 *flit_cnt, int *wr_cnt, struct t3_wq *wq) argument [all...] |
H A D | cxio_hal.c | 275 struct t3_wq *wq, struct cxio_ucontext *uctx) 277 int depth = 1UL << wq->size_log2; 278 int rqsize = 1UL << wq->rq_size_log2; 280 wq->qpid = get_qpid(rdev_p, uctx); 281 if (!wq->qpid) 284 wq->rq = kzalloc(depth * sizeof(struct t3_swrq), GFP_KERNEL); 285 if (!wq->rq) 288 wq->rq_addr = cxio_hal_rqtpool_alloc(rdev_p, rqsize); 289 if (!wq->rq_addr) 292 wq 274 cxio_create_qp(struct cxio_rdev *rdev_p, u32 kernel_domain, struct t3_wq *wq, struct cxio_ucontext *uctx) argument 336 cxio_destroy_qp(struct cxio_rdev *rdev_p, struct t3_wq *wq, struct cxio_ucontext *uctx) argument 350 insert_recv_cqe(struct t3_wq *wq, struct t3_cq *cq) argument 368 cxio_flush_rq(struct t3_wq *wq, struct t3_cq *cq, int count) argument 386 insert_sq_cqe(struct t3_wq *wq, struct t3_cq *cq, struct t3_swsq *sqp) argument 407 cxio_flush_sq(struct t3_wq *wq, struct t3_cq *cq, int count) argument 446 cqe_completes_wr(struct t3_cqe *cqe, struct t3_wq *wq) argument 464 cxio_count_scqes(struct t3_cq *cq, struct t3_wq *wq, int *count) argument 482 cxio_count_rcqes(struct t3_cq *cq, struct t3_wq *wq, int *count) argument 1075 flush_completed_wrs(struct t3_wq *wq, struct t3_cq *cq) argument 1104 create_read_req_cqe(struct t3_wq *wq, struct t3_cqe *hw_cqe, struct t3_cqe *read_cqe) argument 1118 advance_oldest_read(struct t3_wq *wq) argument 1149 cxio_poll_cq(struct t3_wq *wq, struct t3_cq *cq, struct t3_cqe *cqe, u8 *cqe_flushed, u64 *cookie, u32 *credit) argument [all...] |
H A D | iwch_provider.h | 164 struct t3_wq wq; member in struct:iwch_qp
|
/drivers/infiniband/hw/cxgb4/ |
H A D | cq.c | 177 static void insert_recv_cqe(struct t4_wq *wq, struct t4_cq *cq) argument 181 PDBG("%s wq %p cq %p sw_cidx %u sw_pidx %u\n", __func__, 182 wq, cq, cq->sw_cidx, cq->sw_pidx); 188 V_CQE_QPID(wq->sq.qid)); 194 int c4iw_flush_rq(struct t4_wq *wq, struct t4_cq *cq, int count) argument 197 int in_use = wq->rq.in_use - count; 200 PDBG("%s wq %p cq %p rq.in_use %u skip count %u\n", __func__, 201 wq, cq, wq->rq.in_use, count); 203 insert_recv_cqe(wq, c 209 insert_sq_cqe(struct t4_wq *wq, struct t4_cq *cq, struct t4_swsqe *swcqe) argument 228 c4iw_flush_sq(struct t4_wq *wq, struct t4_cq *cq, int count) argument 268 cqe_completes_wr(struct t4_cqe *cqe, struct t4_wq *wq) argument 284 c4iw_count_scqes(struct t4_cq *cq, struct t4_wq *wq, int *count) argument 303 c4iw_count_rcqes(struct t4_cq *cq, struct t4_wq *wq, int *count) argument 322 flush_completed_wrs(struct t4_wq *wq, struct t4_cq *cq) argument 353 create_read_req_cqe(struct t4_wq *wq, struct t4_cqe *hw_cqe, struct t4_cqe *read_cqe) argument 368 advance_oldest_read(struct t4_wq *wq) argument 402 poll_cq(struct t4_wq *wq, struct t4_cq *cq, struct t4_cqe *cqe, u8 *cqe_flushed, u64 *cookie, u32 *credit) argument 569 struct t4_wq *wq; local [all...] |
H A D | qp.c | 94 static int destroy_qp(struct c4iw_rdev *rdev, struct t4_wq *wq, argument 102 wq->rq.memsize, wq->rq.queue, 103 dma_unmap_addr(&wq->rq, mapping)); 104 dealloc_sq(rdev, &wq->sq); 105 c4iw_rqtpool_free(rdev, wq->rq.rqt_hwaddr, wq->rq.rqt_size); 106 kfree(wq->rq.sw_rq); 107 kfree(wq->sq.sw_sq); 108 c4iw_put_qpid(rdev, wq 113 create_qp(struct c4iw_rdev *rdev, struct t4_wq *wq, struct t4_cq *rcq, struct t4_cq *scq, struct c4iw_dev_ucontext *uctx) argument [all...] |
/drivers/infiniband/hw/ipath/ |
H A D | ipath_qp.c | 360 if (qp->r_rq.wq) { 361 qp->r_rq.wq->head = 0; 362 qp->r_rq.wq->tail = 0; 410 if (qp->r_rq.wq) { 411 struct ipath_rwq *wq; local 418 wq = qp->r_rq.wq; 419 head = wq->head; 422 tail = wq->tail; 431 wq 690 struct ipath_rwq *wq = qp->r_rq.wq; local [all...] |
/drivers/infiniband/hw/mlx4/ |
H A D | cq.c | 553 struct mlx4_ib_wq *wq; local 620 wq = &(*cur_qp)->sq; 623 wq->tail += (u16) (wqe_ctr - (u16) wq->tail); 625 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; 626 ++wq->tail; 633 wq = &(*cur_qp)->rq; 634 wc->wr_id = wq [all...] |
H A D | main.c | 73 static struct workqueue_struct *wq; variable in typeref:struct:workqueue_struct 1003 queue_work(wq, &work->work); 1257 flush_workqueue(wq); 1350 wq = create_singlethread_workqueue("mlx4_ib"); 1351 if (!wq) 1356 destroy_workqueue(wq); 1366 destroy_workqueue(wq);
|
/drivers/infiniband/hw/qib/ |
H A D | qib_qp.c | 397 if (qp->r_rq.wq) { 398 qp->r_rq.wq->head = 0; 399 qp->r_rq.wq->tail = 0; 524 if (qp->r_rq.wq) { 525 struct qib_rwq *wq; local 532 wq = qp->r_rq.wq; 533 head = wq->head; 536 tail = wq->tail; 545 wq 919 struct qib_rwq *wq = qp->r_rq.wq; local [all...] |
/drivers/media/dvb/ddbridge/ |
H A D | ddbridge.h | 85 wait_queue_head_t wq; member in struct:ddb_input 113 wait_queue_head_t wq; member in struct:ddb_output 132 wait_queue_head_t wq; member in struct:ddb_i2c
|
/drivers/media/video/ |
H A D | msp3400-driver.h | 102 wait_queue_head_t wq; member in struct:msp_state
|
/drivers/mtd/ubi/ |
H A D | io.c | 344 wait_queue_head_t wq; local 355 init_waitqueue_head(&wq); 362 ei.priv = (unsigned long)&wq; 377 err = wait_event_interruptible(wq, ei.state == MTD_ERASE_DONE ||
|
/drivers/net/ethernet/cisco/enic/ |
H A D | enic.h | 102 ____cacheline_aligned struct vnic_wq wq[ENIC_WQ_MAX]; member in struct:enic
|
/drivers/net/wireless/brcm80211/brcmfmac/ |
H A D | bcmsdh_sdmmc.c | 69 brcmf_pm_resume_wait(struct brcmf_sdio_dev *sdiodev, wait_queue_head_t *wq) argument 74 wait_event_timeout(*wq, false, HZ/100);
|
/drivers/staging/android/ |
H A D | logger.c | 42 wait_queue_head_t wq; /* wait queue for readers */ member in struct:logger_log 266 prepare_to_wait(&log->wq, &wait, TASK_INTERRUPTIBLE); 286 finish_wait(&log->wq, &wait); 503 wake_up_interruptible(&log->wq); 596 poll_wait(file, &log->wq, wait); 723 .wq = __WAIT_QUEUE_HEAD_INITIALIZER(VAR .wq), \
|
/drivers/target/tcm_fc/ |
H A D | tfc_conf.c | 303 struct workqueue_struct *wq; local 325 wq = alloc_workqueue("tcm_fc", 0, 1); 326 if (!wq) { 334 destroy_workqueue(wq); 338 tpg->workqueue = wq;
|
/drivers/video/msm/ |
H A D | mdp.c | 153 static int mdp_wait(struct mdp_info *mdp, uint32_t mask, wait_queue_head_t *wq) argument 158 wait_event_timeout(*wq, !mdp_check_mask(mask), HZ);
|
/drivers/ata/ |
H A D | pata_octeon_cf.c | 41 struct workqueue_struct *wq; member in struct:octeon_cf_port 692 queue_delayed_work(cf_port->wq, 729 queue_delayed_work(cf_port->wq, 884 cf_port->wq = create_singlethread_workqueue(DRV_NAME); 885 if (!cf_port->wq)
|
/drivers/infiniband/core/ |
H A D | mad_priv.h | 211 struct workqueue_struct *wq; member in struct:ib_mad_port_private
|
/drivers/iommu/ |
H A D | amd_iommu_v2.c | 55 wait_queue_head_t wq; /* To wait for count == 0 */ member in struct:pasid_state 68 wait_queue_head_t wq; member in struct:device_state 146 wake_up(&dev_state->wq); 153 prepare_to_wait(&dev_state->wq, &wait, TASK_UNINTERRUPTIBLE); 156 finish_wait(&dev_state->wq, &wait); 295 wake_up(&pasid_state->wq); 303 prepare_to_wait(&pasid_state->wq, &wait, TASK_UNINTERRUPTIBLE); 310 finish_wait(&pasid_state->wq, &wait); 683 init_waitqueue_head(&pasid_state->wq); 771 init_waitqueue_head(&dev_state->wq); [all...] |
/drivers/scsi/bfa/ |
H A D | bfad_im.c | 156 wait_queue_head_t *wq; local 160 wq = (wait_queue_head_t *) cmnd->SCp.ptr; 163 if (wq) 164 wake_up(wq); 293 DECLARE_WAIT_QUEUE_HEAD_ONSTACK(wq); 321 cmnd->SCp.ptr = (char *)&wq; 329 wait_event(wq, test_bit(IO_DONE_BIT, 356 DECLARE_WAIT_QUEUE_HEAD_ONSTACK(wq); 363 cmnd->SCp.ptr = (char *)&wq; 372 wait_event(wq, test_bi [all...] |
/drivers/scsi/fnic/ |
H A D | fnic_fcs.c | 471 struct vnic_wq *wq = &fnic->wq[0]; local 490 if (!vnic_wq_desc_avail(wq)) { 497 fnic_queue_wq_eth_desc(wq, skb, pa, skb->len, 507 struct vnic_wq *wq = &fnic->wq[0]; local 558 if (!vnic_wq_desc_avail(wq)) { 565 fnic_queue_wq_desc(wq, skb, pa, tot_len, fr_eof(fp), 664 static void fnic_wq_complete_frame_send(struct vnic_wq *wq, argument 670 struct fnic *fnic = vnic_dev_priv(wq 710 fnic_free_wq_buf(struct vnic_wq *wq, struct vnic_wq_buf *buf) argument [all...] |
H A D | fnic_scsi.c | 136 static int free_wq_copy_descs(struct fnic *fnic, struct vnic_wq_copy *wq) argument 146 if (wq->to_clean_index <= fnic->fw_ack_index[0]) 147 wq->ring.desc_avail += (fnic->fw_ack_index[0] 148 - wq->to_clean_index + 1); 150 wq->ring.desc_avail += (wq->ring.desc_count 151 - wq->to_clean_index 159 wq->to_clean_index = 160 (fnic->fw_ack_index[0] + 1) % wq->ring.desc_count; 174 struct vnic_wq_copy *wq local 209 struct vnic_wq_copy *wq = &fnic->wq_copy[0]; local 259 fnic_queue_wq_copy_desc(struct fnic *fnic, struct vnic_wq_copy *wq, struct fnic_io_req *io_req, struct scsi_cmnd *sc, int sg_count) argument 358 struct vnic_wq_copy *wq; local 593 is_ack_index_in_range(struct vnic_wq_copy *wq, u16 request_out) argument 622 struct vnic_wq_copy *wq; local 992 fnic_wq_copy_cleanup_handler(struct vnic_wq_copy *wq, struct fcpio_host_req *desc) argument 1046 struct vnic_wq_copy *wq = &fnic->wq_copy[0]; local 1377 struct vnic_wq_copy *wq = &fnic->wq_copy[0]; local [all...] |
/drivers/infiniband/hw/mthca/ |
H A D | mthca_cq.c | 489 struct mthca_wq *wq; local 539 wq = &(*cur_qp)->sq; 541 >> wq->wqe_shift); 547 wq = NULL; 553 wq = &(*cur_qp)->rq; 555 wqe_index = wqe >> wq->wqe_shift; 562 wqe_index = wq->max - 1; 566 if (wq) { 567 if (wq->last_comp < wqe_index) 568 wq [all...] |
/drivers/media/video/au0828/ |
H A D | au0828.h | 179 wait_queue_head_t wq; member in struct:au0828_dmaqueue
|