/drivers/block/paride/ |
H A D | pd.c | 721 struct request *rq; local 724 rq = blk_get_request(disk->gd->queue, READ, __GFP_WAIT); 725 if (IS_ERR(rq)) 726 return PTR_ERR(rq); 728 rq->cmd_type = REQ_TYPE_SPECIAL; 729 rq->special = func; 731 err = blk_execute_rq(disk->gd->queue, disk->gd, rq, 0); 733 blk_put_request(rq);
|
/drivers/s390/char/ |
H A D | raw3270.h | 121 raw3270_request_final(struct raw3270_request *rq) argument 123 return list_empty(&rq->list);
|
H A D | tty3270.c | 325 tty3270_write_callback(struct raw3270_request *rq, void *data) argument 327 struct tty3270 *tp = container_of(rq->view, struct tty3270, view); 329 if (rq->rc != 0) { 334 raw3270_request_reset(rq); 335 xchg(&tp->write, rq); 588 tty3270_read_callback(struct raw3270_request *rq, void *data) argument 590 struct tty3270 *tp = container_of(rq->view, struct tty3270, view); 591 raw3270_get_view(rq->view); 646 tty3270_irq(struct tty3270 *tp, struct raw3270_request *rq, struct irb *irb) argument 656 if (rq) { [all...] |
/drivers/isdn/hardware/mISDN/ |
H A D | speedfax.c | 254 struct channel_req *rq; local 260 rq = arg; 261 if (rq->protocol == ISDN_P_TE_S0) 262 err = sf->isac.open(&sf->isac, rq); 264 err = sf->isar.open(&sf->isar, rq);
|
H A D | mISDNipac.c | 757 open_dchannel(struct isac_hw *isac, struct channel_req *rq) argument 761 if (rq->protocol != ISDN_P_TE_S0) 763 if (rq->adr.channel == 1) 766 rq->ch = &isac->dch.dev.D; 767 rq->ch->protocol = rq->protocol; 769 _queue_data(rq->ch, PH_ACTIVATE_IND, MISDN_ID_ANY, 1491 open_bchannel(struct ipac_hw *ipac, struct channel_req *rq) argument 1495 if (rq->adr.channel == 0 || rq 1543 struct channel_req *rq; local [all...] |
/drivers/scsi/fnic/ |
H A D | fnic_fcs.c | 785 static void fnic_rq_cmpl_frame_recv(struct vnic_rq *rq, struct cq_desc argument 790 struct fnic *fnic = vnic_dev_priv(rq->vdev); 897 vnic_rq_service(&fnic->rq[q_number], cq_desc, completed_index, 914 err = vnic_rq_fill(&fnic->rq[i], fnic_alloc_rq_frame); 931 int fnic_alloc_rq_frame(struct vnic_rq *rq) argument 933 struct fnic *fnic = vnic_dev_priv(rq->vdev); 950 fnic_queue_rq_desc(rq, skb, pa, len); 954 void fnic_free_rq_buf(struct vnic_rq *rq, struct vnic_rq_buf *buf) argument 957 struct fnic *fnic = vnic_dev_priv(rq->vdev);
|
H A D | fnic_res.c | 148 "wq/wq_copy/rq %d/%d/%d\n", 221 vnic_rq_free(&fnic->rq[i]); 250 "wq %d cp_wq %d raw_wq %d rq %d cq %d intr %d\n", 275 err = vnic_rq_alloc(fnic->vdev, &fnic->rq[i], i, 357 vnic_rq_init(&fnic->rq[i],
|
H A D | fnic_main.c | 344 error_status = ioread32(&fnic->rq[i].ctrl->error_status); 470 err = vnic_rq_disable(&fnic->rq[i]); 494 vnic_rq_clean(&fnic->rq[i], fnic_free_rq_buf); 806 err = vnic_rq_fill(&fnic->rq[i], fnic_alloc_rq_frame); 881 vnic_rq_enable(&fnic->rq[i]); 910 vnic_rq_clean(&fnic->rq[i], fnic_free_rq_buf);
|
/drivers/atm/ |
H A D | firestream.c | 679 long rq; local 683 while (!((rq = read_fs (dev, Q_RP(q->offset))) & Q_EMPTY)) { 684 fs_dprintk (FS_DEBUG_QUEUE, "reaping return queue entry at %lx\n", rq); 685 qe = bus_to_virt (rq); 705 long rq; local 711 while (!((rq = read_fs (dev, Q_RP(q->offset))) & Q_EMPTY)) { 712 fs_dprintk (FS_DEBUG_QUEUE, "reaping txdone entry at %lx\n", rq); 713 qe = bus_to_virt (rq); 775 long rq; local 782 while (!((rq [all...] |
/drivers/ide/ |
H A D | ide-dma.c | 108 blk_rq_sectors(cmd->rq) << 9); 492 if (hwif->rq) 493 hwif->rq->errors = 0;
|
/drivers/block/ |
H A D | pktcdvd.c | 702 struct request *rq; local 705 rq = blk_get_request(q, (cgc->data_direction == CGC_DATA_WRITE) ? 707 if (IS_ERR(rq)) 708 return PTR_ERR(rq); 709 blk_rq_set_block_pc(rq); 712 ret = blk_rq_map_kern(q, rq, cgc->buffer, cgc->buflen, 718 rq->cmd_len = COMMAND_SIZE(cgc->cmd[0]); 719 memcpy(rq->cmd, cgc->cmd, CDROM_PACKET_SIZE); 721 rq->timeout = 60*HZ; 723 rq [all...] |
H A D | rbd.c | 298 struct request *rq; /* block request */ member in union:rbd_img_request::__anon358 336 struct gendisk *disk; /* blkdev's gendisk and rq */ 343 struct list_head rq_queue; /* incoming rq queue */ 2146 img_request->rq = NULL; 2283 rbd_assert(img_request->rq != NULL); 2284 more = blk_end_request(img_request->rq, result, xferred); 3309 static void rbd_handle_request(struct rbd_device *rbd_dev, struct request *rq) argument 3313 u64 offset = (u64)blk_rq_pos(rq) << SECTOR_SHIFT; 3314 u64 length = blk_rq_bytes(rq); 3319 if (rq 3417 struct request *rq, *next; local 3437 struct request *rq; local [all...] |
/drivers/staging/lustre/lustre/ptlrpc/ |
H A D | service.c | 1173 struct ptlrpc_request *rq = NULL; local 1192 list_for_each_entry_reverse(rq, 1195 if (req->rq_deadline >= rq->rq_deadline) { 1197 &rq->rq_timed_list); 1388 struct ptlrpc_request *rq, *n; local 1426 list_for_each_entry_safe(rq, n, 1429 if (rq->rq_deadline > now + at_early_margin) { 1432 rq->rq_deadline < deadline) 1433 deadline = rq->rq_deadline; 1437 ptlrpc_at_remove_timed(rq); [all...] |
/drivers/block/aoe/ |
H A D | aoeblk.c | 281 struct request *rq; local 287 while ((rq = blk_peek_request(q))) { 288 blk_start_request(rq); 289 aoe_end_request(d, rq, 1);
|
/drivers/gpu/drm/nouveau/core/subdev/i2c/ |
H A D | base.c | 369 u32 hi, lo, rq, tx, e; local 372 impl->aux_stat(i2c, &hi, &lo, &rq, &tx); 373 if (hi || lo || rq || tx) { 380 if (rq & (1 << port->aux)) e |= NVKM_I2C_IRQ;
|
/drivers/infiniband/hw/mthca/ |
H A D | mthca_provider.h | 275 struct mthca_wq rq; member in struct:mthca_qp
|
/drivers/net/ethernet/stmicro/stmmac/ |
H A D | stmmac_ptp.c | 154 struct ptp_clock_request *rq, int on) 153 stmmac_enable(struct ptp_clock_info *ptp, struct ptp_clock_request *rq, int on) argument
|
/drivers/net/irda/ |
H A D | mcs7780.h | 153 static int mcs_net_ioctl(struct net_device *netdev, struct ifreq *rq, int cmd);
|
/drivers/scsi/osd/ |
H A D | osd_initiator.c | 442 static void _put_request(struct request *rq) argument 450 if (unlikely(rq->bio)) 451 blk_end_request(rq, -ENOMEM, blk_rq_bytes(rq)); 453 blk_put_request(rq); 458 struct request *rq = or->request; local 460 if (rq) { 461 if (rq->next_rq) { 462 _put_request(rq->next_rq); 463 rq [all...] |
/drivers/net/usb/ |
H A D | ax88172a.c | 54 static int ax88172a_ioctl(struct net_device *net, struct ifreq *rq, int cmd) argument 62 return phy_mii_ioctl(net->phydev, rq, cmd);
|
/drivers/net/vmxnet3/ |
H A D | vmxnet3_int.h | 378 #define VMXNET3_RX_ALLOC_THRESHOLD(rq, ring_idx, adapter) \ 379 ((rq)->rx_ring[ring_idx].size >> 3)
|
/drivers/infiniband/hw/cxgb4/ |
H A D | cq.c | 198 int in_use = wq->rq.in_use - count; 201 PDBG("%s wq %p cq %p rq.in_use %u skip count %u\n", __func__, 202 wq, cq, wq->rq.in_use, count); 577 if (unlikely((CQE_WRID_MSN(hw_cqe) != (wq->rq.msn)))) { 640 PDBG("%s completing rq idx %u\n", __func__, wq->rq.cidx); 641 *cookie = wq->rq.sw_rq[wq->rq.cidx].wr_id;
|
/drivers/scsi/bnx2fc/ |
H A D | bnx2fc_tgt.c | 709 tgt->rq = dma_alloc_coherent(&hba->pcidev->dev, tgt->rq_mem_size, 711 if (!tgt->rq) { 716 memset(tgt->rq, 0, tgt->rq_mem_size); 889 if (tgt->rq) { 891 tgt->rq, tgt->rq_dma); 892 tgt->rq = NULL;
|
/drivers/cdrom/ |
H A D | cdrom.c | 2163 struct request *rq; local 2182 rq = blk_get_request(q, READ, GFP_KERNEL); 2183 if (IS_ERR(rq)) { 2184 ret = PTR_ERR(rq); 2187 blk_rq_set_block_pc(rq); 2189 ret = blk_rq_map_user(q, rq, NULL, ubuf, len, GFP_KERNEL); 2191 blk_put_request(rq); 2195 rq->cmd[0] = GPCMD_READ_CD; 2196 rq->cmd[1] = 1 << 2; 2197 rq [all...] |
/drivers/net/ppp/ |
H A D | ppp_generic.c | 83 struct sk_buff_head rq; /* receive queue for pppd */ member in struct:ppp_file 228 /* We limit the length of ppp->file.rq to this (arbitrary) value */ 428 skb = skb_dequeue(&pf->rq); 522 if (skb_peek(&pf->rq)) 1291 if (ppp->file.rq.qlen > PPP_MAX_RQLEN) 1293 skb_queue_tail(&ppp->file.rq, skb); 1672 skb_queue_tail(&pch->file.rq, skb); 1674 while (pch->file.rq.qlen > PPP_MAX_RQLEN && 1675 (skb = skb_dequeue(&pch->file.rq))) 1821 skb_queue_tail(&ppp->file.rq, sk [all...] |