/drivers/staging/octeon/ |
H A D | ethernet-mdio.c | 100 * @rq: the request 105 int cvm_oct_ioctl(struct net_device *dev, struct ifreq *rq, int cmd) argument 115 return phy_mii_ioctl(priv->phydev, rq, cmd);
|
/drivers/ide/ |
H A D | ide-lib.c | 94 struct request *rq = drive->hwif->rq; local 98 if (rq) 100 (unsigned long long)blk_rq_pos(rq));
|
H A D | ide-devsets.c | 162 struct request *rq; local 168 rq = blk_get_request(q, READ, __GFP_WAIT); 169 rq->cmd_type = REQ_TYPE_SPECIAL; 170 rq->cmd_len = 5; 171 rq->cmd[0] = REQ_DEVSET_EXEC; 172 *(int *)&rq->cmd[1] = arg; 173 rq->special = setting->set; 175 if (blk_execute_rq(q, NULL, rq, 0)) 176 ret = rq->errors; 177 blk_put_request(rq); 182 ide_do_devset(ide_drive_t *drive, struct request *rq) argument [all...] |
H A D | ide-ioctls.c | 126 struct request *rq; local 128 rq = blk_get_request(drive->queue, READ, __GFP_WAIT); 129 rq->cmd_type = REQ_TYPE_ATA_TASKFILE; 130 err = blk_execute_rq(drive->queue, NULL, rq, 0); 131 blk_put_request(rq); 221 struct request *rq; local 224 rq = blk_get_request(drive->queue, READ, __GFP_WAIT); 225 rq->cmd_type = REQ_TYPE_SPECIAL; 226 rq->cmd_len = 1; 227 rq [all...] |
H A D | ide-park.c | 13 struct request *rq; local 34 rq = blk_get_request(q, READ, __GFP_WAIT); 35 rq->cmd[0] = REQ_PARK_HEADS; 36 rq->cmd_len = 1; 37 rq->cmd_type = REQ_TYPE_SPECIAL; 38 rq->special = &timeout; 39 rc = blk_execute_rq(q, NULL, rq, 1); 40 blk_put_request(rq); 48 rq = blk_get_request(q, READ, GFP_NOWAIT); 49 if (unlikely(!rq)) 61 ide_do_park_unpark(ide_drive_t *drive, struct request *rq) argument [all...] |
H A D | ide-pm.c | 10 struct request *rq; local 21 rq = blk_get_request(drive->queue, READ, __GFP_WAIT); 22 rq->cmd_type = REQ_TYPE_PM_SUSPEND; 23 rq->special = &rqpm; 29 ret = blk_execute_rq(drive->queue, NULL, rq, 0); 30 blk_put_request(rq); 46 struct request *rq; local 61 rq = blk_get_request(drive->queue, READ, __GFP_WAIT); 62 rq->cmd_type = REQ_TYPE_PM_RESUME; 63 rq 81 ide_complete_power_step(ide_drive_t *drive, struct request *rq) argument 111 ide_start_power_step(ide_drive_t *drive, struct request *rq) argument 182 ide_complete_pm_rq(ide_drive_t *drive, struct request *rq) argument 209 ide_check_pm_state(ide_drive_t *drive, struct request *rq) argument [all...] |
H A D | ide-eh.c | 7 static ide_startstop_t ide_ata_error(ide_drive_t *drive, struct request *rq, argument 15 rq->errors |= ERROR_RESET; 28 rq->errors = ERROR_MAX; 31 rq->errors |= ERROR_RECAL; 35 if ((stat & ATA_DRQ) && rq_data_dir(rq) == READ && 42 if (rq->errors >= ERROR_MAX || blk_noretry_request(rq)) { 43 ide_kill_rq(drive, rq); 48 rq->errors |= ERROR_RESET; 50 if ((rq 63 ide_atapi_error(ide_drive_t *drive, struct request *rq, u8 stat, u8 err) argument 93 __ide_error(ide_drive_t *drive, struct request *rq, u8 stat, u8 err) argument 116 struct request *rq; local 148 struct request *rq = drive->hwif->rq; local [all...] |
H A D | ide-cd_ioctl.c | 303 struct request *rq; local 306 rq = blk_get_request(drive->queue, READ, __GFP_WAIT); 307 rq->cmd_type = REQ_TYPE_SPECIAL; 308 rq->cmd_flags = REQ_QUIET; 309 ret = blk_execute_rq(drive->queue, cd->disk, rq, 0); 310 blk_put_request(rq);
|
H A D | ide-floppy.c | 66 struct request *rq = pc->rq; local 75 rq->cmd_type == REQ_TYPE_BLOCK_PC) 79 u8 *buf = bio_data(rq->bio); 100 if (rq->cmd_type == REQ_TYPE_SPECIAL) 101 rq->errors = uptodate ? 0 : IDE_DRV_ERROR_GENERAL; 136 unsigned int done = blk_rq_bytes(drive->hwif->rq); 191 struct ide_atapi_pc *pc, struct request *rq, 196 int blocks = blk_rq_sectors(rq) / floppy->bs_factor; 197 int cmd = rq_data_dir(rq); 190 idefloppy_create_rw_cmd(ide_drive_t *drive, struct ide_atapi_pc *pc, struct request *rq, unsigned long sector) argument 215 idefloppy_blockpc_cmd(struct ide_disk_obj *floppy, struct ide_atapi_pc *pc, struct request *rq) argument 228 ide_floppy_do_request(ide_drive_t *drive, struct request *rq, sector_t block) argument [all...] |
H A D | ide-gd.c | 160 struct request *rq, sector_t sector) 162 return drive->disk_ops->do_request(drive, rq, sector); 159 ide_gd_do_request(ide_drive_t *drive, struct request *rq, sector_t sector) argument
|
H A D | ide-taskfile.c | 187 struct request *rq = hwif->rq; local 189 if (blk_pm_request(rq)) 190 ide_complete_pm_rq(drive, rq); 290 cmd->rq->errors = 0; 327 struct request *rq = drive->hwif->rq; local 332 rq->errors = err; 339 ide_complete_rq(drive, err ? -EIO : 0, blk_rq_bytes(rq)); 397 ide_complete_rq(drive, 0, blk_rq_sectors(cmd->rq) << 430 struct request *rq; local [all...] |
H A D | pdc202xx_old.c | 152 struct request *rq = hwif->rq; local 159 word_count = (blk_rq_sectors(rq) << 8); 160 word_count = (rq_data_dir(rq) == READ) ?
|
/drivers/net/ethernet/cisco/enic/ |
H A D | vnic_rq.c | 30 static int vnic_rq_alloc_bufs(struct vnic_rq *rq) argument 34 unsigned int i, j, count = rq->ring.desc_count; 37 vdev = rq->vdev; 40 rq->bufs[i] = kzalloc(VNIC_RQ_BUF_BLK_SZ(count), GFP_ATOMIC); 41 if (!rq->bufs[i]) { 48 buf = rq->bufs[i]; 51 buf->desc = (u8 *)rq->ring.descs + 52 rq->ring.desc_size * buf->index; 54 buf->next = rq->bufs[0]; 57 buf->next = rq 70 vnic_rq_free(struct vnic_rq *rq) argument 89 vnic_rq_alloc(struct vnic_dev *vdev, struct vnic_rq *rq, unsigned int index, unsigned int desc_count, unsigned int desc_size) argument 118 vnic_rq_init_start(struct vnic_rq *rq, unsigned int cq_index, unsigned int fetch_index, unsigned int posted_index, unsigned int error_interrupt_enable, unsigned int error_interrupt_offset) argument 142 vnic_rq_init(struct vnic_rq *rq, unsigned int cq_index, unsigned int error_interrupt_enable, unsigned int error_interrupt_offset) argument 162 vnic_rq_error_status(struct vnic_rq *rq) argument 167 vnic_rq_enable(struct vnic_rq *rq) argument 172 vnic_rq_disable(struct vnic_rq *rq) argument 190 vnic_rq_clean(struct vnic_rq *rq, void (*buf_clean)(struct vnic_rq *rq, struct vnic_rq_buf *buf)) argument [all...] |
H A D | enic_res.h | 118 static inline void enic_queue_rq_desc(struct vnic_rq *rq, argument 122 struct rq_enet_desc *desc = vnic_rq_next_desc(rq); 130 vnic_rq_post(rq, os_buf, os_buf_index, dma_addr, len);
|
/drivers/scsi/fnic/ |
H A D | vnic_rq.c | 27 static int vnic_rq_alloc_bufs(struct vnic_rq *rq) argument 31 unsigned int i, j, count = rq->ring.desc_count; 34 vdev = rq->vdev; 37 rq->bufs[i] = kzalloc(VNIC_RQ_BUF_BLK_SZ, GFP_ATOMIC); 38 if (!rq->bufs[i]) { 45 buf = rq->bufs[i]; 48 buf->desc = (u8 *)rq->ring.descs + 49 rq->ring.desc_size * buf->index; 51 buf->next = rq->bufs[0]; 54 buf->next = rq 68 vnic_rq_free(struct vnic_rq *rq) argument 85 vnic_rq_alloc(struct vnic_dev *vdev, struct vnic_rq *rq, unsigned int index, unsigned int desc_count, unsigned int desc_size) argument 114 vnic_rq_init(struct vnic_rq *rq, unsigned int cq_index, unsigned int error_interrupt_enable, unsigned int error_interrupt_offset) argument 140 vnic_rq_error_status(struct vnic_rq *rq) argument 145 vnic_rq_enable(struct vnic_rq *rq) argument 150 vnic_rq_disable(struct vnic_rq *rq) argument 168 vnic_rq_clean(struct vnic_rq *rq, void (*buf_clean)(struct vnic_rq *rq, struct vnic_rq_buf *buf)) argument [all...] |
/drivers/input/misc/ |
H A D | xen-kbdfront.c | 52 static irqreturn_t input_handler(int rq, void *dev_id) argument
|
/drivers/net/ |
H A D | ifb.c | 46 struct sk_buff_head rq; member in struct:ifb_private 73 skb_queue_splice_tail_init(&dp->rq, &dp->tq); 115 if ((skb = skb_peek(&dp->rq)) == NULL) { 206 if (skb_queue_len(&dp->rq) >= dev->tx_queue_len) { 210 __skb_queue_tail(&dp->rq, skb); 225 __skb_queue_purge(&dp->rq); 235 __skb_queue_head_init(&dp->rq);
|
/drivers/staging/vt6655/ |
H A D | ioctl.c | 47 int private_ioctl(PSDevice pDevice, struct ifreq *rq) argument 49 PSCmdRequest pReq = (PSCmdRequest)rq;
|
/drivers/staging/vt6656/ |
H A D | ioctl.c | 46 int private_ioctl(PSDevice pDevice, struct ifreq *rq) argument 49 PSCmdRequest pReq = (PSCmdRequest)rq;
|
/drivers/char/agp/ |
H A D | isoch.c | 75 u32 rq; member in struct:isoch_data 126 target.rq = (tstatus >> 24) & 0xff; 216 master[cdev].rq = master[cdev].n; 218 master[cdev].rq *= (1 << (master[cdev].y - 1)); 220 tot_rq += master[cdev].rq; 227 rq_async = target.rq - rq_isoch; 254 master[cdev].rq += (cdev == ndevs - 1) 266 mcmd |= master[cdev].rq << 24;
|
/drivers/char/ |
H A D | raw.c | 212 struct raw_config_request rq; local 218 if (copy_from_user(&rq, (void __user *) arg, sizeof(rq))) 221 return bind_set(rq.raw_minor, rq.block_major, rq.block_minor); 224 if (copy_from_user(&rq, (void __user *) arg, sizeof(rq))) 227 err = bind_get(rq.raw_minor, &dev); 231 rq 254 struct raw32_config_request rq; local [all...] |
/drivers/infiniband/hw/ipath/ |
H A D | ipath_ruc.c | 169 struct ipath_rq *rq; local 180 rq = &srq->rq; 184 rq = &qp->r_rq; 187 spin_lock_irqsave(&rq->lock, flags); 193 wq = rq->wq; 196 if (tail >= rq->size) 205 wqe = get_rwqe_ptr(rq, tail); 206 if (++tail >= rq->size) 225 if (n >= rq [all...] |
H A D | ipath_ud.c | 56 struct ipath_rq *rq; local 110 rq = &srq->rq; 114 rq = &qp->r_rq; 119 * Note that it is safe to drop the lock after changing rq->tail 122 spin_lock_irqsave(&rq->lock, flags); 123 wq = rq->wq; 126 if (tail >= rq->size) 129 spin_unlock_irqrestore(&rq->lock, flags); 133 wqe = get_rwqe_ptr(rq, tai [all...] |
/drivers/infiniband/hw/qib/ |
H A D | qib_ruc.c | 141 struct qib_rq *rq; local 152 rq = &srq->rq; 156 rq = &qp->r_rq; 159 spin_lock_irqsave(&rq->lock, flags); 165 wq = rq->wq; 168 if (tail >= rq->size) 176 wqe = get_rwqe_ptr(rq, tail); 182 if (++tail >= rq->size) 201 if (n >= rq [all...] |
/drivers/isdn/mISDN/ |
H A D | stack.c | 425 struct channel_req rq; local 441 rq.protocol = protocol; 442 rq.adr.channel = adr->channel; 443 err = dev->D.ctrl(&dev->D, OPEN_CHANNEL, &rq); 462 struct channel_req rq, rq2; local 474 rq.protocol = protocol; 475 rq.adr = *adr; 476 err = dev->D.ctrl(&dev->D, OPEN_CHANNEL, &rq); 479 ch->recv = rq.ch->send; 480 ch->peer = rq 519 struct channel_req rq; local [all...] |