/drivers/s390/block/ |
H A D | xpram.c | 187 struct bio_vec bvec; local 203 bio_for_each_segment(bvec, bio, iter) { 205 kmap(bvec.bv_page) + bvec.bv_offset; 206 bytes = bvec.bv_len;
|
H A D | dcssblk.c | 811 struct bio_vec bvec; local 847 bio_for_each_segment(bvec, bio, iter) { 849 page_address(bvec.bv_page) + bvec.bv_offset; 851 if (unlikely((page_addr & 4095) != 0) || (bvec.bv_len & 4095) != 0) 856 bvec.bv_len); 859 bvec.bv_len); 861 bytes_done += bvec.bv_len;
|
/drivers/block/ |
H A D | brd.c | 297 * Process a single bvec of a bio. 331 struct bio_vec bvec; local 350 bio_for_each_segment(bvec, bio, iter) { 351 unsigned int len = bvec.bv_len; 352 err = brd_do_bvec(brd, bvec.bv_page, len, 353 bvec.bv_offset, rw, sector);
|
H A D | nbd.c | 228 static inline int sock_send_bvec(struct nbd_device *nbd, struct bio_vec *bvec, argument 232 void *kaddr = kmap(bvec->bv_page); 233 result = sock_xmit(nbd, 1, kaddr + bvec->bv_offset, 234 bvec->bv_len, flags); 235 kunmap(bvec->bv_page); 271 struct bio_vec bvec; local 276 rq_for_each_segment(bvec, req, iter) { 278 if (!rq_iter_last(bvec, iter)) 281 nbd->disk->disk_name, req, bvec.bv_len); 282 result = sock_send_bvec(nbd, &bvec, flag 323 sock_recv_bvec(struct nbd_device *nbd, struct bio_vec *bvec) argument 378 struct bio_vec bvec; local [all...] |
H A D | ps3disk.c | 97 struct bio_vec bvec; local 102 rq_for_each_segment(bvec, req, iter) { 108 size = bvec.bv_len; 109 buf = bvec_kmap_irq(&bvec, &flags); 115 flush_kernel_dcache_page(bvec.bv_page);
|
H A D | ps3vram.c | 558 struct bio_vec bvec; local 562 bio_for_each_segment(bvec, bio, iter) { 564 char *ptr = page_address(bvec.bv_page) + bvec.bv_offset; 565 size_t len = bvec.bv_len, retlen;
|
H A D | loop.c | 254 struct bio_vec *bvec, loff_t pos, struct page *page) 257 kmap(bvec->bv_page) + bvec->bv_offset, 258 bvec->bv_len, pos); 259 kunmap(bvec->bv_page); 271 static int do_lo_send_write(struct loop_device *lo, struct bio_vec *bvec, argument 274 int ret = lo_do_transfer(lo, WRITE, page, 0, bvec->bv_page, 275 bvec->bv_offset, bvec->bv_len, pos >> 9); 278 page_address(page), bvec 253 do_lo_send_direct_write(struct loop_device *lo, struct bio_vec *bvec, loff_t pos, struct page *page) argument 291 struct bio_vec bvec; local 368 do_lo_receive(struct loop_device *lo, struct bio_vec *bvec, int bsize, loff_t pos) argument 396 struct bio_vec bvec; local [all...] |
H A D | pktcdvd.c | 953 static void pkt_make_local_copy(struct packet_data *pkt, struct bio_vec *bvec) argument 961 if (bvec[f].bv_page != pkt->pages[p]) { 962 void *vfrom = kmap_atomic(bvec[f].bv_page) + bvec[f].bv_offset; 966 bvec[f].bv_page = pkt->pages[p]; 967 bvec[f].bv_offset = offs; 969 BUG_ON(bvec[f].bv_offset != offs); 1300 struct bio_vec *bvec = pkt->w_bio->bi_io_vec; local 1310 bvec[f].bv_page = pkt->pages[(f * CD_FRAMESIZE) / PAGE_SIZE]; 1311 bvec[ 2506 pkt_merge_bvec(struct request_queue *q, struct bvec_merge_data *bmd, struct bio_vec *bvec) argument [all...] |
H A D | nvme-core.c | 574 struct bio_vec bvec, bvprv; local 586 bio_for_each_segment(bvec, bio, iter) { 587 if (!first && BIOVEC_PHYS_MERGEABLE(&bvprv, &bvec)) { 588 sg->length += bvec.bv_len; 590 if (!first && BIOVEC_NOT_VIRT_MERGEABLE(&bvprv, &bvec)) 595 sg_set_page(sg, bvec.bv_page, 596 bvec.bv_len, bvec.bv_offset); 600 if (split_len - length < bvec.bv_len) 602 length += bvec [all...] |
H A D | rbd.c | 3465 struct bio_vec *bvec) 3498 rbd_assert(bvec->bv_len <= PAGE_SIZE); 3499 if (ret > (int) bvec->bv_len || !bmd->bi_size) 3500 ret = (int) bvec->bv_len; 3464 rbd_merge_bvec(struct request_queue *q, struct bvec_merge_data *bmd, struct bio_vec *bvec) argument
|
/drivers/md/ |
H A D | dm-io.c | 201 * Functions for getting the pages from a bvec. 206 struct bio_vec *bvec = dp->context_ptr; local 207 *p = bvec->bv_page; 208 *len = bvec->bv_len - dp->context_u; 209 *offset = bvec->bv_offset + dp->context_u; 214 struct bio_vec *bvec = dp->context_ptr; local 215 dp->context_ptr = bvec + 1;
|
H A D | raid1.c | 1008 struct bio_vec *bvec; local 1014 bio_for_each_segment_all(bvec, bio, i) { 1015 bvecs[i] = *bvec; 1019 memcpy(kmap(bvecs[i].bv_page) + bvec->bv_offset, 1020 kmap(bvec->bv_page) + bvec->bv_offset, bvec->bv_len); 1022 kunmap(bvec->bv_page); 1391 struct bio_vec *bvec; local 1397 bio_for_each_segment_all(bvec, mbi [all...] |
H A D | raid10.c | 4586 struct bio_vec *bvec = r10_bio->master_bio->bi_io_vec; local 4612 bvec[idx].bv_page,
|
/drivers/block/drbd/ |
H A D | drbd_req.c | 1533 * As long as the BIO is empty we have to allow at least one bvec, 1536 int drbd_merge_bvec(struct request_queue *q, struct bvec_merge_data *bvm, struct bio_vec *bvec) argument 1548 backing_limit = b->merge_bvec_fn(b, bvm, bvec);
|
H A D | drbd_worker.c | 325 struct bio_vec bvec; local 334 bio_for_each_segment(bvec, bio, iter) { 335 sg_set_page(&sg, bvec.bv_page, bvec.bv_len, bvec.bv_offset);
|
H A D | drbd_main.c | 1550 struct bio_vec bvec; local 1554 bio_for_each_segment(bvec, bio, iter) { 1557 err = _drbd_no_send_page(peer_device, bvec.bv_page, 1558 bvec.bv_offset, bvec.bv_len, 1559 bio_iter_last(bvec, iter) 1569 struct bio_vec bvec; local 1573 bio_for_each_segment(bvec, bio, iter) { 1576 err = _drbd_send_page(peer_device, bvec.bv_page, 1577 bvec [all...] |
H A D | drbd_receiver.c | 1708 struct bio_vec bvec; local 1731 bio_for_each_segment(bvec, bio, iter) { 1732 void *mapped = kmap(bvec.bv_page) + bvec.bv_offset; 1733 expect = min_t(int, data_size, bvec.bv_len); 1735 kunmap(bvec.bv_page);
|
/drivers/staging/lustre/lustre/llite/ |
H A D | lloop.c | 196 struct bio_vec bvec; local 221 bio_for_each_segment(bvec, bio, iter) { 222 BUG_ON(bvec.bv_offset != 0); 223 BUG_ON(bvec.bv_len != PAGE_CACHE_SIZE); 225 pages[page_count] = bvec.bv_page; 228 offset += bvec.bv_len;
|
/drivers/block/rsxx/ |
H A D | dma.c | 687 struct bio_vec bvec; local 726 bio_for_each_segment(bvec, bio, iter) { 727 bv_len = bvec.bv_len; 728 bv_off = bvec.bv_offset; 740 laddr, bvec.bv_page,
|
/drivers/block/zram/ |
H A D | zram_drv.c | 282 static inline int is_partial_io(struct bio_vec *bvec) argument 284 return bvec->bv_len != PAGE_SIZE; 350 static void update_position(u32 *index, int *offset, struct bio_vec *bvec) argument 352 if (*offset + bvec->bv_len >= PAGE_SIZE) 354 *offset = (*offset + bvec->bv_len) % PAGE_SIZE; 372 static void handle_zero_page(struct bio_vec *bvec) argument 374 struct page *page = bvec->bv_page; 378 if (is_partial_io(bvec)) 379 memset(user_mem + bvec->bv_offset, 0, bvec 455 zram_bvec_read(struct zram *zram, struct bio_vec *bvec, u32 index, int offset, struct bio *bio) argument 520 zram_bvec_write(struct zram *zram, struct bio_vec *bvec, u32 index, int offset) argument 647 zram_bvec_rw(struct zram *zram, struct bio_vec *bvec, u32 index, int offset, struct bio *bio) argument 858 struct bio_vec bvec; local [all...] |
/drivers/scsi/mpt2sas/ |
H A D | mpt2sas_transport.c | 1919 struct bio_vec bvec; local 1959 bio_for_each_segment(bvec, req->bio, iter) { 1961 page_address(bvec.bv_page) + bvec.bv_offset, 1962 bvec.bv_len); 1963 offset += bvec.bv_len; 2110 bio_for_each_segment(bvec, rsp->bio, iter) { 2111 if (bytes_to_copy <= bvec.bv_len) { 2112 memcpy(page_address(bvec.bv_page) + 2113 bvec [all...] |
/drivers/scsi/mpt3sas/ |
H A D | mpt3sas_transport.c | 1901 struct bio_vec bvec; local 1942 bio_for_each_segment(bvec, req->bio, iter) { 1944 page_address(bvec.bv_page) + bvec.bv_offset, 1945 bvec.bv_len); 1946 offset += bvec.bv_len; 2071 bio_for_each_segment(bvec, rsp->bio, iter) { 2072 if (bytes_to_copy <= bvec.bv_len) { 2073 memcpy(page_address(bvec.bv_page) + 2074 bvec [all...] |