/drivers/staging/zram/ |
H A D | zram_drv.c | 173 static void handle_zero_page(struct bio_vec *bvec) argument 175 struct page *page = bvec->bv_page; 179 memset(user_mem + bvec->bv_offset, 0, bvec->bv_len); 185 static void handle_uncompressed_page(struct zram *zram, struct bio_vec *bvec, argument 188 struct page *page = bvec->bv_page; 194 memcpy(user_mem + bvec->bv_offset, cmem + offset, bvec->bv_len); 201 static inline int is_partial_io(struct bio_vec *bvec) argument 203 return bvec 206 zram_bvec_read(struct zram *zram, struct bio_vec *bvec, u32 index, int offset, struct bio *bio) argument 314 zram_bvec_write(struct zram *zram, struct bio_vec *bvec, u32 index, int offset) argument 453 zram_bvec_rw(struct zram *zram, struct bio_vec *bvec, u32 index, int offset, struct bio *bio, int rw) argument 471 update_position(u32 *index, int *offset, struct bio_vec *bvec) argument 482 struct bio_vec *bvec; local [all...] |
/drivers/md/ |
H A D | dm-io.c | 203 * Functions for getting the pages from a bvec. 208 struct bio_vec *bvec = (struct bio_vec *) dp->context_ptr; local 209 *p = bvec->bv_page; 210 *len = bvec->bv_len; 211 *offset = bvec->bv_offset; 216 struct bio_vec *bvec = (struct bio_vec *) dp->context_ptr; local 217 dp->context_ptr = bvec + 1; 220 static void bvec_dp_init(struct dpages *dp, struct bio_vec *bvec) argument 224 dp->context_ptr = bvec; 460 bvec_dp_init(dp, io_req->mem.ptr.bvec); [all...] |
H A D | raid1.c | 844 struct bio_vec *bvec; local 850 bio_for_each_segment(bvec, bio, i) { 851 bvecs[i] = *bvec; 855 memcpy(kmap(bvecs[i].bv_page) + bvec->bv_offset, 856 kmap(bvec->bv_page) + bvec->bv_offset, bvec->bv_len); 858 kunmap(bvec->bv_page); 1164 struct bio_vec *bvec; local 1174 __bio_for_each_segment(bvec, mbi [all...] |
H A D | dm-raid1.c | 529 .mem.ptr.bvec = bio->bi_io_vec + bio->bi_idx, 632 .mem.ptr.bvec = bio->bi_io_vec + bio->bi_idx,
|
H A D | md.c | 227 struct bio_vec *bvec; local 257 bio_for_each_segment(bvec, bio, i) { 258 if (sofar + bvec->bv_len > size) 259 bvec->bv_len = size - sofar; 260 if (bvec->bv_len == 0) { 264 sofar += bvec->bv_len;
|
/drivers/block/ |
H A D | nbd.c | 221 static inline int sock_send_bvec(struct nbd_device *nbd, struct bio_vec *bvec, argument 225 void *kaddr = kmap(bvec->bv_page); 226 result = sock_xmit(nbd, 1, kaddr + bvec->bv_offset, 227 bvec->bv_len, flags); 228 kunmap(bvec->bv_page); 260 struct bio_vec *bvec; local 265 rq_for_each_segment(bvec, req, iter) { 270 nbd->disk->disk_name, req, bvec->bv_len); 271 result = sock_send_bvec(nbd, bvec, flags); 312 static inline int sock_recv_bvec(struct nbd_device *nbd, struct bio_vec *bvec) argument 367 struct bio_vec *bvec; local [all...] |
H A D | loop.c | 247 struct bio_vec *bvec, loff_t pos, struct page *page) 250 kmap(bvec->bv_page) + bvec->bv_offset, 251 bvec->bv_len, pos); 252 kunmap(bvec->bv_page); 264 static int do_lo_send_write(struct loop_device *lo, struct bio_vec *bvec, argument 267 int ret = lo_do_transfer(lo, WRITE, page, 0, bvec->bv_page, 268 bvec->bv_offset, bvec->bv_len, pos >> 9); 271 page_address(page), bvec 246 do_lo_send_direct_write(struct loop_device *lo, struct bio_vec *bvec, loff_t pos, struct page *page) argument 284 struct bio_vec *bvec; local 360 do_lo_receive(struct loop_device *lo, struct bio_vec *bvec, int bsize, loff_t pos) argument 388 struct bio_vec *bvec; local [all...] |
H A D | brd.c | 297 * Process a single bvec of a bio. 331 struct bio_vec *bvec; local 351 bio_for_each_segment(bvec, bio, i) { 352 unsigned int len = bvec->bv_len; 353 err = brd_do_bvec(brd, bvec->bv_page, len, 354 bvec->bv_offset, rw, sector);
|
H A D | ps3disk.c | 97 struct bio_vec *bvec; local 102 rq_for_each_segment(bvec, req, iter) { 109 size = bvec->bv_len; 110 buf = bvec_kmap_irq(bvec, &flags); 116 flush_kernel_dcache_page(bvec->bv_page);
|
H A D | pktcdvd.c | 1013 static void pkt_make_local_copy(struct packet_data *pkt, struct bio_vec *bvec) argument 1021 if (bvec[f].bv_page != pkt->pages[p]) { 1022 void *vfrom = kmap_atomic(bvec[f].bv_page) + bvec[f].bv_offset; 1026 bvec[f].bv_page = pkt->pages[p]; 1027 bvec[f].bv_offset = offs; 1029 BUG_ON(bvec[f].bv_offset != offs); 1366 struct bio_vec *bvec = pkt->w_bio->bi_io_vec; local 1369 bvec[f].bv_page = pkt->pages[(f * CD_FRAMESIZE) / PAGE_SIZE]; 1370 bvec[ 2594 pkt_merge_bvec(struct request_queue *q, struct bvec_merge_data *bmd, struct bio_vec *bvec) argument [all...] |
H A D | nvme.c | 478 struct bio_vec *bvec, *bvprv = NULL; local 484 bio_for_each_segment(bvec, bio, i) { 485 if (bvprv && BIOVEC_PHYS_MERGEABLE(bvprv, bvec)) { 486 sg->length += bvec->bv_len; 488 if (bvprv && BIOVEC_NOT_VIRT_MERGEABLE(bvprv, bvec)) 491 sg_set_page(sg, bvec->bv_page, bvec->bv_len, 492 bvec->bv_offset); 495 length += bvec->bv_len; 496 bvprv = bvec; [all...] |
H A D | ps3vram.c | 558 struct bio_vec *bvec; local 562 bio_for_each_segment(bvec, bio, i) { 564 char *ptr = page_address(bvec->bv_page) + bvec->bv_offset; 565 size_t len = bvec->bv_len, retlen;
|
H A D | rbd.c | 1550 struct bio_vec *bvec) 1566 if (max <= bvec->bv_len && bio_sectors == 0) 1567 return bvec->bv_len; 1549 rbd_merge_bvec(struct request_queue *q, struct bvec_merge_data *bmd, struct bio_vec *bvec) argument
|
/drivers/s390/block/ |
H A D | xpram.c | 187 struct bio_vec *bvec; local 202 bio_for_each_segment(bvec, bio, i) { 204 kmap(bvec->bv_page) + bvec->bv_offset; 205 bytes = bvec->bv_len;
|
H A D | dcssblk.c | 821 struct bio_vec *bvec; local 857 bio_for_each_segment(bvec, bio, i) { 859 page_address(bvec->bv_page) + bvec->bv_offset; 861 if (unlikely((page_addr & 4095) != 0) || (bvec->bv_len & 4095) != 0) 866 bvec->bv_len); 869 bvec->bv_len); 871 bytes_done += bvec->bv_len;
|
/drivers/block/drbd/ |
H A D | drbd_req.c | 1162 * As long as the BIO is empty we have to allow at least one bvec, 1167 int drbd_merge_bvec(struct request_queue *q, struct bvec_merge_data *bvm, struct bio_vec *bvec) argument 1180 if (limit <= bvec->bv_len) 1181 limit = bvec->bv_len; 1186 backing_limit = b->merge_bvec_fn(b, bvm, bvec);
|
H A D | drbd_worker.c | 284 struct bio_vec *bvec; local 293 __bio_for_each_segment(bvec, bio, i, 0) { 294 sg_set_page(&sg, bvec->bv_page, bvec->bv_len, bvec->bv_offset);
|
H A D | drbd_main.c | 2615 struct bio_vec *bvec; local 2618 __bio_for_each_segment(bvec, bio, i, 0) { 2619 if (!_drbd_no_send_page(mdev, bvec->bv_page, 2620 bvec->bv_offset, bvec->bv_len, 2629 struct bio_vec *bvec; local 2632 __bio_for_each_segment(bvec, bio, i, 0) { 2633 if (!_drbd_send_page(mdev, bvec->bv_page, 2634 bvec->bv_offset, bvec [all...] |
H A D | drbd_int.h | 1511 extern int drbd_merge_bvec(struct request_queue *q, struct bvec_merge_data *bvm, struct bio_vec *bvec); 1759 struct bio_vec *bvec; local 1762 __bio_for_each_segment(bvec, bio, i, 0) { 1763 if (page_count(bvec->bv_page) > 1)
|
H A D | drbd_receiver.c | 1351 struct bio_vec *bvec; local 1380 bio_for_each_segment(bvec, bio, i) { 1381 expect = min_t(int, data_size, bvec->bv_len); 1383 kmap(bvec->bv_page)+bvec->bv_offset, 1385 kunmap(bvec->bv_page);
|
/drivers/block/mtip32xx/ |
H A D | mtip32xx.c | 3483 struct bio_vec *bvec; local 3524 bio_for_each_segment(bvec, bio, nents) { 3526 bvec->bv_page, 3527 bvec->bv_len, 3528 bvec->bv_offset);
|