Searched refs:bvec (Results 1 - 21 of 21) sorted by relevance

/drivers/staging/zram/
H A Dzram_drv.c173 static void handle_zero_page(struct bio_vec *bvec) argument
175 struct page *page = bvec->bv_page;
179 memset(user_mem + bvec->bv_offset, 0, bvec->bv_len);
185 static void handle_uncompressed_page(struct zram *zram, struct bio_vec *bvec, argument
188 struct page *page = bvec->bv_page;
194 memcpy(user_mem + bvec->bv_offset, cmem + offset, bvec->bv_len);
201 static inline int is_partial_io(struct bio_vec *bvec) argument
203 return bvec
206 zram_bvec_read(struct zram *zram, struct bio_vec *bvec, u32 index, int offset, struct bio *bio) argument
314 zram_bvec_write(struct zram *zram, struct bio_vec *bvec, u32 index, int offset) argument
453 zram_bvec_rw(struct zram *zram, struct bio_vec *bvec, u32 index, int offset, struct bio *bio, int rw) argument
471 update_position(u32 *index, int *offset, struct bio_vec *bvec) argument
482 struct bio_vec *bvec; local
[all...]
/drivers/md/
H A Ddm-io.c203 * Functions for getting the pages from a bvec.
208 struct bio_vec *bvec = (struct bio_vec *) dp->context_ptr; local
209 *p = bvec->bv_page;
210 *len = bvec->bv_len;
211 *offset = bvec->bv_offset;
216 struct bio_vec *bvec = (struct bio_vec *) dp->context_ptr; local
217 dp->context_ptr = bvec + 1;
220 static void bvec_dp_init(struct dpages *dp, struct bio_vec *bvec) argument
224 dp->context_ptr = bvec;
460 bvec_dp_init(dp, io_req->mem.ptr.bvec);
[all...]
H A Draid1.c844 struct bio_vec *bvec; local
850 bio_for_each_segment(bvec, bio, i) {
851 bvecs[i] = *bvec;
855 memcpy(kmap(bvecs[i].bv_page) + bvec->bv_offset,
856 kmap(bvec->bv_page) + bvec->bv_offset, bvec->bv_len);
858 kunmap(bvec->bv_page);
1164 struct bio_vec *bvec; local
1174 __bio_for_each_segment(bvec, mbi
[all...]
H A Ddm-raid1.c529 .mem.ptr.bvec = bio->bi_io_vec + bio->bi_idx,
632 .mem.ptr.bvec = bio->bi_io_vec + bio->bi_idx,
H A Dmd.c227 struct bio_vec *bvec; local
257 bio_for_each_segment(bvec, bio, i) {
258 if (sofar + bvec->bv_len > size)
259 bvec->bv_len = size - sofar;
260 if (bvec->bv_len == 0) {
264 sofar += bvec->bv_len;
/drivers/block/
H A Dnbd.c221 static inline int sock_send_bvec(struct nbd_device *nbd, struct bio_vec *bvec, argument
225 void *kaddr = kmap(bvec->bv_page);
226 result = sock_xmit(nbd, 1, kaddr + bvec->bv_offset,
227 bvec->bv_len, flags);
228 kunmap(bvec->bv_page);
260 struct bio_vec *bvec; local
265 rq_for_each_segment(bvec, req, iter) {
270 nbd->disk->disk_name, req, bvec->bv_len);
271 result = sock_send_bvec(nbd, bvec, flags);
312 static inline int sock_recv_bvec(struct nbd_device *nbd, struct bio_vec *bvec) argument
367 struct bio_vec *bvec; local
[all...]
H A Dloop.c247 struct bio_vec *bvec, loff_t pos, struct page *page)
250 kmap(bvec->bv_page) + bvec->bv_offset,
251 bvec->bv_len, pos);
252 kunmap(bvec->bv_page);
264 static int do_lo_send_write(struct loop_device *lo, struct bio_vec *bvec, argument
267 int ret = lo_do_transfer(lo, WRITE, page, 0, bvec->bv_page,
268 bvec->bv_offset, bvec->bv_len, pos >> 9);
271 page_address(page), bvec
246 do_lo_send_direct_write(struct loop_device *lo, struct bio_vec *bvec, loff_t pos, struct page *page) argument
284 struct bio_vec *bvec; local
360 do_lo_receive(struct loop_device *lo, struct bio_vec *bvec, int bsize, loff_t pos) argument
388 struct bio_vec *bvec; local
[all...]
H A Dbrd.c297 * Process a single bvec of a bio.
331 struct bio_vec *bvec; local
351 bio_for_each_segment(bvec, bio, i) {
352 unsigned int len = bvec->bv_len;
353 err = brd_do_bvec(brd, bvec->bv_page, len,
354 bvec->bv_offset, rw, sector);
H A Dps3disk.c97 struct bio_vec *bvec; local
102 rq_for_each_segment(bvec, req, iter) {
109 size = bvec->bv_len;
110 buf = bvec_kmap_irq(bvec, &flags);
116 flush_kernel_dcache_page(bvec->bv_page);
H A Dpktcdvd.c1013 static void pkt_make_local_copy(struct packet_data *pkt, struct bio_vec *bvec) argument
1021 if (bvec[f].bv_page != pkt->pages[p]) {
1022 void *vfrom = kmap_atomic(bvec[f].bv_page) + bvec[f].bv_offset;
1026 bvec[f].bv_page = pkt->pages[p];
1027 bvec[f].bv_offset = offs;
1029 BUG_ON(bvec[f].bv_offset != offs);
1366 struct bio_vec *bvec = pkt->w_bio->bi_io_vec; local
1369 bvec[f].bv_page = pkt->pages[(f * CD_FRAMESIZE) / PAGE_SIZE];
1370 bvec[
2594 pkt_merge_bvec(struct request_queue *q, struct bvec_merge_data *bmd, struct bio_vec *bvec) argument
[all...]
H A Dnvme.c478 struct bio_vec *bvec, *bvprv = NULL; local
484 bio_for_each_segment(bvec, bio, i) {
485 if (bvprv && BIOVEC_PHYS_MERGEABLE(bvprv, bvec)) {
486 sg->length += bvec->bv_len;
488 if (bvprv && BIOVEC_NOT_VIRT_MERGEABLE(bvprv, bvec))
491 sg_set_page(sg, bvec->bv_page, bvec->bv_len,
492 bvec->bv_offset);
495 length += bvec->bv_len;
496 bvprv = bvec;
[all...]
H A Dps3vram.c558 struct bio_vec *bvec; local
562 bio_for_each_segment(bvec, bio, i) {
564 char *ptr = page_address(bvec->bv_page) + bvec->bv_offset;
565 size_t len = bvec->bv_len, retlen;
H A Drbd.c1550 struct bio_vec *bvec)
1566 if (max <= bvec->bv_len && bio_sectors == 0)
1567 return bvec->bv_len;
1549 rbd_merge_bvec(struct request_queue *q, struct bvec_merge_data *bmd, struct bio_vec *bvec) argument
/drivers/s390/block/
H A Dxpram.c187 struct bio_vec *bvec; local
202 bio_for_each_segment(bvec, bio, i) {
204 kmap(bvec->bv_page) + bvec->bv_offset;
205 bytes = bvec->bv_len;
H A Ddcssblk.c821 struct bio_vec *bvec; local
857 bio_for_each_segment(bvec, bio, i) {
859 page_address(bvec->bv_page) + bvec->bv_offset;
861 if (unlikely((page_addr & 4095) != 0) || (bvec->bv_len & 4095) != 0)
866 bvec->bv_len);
869 bvec->bv_len);
871 bytes_done += bvec->bv_len;
/drivers/block/drbd/
H A Ddrbd_req.c1162 * As long as the BIO is empty we have to allow at least one bvec,
1167 int drbd_merge_bvec(struct request_queue *q, struct bvec_merge_data *bvm, struct bio_vec *bvec) argument
1180 if (limit <= bvec->bv_len)
1181 limit = bvec->bv_len;
1186 backing_limit = b->merge_bvec_fn(b, bvm, bvec);
H A Ddrbd_worker.c284 struct bio_vec *bvec; local
293 __bio_for_each_segment(bvec, bio, i, 0) {
294 sg_set_page(&sg, bvec->bv_page, bvec->bv_len, bvec->bv_offset);
H A Ddrbd_main.c2615 struct bio_vec *bvec; local
2618 __bio_for_each_segment(bvec, bio, i, 0) {
2619 if (!_drbd_no_send_page(mdev, bvec->bv_page,
2620 bvec->bv_offset, bvec->bv_len,
2629 struct bio_vec *bvec; local
2632 __bio_for_each_segment(bvec, bio, i, 0) {
2633 if (!_drbd_send_page(mdev, bvec->bv_page,
2634 bvec->bv_offset, bvec
[all...]
H A Ddrbd_int.h1511 extern int drbd_merge_bvec(struct request_queue *q, struct bvec_merge_data *bvm, struct bio_vec *bvec);
1759 struct bio_vec *bvec; local
1762 __bio_for_each_segment(bvec, bio, i, 0) {
1763 if (page_count(bvec->bv_page) > 1)
H A Ddrbd_receiver.c1351 struct bio_vec *bvec; local
1380 bio_for_each_segment(bvec, bio, i) {
1381 expect = min_t(int, data_size, bvec->bv_len);
1383 kmap(bvec->bv_page)+bvec->bv_offset,
1385 kunmap(bvec->bv_page);
/drivers/block/mtip32xx/
H A Dmtip32xx.c3483 struct bio_vec *bvec; local
3524 bio_for_each_segment(bvec, bio, nents) {
3526 bvec->bv_page,
3527 bvec->bv_len,
3528 bvec->bv_offset);

Completed in 300 milliseconds