Searched defs:bvec (Results 1 - 22 of 22) sorted by relevance

/drivers/s390/block/
H A Dxpram.c187 struct bio_vec bvec; local
203 bio_for_each_segment(bvec, bio, iter) {
205 kmap(bvec.bv_page) + bvec.bv_offset;
206 bytes = bvec.bv_len;
H A Ddcssblk.c811 struct bio_vec bvec; local
847 bio_for_each_segment(bvec, bio, iter) {
849 page_address(bvec.bv_page) + bvec.bv_offset;
851 if (unlikely((page_addr & 4095) != 0) || (bvec.bv_len & 4095) != 0)
856 bvec.bv_len);
859 bvec.bv_len);
861 bytes_done += bvec.bv_len;
/drivers/block/
H A Dbrd.c297 * Process a single bvec of a bio.
331 struct bio_vec bvec; local
350 bio_for_each_segment(bvec, bio, iter) {
351 unsigned int len = bvec.bv_len;
352 err = brd_do_bvec(brd, bvec.bv_page, len,
353 bvec.bv_offset, rw, sector);
H A Dnbd.c228 static inline int sock_send_bvec(struct nbd_device *nbd, struct bio_vec *bvec, argument
232 void *kaddr = kmap(bvec->bv_page);
233 result = sock_xmit(nbd, 1, kaddr + bvec->bv_offset,
234 bvec->bv_len, flags);
235 kunmap(bvec->bv_page);
271 struct bio_vec bvec; local
276 rq_for_each_segment(bvec, req, iter) {
278 if (!rq_iter_last(bvec, iter))
281 nbd->disk->disk_name, req, bvec.bv_len);
282 result = sock_send_bvec(nbd, &bvec, flag
323 sock_recv_bvec(struct nbd_device *nbd, struct bio_vec *bvec) argument
378 struct bio_vec bvec; local
[all...]
H A Dps3disk.c97 struct bio_vec bvec; local
102 rq_for_each_segment(bvec, req, iter) {
108 size = bvec.bv_len;
109 buf = bvec_kmap_irq(&bvec, &flags);
115 flush_kernel_dcache_page(bvec.bv_page);
H A Dps3vram.c558 struct bio_vec bvec; local
562 bio_for_each_segment(bvec, bio, iter) {
564 char *ptr = page_address(bvec.bv_page) + bvec.bv_offset;
565 size_t len = bvec.bv_len, retlen;
H A Dloop.c254 struct bio_vec *bvec, loff_t pos, struct page *page)
257 kmap(bvec->bv_page) + bvec->bv_offset,
258 bvec->bv_len, pos);
259 kunmap(bvec->bv_page);
271 static int do_lo_send_write(struct loop_device *lo, struct bio_vec *bvec, argument
274 int ret = lo_do_transfer(lo, WRITE, page, 0, bvec->bv_page,
275 bvec->bv_offset, bvec->bv_len, pos >> 9);
278 page_address(page), bvec
253 do_lo_send_direct_write(struct loop_device *lo, struct bio_vec *bvec, loff_t pos, struct page *page) argument
291 struct bio_vec bvec; local
368 do_lo_receive(struct loop_device *lo, struct bio_vec *bvec, int bsize, loff_t pos) argument
396 struct bio_vec bvec; local
[all...]
H A Dpktcdvd.c953 static void pkt_make_local_copy(struct packet_data *pkt, struct bio_vec *bvec) argument
961 if (bvec[f].bv_page != pkt->pages[p]) {
962 void *vfrom = kmap_atomic(bvec[f].bv_page) + bvec[f].bv_offset;
966 bvec[f].bv_page = pkt->pages[p];
967 bvec[f].bv_offset = offs;
969 BUG_ON(bvec[f].bv_offset != offs);
1300 struct bio_vec *bvec = pkt->w_bio->bi_io_vec; local
1310 bvec[f].bv_page = pkt->pages[(f * CD_FRAMESIZE) / PAGE_SIZE];
1311 bvec[
2506 pkt_merge_bvec(struct request_queue *q, struct bvec_merge_data *bmd, struct bio_vec *bvec) argument
[all...]
H A Dnvme-core.c574 struct bio_vec bvec, bvprv; local
586 bio_for_each_segment(bvec, bio, iter) {
587 if (!first && BIOVEC_PHYS_MERGEABLE(&bvprv, &bvec)) {
588 sg->length += bvec.bv_len;
590 if (!first && BIOVEC_NOT_VIRT_MERGEABLE(&bvprv, &bvec))
595 sg_set_page(sg, bvec.bv_page,
596 bvec.bv_len, bvec.bv_offset);
600 if (split_len - length < bvec.bv_len)
602 length += bvec
[all...]
H A Drbd.c3465 struct bio_vec *bvec)
3498 rbd_assert(bvec->bv_len <= PAGE_SIZE);
3499 if (ret > (int) bvec->bv_len || !bmd->bi_size)
3500 ret = (int) bvec->bv_len;
3464 rbd_merge_bvec(struct request_queue *q, struct bvec_merge_data *bmd, struct bio_vec *bvec) argument
/drivers/md/
H A Ddm-io.c201 * Functions for getting the pages from a bvec.
206 struct bio_vec *bvec = dp->context_ptr; local
207 *p = bvec->bv_page;
208 *len = bvec->bv_len - dp->context_u;
209 *offset = bvec->bv_offset + dp->context_u;
214 struct bio_vec *bvec = dp->context_ptr; local
215 dp->context_ptr = bvec + 1;
H A Draid1.c1008 struct bio_vec *bvec; local
1014 bio_for_each_segment_all(bvec, bio, i) {
1015 bvecs[i] = *bvec;
1019 memcpy(kmap(bvecs[i].bv_page) + bvec->bv_offset,
1020 kmap(bvec->bv_page) + bvec->bv_offset, bvec->bv_len);
1022 kunmap(bvec->bv_page);
1391 struct bio_vec *bvec; local
1397 bio_for_each_segment_all(bvec, mbi
[all...]
H A Draid10.c4586 struct bio_vec *bvec = r10_bio->master_bio->bi_io_vec; local
4612 bvec[idx].bv_page,
/drivers/block/drbd/
H A Ddrbd_req.c1533 * As long as the BIO is empty we have to allow at least one bvec,
1536 int drbd_merge_bvec(struct request_queue *q, struct bvec_merge_data *bvm, struct bio_vec *bvec) argument
1548 backing_limit = b->merge_bvec_fn(b, bvm, bvec);
H A Ddrbd_worker.c325 struct bio_vec bvec; local
334 bio_for_each_segment(bvec, bio, iter) {
335 sg_set_page(&sg, bvec.bv_page, bvec.bv_len, bvec.bv_offset);
H A Ddrbd_main.c1550 struct bio_vec bvec; local
1554 bio_for_each_segment(bvec, bio, iter) {
1557 err = _drbd_no_send_page(peer_device, bvec.bv_page,
1558 bvec.bv_offset, bvec.bv_len,
1559 bio_iter_last(bvec, iter)
1569 struct bio_vec bvec; local
1573 bio_for_each_segment(bvec, bio, iter) {
1576 err = _drbd_send_page(peer_device, bvec.bv_page,
1577 bvec
[all...]
H A Ddrbd_receiver.c1708 struct bio_vec bvec; local
1731 bio_for_each_segment(bvec, bio, iter) {
1732 void *mapped = kmap(bvec.bv_page) + bvec.bv_offset;
1733 expect = min_t(int, data_size, bvec.bv_len);
1735 kunmap(bvec.bv_page);
/drivers/staging/lustre/lustre/llite/
H A Dlloop.c196 struct bio_vec bvec; local
221 bio_for_each_segment(bvec, bio, iter) {
222 BUG_ON(bvec.bv_offset != 0);
223 BUG_ON(bvec.bv_len != PAGE_CACHE_SIZE);
225 pages[page_count] = bvec.bv_page;
228 offset += bvec.bv_len;
/drivers/block/rsxx/
H A Ddma.c687 struct bio_vec bvec; local
726 bio_for_each_segment(bvec, bio, iter) {
727 bv_len = bvec.bv_len;
728 bv_off = bvec.bv_offset;
740 laddr, bvec.bv_page,
/drivers/block/zram/
H A Dzram_drv.c282 static inline int is_partial_io(struct bio_vec *bvec) argument
284 return bvec->bv_len != PAGE_SIZE;
350 static void update_position(u32 *index, int *offset, struct bio_vec *bvec) argument
352 if (*offset + bvec->bv_len >= PAGE_SIZE)
354 *offset = (*offset + bvec->bv_len) % PAGE_SIZE;
372 static void handle_zero_page(struct bio_vec *bvec) argument
374 struct page *page = bvec->bv_page;
378 if (is_partial_io(bvec))
379 memset(user_mem + bvec->bv_offset, 0, bvec
455 zram_bvec_read(struct zram *zram, struct bio_vec *bvec, u32 index, int offset, struct bio *bio) argument
520 zram_bvec_write(struct zram *zram, struct bio_vec *bvec, u32 index, int offset) argument
647 zram_bvec_rw(struct zram *zram, struct bio_vec *bvec, u32 index, int offset, struct bio *bio) argument
858 struct bio_vec bvec; local
[all...]
/drivers/scsi/mpt2sas/
H A Dmpt2sas_transport.c1919 struct bio_vec bvec; local
1959 bio_for_each_segment(bvec, req->bio, iter) {
1961 page_address(bvec.bv_page) + bvec.bv_offset,
1962 bvec.bv_len);
1963 offset += bvec.bv_len;
2110 bio_for_each_segment(bvec, rsp->bio, iter) {
2111 if (bytes_to_copy <= bvec.bv_len) {
2112 memcpy(page_address(bvec.bv_page) +
2113 bvec
[all...]
/drivers/scsi/mpt3sas/
H A Dmpt3sas_transport.c1901 struct bio_vec bvec; local
1942 bio_for_each_segment(bvec, req->bio, iter) {
1944 page_address(bvec.bv_page) + bvec.bv_offset,
1945 bvec.bv_len);
1946 offset += bvec.bv_len;
2071 bio_for_each_segment(bvec, rsp->bio, iter) {
2072 if (bytes_to_copy <= bvec.bv_len) {
2073 memcpy(page_address(bvec.bv_page) +
2074 bvec
[all...]

Completed in 625 milliseconds