/external/qemu/block/ |
H A D | nbd.c | 96 static int nbd_read(BlockDriverState *bs, int64_t sector_num, argument 105 request.from = sector_num * 512;; 126 static int nbd_write(BlockDriverState *bs, int64_t sector_num, argument 135 request.from = sector_num * 512;;
|
H A D | cloop.c | 131 static int cloop_read(BlockDriverState *bs, int64_t sector_num, argument 138 uint32_t sector_offset_in_block=((sector_num+i)%s->sectors_per_block), 139 block_num=(sector_num+i)/s->sectors_per_block;
|
H A D | parallels.c | 105 static int64_t seek_to_sector(BlockDriverState *bs, int64_t sector_num) argument 110 index = sector_num / s->tracks; 111 offset = sector_num % s->tracks; 119 static int parallels_read(BlockDriverState *bs, int64_t sector_num, argument 123 int64_t position = seek_to_sector(bs, sector_num); 131 sector_num++;
|
H A D | bochs.c | 158 static int64_t seek_to_sector(BlockDriverState *bs, int64_t sector_num) argument 161 int64_t offset = sector_num * 512; 189 static int bochs_read(BlockDriverState *bs, int64_t sector_num, argument 195 int64_t block_offset = seek_to_sector(bs, sector_num); 204 sector_num++;
|
H A D | cow.c | 133 static int cow_is_allocated(BlockDriverState *bs, int64_t sector_num, argument 143 changed = is_bit_set(bs, sector_num); 149 if (is_bit_set(bs, sector_num + *num_same) != changed) 156 static int cow_update_bitmap(BlockDriverState *bs, int64_t sector_num, argument 163 error = cow_set_bit(bs, sector_num + i); 172 static int cow_read(BlockDriverState *bs, int64_t sector_num, argument 179 if (cow_is_allocated(bs, sector_num, nb_sectors, &n)) { 181 s->cow_sectors_offset + sector_num * 512, 188 ret = bdrv_read(bs->backing_hd, sector_num, buf, n); 196 sector_num 202 cow_write(BlockDriverState *bs, int64_t sector_num, const uint8_t *buf, int nb_sectors) argument [all...] |
H A D | dmg.c | 186 uint32_t chunk_num,int sector_num) 188 if(chunk_num>=s->n_chunks || s->sectors[chunk_num]>sector_num || 189 s->sectors[chunk_num]+s->sectorcounts[chunk_num]<=sector_num) 195 static inline uint32_t search_chunk(BDRVDMGState* s,int sector_num) argument 201 if(s->sectors[chunk3]>sector_num) 203 else if(s->sectors[chunk3]+s->sectorcounts[chunk3]>sector_num) 211 static inline int dmg_read_chunk(BlockDriverState *bs, int sector_num) argument 215 if(!is_sector_in_chunk(s,s->current_chunk,sector_num)) { 217 uint32_t chunk = search_chunk(s,sector_num); 267 static int dmg_read(BlockDriverState *bs, int64_t sector_num, argument 185 is_sector_in_chunk(BDRVDMGState* s, uint32_t chunk_num,int sector_num) argument [all...] |
H A D | raw-win32.c | 108 static int raw_read(BlockDriverState *bs, int64_t sector_num, argument 115 int64_t offset = sector_num * 512; 129 static int raw_write(BlockDriverState *bs, int64_t sector_num, argument 136 int64_t offset = sector_num * 512;
|
H A D | raw.c | 35 static int check_write_unsafe(BlockDriverState *bs, int64_t sector_num, argument 44 if (sector_num == 0 && nb_sectors > 0) { 51 static int raw_read(BlockDriverState *bs, int64_t sector_num, argument 54 return bdrv_read(bs->file, sector_num, buf, nb_sectors); 69 static int raw_write(BlockDriverState *bs, int64_t sector_num, argument 72 if (check_write_unsafe(bs, sector_num, buf, nb_sectors)) { 88 return bdrv_write(bs->file, sector_num, buf, nb_sectors); 92 int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, 95 return bdrv_aio_readv(bs->file, sector_num, qiov, nb_sectors, cb, opaque); 120 int64_t sector_num, QEMUIOVecto 91 raw_aio_readv(BlockDriverState *bs, int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, BlockDriverCompletionFunc *cb, void *opaque) argument 119 raw_aio_writev(BlockDriverState *bs, int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, BlockDriverCompletionFunc *cb, void *opaque) argument [all...] |
H A D | qcow2-cluster.c | 322 void qcow2_encrypt_sectors(BDRVQcowState *s, int64_t sector_num, argument 334 ivec.ll[0] = cpu_to_le64(sector_num); 338 sector_num++; 345 static int qcow_read(BlockDriverState *bs, int64_t sector_num, argument 355 ret = qcow2_get_cluster_offset(bs, sector_num << 9, &n, 361 index_in_cluster = sector_num & (s->cluster_sectors - 1); 365 n1 = qcow2_backing_read1(bs->backing_hd, sector_num, buf, n); 368 ret = bdrv_read(bs->backing_hd, sector_num, buf, n1); 385 qcow2_encrypt_sectors(s, sector_num, buf, buf, n, 0, 390 sector_num [all...] |
H A D | raw-posix.c | 385 static int raw_read(BlockDriverState *bs, int64_t sector_num, argument 390 ret = raw_pread(bs, sector_num * BDRV_SECTOR_SIZE, buf, 475 static int raw_write(BlockDriverState *bs, int64_t sector_num, argument 479 ret = raw_pwrite(bs, sector_num * BDRV_SECTOR_SIZE, buf, 503 int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, 521 return laio_submit(bs, s->aio_ctx, s->fd, sector_num, qiov, 527 return paio_submit(bs, s->fd, sector_num, qiov, nb_sectors, 532 int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, 535 return raw_aio_submit(bs, sector_num, qiov, nb_sectors, 540 int64_t sector_num, QEMUIOVecto 502 raw_aio_submit(BlockDriverState *bs, int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, BlockDriverCompletionFunc *cb, void *opaque, int type) argument 531 raw_aio_readv(BlockDriverState *bs, int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, BlockDriverCompletionFunc *cb, void *opaque) argument 539 raw_aio_writev(BlockDriverState *bs, int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, BlockDriverCompletionFunc *cb, void *opaque) argument [all...] |
H A D | vpc.c | 238 int64_t sector_num, int write) 241 uint64_t offset = sector_num * 512; 268 // sector_num, pagetable_index, pageentry_index, 328 static int64_t alloc_block(BlockDriverState* bs, int64_t sector_num) argument 336 // Check if sector_num is valid 337 if ((sector_num < 0) || (sector_num > bs->total_sectors)) 341 index = (sector_num * 512) / s->block_size; 365 return get_sector_offset(bs, sector_num, 0); 372 static int vpc_read(BlockDriverState *bs, int64_t sector_num, argument 237 get_sector_offset(BlockDriverState *bs, int64_t sector_num, int write) argument 406 vpc_write(BlockDriverState *bs, int64_t sector_num, const uint8_t *buf, int nb_sectors) argument [all...] |
H A D | qcow.c | 217 static void encrypt_sectors(BDRVQcowState *s, int64_t sector_num, argument 229 ivec.ll[0] = cpu_to_le64(sector_num); 233 sector_num++; 379 static int qcow_is_allocated(BlockDriverState *bs, int64_t sector_num, argument 386 cluster_offset = get_cluster_offset(bs, sector_num << 9, 0, 0, 0, 0); 387 index_in_cluster = sector_num & (s->cluster_sectors - 1); 446 static int qcow_read(BlockDriverState *bs, int64_t sector_num, 454 cluster_offset = get_cluster_offset(bs, sector_num << 9, 0, 0, 0, 0); 455 index_in_cluster = sector_num & (s->cluster_sectors - 1); 462 ret = bdrv_read(bs->backing_hd, sector_num, bu 491 int64_t sector_num; member in struct:QCowAIOCB 517 qcow_aio_setup(BlockDriverState *bs, int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, BlockDriverCompletionFunc *cb, void *opaque, int is_write) argument 633 qcow_aio_readv(BlockDriverState *bs, int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, BlockDriverCompletionFunc *cb, void *opaque) argument 715 qcow_aio_writev(BlockDriverState *bs, int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, BlockDriverCompletionFunc *cb, void *opaque) argument 855 qcow_write_compressed(BlockDriverState *bs, int64_t sector_num, const uint8_t *buf, int nb_sectors) argument [all...] |
H A D | qcow2.c | 296 static int qcow_is_allocated(BlockDriverState *bs, int64_t sector_num, argument 305 ret = qcow2_get_cluster_offset(bs, sector_num << 9, pnum, &cluster_offset); 315 int64_t sector_num, uint8_t *buf, int nb_sectors) 318 if ((sector_num + nb_sectors) <= bs->total_sectors) 320 if (sector_num >= bs->total_sectors) 323 n1 = bs->total_sectors - sector_num; 330 int64_t sector_num; member in struct:QCowAIOCB 400 qcow2_encrypt_sectors(s, acb->sector_num, acb->buf, acb->buf, 407 acb->sector_num += acb->cur_nr_sectors; 418 ret = qcow2_get_cluster_offset(bs, acb->sector_num << 314 qcow2_backing_read1(BlockDriverState *bs, int64_t sector_num, uint8_t *buf, int nb_sectors) argument 492 qcow_aio_setup(BlockDriverState *bs, int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, BlockDriverCompletionFunc *cb, void *opaque, int is_write) argument 519 qcow_aio_readv(BlockDriverState *bs, int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, BlockDriverCompletionFunc *cb, void *opaque) argument 644 qcow_aio_writev(BlockDriverState *bs, int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, BlockDriverCompletionFunc *cb, void *opaque) argument 1149 qcow_write_compressed(BlockDriverState *bs, int64_t sector_num, const uint8_t *buf, int nb_sectors) argument [all...] |
H A D | vmdk.c | 570 static int vmdk_is_allocated(BlockDriverState *bs, int64_t sector_num, argument 577 cluster_offset = get_cluster_offset(bs, NULL, sector_num << 9, 0); 578 index_in_cluster = sector_num % s->cluster_sectors; 586 static int vmdk_read(BlockDriverState *bs, int64_t sector_num, argument 594 cluster_offset = get_cluster_offset(bs, NULL, sector_num << 9, 0); 595 index_in_cluster = sector_num % s->cluster_sectors; 604 ret = bdrv_read(bs->backing_hd, sector_num, buf, n); 615 sector_num += n; 621 static int vmdk_write(BlockDriverState *bs, int64_t sector_num, argument 630 if (sector_num > b [all...] |
H A D | vvfat.c | 817 static inline uint32_t sector2cluster(BDRVVVFATState* s,off_t sector_num) argument 819 return (sector_num-s->faked_sectors)/s->sectors_per_cluster; 827 static inline uint32_t sector_offset_in_cluster(BDRVVVFATState* s,off_t sector_num) argument 829 return (sector_num-s->first_sectors_number-2*s->sectors_per_fat)%s->sectors_per_cluster; 1288 static int vvfat_read(BlockDriverState *bs, int64_t sector_num, argument 1294 for(i=0;i<nb_sectors;i++,sector_num++) { 1295 if (sector_num >= s->sector_count) 1300 sector_num, nb_sectors-i, &n)) { 1301 DLOG(fprintf(stderr, "sectors %d+%d allocated\n", (int)sector_num, n)); 1302 if (s->qcow->drv->bdrv_read(s->qcow, sector_num, bu 2660 vvfat_write(BlockDriverState *bs, int64_t sector_num, const uint8_t *buf, int nb_sectors) argument 2752 vvfat_is_allocated(BlockDriverState *bs, int64_t sector_num, int nb_sectors, int* n) argument 2764 write_target_commit(BlockDriverState *bs, int64_t sector_num, const uint8_t* buffer, int nb_sectors) argument [all...] |
/external/qemu/ |
H A D | dma-helpers.c | 44 uint64_t sector_num; member in struct:__anon9401 89 dbs->sector_num += dbs->iov.size / 512; 120 dbs->acb = bdrv_aio_writev(dbs->bs, dbs->sector_num, &dbs->iov, 123 dbs->acb = bdrv_aio_readv(dbs->bs, dbs->sector_num, &dbs->iov, 148 BlockDriverState *bs, QEMUSGList *sg, uint64_t sector_num, 157 dbs->sector_num = sector_num; 147 dma_bdrv_io( BlockDriverState *bs, QEMUSGList *sg, uint64_t sector_num, BlockDriverCompletionFunc *cb, void *opaque, int is_write) argument
|
H A D | posix-aio-compat.c | 569 int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, 587 acb->aio_offset = sector_num * 512; 592 //trace_paio_submit(acb, opaque, sector_num, nb_sectors, type); 568 paio_submit(BlockDriverState *bs, int fd, int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, BlockDriverCompletionFunc *cb, void *opaque, int type) argument
|
H A D | block.c | 46 int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, 49 int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, 55 static int bdrv_read_em(BlockDriverState *bs, int64_t sector_num, 57 static int bdrv_write_em(BlockDriverState *bs, int64_t sector_num, 896 static int bdrv_check_request(BlockDriverState *bs, int64_t sector_num, argument 899 return bdrv_check_byte_request(bs, sector_num * BDRV_SECTOR_SIZE, 904 int bdrv_read(BlockDriverState *bs, int64_t sector_num, argument 911 if (bdrv_check_request(bs, sector_num, nb_sectors)) 914 return drv->bdrv_read(bs, sector_num, buf, nb_sectors); 917 static void set_dirty_bitmap(BlockDriverState *bs, int64_t sector_num, argument 951 bdrv_write(BlockDriverState *bs, int64_t sector_num, const uint8_t *buf, int nb_sectors) argument 978 int64_t sector_num; local 1023 int64_t sector_num; local 1097 bdrv_write_sync(BlockDriverState *bs, int64_t sector_num, const uint8_t *buf, int nb_sectors) argument 1498 bdrv_is_allocated(BlockDriverState *bs, int64_t sector_num, int nb_sectors, int *pnum) argument 1726 bdrv_write_compressed(BlockDriverState *bs, int64_t sector_num, const uint8_t *buf, int nb_sectors) argument 1977 bdrv_aio_readv(BlockDriverState *bs, int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, BlockDriverCompletionFunc *cb, void *opaque) argument 2001 bdrv_aio_writev(BlockDriverState *bs, int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, BlockDriverCompletionFunc *cb, void *opaque) argument 2317 bdrv_aio_rw_vector(BlockDriverState *bs, int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, BlockDriverCompletionFunc *cb, void *opaque, int is_write) argument 2348 bdrv_aio_readv_em(BlockDriverState *bs, int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, BlockDriverCompletionFunc *cb, void *opaque) argument 2355 bdrv_aio_writev_em(BlockDriverState *bs, int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, BlockDriverCompletionFunc *cb, void *opaque) argument 2410 bdrv_read_em(BlockDriverState *bs, int64_t sector_num, uint8_t *buf, int nb_sectors) argument 2441 bdrv_write_em(BlockDriverState *bs, int64_t sector_num, const uint8_t *buf, int nb_sectors) argument [all...] |
/external/grub/stage2/ |
H A D | disk_io.c | 264 int sector_num = sector; local 268 (*disk_read_func) (sector_num++, byte_offset, length); 274 (*disk_read_func) (sector_num++, 0, buf_geom.sector_size); 277 (*disk_read_func) (sector_num, 0, length);
|