/drivers/block/ |
H A D | amiflop.c | 16 * - currently only 9 and 18 sector disks 732 printk (KERN_INFO "can't find sync for sector %d\n", scnt); 1049 "track %d, unit %d for sector %d\n", 1086 printk(KERN_INFO "dos_read: unknown sector len " 1095 "%d, unit %d for sector%d, disk sector %d\n", 1376 unsigned int cnt, block, track, sector; local 1397 printk("fd: sector %ld + %d requested for %s\n", 1408 sector = block % (floppy->dtype->sects * floppy->type->sect_mult); 1411 printk("access to track %d, sector [all...] |
H A D | brd.c | 52 * Look up and return a brd's page for a given sector. 55 static struct page *brd_lookup_page(struct brd_device *brd, sector_t sector) argument 72 idx = sector >> PAGE_SECTORS_SHIFT; /* sector to page index */ 82 * Look up and return a brd's page for a given sector. 86 static struct page *brd_insert_page(struct brd_device *brd, sector_t sector) argument 92 page = brd_lookup_page(brd, sector); 119 idx = sector >> PAGE_SECTORS_SHIFT; 134 static void brd_free_page(struct brd_device *brd, sector_t sector) argument 140 idx = sector >> PAGE_SECTORS_SHIF 147 brd_zero_page(struct brd_device *brd, sector_t sector) argument 196 copy_to_brd_setup(struct brd_device *brd, sector_t sector, size_t n) argument 212 discard_from_brd(struct brd_device *brd, sector_t sector, size_t n) argument 233 copy_to_brd(struct brd_device *brd, const void *src, sector_t sector, size_t n) argument 265 copy_from_brd(void *dst, struct brd_device *brd, sector_t sector, size_t n) argument 299 brd_do_bvec(struct brd_device *brd, struct page *page, unsigned int len, unsigned int off, int rw, sector_t sector) argument 332 sector_t sector; local 363 brd_rw_page(struct block_device *bdev, sector_t sector, struct page *page, int rw) argument 373 brd_direct_access(struct block_device *bdev, sector_t sector, void **kaddr, unsigned long *pfn) argument [all...] |
H A D | pktcdvd.c | 109 static sector_t get_zone(sector_t sector, struct pktcdvd_device *pd) argument 111 return (sector + pd->offset) & ~(sector_t)(pd->settings.size - 1); 639 * Find the first node in the pd->bio_queue rb tree with a starting sector >= s. 986 bio, (unsigned long long)pkt->sector, 1035 int first_frame = (bio->bi_iter.bi_sector - pkt->sector) / 1048 (unsigned long long)pkt->sector); 1063 bio->bi_iter.bi_sector = pkt->sector + f * (CD_FRAMESIZE >> 9); 1083 frames_read, (unsigned long long)pkt->sector); 1097 if (pkt->sector == zone || pkt->list.next == &pd->cdrw.pkt_free_list) { 1099 if (pkt->sector ! [all...] |
H A D | swim.c | 36 unsigned char sector; member in struct:sector_header 461 int sector, unsigned char *buffer) 478 if (!ret && (header.sector == sector)) { 490 (header.sector != sector)) 502 int side, track, sector; local 512 sector = x % fs->secpertrack + 1; 516 ret = swim_read_sector(fs, side, track, sector, 459 swim_read_sector(struct floppy_state *fs, int side, int track, int sector, unsigned char *buffer) argument
|
/drivers/block/drbd/ |
H A D | drbd_actlog.c | 140 sector_t sector, int rw) 156 bio->bi_iter.bi_sector = sector; 191 sector_t sector, int rw) 200 (unsigned long long)sector, (rw & WRITE) ? "WRITE" : "READ", 203 if (sector < drbd_md_first_sector(bdev) || 204 sector + 7 > drbd_md_last_sector(bdev)) 207 (unsigned long long)sector, (rw & WRITE) ? "WRITE" : "READ"); 209 err = _drbd_md_sync_page_io(device, bdev, sector, rw); 212 (unsigned long long)sector, (rw & WRITE) ? "WRITE" : "READ", err); 256 unsigned first = i->sector >> (AL_EXTENT_SHIF 138 _drbd_md_sync_page_io(struct drbd_device *device, struct drbd_backing_dev *bdev, sector_t sector, int rw) argument 190 drbd_md_sync_page_io(struct drbd_device *device, struct drbd_backing_dev *bdev, sector_t sector, int rw) argument 463 sector_t sector; local 829 __drbd_change_sync(struct drbd_device *device, sector_t sector, int size, enum update_sync_bits_mode mode, const char *file, const unsigned int line) argument 944 drbd_rs_begin_io(struct drbd_device *device, sector_t sector) argument 996 drbd_try_rs_begin_io(struct drbd_device *device, sector_t sector) argument 1118 drbd_rs_complete_io(struct drbd_device *device, sector_t sector) argument [all...] |
H A D | drbd_interval.c | 25 sector_t max = node->sector + (node->size >> 9); 50 sector_t this_end = this->sector + (this->size >> 9); 61 if (this->sector < here->sector) 63 else if (this->sector > here->sector) 81 * @sector: start sector of @interval 84 * Returns if the tree contains the node @interval with start sector @start. 87 * sector numbe 90 drbd_contains_interval(struct rb_root *root, sector_t sector, struct drbd_interval *interval) argument 134 drbd_find_overlap(struct rb_root *root, sector_t sector, unsigned int size) argument 164 drbd_next_overlap(struct drbd_interval *i, sector_t sector, unsigned int size) argument [all...] |
H A D | drbd_interval.h | 9 sector_t sector; /* start sector of the interval */ member in struct:drbd_interval 37 #define drbd_for_each_overlap(i, root, sector, size) \ 38 for (i = drbd_find_overlap(root, sector, size); \ 40 i = drbd_next_overlap(i, sector, size))
|
H A D | drbd_main.c | 1311 * @sector: sector, needs to be in big endian byte order 1316 u64 sector, u32 blksize, u64 block_id) 1328 p->sector = sector; 1335 /* dp->sector and dp->block_id already/still in network byte order, 1343 _drbd_send_ack(peer_device, cmd, dp->sector, cpu_to_be32(data_size), 1350 _drbd_send_ack(peer_device, cmd, rp->sector, rp->blksize, rp->block_id); 1363 cpu_to_be64(peer_req->i.sector), 1371 sector_t sector, in 1315 _drbd_send_ack(struct drbd_peer_device *peer_device, enum drbd_packet cmd, u64 sector, u32 blksize, u64 block_id) argument 1370 drbd_send_ack_ex(struct drbd_peer_device *peer_device, enum drbd_packet cmd, sector_t sector, int blksize, u64 block_id) argument 1379 drbd_send_drequest(struct drbd_peer_device *peer_device, int cmd, sector_t sector, int size, u64 block_id) argument 1395 drbd_send_drequest_csum(struct drbd_peer_device *peer_device, sector_t sector, int size, void *digest, int digest_size, enum drbd_packet cmd) argument 1413 drbd_send_ov_request(struct drbd_peer_device *peer_device, sector_t sector, int size) argument 3070 sector_t sector; local [all...] |
H A D | drbd_protocol.h | 121 u64 sector; /* 64 bits sector number */ member in struct:p_data 141 u64 sector; member in struct:p_block_ack 148 u64 sector; member in struct:p_block_req 271 u64 sector; member in struct:p_block_desc
|
H A D | drbd_receiver.c | 341 drbd_alloc_peer_req(struct drbd_peer_device *peer_device, u64 id, sector_t sector, 369 peer_req->i.sector = sector; 1373 sector_t sector = peer_req->i.sector; local 1391 sector, data_size >> 9, GFP_NOIO)) 1417 /* > peer_req->i.sector, unless this is the first bio */ 1418 bio->bi_iter.bi_sector = sector; 1450 sector += len >> 9; 1587 read_in_block(struct drbd_peer_device *peer_device, u64 id, sector_t sector, 1705 recv_dless_read(struct drbd_peer_device *peer_device, struct drbd_request *req, sector_t sector, int data_size) argument 1763 sector_t sector = peer_req->i.sector; local 1822 find_request(struct drbd_device *device, struct rb_root *root, u64 id, sector_t sector, bool missing_ok, const char *func) argument 1843 sector_t sector; local 1877 sector_t sector; local 1908 restart_conflicting_writes(struct drbd_device *device, sector_t sector, int size) argument 1936 sector_t sector = peer_req->i.sector; local 2136 fail_postponed_requests(struct drbd_device *device, sector_t sector, unsigned int size) argument 2166 sector_t sector = peer_req->i.sector; local 2275 sector_t sector; local 2442 drbd_rs_should_slow_down(struct drbd_device *device, sector_t sector, bool throttle_if_app_is_waiting) argument 2516 sector_t sector; local 5165 sector_t sector = be64_to_cpu(p->sector); local 5191 validate_req_change_req_state(struct drbd_device *device, u64 id, sector_t sector, struct rb_root *root, const char *func, enum drbd_req_event what, bool missing_ok) argument 5217 sector_t sector = be64_to_cpu(p->sector); local 5263 sector_t sector = be64_to_cpu(p->sector); local 5299 sector_t sector = be64_to_cpu(p->sector); local 5320 sector_t sector; local 5382 sector_t sector; local [all...] |
H A D | drbd_req.c | 34 static bool drbd_may_do_local_read(struct drbd_device *device, sector_t sector, int size); 80 req->i.sector = bio_src->bi_iter.bi_sector; 147 s, (unsigned long long)req->i.sector, req->i.size); 166 drbd_set_out_of_sync(device, req->i.sector, req->i.size); 169 drbd_set_in_sync(device, req->i.sector, req->i.size); 189 (unsigned long long) req->i.sector, req->i.size); 558 drbd_warn(device, "local %s IO error sector %llu+%u on %s\n", 560 (unsigned long long)req->i.sector, 654 drbd_set_out_of_sync(device, req->i.sector, req->i.size); 924 static bool drbd_may_do_local_read(struct drbd_device *device, sector_t sector, in argument 944 remote_due_to_read_balancing(struct drbd_device *device, sector_t sector, enum drbd_read_balancing rbm) argument 989 sector_t sector = req->i.sector; local [all...] |
H A D | drbd_worker.c | 158 drbd_rs_complete_io(device, i.sector); 185 (unsigned long long)peer_req->i.sector); 190 (unsigned long long)peer_req->i.sector); 360 sector_t sector = peer_req->i.sector; local 371 err = drbd_send_drequest_csum(peer_device, sector, size, 391 static int read_for_csum(struct drbd_peer_device *peer_device, sector_t sector, int size) argument 401 peer_req = drbd_alloc_peer_req(peer_device, ID_SYNCER /* unused */, sector, 591 sector_t sector; local 650 sector 758 sector_t sector; local 1174 sector_t sector = peer_req->i.sector; local 1215 drbd_ov_out_of_sync_found(struct drbd_device *device, sector_t sector, int size) argument 1233 sector_t sector = peer_req->i.sector; local [all...] |
/drivers/block/mtip32xx/ |
H A D | mtip32xx.h | 225 unsigned char sector; member in union:host_to_dev_fis::__anon336 406 * Pointer to the beginning of a sector buffer that is used 411 * Pointer to the beginning of a sector buffer that is used
|
/drivers/char/ |
H A D | ps3flash.c | 39 int tag; /* Start sector of buffer, -1 if invalid */ 110 u64 size, sector, offset; local 130 sector = *pos / dev->bounce_size * priv->chunk_sectors; 140 res = ps3flash_fetch(dev, sector); 163 sector += priv->chunk_sectors; 179 u64 size, sector, offset; local 199 sector = *pos / dev->bounce_size * priv->chunk_sectors; 210 res = ps3flash_fetch(dev, sector); 211 else if (sector != priv->tag) 231 priv->tag = sector; [all...] |
/drivers/ide/ |
H A D | ide-cd.c | 148 unsigned long sector; local 181 sector = (sense->information[0] << 24) | 187 /* device sector size is 2K */ 188 sector <<= 2; 191 sector &= ~(bio_sectors - 1); 201 if (sector < get_capacity(info->disk) && 202 drive->probed_capacity - sector < 4 * 75) 203 set_capacity(info->disk, sector); 355 * No point in re-trying a zillion times on a bad sector. 360 "(bad sector)", sta [all...] |
H A D | ide-floppy.c | 192 unsigned long sector) 195 int block = sector / floppy->bs_factor; 346 "%d sector size, %d rpm\n", 405 "%d sector size", 428 "sector size\n", 190 idefloppy_create_rw_cmd(ide_drive_t *drive, struct ide_atapi_pc *pc, struct request *rq, unsigned long sector) argument
|
H A D | ide-gd.c | 160 struct request *rq, sector_t sector) 162 return drive->disk_ops->do_request(drive, rq, sector); 159 ide_gd_do_request(ide_drive_t *drive, struct request *rq, sector_t sector) argument
|
/drivers/infiniband/ulp/iser/ |
H A D | iscsi_iser.c | 354 * @sector: error sector if exsists (output) 361 * In addition the error sector is marked. 364 iscsi_iser_check_protection(struct iscsi_task *task, sector_t *sector) argument 370 sector); 373 sector);
|
H A D | iser_verbs.c | 1265 enum iser_data_dir cmd_dir, sector_t *sector) 1286 *sector = scsi_get_lba(iser_task->sc) + sector_off; 1288 pr_err("PI error found type %d at sector %llx " 1291 (unsigned long long)*sector, 1264 iser_check_task_pi_status(struct iscsi_iser_task *iser_task, enum iser_data_dir cmd_dir, sector_t *sector) argument
|
/drivers/md/ |
H A D | bitmap.c | 1463 sector_t sector = 0; local 1467 while (sector < bitmap->mddev->resync_max_sectors) { 1468 bitmap_end_sync(bitmap, sector, &blocks, 0); 1469 sector += blocks; 1474 void bitmap_cond_end_sync(struct bitmap *bitmap, sector_t sector) argument 1481 if (sector == 0) { 1491 bitmap->mddev->curr_resync_completed = sector; 1493 sector &= ~((1ULL << bitmap->counts.chunkshift) - 1); 1495 while (s < sector && s < bitmap->mddev->resync_max_sectors) { 1706 sector_t sector local [all...] |
H A D | dm-crypt.c | 60 sector_t sector; member in struct:dm_crypt_io 194 * plain: the initial vector is the 32-bit little-endian version of the sector 197 * plain64: the initial vector is the 64-bit little-endian version of the sector 200 * essiv: "encrypted sector|salt initial vector", the sector number is 214 * with an IV derived from the sector number, the data and 217 * of sector must be tweaked according to decrypted data. 228 * with an IV derived from initial key and the sector number. 229 * In addition, whitening value is applied on every sector, whitening 230 * is calculated from initial key, sector numbe 677 u64 sector = cpu_to_le64((u64)dmreq->iv_sector); local 716 u64 sector = cpu_to_le64((u64)dmreq->iv_sector); local 797 crypt_convert_init(struct crypt_config *cc, struct convert_context *ctx, struct bio *bio_out, struct bio *bio_in, sector_t sector) argument 1016 crypt_io_init(struct dm_crypt_io *io, struct crypt_config *cc, struct bio *bio, sector_t sector) argument 1199 sector_t sector = io->sector; local [all...] |
H A D | dm-exception-store.h | 197 sector_t sector) 199 return sector >> store->chunk_shift; 196 sector_to_chunk(struct dm_exception_store *store, sector_t sector) argument
|
H A D | dm-raid1.c | 269 io[i].sector = 0; 339 from.sector = m->offset + dm_rh_region_to_sector(ms->rh, key); 358 dest->sector = m->offset + dm_rh_region_to_sector(ms->rh, key); 407 static struct mirror *choose_mirror(struct mirror_set *ms, sector_t sector) argument 460 io->sector = map_sector(m, bio);
|
H A D | dm-region-hash.c | 116 static region_t dm_rh_sector_to_region(struct dm_region_hash *rh, sector_t sector) argument 118 return sector >> rh->region_shift;
|
H A D | dm-snap.c | 595 * Return the exception data for a sector, or NULL if not 881 sector_t sector, unsigned chunk_size); 950 dest.sector = chunk_to_sector(s->store, old_chunk); 951 dest.count = min(io_size, get_dev_size(dest.bdev) - dest.sector); 954 src.sector = chunk_to_sector(s->store, new_chunk); 967 while (origin_write_extent(s, dest.sector, io_size)) { 1525 src.sector = chunk_to_sector(s->store, pe->e.old_chunk); 1526 src.count = min((sector_t)s->store->chunk_size, dev_size - src.sector); 1529 dest.sector = chunk_to_sector(s->store, pe->e.new_chunk); 1970 static int __origin_write(struct list_head *snapshots, sector_t sector, argument 2110 origin_write_extent(struct dm_snapshot *merging_snap, sector_t sector, unsigned size) argument [all...] |