Searched defs:sector (Results 1 - 25 of 60) sorted by path

123

/drivers/block/
H A Damiflop.c16 * - currently only 9 and 18 sector disks
732 printk (KERN_INFO "can't find sync for sector %d\n", scnt);
1049 "track %d, unit %d for sector %d\n",
1086 printk(KERN_INFO "dos_read: unknown sector len "
1095 "%d, unit %d for sector%d, disk sector %d\n",
1376 unsigned int cnt, block, track, sector; local
1397 printk("fd: sector %ld + %d requested for %s\n",
1408 sector = block % (floppy->dtype->sects * floppy->type->sect_mult);
1411 printk("access to track %d, sector
[all...]
H A Dbrd.c52 * Look up and return a brd's page for a given sector.
55 static struct page *brd_lookup_page(struct brd_device *brd, sector_t sector) argument
72 idx = sector >> PAGE_SECTORS_SHIFT; /* sector to page index */
82 * Look up and return a brd's page for a given sector.
86 static struct page *brd_insert_page(struct brd_device *brd, sector_t sector) argument
92 page = brd_lookup_page(brd, sector);
119 idx = sector >> PAGE_SECTORS_SHIFT;
134 static void brd_free_page(struct brd_device *brd, sector_t sector) argument
140 idx = sector >> PAGE_SECTORS_SHIF
147 brd_zero_page(struct brd_device *brd, sector_t sector) argument
196 copy_to_brd_setup(struct brd_device *brd, sector_t sector, size_t n) argument
212 discard_from_brd(struct brd_device *brd, sector_t sector, size_t n) argument
233 copy_to_brd(struct brd_device *brd, const void *src, sector_t sector, size_t n) argument
265 copy_from_brd(void *dst, struct brd_device *brd, sector_t sector, size_t n) argument
299 brd_do_bvec(struct brd_device *brd, struct page *page, unsigned int len, unsigned int off, int rw, sector_t sector) argument
332 sector_t sector; local
363 brd_rw_page(struct block_device *bdev, sector_t sector, struct page *page, int rw) argument
373 brd_direct_access(struct block_device *bdev, sector_t sector, void **kaddr, unsigned long *pfn) argument
[all...]
H A Dpktcdvd.c109 static sector_t get_zone(sector_t sector, struct pktcdvd_device *pd) argument
111 return (sector + pd->offset) & ~(sector_t)(pd->settings.size - 1);
639 * Find the first node in the pd->bio_queue rb tree with a starting sector >= s.
986 bio, (unsigned long long)pkt->sector,
1035 int first_frame = (bio->bi_iter.bi_sector - pkt->sector) /
1048 (unsigned long long)pkt->sector);
1063 bio->bi_iter.bi_sector = pkt->sector + f * (CD_FRAMESIZE >> 9);
1083 frames_read, (unsigned long long)pkt->sector);
1097 if (pkt->sector == zone || pkt->list.next == &pd->cdrw.pkt_free_list) {
1099 if (pkt->sector !
[all...]
H A Dswim.c36 unsigned char sector; member in struct:sector_header
461 int sector, unsigned char *buffer)
478 if (!ret && (header.sector == sector)) {
490 (header.sector != sector))
502 int side, track, sector; local
512 sector = x % fs->secpertrack + 1;
516 ret = swim_read_sector(fs, side, track, sector,
459 swim_read_sector(struct floppy_state *fs, int side, int track, int sector, unsigned char *buffer) argument
/drivers/block/drbd/
H A Ddrbd_actlog.c140 sector_t sector, int rw)
156 bio->bi_iter.bi_sector = sector;
191 sector_t sector, int rw)
200 (unsigned long long)sector, (rw & WRITE) ? "WRITE" : "READ",
203 if (sector < drbd_md_first_sector(bdev) ||
204 sector + 7 > drbd_md_last_sector(bdev))
207 (unsigned long long)sector, (rw & WRITE) ? "WRITE" : "READ");
209 err = _drbd_md_sync_page_io(device, bdev, sector, rw);
212 (unsigned long long)sector, (rw & WRITE) ? "WRITE" : "READ", err);
256 unsigned first = i->sector >> (AL_EXTENT_SHIF
138 _drbd_md_sync_page_io(struct drbd_device *device, struct drbd_backing_dev *bdev, sector_t sector, int rw) argument
190 drbd_md_sync_page_io(struct drbd_device *device, struct drbd_backing_dev *bdev, sector_t sector, int rw) argument
463 sector_t sector; local
829 __drbd_change_sync(struct drbd_device *device, sector_t sector, int size, enum update_sync_bits_mode mode, const char *file, const unsigned int line) argument
944 drbd_rs_begin_io(struct drbd_device *device, sector_t sector) argument
996 drbd_try_rs_begin_io(struct drbd_device *device, sector_t sector) argument
1118 drbd_rs_complete_io(struct drbd_device *device, sector_t sector) argument
[all...]
H A Ddrbd_interval.c25 sector_t max = node->sector + (node->size >> 9);
50 sector_t this_end = this->sector + (this->size >> 9);
61 if (this->sector < here->sector)
63 else if (this->sector > here->sector)
81 * @sector: start sector of @interval
84 * Returns if the tree contains the node @interval with start sector @start.
87 * sector numbe
90 drbd_contains_interval(struct rb_root *root, sector_t sector, struct drbd_interval *interval) argument
134 drbd_find_overlap(struct rb_root *root, sector_t sector, unsigned int size) argument
164 drbd_next_overlap(struct drbd_interval *i, sector_t sector, unsigned int size) argument
[all...]
H A Ddrbd_interval.h9 sector_t sector; /* start sector of the interval */ member in struct:drbd_interval
37 #define drbd_for_each_overlap(i, root, sector, size) \
38 for (i = drbd_find_overlap(root, sector, size); \
40 i = drbd_next_overlap(i, sector, size))
H A Ddrbd_main.c1311 * @sector: sector, needs to be in big endian byte order
1316 u64 sector, u32 blksize, u64 block_id)
1328 p->sector = sector;
1335 /* dp->sector and dp->block_id already/still in network byte order,
1343 _drbd_send_ack(peer_device, cmd, dp->sector, cpu_to_be32(data_size),
1350 _drbd_send_ack(peer_device, cmd, rp->sector, rp->blksize, rp->block_id);
1363 cpu_to_be64(peer_req->i.sector),
1371 sector_t sector, in
1315 _drbd_send_ack(struct drbd_peer_device *peer_device, enum drbd_packet cmd, u64 sector, u32 blksize, u64 block_id) argument
1370 drbd_send_ack_ex(struct drbd_peer_device *peer_device, enum drbd_packet cmd, sector_t sector, int blksize, u64 block_id) argument
1379 drbd_send_drequest(struct drbd_peer_device *peer_device, int cmd, sector_t sector, int size, u64 block_id) argument
1395 drbd_send_drequest_csum(struct drbd_peer_device *peer_device, sector_t sector, int size, void *digest, int digest_size, enum drbd_packet cmd) argument
1413 drbd_send_ov_request(struct drbd_peer_device *peer_device, sector_t sector, int size) argument
3070 sector_t sector; local
[all...]
H A Ddrbd_protocol.h121 u64 sector; /* 64 bits sector number */ member in struct:p_data
141 u64 sector; member in struct:p_block_ack
148 u64 sector; member in struct:p_block_req
271 u64 sector; member in struct:p_block_desc
H A Ddrbd_receiver.c341 drbd_alloc_peer_req(struct drbd_peer_device *peer_device, u64 id, sector_t sector,
369 peer_req->i.sector = sector;
1373 sector_t sector = peer_req->i.sector; local
1391 sector, data_size >> 9, GFP_NOIO))
1417 /* > peer_req->i.sector, unless this is the first bio */
1418 bio->bi_iter.bi_sector = sector;
1450 sector += len >> 9;
1587 read_in_block(struct drbd_peer_device *peer_device, u64 id, sector_t sector,
1705 recv_dless_read(struct drbd_peer_device *peer_device, struct drbd_request *req, sector_t sector, int data_size) argument
1763 sector_t sector = peer_req->i.sector; local
1822 find_request(struct drbd_device *device, struct rb_root *root, u64 id, sector_t sector, bool missing_ok, const char *func) argument
1843 sector_t sector; local
1877 sector_t sector; local
1908 restart_conflicting_writes(struct drbd_device *device, sector_t sector, int size) argument
1936 sector_t sector = peer_req->i.sector; local
2136 fail_postponed_requests(struct drbd_device *device, sector_t sector, unsigned int size) argument
2166 sector_t sector = peer_req->i.sector; local
2275 sector_t sector; local
2442 drbd_rs_should_slow_down(struct drbd_device *device, sector_t sector, bool throttle_if_app_is_waiting) argument
2516 sector_t sector; local
5165 sector_t sector = be64_to_cpu(p->sector); local
5191 validate_req_change_req_state(struct drbd_device *device, u64 id, sector_t sector, struct rb_root *root, const char *func, enum drbd_req_event what, bool missing_ok) argument
5217 sector_t sector = be64_to_cpu(p->sector); local
5263 sector_t sector = be64_to_cpu(p->sector); local
5299 sector_t sector = be64_to_cpu(p->sector); local
5320 sector_t sector; local
5382 sector_t sector; local
[all...]
H A Ddrbd_req.c34 static bool drbd_may_do_local_read(struct drbd_device *device, sector_t sector, int size);
80 req->i.sector = bio_src->bi_iter.bi_sector;
147 s, (unsigned long long)req->i.sector, req->i.size);
166 drbd_set_out_of_sync(device, req->i.sector, req->i.size);
169 drbd_set_in_sync(device, req->i.sector, req->i.size);
189 (unsigned long long) req->i.sector, req->i.size);
558 drbd_warn(device, "local %s IO error sector %llu+%u on %s\n",
560 (unsigned long long)req->i.sector,
654 drbd_set_out_of_sync(device, req->i.sector, req->i.size);
924 static bool drbd_may_do_local_read(struct drbd_device *device, sector_t sector, in argument
944 remote_due_to_read_balancing(struct drbd_device *device, sector_t sector, enum drbd_read_balancing rbm) argument
989 sector_t sector = req->i.sector; local
[all...]
H A Ddrbd_worker.c158 drbd_rs_complete_io(device, i.sector);
185 (unsigned long long)peer_req->i.sector);
190 (unsigned long long)peer_req->i.sector);
360 sector_t sector = peer_req->i.sector; local
371 err = drbd_send_drequest_csum(peer_device, sector, size,
391 static int read_for_csum(struct drbd_peer_device *peer_device, sector_t sector, int size) argument
401 peer_req = drbd_alloc_peer_req(peer_device, ID_SYNCER /* unused */, sector,
591 sector_t sector; local
650 sector
758 sector_t sector; local
1174 sector_t sector = peer_req->i.sector; local
1215 drbd_ov_out_of_sync_found(struct drbd_device *device, sector_t sector, int size) argument
1233 sector_t sector = peer_req->i.sector; local
[all...]
/drivers/block/mtip32xx/
H A Dmtip32xx.h225 unsigned char sector; member in union:host_to_dev_fis::__anon336
406 * Pointer to the beginning of a sector buffer that is used
411 * Pointer to the beginning of a sector buffer that is used
/drivers/char/
H A Dps3flash.c39 int tag; /* Start sector of buffer, -1 if invalid */
110 u64 size, sector, offset; local
130 sector = *pos / dev->bounce_size * priv->chunk_sectors;
140 res = ps3flash_fetch(dev, sector);
163 sector += priv->chunk_sectors;
179 u64 size, sector, offset; local
199 sector = *pos / dev->bounce_size * priv->chunk_sectors;
210 res = ps3flash_fetch(dev, sector);
211 else if (sector != priv->tag)
231 priv->tag = sector;
[all...]
/drivers/ide/
H A Dide-cd.c148 unsigned long sector; local
181 sector = (sense->information[0] << 24) |
187 /* device sector size is 2K */
188 sector <<= 2;
191 sector &= ~(bio_sectors - 1);
201 if (sector < get_capacity(info->disk) &&
202 drive->probed_capacity - sector < 4 * 75)
203 set_capacity(info->disk, sector);
355 * No point in re-trying a zillion times on a bad sector.
360 "(bad sector)", sta
[all...]
H A Dide-floppy.c192 unsigned long sector)
195 int block = sector / floppy->bs_factor;
346 "%d sector size, %d rpm\n",
405 "%d sector size",
428 "sector size\n",
190 idefloppy_create_rw_cmd(ide_drive_t *drive, struct ide_atapi_pc *pc, struct request *rq, unsigned long sector) argument
H A Dide-gd.c160 struct request *rq, sector_t sector)
162 return drive->disk_ops->do_request(drive, rq, sector);
159 ide_gd_do_request(ide_drive_t *drive, struct request *rq, sector_t sector) argument
/drivers/infiniband/ulp/iser/
H A Discsi_iser.c354 * @sector: error sector if exsists (output)
361 * In addition the error sector is marked.
364 iscsi_iser_check_protection(struct iscsi_task *task, sector_t *sector) argument
370 sector);
373 sector);
H A Diser_verbs.c1265 enum iser_data_dir cmd_dir, sector_t *sector)
1286 *sector = scsi_get_lba(iser_task->sc) + sector_off;
1288 pr_err("PI error found type %d at sector %llx "
1291 (unsigned long long)*sector,
1264 iser_check_task_pi_status(struct iscsi_iser_task *iser_task, enum iser_data_dir cmd_dir, sector_t *sector) argument
/drivers/md/
H A Dbitmap.c1463 sector_t sector = 0; local
1467 while (sector < bitmap->mddev->resync_max_sectors) {
1468 bitmap_end_sync(bitmap, sector, &blocks, 0);
1469 sector += blocks;
1474 void bitmap_cond_end_sync(struct bitmap *bitmap, sector_t sector) argument
1481 if (sector == 0) {
1491 bitmap->mddev->curr_resync_completed = sector;
1493 sector &= ~((1ULL << bitmap->counts.chunkshift) - 1);
1495 while (s < sector && s < bitmap->mddev->resync_max_sectors) {
1706 sector_t sector local
[all...]
H A Ddm-crypt.c60 sector_t sector; member in struct:dm_crypt_io
194 * plain: the initial vector is the 32-bit little-endian version of the sector
197 * plain64: the initial vector is the 64-bit little-endian version of the sector
200 * essiv: "encrypted sector|salt initial vector", the sector number is
214 * with an IV derived from the sector number, the data and
217 * of sector must be tweaked according to decrypted data.
228 * with an IV derived from initial key and the sector number.
229 * In addition, whitening value is applied on every sector, whitening
230 * is calculated from initial key, sector numbe
677 u64 sector = cpu_to_le64((u64)dmreq->iv_sector); local
716 u64 sector = cpu_to_le64((u64)dmreq->iv_sector); local
797 crypt_convert_init(struct crypt_config *cc, struct convert_context *ctx, struct bio *bio_out, struct bio *bio_in, sector_t sector) argument
1016 crypt_io_init(struct dm_crypt_io *io, struct crypt_config *cc, struct bio *bio, sector_t sector) argument
1199 sector_t sector = io->sector; local
[all...]
H A Ddm-exception-store.h197 sector_t sector)
199 return sector >> store->chunk_shift;
196 sector_to_chunk(struct dm_exception_store *store, sector_t sector) argument
H A Ddm-raid1.c269 io[i].sector = 0;
339 from.sector = m->offset + dm_rh_region_to_sector(ms->rh, key);
358 dest->sector = m->offset + dm_rh_region_to_sector(ms->rh, key);
407 static struct mirror *choose_mirror(struct mirror_set *ms, sector_t sector) argument
460 io->sector = map_sector(m, bio);
H A Ddm-region-hash.c116 static region_t dm_rh_sector_to_region(struct dm_region_hash *rh, sector_t sector) argument
118 return sector >> rh->region_shift;
H A Ddm-snap.c595 * Return the exception data for a sector, or NULL if not
881 sector_t sector, unsigned chunk_size);
950 dest.sector = chunk_to_sector(s->store, old_chunk);
951 dest.count = min(io_size, get_dev_size(dest.bdev) - dest.sector);
954 src.sector = chunk_to_sector(s->store, new_chunk);
967 while (origin_write_extent(s, dest.sector, io_size)) {
1525 src.sector = chunk_to_sector(s->store, pe->e.old_chunk);
1526 src.count = min((sector_t)s->store->chunk_size, dev_size - src.sector);
1529 dest.sector = chunk_to_sector(s->store, pe->e.new_chunk);
1970 static int __origin_write(struct list_head *snapshots, sector_t sector, argument
2110 origin_write_extent(struct dm_snapshot *merging_snap, sector_t sector, unsigned size) argument
[all...]

Completed in 349 milliseconds

123