/fs/xfs/libxfs/ |
H A D | xfs_alloc_btree.h | 41 #define XFS_ALLOC_REC_ADDR(mp, block, index) \ 45 (((index) - 1) * sizeof(xfs_alloc_rec_t)))) 47 #define XFS_ALLOC_KEY_ADDR(mp, block, index) \ 51 ((index) - 1) * sizeof(xfs_alloc_key_t))) 53 #define XFS_ALLOC_PTR_ADDR(mp, block, index, maxrecs) \ 58 ((index) - 1) * sizeof(xfs_alloc_ptr_t)))
|
H A D | xfs_ialloc_btree.h | 41 #define XFS_INOBT_REC_ADDR(mp, block, index) \ 45 (((index) - 1) * sizeof(xfs_inobt_rec_t)))) 47 #define XFS_INOBT_KEY_ADDR(mp, block, index) \ 51 ((index) - 1) * sizeof(xfs_inobt_key_t))) 53 #define XFS_INOBT_PTR_ADDR(mp, block, index, maxrecs) \ 58 ((index) - 1) * sizeof(xfs_inobt_ptr_t)))
|
H A D | xfs_bmap_btree.h | 42 #define XFS_BMBT_REC_ADDR(mp, block, index) \ 46 ((index) - 1) * sizeof(xfs_bmbt_rec_t))) 48 #define XFS_BMBT_KEY_ADDR(mp, block, index) \ 52 ((index) - 1) * sizeof(xfs_bmbt_key_t))) 54 #define XFS_BMBT_PTR_ADDR(mp, block, index, maxrecs) \ 59 ((index) - 1) * sizeof(xfs_bmbt_ptr_t))) 61 #define XFS_BMDR_REC_ADDR(block, index) \ 65 ((index) - 1) * sizeof(xfs_bmdr_rec_t))) 67 #define XFS_BMDR_KEY_ADDR(block, index) \ 71 ((index) [all...] |
/fs/fscache/ |
H A D | histogram.c | 29 unsigned long index; local 40 index = (unsigned long) v - 3; 41 n[0] = atomic_read(&fscache_obj_instantiate_histogram[index]); 42 n[1] = atomic_read(&fscache_ops_histogram[index]); 43 n[2] = atomic_read(&fscache_objs_histogram[index]); 44 n[3] = atomic_read(&fscache_retrieval_delay_histogram[index]); 45 n[4] = atomic_read(&fscache_retrieval_histogram[index]); 49 t = (index * 1000) / HZ; 52 index, t, n[0], n[1], n[2], n[3], n[4]);
|
/fs/squashfs/ |
H A D | namei.c | 42 * Directories are sorted, and can contain a directory index to speed up 44 * storing the index/filename mapping to the first directory header 46 * and at lookup the index is scanned linearly looking for the first filename 49 * The general idea of the index is ensure only one metadata block needs to be 69 * Lookup name in the directory index, returning the location of the metadata 70 * block containing it, and the directory index this represents. 72 * If we get an error reading the index then return the part of the index 73 * (if any) we have managed to read - the index isn't essential, just 84 struct squashfs_dir_index *index; local [all...] |
H A D | symlink.c | 51 int index = page->index << PAGE_CACHE_SHIFT; local 54 int length = min_t(int, i_size_read(inode) - index, PAGE_CACHE_SIZE); 59 TRACE("Entered squashfs_symlink_readpage, page index %ld, start block " 60 "%llx, offset %x\n", page->index, block, offset); 63 * Skip index bytes into symlink metadata. 65 if (index) { 67 index);
|
H A D | block.c | 90 int squashfs_read_data(struct super_block *sb, u64 index, int length, argument 95 int offset = index & ((1 << msblk->devblksize_log2) - 1); 96 u64 cur_index = index >> msblk->devblksize_log2; 112 *next_index = index + length; 115 index, compressed ? "" : "un", length, output->length); 118 (index + length) > msblk->bytes_used) 132 if ((index + 2) > msblk->bytes_used) 144 *next_index = index + length + 2; 146 TRACE("Block @ 0x%llx, %scompressed size %d\n", index, 150 (index [all...] |
H A D | file.c | 32 * larger), the code implements an index cache that caches the mapping from 33 * block index to datablock location on disk. 35 * The index cache allows Squashfs to handle large files (up to 1.75 TiB) while 39 * The index cache is designed to be memory efficient, and by default uses 57 * Locate cache slot in range [offset, index] for specified inode. If 58 * there's more than one return the slot closest to index. 61 int index) 69 TRACE("locate_meta_index: index %d, offset %d\n", index, offset); 77 msblk->meta_index[i].offset <= index 60 locate_meta_index(struct inode *inode, int offset, int index) argument 236 fill_meta_index(struct inode *inode, int index, u64 *index_block, int *index_offset, u64 *data_block) argument 335 read_blocklist(struct inode *inode, int index, u64 *block) argument 441 squashfs_readpage_sparse(struct page *page, int index, int file_end) argument 457 int index = page->index >> (msblk->block_log - PAGE_CACHE_SHIFT); local [all...] |
H A D | xattr.h | 39 unsigned int index, int *count, unsigned int *size, 38 squashfs_xattr_lookup(struct super_block *sb, unsigned int index, int *count, unsigned int *size, unsigned long long *xattr) argument
|
H A D | id.c | 29 * stored compressed into metadata blocks. A second index table is used to 30 * locate these. This second index table for speed of access (and because it 43 * Map uid/gid index into real 32-bit uid/gid using the id look up table 45 int squashfs_get_id(struct super_block *sb, unsigned int index, argument 49 int block = SQUASHFS_ID_BLOCK(index); 50 int offset = SQUASHFS_ID_BLOCK_OFFSET(index);
|
H A D | xattr_id.c | 41 int squashfs_xattr_lookup(struct super_block *sb, unsigned int index, argument 45 int block = SQUASHFS_XATTR_BLOCK(index); 46 int offset = SQUASHFS_XATTR_BLOCK_OFFSET(index);
|
/fs/cachefiles/ |
H A D | proc.c | 26 unsigned long index; local 37 index = (unsigned long) v - 3; 38 x = atomic_read(&cachefiles_lookup_histogram[index]); 39 y = atomic_read(&cachefiles_mkdir_histogram[index]); 40 z = atomic_read(&cachefiles_create_histogram[index]); 44 t = (index * 1000) / HZ; 46 seq_printf(m, "%4lu 0.%03u %9u %9u %9u\n", index, t, x, y, z);
|
/fs/ntfs/ |
H A D | aops.h | 49 * @index: index into the page cache for @mapping of the page to map 52 * @index, where @index is in units of PAGE_CACHE_SIZE, and not in bytes. 87 unsigned long index) 89 struct page *page = read_mapping_page(mapping, index, NULL); 86 ntfs_map_page(struct address_space *mapping, unsigned long index) argument
|
H A D | index.c | 2 * index.c - NTFS kernel index handling. Part of the Linux-NTFS project. 27 #include "index.h" 31 * ntfs_index_ctx_get - allocate and initialize a new index context 32 * @idx_ni: ntfs index inode with which to initialize the context 34 * Allocate a new index context, initialize it with @idx_ni and return it. 37 * Locking: Caller must hold i_mutex on the index inode. 50 * ntfs_index_ctx_put - release an index context 51 * @ictx: index context to free 53 * Release the index contex [all...] |
H A D | debug.c | 152 int index = -lcn - 1; local 154 if (index > -LCN_ENOENT - 1) 155 index = 3; 157 (long long)(rl + i)->vcn, lcn_str[index],
|
H A D | dir.c | 73 * - Each page cache page in the index allocation mapping must be 113 /* Find the index root attribute in the mft record. */ 125 /* Get to the index root value (it's been verified in read_inode). */ 128 index_end = (u8*)&ir->index + le32_to_cpu(ir->index.index_length); 129 /* The first index entry. */ 130 ie = (INDEX_ENTRY*)((u8*)&ir->index + 131 le32_to_cpu(ir->index.entries_offset)); 253 * is definitely no such name in this index but we might need to 282 * We have finished with this index withou [all...] |
H A D | bitmap.c | 51 pgoff_t index, end_index; local 70 index = start_bit >> (3 + PAGE_CACHE_SHIFT); 75 page = ntfs_map_page(mapping, index); 120 while (index < end_index) { 123 /* Update @index and get the next page. */ 127 page = ntfs_map_page(mapping, ++index);
|
/fs/udf/ |
H A D | unicode.c | 411 int index, newIndex = 0, needsCRC = 0; local 422 for (index = 0; index < udfLen; index++) { 423 curr = udfName[index]; 427 while (index + 1 < udfLen && 428 (udfName[index + 1] == '/' || 429 udfName[index + 1] == 0)) 430 index++; 433 (udfLen - index [all...] |
/fs/jfs/ |
H A D | jfs_dtree.c | 37 * each directory page maintains a sorted entry index table 38 * which stores the start slot index of sorted entries 117 s16 index; member in struct:dtsplit 185 static void dtInsertEntry(dtpage_t * p, int index, struct component_name * key, 203 * Reads a page of a directory's index table. 245 * index and pointer to its entry. 249 static struct dir_table_slot *find_index(struct inode *ip, u32 index, argument 259 if (index < 2) { 261 jfs_warn("find_entry called with index = %d", index); 304 lock_index(tid_t tid, struct inode *ip, struct metapage * mp, u32 index) argument 341 u32 index; local 496 free_index(tid_t tid, struct inode *ip, u32 index, u32 next) argument 524 modify_index(tid_t tid, struct inode *ip, u32 index, s64 bn, int slot, struct metapage ** mp, s64 *lblock) argument 549 read_index(struct inode *ip, u32 index, struct dir_table_slot * dirtab_slot) argument 589 int base, index, lim; local 828 int index; local 2089 int index; local 2243 int index, nextindex; local 2452 int index; local 2923 uint index; local 3013 s16 index; member in struct:dtoffset 3019 int index; local 3422 s16 index; member in struct:dtoffset 3428 int index; local 3907 dtInsertEntry(dtpage_t * p, int index, struct component_name * key, ddata_t * data, struct dt_lock ** dtlock) argument 4528 int index; local [all...] |
H A D | ioctl.c | 42 int index=0; local 45 while (jfs_map[index].jfs_flag) { 47 if (jfs_map[index].ext2_flag & flags) 48 mapped |= jfs_map[index].jfs_flag; 50 if (jfs_map[index].jfs_flag & flags) 51 mapped |= jfs_map[index].ext2_flag; 53 index++;
|
H A D | jfs_xtree.c | 93 s16 index; member in struct:xtsplit 155 int index; local 185 * and returned the index of the next entry. 188 XT_GETSEARCH(ip, btstack.top, bn, mp, p, index); 203 xad = &p->xad[index]; 235 * btstack contains (bn, index) of search path traversed to the entry. 249 int base, index, lim, btindex; local 271 * greater than K so that the returned index is the position of 292 (index = jfs_ip->btindex) < 294 xad = &p->xad[index]; 546 int index, nextindex; local 1370 int index, nextindex, len; local 1533 int index, nextindex, llen, rlen; local 1712 int index0, index, newindex, nextindex; local 2179 int index, nextindex; local 2347 int index, nextindex; local 2416 int index, nextindex; local 2579 int index; local 2901 int base, index, lim; local 3169 int index, nextindex; local 3723 int index; local [all...] |
/fs/nilfs2/ |
H A D | page.c | 43 __nilfs_get_page_block(struct page *page, unsigned long block, pgoff_t index, argument 53 first_block = (unsigned long)index << (PAGE_CACHE_SHIFT - blkbits); 67 pgoff_t index = blkoff >> (PAGE_CACHE_SHIFT - blkbits); local 71 page = grab_cache_page(mapping, index); 75 bh = __nilfs_get_page_block(page, blkoff, index, blkbits, b_state); 183 printk(KERN_CRIT "NILFS_PAGE_BUG(%p): cnt=%d index#=%llu flags=0x%lx " 186 (unsigned long long)page->index, page->flags, m, ino); 262 pgoff_t index = 0; local 267 if (!pagevec_lookup_tag(&pvec, smap, &index, PAGECACHE_TAG_DIRTY, 278 dpage = grab_cache_page(dmap, page->index); 317 pgoff_t index = 0; local 383 pgoff_t index = 0; local 522 pgoff_t index; local [all...] |
/fs/xfs/ |
H A D | xfs_attr.h | 28 * then that int is used as the index into the Btree. Since the hashval 91 * and an index, return a pointer to the indicated attribute in the buffer. 93 #define ATTR_ENTRY(buffer, index) \ 95 &((char *)buffer)[ ((attrlist_t *)(buffer))->al_offset[index] ]) 131 int index; /* index into output buffer */ member in struct:xfs_attr_list_context
|
/fs/ |
H A D | filesystems.c | 129 int err, index; local 138 for (tmp=file_systems, index=0 ; tmp ; tmp=tmp->next, index++) { 140 err = index; 149 static int fs_name(unsigned int index, char __user * buf) argument 155 for (tmp = file_systems; tmp; tmp = tmp->next, index--) 156 if (index <= 0 && try_module_get(tmp->owner)) 172 int index; local 175 for (tmp = file_systems, index = 0 ; tmp ; tmp = tmp->next, index [all...] |
/fs/ecryptfs/ |
H A D | mmap.c | 47 struct page *ecryptfs_get_locked_page(struct inode *inode, loff_t index) argument 49 struct page *page = read_mapping_page(inode->i_mapping, index, NULL); 72 "page (upper index [0x%.16lx])\n", page->index); 131 loff_t view_extent_num = ((((loff_t)page->index) 201 rc = ecryptfs_read_lower_page_segment(page, page->index, 0, 219 page, page->index, 0, PAGE_CACHE_SIZE, 240 ecryptfs_printk(KERN_DEBUG, "Unlocking page with index = [0x%.16lx]\n", 241 page->index); 254 if ((i_size_read(inode) / PAGE_CACHE_SIZE) != page->index) 283 pgoff_t index = pos >> PAGE_CACHE_SHIFT; local 483 pgoff_t index = pos >> PAGE_CACHE_SHIFT; local [all...] |