/fs/jffs2/ |
H A D | xattr.c | 424 * verify_xattr_ref(c, ref) 427 * save_xattr_ref(c, ref) 432 * delete_xattr_ref(c, ref) 444 static int verify_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref) argument 454 if (ref_flags(ref->node) != REF_UNCHECKED) 456 offset = ref_offset(ref->node); 482 ref->ino = je32_to_cpu(rr.ino); 483 ref->xid = je32_to_cpu(rr.xid); 484 ref->xseqno = je32_to_cpu(rr.xseqno); 485 if (ref 506 save_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref) argument 554 struct jffs2_xattr_ref *ref; local 576 delete_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref) argument 600 struct jffs2_xattr_ref *ref, *_ref; local 618 struct jffs2_xattr_ref *ref, *_ref; local 640 struct jffs2_xattr_ref *ref, *cmp, **pref, **pcmp; local 743 struct jffs2_xattr_ref *ref, *_ref; local 777 struct jffs2_xattr_ref *ref, *_ref; local 967 struct jffs2_xattr_ref *ref, **pref; local 1031 struct jffs2_xattr_ref *ref, **pref; local 1093 struct jffs2_xattr_ref *ref, *newref, **pref; local 1248 jffs2_garbage_collect_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref, struct jffs2_raw_node_ref *raw) argument 1326 jffs2_release_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref) argument [all...] |
H A D | malloc.c | 213 struct jffs2_raw_node_ref **p, *ref; local 219 ref = *p; 224 if (ref && ref->flash_offset != REF_EMPTY_NODE) 225 ref++; 228 if (!ref) { 230 ref = *p = jffs2_alloc_refblock(); 231 if (!ref) 234 if (ref->flash_offset == REF_LINK_NODE) { 235 p = &ref 308 struct jffs2_xattr_ref *ref; local 319 jffs2_free_xattr_ref(struct jffs2_xattr_ref *ref) argument [all...] |
H A D | readinode.c | 33 struct jffs2_raw_node_ref *ref = tn->fn->raw; local 43 ofs = ref_offset(ref) + sizeof(struct jffs2_raw_inode); 53 ref_offset(ref), tn->csize, ofs); 62 ref_offset(ref), tn->csize, tn->partial_crc, tn->data_crc, ofs - len, ofs, len); 109 ref_offset(ref), tn->data_crc, crc); 114 jeb = &c->blocks[ref->flash_offset / c->sector_size]; 115 len = ref_totlen(c, jeb, ref); 120 ref->flash_offset |= REF_PRISTINE; 567 /* Returns first valid node after 'ref'. May return 'ref' */ 568 jffs2_first_valid_node(struct jffs2_raw_node_ref *ref) argument 586 read_direntry(struct jffs2_sb_info *c, struct jffs2_raw_node_ref *ref, struct jffs2_raw_dirent *rd, size_t read, struct jffs2_readinode_info *rii) argument 693 read_dnode(struct jffs2_sb_info *c, struct jffs2_raw_node_ref *ref, struct jffs2_raw_inode *rd, int rdlen, struct jffs2_readinode_info *rii) argument 868 read_unknown(struct jffs2_sb_info *c, struct jffs2_raw_node_ref *ref, struct jffs2_unknown_node *un) argument 920 read_more(struct jffs2_sb_info *c, struct jffs2_raw_node_ref *ref, int needed_len, int *rdlen, unsigned char *buf) argument 964 struct jffs2_raw_node_ref *ref, *valid_ref; local [all...] |
H A D | nodelist.c | 590 struct jffs2_raw_node_ref *ref; local 595 ref = jeb->last_node; 597 dbg_noderef("Last node at %p is (%08x,%p)\n", ref, ref->flash_offset, 598 ref->next_in_ino); 600 while (ref->flash_offset != REF_EMPTY_NODE) { 601 if (ref->flash_offset == REF_LINK_NODE) 602 ref = ref->next_in_ino; 604 ref 692 __ref_totlen(struct jffs2_sb_info *c, struct jffs2_eraseblock *jeb, struct jffs2_raw_node_ref *ref) argument 718 __jffs2_ref_totlen(struct jffs2_sb_info *c, struct jffs2_eraseblock *jeb, struct jffs2_raw_node_ref *ref) argument [all...] |
H A D | nodelist.h | 101 static inline struct jffs2_raw_node_ref *ref_next(struct jffs2_raw_node_ref *ref) argument 103 ref++; 106 if (ref->flash_offset == REF_LINK_NODE) { 107 ref = ref->next_in_ino; 108 if (!ref) 109 return ref; 113 if (ref->flash_offset == REF_EMPTY_NODE) 116 return ref; 136 #define ref_flags(ref) ((re [all...] |
H A D | xattr.h | 66 static inline int is_xattr_ref_dead(struct jffs2_xattr_ref *ref) argument 68 return ((ref->xseqno & XREF_DELETE_MARKER) != 0); 86 extern int jffs2_garbage_collect_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref, 90 extern void jffs2_release_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref);
|
H A D | nodemgmt.c | 579 void jffs2_mark_node_obsolete(struct jffs2_sb_info *c, struct jffs2_raw_node_ref *ref) argument 588 if(unlikely(!ref)) { 592 if (ref_obsolete(ref)) { 594 __func__, ref_offset(ref)); 597 blocknr = ref->flash_offset / c->sector_size; 600 ref->flash_offset); 618 freed_len = ref_totlen(c, jeb, ref); 620 if (ref_flags(ref) == REF_UNCHECKED) { 624 ref->flash_offset, jeb->used_size); 628 ref_offset(ref), freed_le [all...] |
H A D | erase.c | 236 struct jffs2_raw_node_ref *ref, struct jffs2_eraseblock *jeb) 241 prev = &ref->next_in_ino; 262 if (this == ref) 318 struct jffs2_raw_node_ref *block, *ref; local 322 block = ref = jeb->first_node; 324 while (ref) { 325 if (ref->flash_offset == REF_LINK_NODE) { 326 ref = ref->next_in_ino; 328 block = ref; 235 jffs2_remove_node_refs_from_ino_list(struct jffs2_sb_info *c, struct jffs2_raw_node_ref *ref, struct jffs2_eraseblock *jeb) argument [all...] |
H A D | scan.c | 387 struct jffs2_xattr_ref *ref; local 409 ref = jffs2_alloc_xattr_ref(); 410 if (!ref) 415 * ref->xid is used to store 32bit xid, xd is not used 416 * ref->ino is used to store 32bit inode-number, ic is not used 418 * are exclusive. In a similar way, ref->next is temporarily 422 ref->ino = je32_to_cpu(rr->ino); 423 ref->xid = je32_to_cpu(rr->xid); 424 ref->xseqno = je32_to_cpu(rr->xseqno); 425 if (ref [all...] |
/fs/btrfs/ |
H A D | delayed-ref.c | 23 #include "delayed-ref.h" 135 * insert a new ref into the rbtree. This returns any existing refs 168 /* insert a new ref to head ref rbtree */ 199 * find an head entry based on bytenr. This returns the delayed ref 264 struct btrfs_delayed_ref_node *ref) 266 if (btrfs_delayed_ref_is_head(ref)) { 267 head = btrfs_delayed_node_to_head(ref); 271 rb_erase(&ref->rb_node, &head->ref_root); 273 ref 261 drop_delayed_ref(struct btrfs_trans_handle *trans, struct btrfs_delayed_ref_root *delayed_refs, struct btrfs_delayed_ref_head *head, struct btrfs_delayed_ref_node *ref) argument 280 merge_ref(struct btrfs_trans_handle *trans, struct btrfs_delayed_ref_root *delayed_refs, struct btrfs_delayed_ref_head *head, struct btrfs_delayed_ref_node *ref, u64 seq) argument 359 struct btrfs_delayed_ref_node *ref; local 496 struct btrfs_delayed_ref_head *ref; local 554 add_delayed_ref_head(struct btrfs_fs_info *fs_info, struct btrfs_trans_handle *trans, struct btrfs_delayed_ref_node *ref, u64 bytenr, u64 num_bytes, int action, int is_data) argument 637 add_delayed_tree_ref(struct btrfs_fs_info *fs_info, struct btrfs_trans_handle *trans, struct btrfs_delayed_ref_head *head_ref, struct btrfs_delayed_ref_node *ref, u64 bytenr, u64 num_bytes, u64 parent, u64 ref_root, int level, int action, int no_quota) argument 699 add_delayed_data_ref(struct btrfs_fs_info *fs_info, struct btrfs_trans_handle *trans, struct btrfs_delayed_ref_head *head_ref, struct btrfs_delayed_ref_node *ref, u64 bytenr, u64 num_bytes, u64 parent, u64 ref_root, u64 owner, u64 offset, int action, int no_quota) argument 772 struct btrfs_delayed_tree_ref *ref; local 821 struct btrfs_delayed_data_ref *ref; local [all...] |
H A D | inode-item.c | 29 struct btrfs_inode_ref *ref; local 40 ref = (struct btrfs_inode_ref *)(ptr + cur_offset); 41 len = btrfs_inode_ref_name_len(leaf, ref); 42 name_ptr = (unsigned long)(ref + 1); 43 cur_offset += len + sizeof(*ref); 47 *ref_ret = ref; 74 * we'll return success and the inode ref object. 172 * Common case only one ref in the item, remove the 200 struct btrfs_inode_ref *ref; local 208 int del_len = name_len + sizeof(*ref); 333 struct btrfs_inode_ref *ref; local [all...] |
H A D | backref.c | 25 #include "delayed-ref.h" 95 * from the shared data ref, we only have the leaf but we need 199 struct __prelim_ref *ref; local 204 ref = kmem_cache_alloc(btrfs_prelim_ref_cache, gfp_mask); 205 if (!ref) 208 ref->root_id = root_id; 210 ref->key_for_search = *key; 212 memset(&ref->key_for_search, 0, sizeof(ref->key_for_search)); 214 ref 224 add_all_parents(struct btrfs_root *root, struct btrfs_path *path, struct ulist *parents, struct __prelim_ref *ref, int level, u64 time_seq, const u64 *extent_item_pos, u64 total_refs) argument 308 __resolve_indirect_ref(struct btrfs_fs_info *fs_info, struct btrfs_path *path, u64 time_seq, struct __prelim_ref *ref, struct ulist *parents, const u64 *extent_item_pos, u64 total_refs) argument 388 struct __prelim_ref *ref; local 484 struct __prelim_ref *ref; local 612 struct btrfs_delayed_tree_ref *ref; local 621 struct btrfs_delayed_tree_ref *ref; local 631 struct btrfs_delayed_data_ref *ref; local 653 struct btrfs_delayed_data_ref *ref; local 899 struct __prelim_ref *ref; local [all...] |
H A D | delayed-ref.h | 25 #define BTRFS_UPDATE_DELAYED_HEAD 4 /* not changing ref count on head ref */ 39 /* ref count on this data structure */ 49 * parent to have more than one ref on an extent 94 * until the delayed ref is processed. must_insert_reserved is 95 * used to flag a delayed ref so the accounting can be updated 124 /* head ref rbtree */ 130 /* how many delayed ref updates we've queued, used by the 172 static inline void btrfs_put_delayed_ref(struct btrfs_delayed_ref_node *ref) argument 174 WARN_ON(atomic_read(&ref [all...] |
H A D | root-tree.c | 355 struct btrfs_root_ref *ref; local 374 ref = btrfs_item_ptr(leaf, path->slots[0], 377 WARN_ON(btrfs_root_ref_dirid(leaf, ref) != dirid); 378 WARN_ON(btrfs_root_ref_name_len(leaf, ref) != name_len); 379 ptr = (unsigned long)(ref + 1); 381 *sequence = btrfs_root_ref_sequence(leaf, ref); 411 * For a forward ref, the root_id is the id of the tree referencing 414 * For a back ref the root_id is the id of the subvol or snapshot and 427 struct btrfs_root_ref *ref; local 440 sizeof(*ref) [all...] |
H A D | print-tree.c | 47 struct btrfs_extent_data_ref *ref) 51 btrfs_extent_data_ref_root(eb, ref), 52 btrfs_extent_data_ref_objectid(eb, ref), 53 btrfs_extent_data_ref_offset(eb, ref), 54 btrfs_extent_data_ref_count(eb, ref)); 144 printk("\t\textent back ref root %llu gen %llu " 46 print_extent_data_ref(struct extent_buffer *eb, struct btrfs_extent_data_ref *ref) argument
|
H A D | Makefile | 10 compression.o delayed-ref.o relocation.o delayed-inode.o scrub.o \
|
H A D | export.c | 159 struct btrfs_root_ref *ref; local 199 ref = btrfs_item_ptr(leaf, path->slots[0], 201 key.objectid = btrfs_root_ref_dirid(leaf, ref);
|
H A D | extent-tree.c | 736 * the head node for delayed ref is used to store the sum of all the 937 * The extent ref structure for the implicit back refs has fields for: 947 * The extent ref structure for the full back refs has field for: 1077 struct btrfs_extent_data_ref *ref) 1079 return hash_extent_data_ref(btrfs_extent_data_ref_root(leaf, ref), 1080 btrfs_extent_data_ref_objectid(leaf, ref), 1081 btrfs_extent_data_ref_offset(leaf, ref)); 1085 struct btrfs_extent_data_ref *ref, 1088 if (btrfs_extent_data_ref_root(leaf, ref) != root_objectid || 1089 btrfs_extent_data_ref_objectid(leaf, ref) ! 1076 hash_extent_data_ref_item(struct extent_buffer *leaf, struct btrfs_extent_data_ref *ref) argument 1084 match_extent_data_ref(struct extent_buffer *leaf, struct btrfs_extent_data_ref *ref, u64 root_objectid, u64 owner, u64 offset) argument 1103 struct btrfs_extent_data_ref *ref; local 1213 struct btrfs_shared_data_ref *ref; local 1224 struct btrfs_extent_data_ref *ref; local 2064 struct btrfs_delayed_data_ref *ref; local 2224 struct btrfs_delayed_tree_ref *ref; local 2329 struct btrfs_delayed_ref_node *ref, *last = NULL;; local 2358 struct btrfs_delayed_ref_node *ref; local 2812 struct btrfs_delayed_ref_node *ref; local 2876 struct btrfs_delayed_ref_node *ref; local 2944 struct btrfs_extent_data_ref *ref; local 7029 struct btrfs_shared_data_ref *ref; local 7034 struct btrfs_extent_data_ref *ref; local [all...] |
/fs/cifs/ |
H A D | cifs_dfs_ref.c | 127 * @ref: server's referral 138 const struct dfs_info3_param *ref, 153 if (strlen(fullpath) - ref->path_consumed) 154 prepath = fullpath + ref->path_consumed; 156 *devname = cifs_build_devname(ref->node_name, prepath); 216 /* copy new IP and ref share name */ 241 * @ref: server's referral 244 const char *fullpath, const struct dfs_info3_param *ref) 252 fullpath + 1, ref, &devname); 264 static void dump_referral(const struct dfs_info3_param *ref) argument 136 cifs_compose_mount_options(const char *sb_mountdata, const char *fullpath, const struct dfs_info3_param *ref, char **devname) argument 243 cifs_dfs_do_refmount(struct cifs_sb_info *cifs_sb, const char *fullpath, const struct dfs_info3_param *ref) argument [all...] |
/fs/nfs/ |
H A D | pnfs_dev.c | 87 if (atomic_read(&d->ref)) 179 atomic_inc(&d->ref); 208 atomic_inc(&new->ref); 241 /* balance the initial ref set in pnfs_insert_deviceid */ 242 if (atomic_dec_and_test(&d->ref)) 257 atomic_set(&d->ref, 1); 268 * Note that since the test for d->ref == 0 is sufficient to establish 274 if (!atomic_dec_and_test(&d->ref)) 314 if (d->nfs_client == clp && atomic_read(&d->ref)) { 328 if (atomic_dec_and_test(&d->ref)) [all...] |
H A D | callback_proc.c | 375 struct referring_call *ref; local 392 ref = &rclist->rcl_refcalls[j]; 400 ref->rc_sequenceid, ref->rc_slotid); 403 status = (test_bit(ref->rc_slotid, tbl->used_slots) && 404 tbl->slots[ref->rc_slotid].seq_nr == 405 ref->rc_sequenceid);
|
/fs/ubifs/ |
H A D | log.c | 180 struct ubifs_ref_node *ref; local 185 ref = kzalloc(c->ref_node_alsz, GFP_NOFS); 186 if (!ref) { 240 ref->ch.node_type = UBIFS_REF_NODE; 241 ref->lnum = cpu_to_le32(bud->lnum); 242 ref->offs = cpu_to_le32(bud->start); 243 ref->jhead = cpu_to_le32(jhead); 271 dbg_log("write ref LEB %d:%d", 273 err = ubifs_write_node(c, ref, UBIFS_REF_NODE_SZ, c->lhead_lnum, 283 kfree(ref); 363 struct ubifs_ref_node *ref; local 657 struct ubifs_ref_node *ref = snod->node; local [all...] |
H A D | replay.c | 778 * @ref: the reference node to validate 786 static int validate_ref(struct ubifs_info *c, const struct ubifs_ref_node *ref) argument 789 int lnum = le32_to_cpu(ref->lnum); 790 unsigned int offs = le32_to_cpu(ref->offs); 791 unsigned int jhead = le32_to_cpu(ref->jhead); 794 * ref->offs may point to the end of LEB when the journal head points 917 const struct ubifs_ref_node *ref = snod->node; local 919 err = validate_ref(c, ref); 925 err = add_replay_bud(c, le32_to_cpu(ref->lnum), 926 le32_to_cpu(ref [all...] |
/fs/quota/ |
H A D | quota_tree.c | 296 __le32 *ref; local 316 ref = (__le32 *)buf; 317 newblk = le32_to_cpu(ref[get_index(info, dquot->dq_id, depth)]); 325 le32_to_cpu(ref[get_index(info, 336 ref[get_index(info, dquot->dq_id, depth)] = 466 __le32 *ref = (__le32 *)buf; local 476 newblk = le32_to_cpu(ref[get_index(info, dquot->dq_id, depth)]); 485 ref[get_index(info, dquot->dq_id, depth)] = cpu_to_le32(0); 487 for (i = 0; i < (info->dqi_usable_bs >> 2) && !ref[i]; i++) 562 __le32 *ref local [all...] |
/fs/xfs/libxfs/ |
H A D | xfs_sb.c | 119 int ref = 0; local 125 ref = atomic_inc_return(&pag->pag_ref); 128 trace_xfs_perag_get(mp, agno, ref, _RET_IP_); 143 int ref; local 152 ref = atomic_inc_return(&pag->pag_ref); 154 trace_xfs_perag_get_tag(mp, pag->pag_agno, ref, _RET_IP_); 162 int ref; local 165 ref = atomic_dec_return(&pag->pag_ref); 166 trace_xfs_perag_put(pag->pag_mount, pag->pag_agno, ref, _RET_IP_);
|