/lib/ |
H A D | idr.c | 84 * @gfp_mask: allocation mask 87 * If @layer_idr is %NULL, directly allocate one using @gfp_mask or fetch 95 static struct idr_layer *idr_layer_alloc(gfp_t gfp_mask, struct idr *layer_idr) argument 110 new = kmem_cache_zalloc(idr_layer_cache, gfp_mask | __GFP_NOWARN); 135 return kmem_cache_zalloc(idr_layer_cache, gfp_mask); 193 static int __idr_pre_get(struct idr *idp, gfp_t gfp_mask) argument 197 new = kmem_cache_zalloc(idr_layer_cache, gfp_mask); 210 * @gfp_mask: allocation mask for idr_layer_alloc() 222 gfp_t gfp_mask, struct idr *layer_idr) 274 new = idr_layer_alloc(gfp_mask, layer_id 221 sub_alloc(struct idr *idp, int *starting_id, struct idr_layer **pa, gfp_t gfp_mask, struct idr *layer_idr) argument 290 idr_get_empty_slot(struct idr *idp, int starting_id, struct idr_layer **pa, gfp_t gfp_mask, struct idr *layer_idr) argument 396 idr_preload(gfp_t gfp_mask) argument 451 idr_alloc(struct idr *idr, void *ptr, int start, int end, gfp_t gfp_mask) argument 489 idr_alloc_cyclic(struct idr *idr, void *ptr, int start, int end, gfp_t gfp_mask) argument 897 ida_pre_get(struct ida *ida, gfp_t gfp_mask) argument 1079 ida_simple_get(struct ida *ida, unsigned int start, unsigned int end, gfp_t gfp_mask) argument [all...] |
H A D | scatterlist.c | 136 static struct scatterlist *sg_kmalloc(unsigned int nents, gfp_t gfp_mask) argument 148 void *ptr = (void *) __get_free_page(gfp_mask); 149 kmemleak_alloc(ptr, PAGE_SIZE, 1, gfp_mask); 152 return kmalloc(nents * sizeof(struct scatterlist), gfp_mask); 233 * @gfp_mask: GFP allocation mask 249 gfp_t gfp_mask, sg_alloc_fn *alloc_fn) 280 sg = alloc_fn(alloc_size, gfp_mask); 324 * @gfp_mask: GFP allocation mask 331 int sg_alloc_table(struct sg_table *table, unsigned int nents, gfp_t gfp_mask) argument 336 NULL, gfp_mask, sg_kmallo 247 __sg_alloc_table(struct sg_table *table, unsigned int nents, unsigned int max_ents, struct scatterlist *first_chunk, gfp_t gfp_mask, sg_alloc_fn *alloc_fn) argument 364 sg_alloc_table_from_pages(struct sg_table *sgt, struct page **pages, unsigned int n_pages, unsigned long offset, unsigned long size, gfp_t gfp_mask) argument [all...] |
H A D | radix-tree.c | 84 return root->gfp_mask & __GFP_BITS_MASK; 107 root->gfp_mask |= (__force gfp_t)(1 << (tag + __GFP_BITS_SHIFT)); 112 root->gfp_mask &= (__force gfp_t)~(1 << (tag + __GFP_BITS_SHIFT)); 117 root->gfp_mask &= __GFP_BITS_MASK; 122 return (__force unsigned)root->gfp_mask & (1 << (tag + __GFP_BITS_SHIFT)); 183 gfp_t gfp_mask = root_gfp_mask(root); local 190 if (!(gfp_mask & __GFP_WAIT) && !in_interrupt()) { 211 ret = kmem_cache_alloc(radix_tree_node_cachep, gfp_mask); 252 static int __radix_tree_preload(gfp_t gfp_mask) argument 262 node = kmem_cache_alloc(radix_tree_node_cachep, gfp_mask); 286 radix_tree_preload(gfp_t gfp_mask) argument 299 radix_tree_maybe_preload(gfp_t gfp_mask) argument [all...] |
H A D | textsearch.c | 248 * @gfp_mask: allocation mask 262 unsigned int len, gfp_t gfp_mask, int flags) 287 conf = ops->init(pattern, len, gfp_mask, flags); 261 textsearch_prepare(const char *algo, const void *pattern, unsigned int len, gfp_t gfp_mask, int flags) argument
|
H A D | ts_kmp.c | 96 gfp_t gfp_mask, int flags) 104 conf = alloc_ts_config(priv_size, gfp_mask); 95 kmp_init(const void *pattern, unsigned int len, gfp_t gfp_mask, int flags) argument
|
H A D | ts_bm.c | 146 gfp_t gfp_mask, int flags) 154 conf = alloc_ts_config(priv_size, gfp_mask); 145 bm_init(const void *pattern, unsigned int len, gfp_t gfp_mask, int flags) argument
|
H A D | kfifo.c | 39 size_t esize, gfp_t gfp_mask) 57 fifo->data = kmalloc(size * esize, gfp_mask); 38 __kfifo_alloc(struct __kfifo *fifo, unsigned int size, size_t esize, gfp_t gfp_mask) argument
|
H A D | kobject.c | 142 * @gfp_mask: the allocation type used to allocate the path 146 char *kobject_get_path(struct kobject *kobj, gfp_t gfp_mask) argument 154 path = kzalloc(len, gfp_mask);
|
H A D | ts_fsm.c | 260 gfp_t gfp_mask, int flags) 286 conf = alloc_ts_config(priv_size, gfp_mask); 259 fsm_init(const void *pattern, unsigned int len, gfp_t gfp_mask, int flags) argument
|
H A D | btree.c | 81 void *btree_alloc(gfp_t gfp_mask, void *pool_data) argument 83 return kmem_cache_alloc(btree_cachep, gfp_mask);
|