Lines Matching refs:nodeid

241 #define MAKE_LIST(cachep, listp, slab, nodeid)				\  argument
244 list_splice(&get_node(cachep, nodeid)->slab, listp); \
247 #define MAKE_ALL_LISTS(cachep, ptr, nodeid) \ argument
249 MAKE_LIST((cachep), (&(ptr)->slabs_full), slabs_full, nodeid); \
250 MAKE_LIST((cachep), (&(ptr)->slabs_partial), slabs_partial, nodeid); \
251 MAKE_LIST((cachep), (&(ptr)->slabs_free), slabs_free, nodeid); \
1150 int nodeid) in init_list() argument
1154 ptr = kmalloc_node(sizeof(struct kmem_cache_node), GFP_NOWAIT, nodeid); in init_list()
1163 MAKE_ALL_LISTS(cachep, ptr, nodeid); in init_list()
1164 cachep->node[nodeid] = ptr; in init_list()
1306 slab_out_of_memory(struct kmem_cache *cachep, gfp_t gfpflags, int nodeid) in slab_out_of_memory() argument
1319 nodeid, gfpflags, &gfpflags); in slab_out_of_memory()
1349 int nodeid) in kmem_getpages() argument
1356 folio = (struct folio *) __alloc_pages_node(nodeid, flags, cachep->gfporder); in kmem_getpages()
1358 slab_out_of_memory(cachep, flags, nodeid); in kmem_getpages()
2275 gfp_t local_flags, int nodeid) in alloc_slabmgmt() argument
2288 local_flags, nodeid); in alloc_slabmgmt()
2523 gfp_t flags, int nodeid) in cache_grow_begin() argument
2550 slab = kmem_getpages(cachep, local_flags, nodeid); in cache_grow_begin()
3118 int nodeid) in ____cache_alloc_node() argument
3125 VM_BUG_ON(nodeid < 0 || nodeid >= MAX_NUMNODES); in ____cache_alloc_node()
3126 n = get_node(cachep, nodeid); in ____cache_alloc_node()
3135 check_spinlock_acquired_node(cachep, nodeid); in ____cache_alloc_node()
3154 slab = cache_grow_begin(cachep, gfp_exact_node(flags), nodeid); in ____cache_alloc_node()
3165 __do_cache_alloc(struct kmem_cache *cachep, gfp_t flags, int nodeid) in __do_cache_alloc() argument
3170 if (nodeid == NUMA_NO_NODE) { in __do_cache_alloc()
3183 nodeid = slab_node; in __do_cache_alloc()
3184 } else if (nodeid == slab_node) { in __do_cache_alloc()
3186 } else if (!get_node(cachep, nodeid)) { in __do_cache_alloc()
3197 objp = ____cache_alloc_node(cachep, flags, nodeid); in __do_cache_alloc()
3204 __do_cache_alloc(struct kmem_cache *cachep, gfp_t flags, int nodeid __maybe_unused) in __do_cache_alloc()
3213 int nodeid, size_t orig_size, unsigned long caller) in slab_alloc_node() argument
3230 objp = __do_cache_alloc(cachep, flags, nodeid); in slab_alloc_node()
3507 void *kmem_cache_alloc_node(struct kmem_cache *cachep, gfp_t flags, int nodeid) in kmem_cache_alloc_node() argument
3509 void *ret = slab_alloc_node(cachep, NULL, flags, nodeid, cachep->object_size, _RET_IP_); in kmem_cache_alloc_node()
3511 trace_kmem_cache_alloc(_RET_IP_, ret, cachep, flags, nodeid); in kmem_cache_alloc_node()
3518 int nodeid, size_t orig_size, in __kmem_cache_alloc_node() argument
3521 return slab_alloc_node(cachep, NULL, flags, nodeid, in __kmem_cache_alloc_node()