Lines Matching refs:pool

155 	struct gen_pool *pool;  in gen_pool_create()  local
157 pool = kmalloc_node(sizeof(struct gen_pool), GFP_KERNEL, nid); in gen_pool_create()
158 if (pool != NULL) { in gen_pool_create()
159 spin_lock_init(&pool->lock); in gen_pool_create()
160 INIT_LIST_HEAD(&pool->chunks); in gen_pool_create()
161 pool->min_alloc_order = min_alloc_order; in gen_pool_create()
162 pool->algo = gen_pool_first_fit; in gen_pool_create()
163 pool->data = NULL; in gen_pool_create()
164 pool->name = NULL; in gen_pool_create()
166 return pool; in gen_pool_create()
184 int gen_pool_add_owner(struct gen_pool *pool, unsigned long virt, phys_addr_t phys, in gen_pool_add_owner() argument
188 unsigned long nbits = size >> pool->min_alloc_order; in gen_pool_add_owner()
202 spin_lock(&pool->lock); in gen_pool_add_owner()
203 list_add_rcu(&chunk->next_chunk, &pool->chunks); in gen_pool_add_owner()
204 spin_unlock(&pool->lock); in gen_pool_add_owner()
217 phys_addr_t gen_pool_virt_to_phys(struct gen_pool *pool, unsigned long addr) in gen_pool_virt_to_phys() argument
223 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) { in gen_pool_virt_to_phys()
242 void gen_pool_destroy(struct gen_pool *pool) in gen_pool_destroy() argument
246 int order = pool->min_alloc_order; in gen_pool_destroy()
249 list_for_each_safe(_chunk, _next_chunk, &pool->chunks) { in gen_pool_destroy()
259 kfree_const(pool->name); in gen_pool_destroy()
260 kfree(pool); in gen_pool_destroy()
277 unsigned long gen_pool_alloc_algo_owner(struct gen_pool *pool, size_t size, in gen_pool_alloc_algo_owner() argument
282 int order = pool->min_alloc_order; in gen_pool_alloc_algo_owner()
297 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) { in gen_pool_alloc_algo_owner()
305 nbits, data, pool, chunk->start_addr); in gen_pool_alloc_algo_owner()
341 void *gen_pool_dma_alloc(struct gen_pool *pool, size_t size, dma_addr_t *dma) in gen_pool_dma_alloc() argument
343 return gen_pool_dma_alloc_algo(pool, size, dma, pool->algo, pool->data); in gen_pool_dma_alloc()
362 void *gen_pool_dma_alloc_algo(struct gen_pool *pool, size_t size, in gen_pool_dma_alloc_algo() argument
367 if (!pool) in gen_pool_dma_alloc_algo()
370 vaddr = gen_pool_alloc_algo(pool, size, algo, data); in gen_pool_dma_alloc_algo()
375 *dma = gen_pool_virt_to_phys(pool, vaddr); in gen_pool_dma_alloc_algo()
395 void *gen_pool_dma_alloc_align(struct gen_pool *pool, size_t size, in gen_pool_dma_alloc_align() argument
400 return gen_pool_dma_alloc_algo(pool, size, dma, in gen_pool_dma_alloc_align()
419 void *gen_pool_dma_zalloc(struct gen_pool *pool, size_t size, dma_addr_t *dma) in gen_pool_dma_zalloc() argument
421 return gen_pool_dma_zalloc_algo(pool, size, dma, pool->algo, pool->data); in gen_pool_dma_zalloc()
440 void *gen_pool_dma_zalloc_algo(struct gen_pool *pool, size_t size, in gen_pool_dma_zalloc_algo() argument
443 void *vaddr = gen_pool_dma_alloc_algo(pool, size, dma, algo, data); in gen_pool_dma_zalloc_algo()
466 void *gen_pool_dma_zalloc_align(struct gen_pool *pool, size_t size, in gen_pool_dma_zalloc_align() argument
471 return gen_pool_dma_zalloc_algo(pool, size, dma, in gen_pool_dma_zalloc_align()
487 void gen_pool_free_owner(struct gen_pool *pool, unsigned long addr, size_t size, in gen_pool_free_owner() argument
491 int order = pool->min_alloc_order; in gen_pool_free_owner()
503 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) { in gen_pool_free_owner()
531 void gen_pool_for_each_chunk(struct gen_pool *pool, in gen_pool_for_each_chunk() argument
532 void (*func)(struct gen_pool *pool, struct gen_pool_chunk *chunk, void *data), in gen_pool_for_each_chunk() argument
538 list_for_each_entry_rcu(chunk, &(pool)->chunks, next_chunk) in gen_pool_for_each_chunk()
539 func(pool, chunk, data); in gen_pool_for_each_chunk()
553 bool gen_pool_has_addr(struct gen_pool *pool, unsigned long start, in gen_pool_has_addr() argument
561 list_for_each_entry_rcu(chunk, &(pool)->chunks, next_chunk) { in gen_pool_has_addr()
580 size_t gen_pool_avail(struct gen_pool *pool) in gen_pool_avail() argument
586 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) in gen_pool_avail()
599 size_t gen_pool_size(struct gen_pool *pool) in gen_pool_size() argument
605 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) in gen_pool_size()
622 void gen_pool_set_algo(struct gen_pool *pool, genpool_algo_t algo, void *data) in gen_pool_set_algo() argument
626 pool->algo = algo; in gen_pool_set_algo()
627 if (!pool->algo) in gen_pool_set_algo()
628 pool->algo = gen_pool_first_fit; in gen_pool_set_algo()
630 pool->data = data; in gen_pool_set_algo()
649 struct gen_pool *pool, unsigned long start_addr) in gen_pool_first_fit() argument
668 struct gen_pool *pool, unsigned long start_addr) in gen_pool_first_fit_align() argument
675 order = pool->min_alloc_order; in gen_pool_first_fit_align()
696 struct gen_pool *pool, unsigned long start_addr) in gen_pool_fixed_alloc() argument
704 order = pool->min_alloc_order; in gen_pool_fixed_alloc()
731 unsigned int nr, void *data, struct gen_pool *pool, in gen_pool_first_fit_order_align() argument
756 struct gen_pool *pool, unsigned long start_addr) in gen_pool_best_fit() argument
832 struct gen_pool **ptr, *pool; in devm_gen_pool_create() local
849 pool = gen_pool_create(min_alloc_order, nid); in devm_gen_pool_create()
850 if (!pool) in devm_gen_pool_create()
853 *ptr = pool; in devm_gen_pool_create()
854 pool->name = pool_name; in devm_gen_pool_create()
857 return pool; in devm_gen_pool_create()
885 struct gen_pool *pool = NULL; in of_gen_pool_get() local
903 pool = gen_pool_get(&pdev->dev, name); in of_gen_pool_get()
906 return pool; in of_gen_pool_get()