Home
last modified time | relevance | path

Searched refs:chunk_list (Results 1 – 19 of 19) sorted by relevance

/linux-5.19.10/arch/powerpc/platforms/ps3/
Dmm.c405 list_for_each_entry(c, &r->chunk_list.head, link) { in dma_find_chunk()
434 list_for_each_entry(c, &r->chunk_list.head, link) { in dma_find_chunk_lpar()
540 list_add(&c->link, &r->chunk_list.head); in dma_sb_map_pages()
574 if (list_empty(&r->chunk_list.head)) { in dma_ioc0_map_pages()
579 last = list_entry(r->chunk_list.head.next, in dma_ioc0_map_pages()
610 list_add(&c->link, &r->chunk_list.head); in dma_ioc0_map_pages()
660 INIT_LIST_HEAD(&r->chunk_list.head); in dma_sb_region_create()
661 spin_lock_init(&r->chunk_list.lock); in dma_sb_region_create()
682 INIT_LIST_HEAD(&r->chunk_list.head); in dma_ioc0_region_create()
683 spin_lock_init(&r->chunk_list.lock); in dma_ioc0_region_create()
[all …]
/linux-5.19.10/drivers/net/ethernet/mellanox/mlx4/
Dicm.h68 struct list_head chunk_list; member
100 iter->chunk = list_empty(&icm->chunk_list) ? in mlx4_icm_first()
101 NULL : list_entry(icm->chunk_list.next, in mlx4_icm_first()
114 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mlx4_icm_next()
Dicm.c86 list_for_each_entry_safe(chunk, tmp, &icm->chunk_list, list) { in mlx4_free_icm()
155 INIT_LIST_HEAD(&icm->chunk_list); in mlx4_alloc_icm()
176 list_add_tail(&chunk->list, &icm->chunk_list); in mlx4_alloc_icm()
335 list_for_each_entry(chunk, &icm->chunk_list, list) { in mlx4_table_find()
/linux-5.19.10/drivers/infiniband/hw/usnic/
Dusnic_uiom.c65 static void usnic_uiom_put_pages(struct list_head *chunk_list, int dirty) in usnic_uiom_put_pages() argument
73 list_for_each_entry_safe(chunk, tmp, chunk_list, list) { in usnic_uiom_put_pages()
87 struct list_head *chunk_list = &uiomr->chunk_list; in usnic_uiom_get_pages() local
116 INIT_LIST_HEAD(chunk_list); in usnic_uiom_get_pages()
176 list_add_tail(&chunk->list, chunk_list); in usnic_uiom_get_pages()
184 usnic_uiom_put_pages(chunk_list, 0); in usnic_uiom_get_pages()
239 usnic_uiom_put_pages(&uiomr->chunk_list, dirty & writable); in __usnic_uiom_reg_release()
259 chunk = list_first_entry(&uiomr->chunk_list, struct usnic_uiom_chunk, in usnic_uiom_map_sorted_intervals()
414 usnic_uiom_put_pages(&uiomr->chunk_list, 0); in usnic_uiom_reg_get()
Dusnic_uiom.h72 struct list_head chunk_list; member
/linux-5.19.10/drivers/infiniband/hw/hns/
Dhns_roce_hem.h84 struct list_head chunk_list; member
141 iter->chunk = list_empty(&hem->chunk_list) ? NULL : in hns_roce_hem_first()
142 list_entry(hem->chunk_list.next, in hns_roce_hem_first()
155 if (iter->chunk->list.next == &iter->hem->chunk_list) { in hns_roce_hem_next()
Dhns_roce_hem.c269 INIT_LIST_HEAD(&hem->chunk_list); in hns_roce_alloc_hem()
284 list_add_tail(&chunk->list, &hem->chunk_list); in hns_roce_alloc_hem()
323 list_for_each_entry_safe(chunk, tmp, &hem->chunk_list, list) { in hns_roce_free_hem()
761 list_for_each_entry(chunk, &hem->chunk_list, list) { in hns_roce_table_find()
/linux-5.19.10/net/sctp/
Doutput.c153 INIT_LIST_HEAD(&packet->chunk_list); in sctp_packet_init()
167 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, list) { in sctp_packet_free()
233 list_add_tail(&pad->list, &pkt->chunk_list); in sctp_packet_bundle_pad()
373 list_add_tail(&chunk->list, &packet->chunk_list); in __sctp_packet_append_chunk()
459 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, in sctp_packet_pack()
480 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, list) { in sctp_packet_pack()
523 if (list_empty(&packet->chunk_list)) in sctp_packet_pack()
527 &packet->chunk_list); in sctp_packet_pack()
534 } while (!list_empty(&packet->chunk_list)); in sctp_packet_pack()
580 if (list_empty(&packet->chunk_list)) in sctp_packet_transmit()
[all …]
/linux-5.19.10/drivers/infiniband/hw/mthca/
Dmthca_memfree.h59 struct list_head chunk_list; member
103 iter->chunk = list_empty(&icm->chunk_list) ? in mthca_icm_first()
104 NULL : list_entry(icm->chunk_list.next, in mthca_icm_first()
117 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mthca_icm_next()
Dmthca_memfree.c95 list_for_each_entry_safe(chunk, tmp, &icm->chunk_list, list) { in mthca_free_icm()
153 INIT_LIST_HEAD(&icm->chunk_list); in mthca_alloc_icm()
167 list_add_tail(&chunk->list, &icm->chunk_list); in mthca_alloc_icm()
297 list_for_each_entry(chunk, &icm->chunk_list, list) { in mthca_table_find()
/linux-5.19.10/drivers/infiniband/hw/efa/
Defa_verbs.c118 struct pbl_chunk_list chunk_list; member
1253 struct pbl_chunk_list *chunk_list = &pbl->phys.indirect.chunk_list; in pbl_chunk_list_create() local
1267 chunk_list->size = chunk_list_size; in pbl_chunk_list_create()
1268 chunk_list->chunks = kcalloc(chunk_list_size, in pbl_chunk_list_create()
1269 sizeof(*chunk_list->chunks), in pbl_chunk_list_create()
1271 if (!chunk_list->chunks) in pbl_chunk_list_create()
1280 chunk_list->chunks[i].buf = kzalloc(EFA_CHUNK_SIZE, GFP_KERNEL); in pbl_chunk_list_create()
1281 if (!chunk_list->chunks[i].buf) in pbl_chunk_list_create()
1284 chunk_list->chunks[i].length = EFA_CHUNK_USED_SIZE; in pbl_chunk_list_create()
1286 chunk_list->chunks[chunk_list_size - 1].length = in pbl_chunk_list_create()
[all …]
/linux-5.19.10/drivers/gpu/drm/qxl/
Dqxl_image.c51 list_add_tail(&chunk->head, &image->chunk_list); in qxl_allocate_chunk()
68 INIT_LIST_HEAD(&image->chunk_list); in qxl_image_alloc_objects()
90 list_for_each_entry_safe(chunk, tmp, &dimage->chunk_list, head) { in qxl_image_free_objects()
120 drv_chunk = list_first_entry(&dimage->chunk_list, struct qxl_drm_chunk, head); in qxl_image_init_helper()
Dqxl_drv.h167 struct list_head chunk_list; member
/linux-5.19.10/drivers/net/ethernet/mellanox/mlx5/core/steering/
Ddr_icm_pool.c212 list_del(&chunk->chunk_list); in dr_icm_chunk_destroy()
301 list_for_each_entry_safe(chunk, next, &buddy->hot_list, chunk_list) in dr_icm_buddy_destroy()
304 list_for_each_entry_safe(chunk, next, &buddy->used_list, chunk_list) in dr_icm_buddy_destroy()
340 INIT_LIST_HEAD(&chunk->chunk_list); in dr_icm_chunk_create()
343 list_add_tail(&chunk->chunk_list, &buddy_mem_pool->used_list); in dr_icm_chunk_create()
375 list_for_each_entry_safe(chunk, tmp_chunk, &buddy->hot_list, chunk_list) { in dr_icm_pool_sync_all_buddy_pools()
475 list_move_tail(&chunk->chunk_list, &buddy->hot_list); in mlx5dr_icm_free_chunk()
Ddr_types.h1091 struct list_head chunk_list; member
/linux-5.19.10/drivers/s390/block/
Ddasd_int.h683 dasd_init_chunklist(struct list_head *chunk_list, void *mem, in dasd_init_chunklist() argument
688 INIT_LIST_HEAD(chunk_list); in dasd_init_chunklist()
691 list_add(&chunk->list, chunk_list); in dasd_init_chunklist()
695 dasd_alloc_chunk(struct list_head *chunk_list, unsigned long size) in dasd_alloc_chunk() argument
700 list_for_each_entry(chunk, chunk_list, list) { in dasd_alloc_chunk()
717 dasd_free_chunk(struct list_head *chunk_list, void *mem) in dasd_free_chunk() argument
725 left = chunk_list; in dasd_free_chunk()
726 list_for_each(p, chunk_list) { in dasd_free_chunk()
732 if (left->next != chunk_list) { in dasd_free_chunk()
740 if (left != chunk_list) { in dasd_free_chunk()
/linux-5.19.10/drivers/crypto/
Dn2_core.c685 struct list_head chunk_list; member
862 INIT_LIST_HEAD(&rctx->chunk_list); in n2_compute_chunks()
895 &rctx->chunk_list); in n2_compute_chunks()
922 list_add_tail(&chunk->entry, &rctx->chunk_list); in n2_compute_chunks()
936 list_for_each_entry_safe(c, tmp, &rctx->chunk_list, entry) { in n2_chunk_complete()
963 list_for_each_entry_safe(c, tmp, &rctx->chunk_list, entry) { in n2_do_ecb()
1020 list_for_each_entry_safe(c, tmp, &rctx->chunk_list, in n2_do_chaining()
1033 list_for_each_entry_safe_reverse(c, tmp, &rctx->chunk_list, in n2_do_chaining()
/linux-5.19.10/arch/powerpc/include/asm/
Dps3.h96 } chunk_list; member
/linux-5.19.10/include/net/sctp/
Dstructs.h712 struct list_head chunk_list; member