/linux-6.1.9/drivers/infiniband/hw/hns/ |
D | hns_roce_main.c | 313 ucontext, &entry->rdma_entry, length, 0); in hns_roce_user_mmap_entry_insert() 317 ucontext, &entry->rdma_entry, length, 1, in hns_roce_user_mmap_entry_insert() 337 &context->db_mmap_entry->rdma_entry); in hns_roce_dealloc_uar_entry() 427 struct rdma_user_mmap_entry *rdma_entry; in hns_roce_mmap() local 433 rdma_entry = rdma_user_mmap_entry_get_pgoff(uctx, vma->vm_pgoff); in hns_roce_mmap() 434 if (!rdma_entry) in hns_roce_mmap() 437 entry = to_hns_mmap(rdma_entry); in hns_roce_mmap() 449 ret = rdma_user_mmap_io(uctx, vma, pfn, rdma_entry->npages * PAGE_SIZE, in hns_roce_mmap() 450 prot, rdma_entry); in hns_roce_mmap() 452 rdma_user_mmap_entry_put(rdma_entry); in hns_roce_mmap() [all …]
|
D | hns_roce_device.h | 194 struct rdma_user_mmap_entry rdma_entry; member 1019 to_hns_mmap(struct rdma_user_mmap_entry *rdma_entry) in to_hns_mmap() argument 1021 return container_of(rdma_entry, struct hns_user_mmap_entry, rdma_entry); in to_hns_mmap()
|
D | hns_roce_qp.c | 348 rdma_user_mmap_entry_remove(&hr_qp->dwqe_mmap_entry->rdma_entry); in qp_user_mmap_entry_remove() 865 struct rdma_user_mmap_entry *rdma_entry; in qp_mmap_entry() local 880 rdma_entry = &hr_qp->dwqe_mmap_entry->rdma_entry; in qp_mmap_entry() 881 resp->dwqe_mmap_key = rdma_user_mmap_get_offset(rdma_entry); in qp_mmap_entry()
|
/linux-6.1.9/drivers/infiniband/hw/mlx5/ |
D | dm.c | 158 context, &mentry->rdma_entry, size, in add_dm_mmap_entry() 180 page_idx = op_entry->mentry.rdma_entry.start_pgoff & 0xFFFF; in copy_op_to_user() 267 rdma_user_mmap_entry_remove(&op_entry->mentry.rdma_entry); in UVERBS_HANDLER() 317 page_idx = dm->mentry.rdma_entry.start_pgoff & 0xFFFF; in handle_alloc_dm_memic() 333 rdma_user_mmap_entry_remove(&dm->mentry.rdma_entry); in handle_alloc_dm_memic() 455 rdma_user_mmap_entry_remove(&entry->mentry.rdma_entry); in dm_memic_remove_ops() 463 rdma_user_mmap_entry_remove(&dm->mentry.rdma_entry); in mlx5_dm_memic_dealloc() 515 page_idx = memic->mentry.rdma_entry.start_pgoff & 0xFFFF; in UVERBS_HANDLER()
|
D | mlx5_ib.h | 604 struct rdma_user_mmap_entry rdma_entry; member 1201 to_mmmap(struct rdma_user_mmap_entry *rdma_entry) in to_mmmap() argument 1203 return container_of(rdma_entry, in to_mmmap() 1204 struct mlx5_user_mmap_entry, rdma_entry); in to_mmmap()
|
D | main.c | 2274 rdma_user_mmap_entry_put(&mentry->rdma_entry); in mlx5_ib_mmap_offset() 2280 u64 cmd = (entry->rdma_entry.start_pgoff >> 16) & 0xFFFF; in mlx5_entry_to_mmap_offset() 2281 u64 index = entry->rdma_entry.start_pgoff & 0xFFFF; in mlx5_entry_to_mmap_offset() 3327 rdma_user_mmap_entry_remove(&obj->rdma_entry); in mmap_obj_cleanup() 3336 &c->ibucontext, &entry->rdma_entry, length, in mlx5_rdma_user_mmap_entry_insert() 3407 length = entry->rdma_entry.npages * PAGE_SIZE; in UVERBS_HANDLER() 3534 length = entry->rdma_entry.npages * PAGE_SIZE; in UVERBS_HANDLER()
|
/linux-6.1.9/drivers/infiniband/hw/efa/ |
D | efa_verbs.c | 31 struct rdma_user_mmap_entry rdma_entry; member 163 to_emmap(struct rdma_user_mmap_entry *rdma_entry) in to_emmap() argument 165 return container_of(rdma_entry, struct efa_user_mmap_entry, rdma_entry); in to_emmap() 478 err = rdma_user_mmap_entry_insert(ucontext, &entry->rdma_entry, in efa_user_mmap_entry_insert() 484 *offset = rdma_user_mmap_get_offset(&entry->rdma_entry); in efa_user_mmap_entry_insert() 486 return &entry->rdma_entry; in efa_user_mmap_entry_insert() 1863 void efa_mmap_free(struct rdma_user_mmap_entry *rdma_entry) in efa_mmap_free() argument 1865 struct efa_user_mmap_entry *entry = to_emmap(rdma_entry); in efa_mmap_free() 1873 struct rdma_user_mmap_entry *rdma_entry; in __efa_mmap() local 1879 rdma_entry = rdma_user_mmap_entry_get(&ucontext->ibucontext, vma); in __efa_mmap() [all …]
|
D | efa.h | 168 void efa_mmap_free(struct rdma_user_mmap_entry *rdma_entry);
|
/linux-6.1.9/drivers/infiniband/hw/erdma/ |
D | erdma_verbs.h | 29 struct rdma_user_mmap_entry rdma_entry; member 288 return container_of(ibmmap, struct erdma_user_mmap_entry, rdma_entry); in to_emmap() 319 void erdma_mmap_free(struct rdma_user_mmap_entry *rdma_entry);
|
D | erdma_verbs.c | 251 ret = rdma_user_mmap_entry_insert(&uctx->ibucontext, &entry->rdma_entry, in erdma_user_mmap_entry_insert() 258 *mmap_offset = rdma_user_mmap_get_offset(&entry->rdma_entry); in erdma_user_mmap_entry_insert() 260 return &entry->rdma_entry; in erdma_user_mmap_entry_insert() 1075 struct rdma_user_mmap_entry *rdma_entry; in erdma_mmap() local 1080 rdma_entry = rdma_user_mmap_entry_get(ctx, vma); in erdma_mmap() 1081 if (!rdma_entry) in erdma_mmap() 1084 entry = to_emmap(rdma_entry); in erdma_mmap() 1096 prot, rdma_entry); in erdma_mmap() 1098 rdma_user_mmap_entry_put(rdma_entry); in erdma_mmap() 1102 void erdma_mmap_free(struct rdma_user_mmap_entry *rdma_entry) in erdma_mmap_free() argument [all …]
|
/linux-6.1.9/drivers/infiniband/hw/qedr/ |
D | qedr.h | 499 struct rdma_user_mmap_entry rdma_entry; member 637 get_qedr_mmap_entry(struct rdma_user_mmap_entry *rdma_entry) in get_qedr_mmap_entry() argument 639 return container_of(rdma_entry, struct qedr_user_mmap_entry, in get_qedr_mmap_entry() 640 rdma_entry); in get_qedr_mmap_entry()
|
D | verbs.h | 49 void qedr_mmap_free(struct rdma_user_mmap_entry *rdma_entry);
|
D | verbs.c | 310 rc = rdma_user_mmap_entry_insert(uctx, &entry->rdma_entry, in qedr_alloc_ucontext() 316 ctx->db_mmap_entry = &entry->rdma_entry; in qedr_alloc_ucontext() 375 void qedr_mmap_free(struct rdma_user_mmap_entry *rdma_entry) in qedr_mmap_free() argument 377 struct qedr_user_mmap_entry *entry = get_qedr_mmap_entry(rdma_entry); in qedr_mmap_free() 392 struct rdma_user_mmap_entry *rdma_entry; in qedr_mmap() local 401 rdma_entry = rdma_user_mmap_entry_get(ucontext, vma); in qedr_mmap() 402 if (!rdma_entry) { in qedr_mmap() 407 entry = get_qedr_mmap_entry(rdma_entry); in qedr_mmap() 417 rdma_entry); in qedr_mmap() 432 rdma_user_mmap_entry_put(rdma_entry); in qedr_mmap() [all …]
|
/linux-6.1.9/drivers/infiniband/sw/siw/ |
D | siw_verbs.c | 38 void siw_mmap_free(struct rdma_user_mmap_entry *rdma_entry) in siw_mmap_free() argument 40 struct siw_user_mmap_entry *entry = to_siw_mmap_entry(rdma_entry); in siw_mmap_free() 49 struct rdma_user_mmap_entry *rdma_entry; in siw_mmap() local 60 rdma_entry = rdma_user_mmap_entry_get(&uctx->base_ucontext, vma); in siw_mmap() 61 if (!rdma_entry) { in siw_mmap() 66 entry = to_siw_mmap_entry(rdma_entry); in siw_mmap() 75 rdma_user_mmap_entry_put(rdma_entry); in siw_mmap() 273 &entry->rdma_entry, in siw_mmap_entry_insert() 280 *offset = rdma_user_mmap_get_offset(&entry->rdma_entry); in siw_mmap_entry_insert() 282 return &entry->rdma_entry; in siw_mmap_entry_insert()
|
D | siw_verbs.h | 84 void siw_mmap_free(struct rdma_user_mmap_entry *rdma_entry);
|
D | siw.h | 488 struct rdma_user_mmap_entry rdma_entry; member 594 return container_of(rdma_mmap, struct siw_user_mmap_entry, rdma_entry); in to_siw_mmap_entry()
|
/linux-6.1.9/drivers/infiniband/hw/irdma/ |
D | main.h | 367 to_irdma_mmap_entry(struct rdma_user_mmap_entry *rdma_entry) in to_irdma_mmap_entry() argument 369 return container_of(rdma_entry, struct irdma_user_mmap_entry, in to_irdma_mmap_entry() 370 rdma_entry); in to_irdma_mmap_entry()
|
D | verbs.h | 220 struct rdma_user_mmap_entry rdma_entry; member
|
D | verbs.c | 135 static void irdma_mmap_free(struct rdma_user_mmap_entry *rdma_entry) in irdma_mmap_free() argument 137 struct irdma_user_mmap_entry *entry = to_irdma_mmap_entry(rdma_entry); in irdma_mmap_free() 156 &entry->rdma_entry, PAGE_SIZE); in irdma_user_mmap_entry_insert() 161 *mmap_offset = rdma_user_mmap_get_offset(&entry->rdma_entry); in irdma_user_mmap_entry_insert() 163 return &entry->rdma_entry; in irdma_user_mmap_entry_insert() 173 struct rdma_user_mmap_entry *rdma_entry; in irdma_mmap() local 185 rdma_entry = rdma_user_mmap_entry_get(&ucontext->ibucontext, vma); in irdma_mmap() 186 if (!rdma_entry) { in irdma_mmap() 193 entry = to_irdma_mmap_entry(rdma_entry); in irdma_mmap() 205 rdma_entry); in irdma_mmap() [all …]
|