/linux-6.6.21/tools/tracing/rtla/src/ |
D | osnoise.c | 24 char *osnoise_get_cpus(struct osnoise_context *context) in osnoise_get_cpus() argument 26 if (context->curr_cpus) in osnoise_get_cpus() 27 return context->curr_cpus; in osnoise_get_cpus() 29 if (context->orig_cpus) in osnoise_get_cpus() 30 return context->orig_cpus; in osnoise_get_cpus() 32 context->orig_cpus = tracefs_instance_file_read(NULL, "osnoise/cpus", NULL); in osnoise_get_cpus() 38 return context->orig_cpus; in osnoise_get_cpus() 48 int osnoise_set_cpus(struct osnoise_context *context, char *cpus) in osnoise_set_cpus() argument 50 char *orig_cpus = osnoise_get_cpus(context); in osnoise_set_cpus() 57 context->curr_cpus = strdup(cpus); in osnoise_set_cpus() [all …]
|
D | osnoise.h | 59 int osnoise_get_context(struct osnoise_context *context); 60 void osnoise_put_context(struct osnoise_context *context); 62 int osnoise_set_cpus(struct osnoise_context *context, char *cpus); 63 void osnoise_restore_cpus(struct osnoise_context *context); 65 int osnoise_set_runtime_period(struct osnoise_context *context, 68 void osnoise_restore_runtime_period(struct osnoise_context *context); 70 int osnoise_set_stop_us(struct osnoise_context *context, 72 void osnoise_restore_stop_us(struct osnoise_context *context); 74 int osnoise_set_stop_total_us(struct osnoise_context *context, 76 void osnoise_restore_stop_total_us(struct osnoise_context *context); [all …]
|
/linux-6.6.21/drivers/misc/vmw_vmci/ |
D | vmci_context.c | 44 static void ctx_signal_notify(struct vmci_ctx *context) in ctx_signal_notify() argument 46 *context->notify = true; in ctx_signal_notify() 49 static void ctx_clear_notify(struct vmci_ctx *context) in ctx_clear_notify() argument 51 *context->notify = false; in ctx_clear_notify() 58 static void ctx_clear_notify_call(struct vmci_ctx *context) in ctx_clear_notify_call() argument 60 if (context->pending_datagrams == 0 && in ctx_clear_notify_call() 61 vmci_handle_arr_get_size(context->pending_doorbell_array) == 0) in ctx_clear_notify_call() 62 ctx_clear_notify(context); in ctx_clear_notify_call() 69 void vmci_ctx_check_signal_notify(struct vmci_ctx *context) in vmci_ctx_check_signal_notify() argument 71 spin_lock(&context->lock); in vmci_ctx_check_signal_notify() [all …]
|
D | vmci_route.c | 42 if (VMCI_INVALID_ID == dst->context) in vmci_route() 46 if (VMCI_HYPERVISOR_CONTEXT_ID == dst->context) { in vmci_route() 64 if (VMCI_HOST_CONTEXT_ID == src->context) in vmci_route() 75 if (VMCI_INVALID_ID == src->context && in vmci_route() 77 src->context = vmci_get_context_id(); in vmci_route() 85 if (VMCI_HOST_CONTEXT_ID == dst->context) { in vmci_route() 94 if (src->context == VMCI_HYPERVISOR_CONTEXT_ID) { in vmci_route() 113 if (VMCI_INVALID_ID == src->context) in vmci_route() 114 src->context = vmci_get_context_id(); in vmci_route() 130 if (VMCI_INVALID_ID == src->context) { in vmci_route() [all …]
|
/linux-6.6.21/security/selinux/ss/ |
D | mls.h | 31 int mls_compute_context_len(struct policydb *p, struct context *context); 32 void mls_sid_to_context(struct policydb *p, struct context *context, 34 int mls_context_isvalid(struct policydb *p, struct context *c); 41 struct context *context, 45 int mls_from_string(struct policydb *p, char *str, struct context *context, 48 int mls_range_set(struct context *context, struct mls_range *range); 52 struct context *oldc, 53 struct context *newc); 56 struct context *scontext, 57 struct context *tcontext, [all …]
|
D | mls.c | 36 int mls_compute_context_len(struct policydb *p, struct context *context) in mls_compute_context_len() argument 48 u32 index_sens = context->range.level[l].sens; in mls_compute_context_len() 54 e = &context->range.level[l].cat; in mls_compute_context_len() 73 if (mls_level_eq(&context->range.level[0], in mls_compute_context_len() 74 &context->range.level[1])) in mls_compute_context_len() 90 struct context *context, in mls_sid_to_context() argument 108 context->range.level[l].sens - 1)); in mls_sid_to_context() 114 e = &context->range.level[l].cat; in mls_sid_to_context() 150 if (mls_level_eq(&context->range.level[0], in mls_sid_to_context() 151 &context->range.level[1])) in mls_sid_to_context() [all …]
|
/linux-6.6.21/drivers/gpu/drm/etnaviv/ |
D | etnaviv_mmu.c | 16 static void etnaviv_context_unmap(struct etnaviv_iommu_context *context, in etnaviv_context_unmap() argument 29 unmapped_page = context->global->ops->unmap(context, iova, in etnaviv_context_unmap() 39 static int etnaviv_context_map(struct etnaviv_iommu_context *context, in etnaviv_context_map() argument 55 ret = context->global->ops->map(context, iova, paddr, pgsize, in etnaviv_context_map() 67 etnaviv_context_unmap(context, orig_iova, orig_size - size); in etnaviv_context_map() 72 static int etnaviv_iommu_map(struct etnaviv_iommu_context *context, u32 iova, in etnaviv_iommu_map() argument 79 if (!context || !sgt) in etnaviv_iommu_map() 88 ret = etnaviv_context_map(context, da, pa, bytes, prot); in etnaviv_iommu_map() 95 context->flush_seq++; in etnaviv_iommu_map() 100 etnaviv_context_unmap(context, iova, da - iova); in etnaviv_iommu_map() [all …]
|
D | etnaviv_iommu.c | 28 to_v1_context(struct etnaviv_iommu_context *context) in to_v1_context() argument 30 return container_of(context, struct etnaviv_iommuv1_context, base); in to_v1_context() 33 static void etnaviv_iommuv1_free(struct etnaviv_iommu_context *context) in etnaviv_iommuv1_free() argument 35 struct etnaviv_iommuv1_context *v1_context = to_v1_context(context); in etnaviv_iommuv1_free() 37 drm_mm_takedown(&context->mm); in etnaviv_iommuv1_free() 39 dma_free_wc(context->global->dev, PT_SIZE, v1_context->pgtable_cpu, in etnaviv_iommuv1_free() 42 context->global->v1.shared_context = NULL; in etnaviv_iommuv1_free() 47 static int etnaviv_iommuv1_map(struct etnaviv_iommu_context *context, in etnaviv_iommuv1_map() argument 51 struct etnaviv_iommuv1_context *v1_context = to_v1_context(context); in etnaviv_iommuv1_map() 62 static size_t etnaviv_iommuv1_unmap(struct etnaviv_iommu_context *context, in etnaviv_iommuv1_unmap() argument [all …]
|
D | etnaviv_iommu_v2.c | 42 to_v2_context(struct etnaviv_iommu_context *context) in to_v2_context() argument 44 return container_of(context, struct etnaviv_iommuv2_context, base); in to_v2_context() 47 static void etnaviv_iommuv2_free(struct etnaviv_iommu_context *context) in etnaviv_iommuv2_free() argument 49 struct etnaviv_iommuv2_context *v2_context = to_v2_context(context); in etnaviv_iommuv2_free() 52 drm_mm_takedown(&context->mm); in etnaviv_iommuv2_free() 56 dma_free_wc(context->global->dev, SZ_4K, in etnaviv_iommuv2_free() 61 dma_free_wc(context->global->dev, SZ_4K, v2_context->mtlb_cpu, in etnaviv_iommuv2_free() 64 clear_bit(v2_context->id, context->global->v2.pta_alloc); in etnaviv_iommuv2_free() 92 static int etnaviv_iommuv2_map(struct etnaviv_iommu_context *context, in etnaviv_iommuv2_map() argument 96 struct etnaviv_iommuv2_context *v2_context = to_v2_context(context); in etnaviv_iommuv2_map() [all …]
|
/linux-6.6.21/kernel/ |
D | auditsc.c | 914 static inline void audit_proctitle_free(struct audit_context *context) in audit_proctitle_free() argument 916 kfree(context->proctitle.value); in audit_proctitle_free() 917 context->proctitle.value = NULL; in audit_proctitle_free() 918 context->proctitle.len = 0; in audit_proctitle_free() 921 static inline void audit_free_module(struct audit_context *context) in audit_free_module() argument 923 if (context->type == AUDIT_KERN_MODULE) { in audit_free_module() 924 kfree(context->module.name); in audit_free_module() 925 context->module.name = NULL; in audit_free_module() 928 static inline void audit_free_names(struct audit_context *context) in audit_free_names() argument 932 list_for_each_entry_safe(n, next, &context->names_list, list) { in audit_free_names() [all …]
|
/linux-6.6.21/fs/xfs/ |
D | xfs_attr_list.c | 54 struct xfs_attr_list_context *context) in xfs_attr_shortform_list() argument 56 struct xfs_attrlist_cursor_kern *cursor = &context->cursor; in xfs_attr_shortform_list() 57 struct xfs_inode *dp = context->dp; in xfs_attr_shortform_list() 69 trace_xfs_attr_list_sf(context); in xfs_attr_shortform_list() 80 if (context->bufsize == 0 || in xfs_attr_shortform_list() 82 (dp->i_af.if_bytes + sf->hdr.count * 16) < context->bufsize)) { in xfs_attr_shortform_list() 84 if (XFS_IS_CORRUPT(context->dp->i_mount, in xfs_attr_shortform_list() 88 context->put_listent(context, in xfs_attr_shortform_list() 97 if (context->seen_enough) in xfs_attr_shortform_list() 101 trace_xfs_attr_list_sf_all(context); in xfs_attr_shortform_list() [all …]
|
D | xfs_xattr.c | 198 struct xfs_attr_list_context *context, in __xfs_xattr_put_listent() argument 207 if (context->count < 0 || context->seen_enough) in __xfs_xattr_put_listent() 210 if (!context->buffer) in __xfs_xattr_put_listent() 213 arraytop = context->count + prefix_len + namelen + 1; in __xfs_xattr_put_listent() 214 if (arraytop > context->firstu) { in __xfs_xattr_put_listent() 215 context->count = -1; /* insufficient space */ in __xfs_xattr_put_listent() 216 context->seen_enough = 1; in __xfs_xattr_put_listent() 219 offset = context->buffer + context->count; in __xfs_xattr_put_listent() 227 context->count += prefix_len + namelen + 1; in __xfs_xattr_put_listent() 233 struct xfs_attr_list_context *context, in xfs_xattr_put_listent() argument [all …]
|
/linux-6.6.21/drivers/gpu/drm/tegra/ |
D | uapi.c | 31 static void tegra_drm_channel_context_close(struct tegra_drm_context *context) in tegra_drm_channel_context_close() argument 36 if (context->memory_context) in tegra_drm_channel_context_close() 37 host1x_memory_context_put(context->memory_context); in tegra_drm_channel_context_close() 39 xa_for_each(&context->mappings, id, mapping) in tegra_drm_channel_context_close() 42 xa_destroy(&context->mappings); in tegra_drm_channel_context_close() 44 host1x_channel_put(context->channel); in tegra_drm_channel_context_close() 46 kfree(context); in tegra_drm_channel_context_close() 51 struct tegra_drm_context *context; in tegra_drm_uapi_close_file() local 55 xa_for_each(&file->contexts, id, context) in tegra_drm_uapi_close_file() 56 tegra_drm_channel_context_close(context); in tegra_drm_uapi_close_file() [all …]
|
D | submit.c | 26 #define SUBMIT_ERR(context, fmt, ...) \ argument 27 dev_err_ratelimited(context->client->base.dev, \ 146 tegra_drm_mapping_get(struct tegra_drm_context *context, u32 id) in tegra_drm_mapping_get() argument 150 xa_lock(&context->mappings); in tegra_drm_mapping_get() 152 mapping = xa_load(&context->mappings, id); in tegra_drm_mapping_get() 156 xa_unlock(&context->mappings); in tegra_drm_mapping_get() 180 struct tegra_drm_context *context, in submit_copy_gather_data() argument 187 SUBMIT_ERR(context, "gather_data_words cannot be zero"); in submit_copy_gather_data() 192 SUBMIT_ERR(context, "gather_data_words is too large"); in submit_copy_gather_data() 198 SUBMIT_ERR(context, "failed to allocate memory for bo info"); in submit_copy_gather_data() [all …]
|
/linux-6.6.21/drivers/net/ethernet/mellanox/mlx4/ |
D | en_resources.c | 42 int user_prio, struct mlx4_qp_context *context) in mlx4_en_fill_qp_context() argument 47 memset(context, 0, sizeof(*context)); in mlx4_en_fill_qp_context() 48 context->flags = cpu_to_be32(7 << 16 | rss << MLX4_RSS_QPC_FLAG_OFFSET); in mlx4_en_fill_qp_context() 49 context->pd = cpu_to_be32(mdev->priv_pdn); in mlx4_en_fill_qp_context() 50 context->mtu_msgmax = 0xff; in mlx4_en_fill_qp_context() 52 context->rq_size_stride = ilog2(size) << 3 | (ilog2(stride) - 4); in mlx4_en_fill_qp_context() 54 context->sq_size_stride = ilog2(size) << 3 | (ilog2(stride) - 4); in mlx4_en_fill_qp_context() 56 context->params2 |= cpu_to_be32(MLX4_QP_BIT_FPP); in mlx4_en_fill_qp_context() 59 context->sq_size_stride = ilog2(TXBB_SIZE) - 4; in mlx4_en_fill_qp_context() 61 context->usr_page = cpu_to_be32(mlx4_to_hw_uar_index(mdev->dev, in mlx4_en_fill_qp_context() [all …]
|
/linux-6.6.21/drivers/platform/x86/intel/int1092/ |
D | intel_sar.c | 38 static void update_sar_data(struct wwan_sar_context *context) in update_sar_data() argument 41 &context->config_data[context->reg_value]; in update_sar_data() 44 context->sar_data.device_mode < config->total_dev_mode) { in update_sar_data() 48 if (context->sar_data.device_mode == in update_sar_data() 53 context->sar_data.antennatable_index = dev_mode->antennatable_index; in update_sar_data() 54 context->sar_data.bandtable_index = dev_mode->bandtable_index; in update_sar_data() 55 context->sar_data.sartable_index = dev_mode->sartable_index; in update_sar_data() 77 static acpi_status parse_package(struct wwan_sar_context *context, union acpi_object *item) in parse_package() argument 89 data = &context->config_data[reg]; in parse_package() 130 struct wwan_sar_context *context = dev_get_drvdata(&device->dev); in sar_get_device_mode() local [all …]
|
/linux-6.6.21/drivers/gpu/drm/amd/display/dc/dml/dcn32/ |
D | dcn32_fpu.c | 263 struct dc_state *context, in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch() argument 269 struct vba_vars_st *vba = &context->bw_ctx.dml.vba; in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch() 271 …enum clock_change_support temp_clock_change_support = vba->DRAMClockChangeSupport[vlevel][context-… in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch() 277 …vba->DRAMClockChangeSupport[vlevel][context->bw_ctx.dml.vba.maxMpcComb] = temp_clock_change_suppor… in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch() 278 context->bw_ctx.dml.soc.dram_clock_change_latency_us = in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch() 280 dcn32_internal_validate_bw(dc, context, pipes, &pipe_cnt, &vlevel, false); in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch() 284 dcn32_subvp_in_use(dc, context)) in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch() 285 …vba->DRAMClockChangeSupport[vlevel][context->bw_ctx.dml.vba.maxMpcComb] = temp_clock_change_suppor… in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch() 287 if (vlevel < context->bw_ctx.dml.vba.soc.num_states && in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch() 321 struct dc_state *context, in dcn32_helper_populate_phantom_dlg_params() argument [all …]
|
/linux-6.6.21/arch/riscv/kernel/ |
D | suspend.c | 11 void suspend_save_csrs(struct suspend_context *context) in suspend_save_csrs() argument 13 context->scratch = csr_read(CSR_SCRATCH); in suspend_save_csrs() 14 context->tvec = csr_read(CSR_TVEC); in suspend_save_csrs() 15 context->ie = csr_read(CSR_IE); in suspend_save_csrs() 28 context->satp = csr_read(CSR_SATP); in suspend_save_csrs() 32 void suspend_restore_csrs(struct suspend_context *context) in suspend_restore_csrs() argument 34 csr_write(CSR_SCRATCH, context->scratch); in suspend_restore_csrs() 35 csr_write(CSR_TVEC, context->tvec); in suspend_restore_csrs() 36 csr_write(CSR_IE, context->ie); in suspend_restore_csrs() 39 csr_write(CSR_SATP, context->satp); in suspend_restore_csrs() [all …]
|
/linux-6.6.21/drivers/net/ethernet/qlogic/qed/ |
D | qed_nvmetcp_fw_funcs.c | 129 init_nvmetcp_task_params(struct e5_nvmetcp_task_context *context, in init_nvmetcp_task_params() argument 133 context->ystorm_st_context.state.cccid = task_params->host_cccid; in init_nvmetcp_task_params() 134 SET_FIELD(context->ustorm_st_context.error_flags, USTORM_NVMETCP_TASK_ST_CTX_NVME_TCP, 1); in init_nvmetcp_task_params() 135 context->ustorm_st_context.nvme_tcp_opaque_lo = cpu_to_le32(task_params->opq.lo); in init_nvmetcp_task_params() 136 context->ustorm_st_context.nvme_tcp_opaque_hi = cpu_to_le32(task_params->opq.hi); in init_nvmetcp_task_params() 145 struct e5_nvmetcp_task_context *context = task_params->context; in init_default_nvmetcp_task() local 146 const u8 val_byte = context->mstorm_ag_context.cdu_validation; in init_default_nvmetcp_task() 149 memset(context, 0, sizeof(*context)); in init_default_nvmetcp_task() 150 init_nvmetcp_task_params(context, task_params, in init_default_nvmetcp_task() 159 context->ystorm_st_context.pdu_hdr.task_hdr.reg[dw_index] = in init_default_nvmetcp_task() [all …]
|
/linux-6.6.21/arch/s390/include/asm/ |
D | mmu_context.h | 24 spin_lock_init(&mm->context.lock); in init_new_context() 25 INIT_LIST_HEAD(&mm->context.pgtable_list); in init_new_context() 26 INIT_LIST_HEAD(&mm->context.gmap_list); in init_new_context() 27 cpumask_clear(&mm->context.cpu_attach_mask); in init_new_context() 28 atomic_set(&mm->context.flush_count, 0); in init_new_context() 29 atomic_set(&mm->context.protected_count, 0); in init_new_context() 30 mm->context.gmap_asce = 0; in init_new_context() 31 mm->context.flush_mm = 0; in init_new_context() 33 mm->context.alloc_pgste = page_table_allocate_pgste || in init_new_context() 35 (current->mm && current->mm->context.alloc_pgste); in init_new_context() [all …]
|
/linux-6.6.21/drivers/gpu/drm/amd/display/dc/dml/dcn30/ |
D | dcn30_fpu.c | 365 void dcn30_fpu_update_soc_for_wm_a(struct dc *dc, struct dc_state *context) in dcn30_fpu_update_soc_for_wm_a() argument 371 if (!context->bw_ctx.bw.dcn.clk.fw_based_mclk_switching || in dcn30_fpu_update_soc_for_wm_a() 372 context->bw_ctx.dml.soc.dram_clock_change_latency_us == 0) in dcn30_fpu_update_soc_for_wm_a() 373 …context->bw_ctx.dml.soc.dram_clock_change_latency_us = dc->clk_mgr->bw_params->wm_table.nv_entries… in dcn30_fpu_update_soc_for_wm_a() 374 …context->bw_ctx.dml.soc.sr_enter_plus_exit_time_us = dc->clk_mgr->bw_params->wm_table.nv_entries[W… in dcn30_fpu_update_soc_for_wm_a() 375 …context->bw_ctx.dml.soc.sr_exit_time_us = dc->clk_mgr->bw_params->wm_table.nv_entries[WM_A].dml_in… in dcn30_fpu_update_soc_for_wm_a() 380 struct dc *dc, struct dc_state *context, in dcn30_fpu_calculate_wm_and_dlg() argument 385 int maxMpcComb = context->bw_ctx.dml.vba.maxMpcComb; in dcn30_fpu_calculate_wm_and_dlg() 387 double dcfclk = context->bw_ctx.dml.vba.DCFCLKState[vlevel][maxMpcComb]; in dcn30_fpu_calculate_wm_and_dlg() 388 …bool pstate_en = context->bw_ctx.dml.vba.DRAMClockChangeSupport[vlevel][maxMpcComb] != dm_dram_clo… in dcn30_fpu_calculate_wm_and_dlg() [all …]
|
/linux-6.6.21/arch/sparc/mm/ |
D | tsb.c | 124 spin_lock_irqsave(&mm->context.lock, flags); in flush_tsb_user() 127 base = (unsigned long) mm->context.tsb_block[MM_TSB_BASE].tsb; in flush_tsb_user() 128 nentries = mm->context.tsb_block[MM_TSB_BASE].tsb_nentries; in flush_tsb_user() 140 else if (mm->context.tsb_block[MM_TSB_HUGE].tsb) { in flush_tsb_user() 141 base = (unsigned long) mm->context.tsb_block[MM_TSB_HUGE].tsb; in flush_tsb_user() 142 nentries = mm->context.tsb_block[MM_TSB_HUGE].tsb_nentries; in flush_tsb_user() 149 spin_unlock_irqrestore(&mm->context.lock, flags); in flush_tsb_user() 157 spin_lock_irqsave(&mm->context.lock, flags); in flush_tsb_user_page() 160 base = (unsigned long) mm->context.tsb_block[MM_TSB_BASE].tsb; in flush_tsb_user_page() 161 nentries = mm->context.tsb_block[MM_TSB_BASE].tsb_nentries; in flush_tsb_user_page() [all …]
|
/linux-6.6.21/arch/powerpc/mm/book3s64/ |
D | mmu_context.c | 99 mm->context.hash_context = kmalloc(sizeof(struct hash_mm_context), in hash__init_new_context() 101 if (!mm->context.hash_context) in hash__init_new_context() 118 if (mm->context.id == 0) { in hash__init_new_context() 119 memset(mm->context.hash_context, 0, sizeof(struct hash_mm_context)); in hash__init_new_context() 123 …memcpy(mm->context.hash_context, current->mm->context.hash_context, sizeof(struct hash_mm_context)… in hash__init_new_context() 126 if (current->mm->context.hash_context->spt) { in hash__init_new_context() 127 mm->context.hash_context->spt = kmalloc(sizeof(struct subpage_prot_table), in hash__init_new_context() 129 if (!mm->context.hash_context->spt) { in hash__init_new_context() 130 kfree(mm->context.hash_context); in hash__init_new_context() 137 index = realloc_context_ids(&mm->context); in hash__init_new_context() [all …]
|
/linux-6.6.21/drivers/usb/image/ |
D | microtek.c | 191 MTS_DEBUG("transfer = 0x%x context = 0x%x\n",(int)transfer,(int)context ); \ 192 …MTS_DEBUG("status = 0x%x data-length = 0x%x sent = 0x%x\n",transfer->status,(int)context->data_len… 193 mts_debug_dump(context->instance);\ 208 struct mts_transfer_context* context = (struct mts_transfer_context*)transfer->context; \ 385 context->instance->usb_dev, in mts_int_submit_urb() 390 context in mts_int_submit_urb() 396 set_host_byte(context->srb, DID_ERROR); in mts_int_submit_urb() 407 if ( likely(context->final_callback != NULL) ) in mts_transfer_cleanup() 408 context->final_callback(context->srb); in mts_transfer_cleanup() 415 context->srb->result &= MTS_SCSI_ERR_MASK; in mts_transfer_done() [all …]
|
/linux-6.6.21/drivers/infiniband/hw/hns/ |
D | hns_roce_cmd.c | 79 struct hns_roce_cmd_context *context = in hns_roce_cmd_event() local 80 &hr_dev->cmd.context[token % hr_dev->cmd.max_cmds]; in hns_roce_cmd_event() 82 if (unlikely(token != context->token)) { in hns_roce_cmd_event() 85 token, context->token); in hns_roce_cmd_event() 89 context->result = (status == HNS_ROCE_CMD_SUCCESS) ? 0 : (-EIO); in hns_roce_cmd_event() 90 context->out_param = out_param; in hns_roce_cmd_event() 91 complete(&context->done); in hns_roce_cmd_event() 98 struct hns_roce_cmd_context *context; in __hns_roce_cmd_mbox_wait() local 105 context = &cmd->context[cmd->free_head]; in __hns_roce_cmd_mbox_wait() 106 cmd->free_head = context->next; in __hns_roce_cmd_mbox_wait() [all …]
|