/linux-6.6.21/rust/macros/ |
D | quote.rs | 42 ($span:expr => $($tt:tt)*) => {{ 47 let span = $span; localVariable 48 quote_spanned!(@proc tokens span $($tt)*); 52 (@proc $v:ident $span:ident) => {}; 53 (@proc $v:ident $span:ident #$id:ident $($tt:tt)*) => { 57 quote_spanned!(@proc $v $span $($tt)*); 59 (@proc $v:ident $span:ident #(#$id:ident)* $($tt:tt)*) => { 65 quote_spanned!(@proc $v $span $($tt)*); 67 (@proc $v:ident $span:ident ( $($inner:tt)* ) $($tt:tt)*) => { 69 quote_spanned!(@proc tokens $span $($inner)*); [all …]
|
D | paste.rs | 8 let mut span = None; in concat() localVariable 12 Some(TokenTree::Literal(lit)) => segments.push((lit.to_string(), lit.span())), in concat() 18 segments.push((value, ident.span())); in concat() 30 span.is_none(), in concat() 33 span = Some(sp); in concat() 46 TokenTree::Ident(Ident::new(&pasted, span.unwrap_or(group_span))) in concat() 53 let span = group.span(); in expand() localVariable 62 *token = concat(&stream[1..stream.len() - 1], span); in expand() 67 group.set_span(span); in expand()
|
/linux-6.6.21/drivers/net/ethernet/marvell/prestera/ |
D | prestera_span.c | 37 list_add_tail(&entry->list, &port->sw->span->entries); in prestera_span_entry_create() 49 prestera_span_entry_find_by_id(struct prestera_span *span, u8 span_id) in prestera_span_entry_find_by_id() argument 53 list_for_each_entry(entry, &span->entries, list) { in prestera_span_entry_find_by_id() 62 prestera_span_entry_find_by_port(struct prestera_span *span, in prestera_span_entry_find_by_port() argument 67 list_for_each_entry(entry, &span->entries, list) { in prestera_span_entry_find_by_port() 82 entry = prestera_span_entry_find_by_port(sw->span, port); in prestera_span_get() 108 entry = prestera_span_entry_find_by_id(sw->span, span_id); in prestera_span_put() 171 struct prestera_span *span; in prestera_span_init() local 173 span = kzalloc(sizeof(*span), GFP_KERNEL); in prestera_span_init() 174 if (!span) in prestera_span_init() [all …]
|
/linux-6.6.21/drivers/net/ethernet/mellanox/mlxsw/ |
D | spectrum_span.c | 46 struct mlxsw_sp_span *span; member 77 return atomic_read(&mlxsw_sp->span->active_entries_count); in mlxsw_sp_span_occ_get() 83 struct mlxsw_sp_span *span; in mlxsw_sp_span_init() local 90 span = kzalloc(struct_size(span, entries, entries_count), GFP_KERNEL); in mlxsw_sp_span_init() 91 if (!span) in mlxsw_sp_span_init() 93 refcount_set(&span->policer_id_base_ref_count, 0); in mlxsw_sp_span_init() 94 span->entries_count = entries_count; in mlxsw_sp_span_init() 95 atomic_set(&span->active_entries_count, 0); in mlxsw_sp_span_init() 96 mutex_init(&span->analyzed_ports_lock); in mlxsw_sp_span_init() 97 INIT_LIST_HEAD(&span->analyzed_ports_list); in mlxsw_sp_span_init() [all …]
|
/linux-6.6.21/drivers/scsi/megaraid/ |
D | megaraid_sas_fp.c | 123 u16 MR_LdSpanArrayGet(u32 ld, u32 span, struct MR_DRV_RAID_MAP_ALL *map) in MR_LdSpanArrayGet() argument 125 return le16_to_cpu(map->raidMap.ldSpanMap[ld].spanBlock[span].span.arrayRef); in MR_LdSpanArrayGet() 148 static struct MR_LD_SPAN *MR_LdSpanPtrGet(u32 ld, u32 span, in MR_LdSpanPtrGet() argument 151 return &map->raidMap.ldSpanMap[ld].spanBlock[span].span; in MR_LdSpanPtrGet() 379 u32 span, j; in MR_GetSpanBlock() local 381 for (span = 0; span < raid->spanDepth; span++, pSpanBlock++) { in MR_GetSpanBlock() 398 return span; in MR_GetSpanBlock() 430 u32 span, info; in mr_spanset_get_span_block() local 442 for (span = 0; span < raid->spanDepth; span++) in mr_spanset_get_span_block() 443 if (le32_to_cpu(map->raidMap.ldSpanMap[ld].spanBlock[span]. in mr_spanset_get_span_block() [all …]
|
/linux-6.6.21/drivers/iommu/iommufd/ |
D | pages.c | 938 struct interval_tree_double_span_iter span; member 960 struct interval_tree_double_span_iter span; in pfn_reader_unpin() local 965 interval_tree_for_each_double_span(&span, &pages->access_itree, in pfn_reader_unpin() 967 if (span.is_used) in pfn_reader_unpin() 970 batch_unpin(&pfns->batch, pages, span.start_hole - start, in pfn_reader_unpin() 971 span.last_hole - span.start_hole + 1); in pfn_reader_unpin() 978 struct interval_tree_double_span_iter *span = &pfns->span; in pfn_reader_fill_span() local 984 WARN_ON(span->last_used < start_index)) in pfn_reader_fill_span() 987 if (span->is_used == 1) { in pfn_reader_fill_span() 989 start_index, span->last_used); in pfn_reader_fill_span() [all …]
|
D | double_span.h | 46 #define interval_tree_for_each_double_span(span, itree1, itree2, first_index, \ argument 48 for (interval_tree_double_span_iter_first(span, itree1, itree2, \ 50 !interval_tree_double_span_iter_done(span); \ 51 interval_tree_double_span_iter_next(span))
|
D | io_pagetable.c | 72 static bool __alloc_iova_check_hole(struct interval_tree_double_span_iter *span, in __alloc_iova_check_hole() argument 77 if (span->is_used || span->last_hole - span->start_hole < length - 1) in __alloc_iova_check_hole() 80 span->start_hole = ALIGN(span->start_hole, iova_alignment) | in __alloc_iova_check_hole() 82 if (span->start_hole > span->last_hole || in __alloc_iova_check_hole() 83 span->last_hole - span->start_hole < length - 1) in __alloc_iova_check_hole() 88 static bool __alloc_iova_check_used(struct interval_tree_span_iter *span, in __alloc_iova_check_used() argument 93 if (span->is_hole || span->last_used - span->start_used < length - 1) in __alloc_iova_check_used() 96 span->start_used = ALIGN(span->start_used, iova_alignment) | in __alloc_iova_check_used() 98 if (span->start_used > span->last_used || in __alloc_iova_check_used() 99 span->last_used - span->start_used < length - 1) in __alloc_iova_check_used()
|
D | ioas.c | 67 struct interval_tree_span_iter span; in iommufd_ioas_iova_ranges() local 83 interval_tree_for_each_span(&span, &ioas->iopt.reserved_itree, 0, in iommufd_ioas_iova_ranges() 85 if (!span.is_hole) in iommufd_ioas_iova_ranges() 89 .start = span.start_hole, in iommufd_ioas_iova_ranges() 90 .last = span.last_hole, in iommufd_ioas_iova_ranges()
|
D | vfio_compat.c | 391 struct interval_tree_span_iter span; in iommufd_fill_cap_iova() local 393 interval_tree_for_each_span(&span, &ioas->iopt.reserved_itree, 0, in iommufd_fill_cap_iova() 397 if (!span.is_hole) in iommufd_fill_cap_iova() 399 range.start = span.start_hole; in iommufd_fill_cap_iova() 400 range.end = span.last_hole; in iommufd_fill_cap_iova()
|
/linux-6.6.21/drivers/iio/dac/ |
D | ltc2688.c | 139 int ret, reg, span; in ltc2688_span_get() local 145 span = FIELD_GET(LTC2688_CH_SPAN_MSK, reg); in ltc2688_span_get() 147 if (span >= LTC2688_SPAN_RANGE_MAX) in ltc2688_span_get() 150 return span; in ltc2688_span_get() 160 int span, fs; in ltc2688_scale_get() local 162 span = ltc2688_span_get(st, c); in ltc2688_scale_get() 163 if (span < 0) in ltc2688_scale_get() 164 return span; in ltc2688_scale_get() 166 fs = ltc2688_span_helper[span][1] - ltc2688_span_helper[span][0]; in ltc2688_scale_get() 177 int span; in ltc2688_offset_get() local [all …]
|
D | ad5791.c | 68 int (*get_lin_comp) (unsigned int span); 210 static int ad5791_get_lin_comp(unsigned int span) in ad5791_get_lin_comp() argument 212 if (span <= 10000) in ad5791_get_lin_comp() 214 else if (span <= 12000) in ad5791_get_lin_comp() 216 else if (span <= 16000) in ad5791_get_lin_comp() 218 else if (span <= 19000) in ad5791_get_lin_comp() 224 static int ad5780_get_lin_comp(unsigned int span) in ad5780_get_lin_comp() argument 226 if (span <= 10000) in ad5780_get_lin_comp()
|
/linux-6.6.21/tools/testing/selftests/resctrl/ |
D | cmt_test.c | 30 static int check_results(struct resctrl_val_param *param, size_t span, int no_of_bits) in check_results() argument 61 return show_cache_info(sum_llc_occu_resc, no_of_bits, span, in check_results() 80 size_t span; in cmt_resctrl_val() local 113 span = cache_size * n / count_of_bits; in cmt_resctrl_val() 121 ret = asprintf(&span_str, "%zu", span); in cmt_resctrl_val() 134 ret = check_results(¶m, span, n); in cmt_resctrl_val()
|
D | cat_test.c | 44 static int check_results(struct resctrl_val_param *param, size_t span) in check_results() argument 79 return show_cache_info(sum_llc_perf_miss, no_of_bits, span / 64, in check_results() 99 size_t span; in cat_perf_miss_val() local 144 span = cache_size * (count_of_bits - n) / count_of_bits; in cat_perf_miss_val() 166 span = cache_size * n / count_of_bits; in cat_perf_miss_val() 174 ret = cat_val(¶m, span); in cat_perf_miss_val() 176 ret = check_results(¶m, span); in cat_perf_miss_val()
|
D | mbm_test.c | 18 show_bw_info(unsigned long *bw_imc, unsigned long *bw_resc, size_t span) in show_bw_info() argument 43 ksft_print_msg("Span (MB): %zu\n", span / MB); in show_bw_info() 50 static int check_results(size_t span) in check_results() argument 82 ret = show_bw_info(bw_imc, bw_resc, span); in check_results()
|
D | resctrl.h | 96 int run_fill_buf(size_t span, int memflush, int op, bool once); 108 int cat_val(struct resctrl_val_param *param, size_t span);
|
D | fill_buf.c | 167 int run_fill_buf(size_t span, int memflush, int op, bool once) in run_fill_buf() argument 169 size_t cache_size = span; in run_fill_buf()
|
/linux-6.6.21/Documentation/translations/zh_CN/scheduler/ |
D | sched-domains.rst | 22 每个调度域管辖数个CPU(存储在->span字段中)。一个调度域的span必须是它的子调度域span的 25 CPU出现永远不会被指定任务运行的情况,直到允许的CPU掩码被显式设定。调度域的span字段意味 29 组织,存储在->groups指针中。这些组的CPU掩码的并集必须和调度域span字段一致。->groups
|
/linux-6.6.21/include/linux/ |
D | interval_tree.h | 82 #define interval_tree_for_each_span(span, itree, first_index, last_index) \ argument 83 for (interval_tree_span_iter_first(span, itree, \ 85 !interval_tree_span_iter_done(span); \ 86 interval_tree_span_iter_next(span))
|
/linux-6.6.21/tools/testing/selftests/bpf/progs/ |
D | test_access_variable_array.c | 8 unsigned long span = 0; variable 14 span = sd->span[0]; in BPF_PROG()
|
/linux-6.6.21/Documentation/scheduler/ |
D | sched-domains.rst | 10 Each scheduling domain spans a number of CPUs (stored in the ->span field). 11 A domain's span MUST be a superset of it child's span (this restriction could 12 be relaxed if the need arises), and a base domain for CPU i MUST span at least 13 i. The top domain for each CPU will generally span all CPUs in the system 16 explicitly set. A sched domain's span means "balance process load among these 22 domain's span. The group pointed to by the ->groups pointer MUST contain the CPU 58 The "base" domain will "span" the first level of the hierarchy. In the case 59 of SMT, you'll span all siblings of the physical CPU, with each group being 62 In SMP, the parent of the base domain will span all physical CPUs in the 64 of the SMP domain will span the entire machine, with each group having the
|
/linux-6.6.21/drivers/platform/surface/aggregator/ |
D | ssh_parser.h | 138 struct ssam_span *span) in sshp_buf_span_from() argument 140 span->ptr = buf->ptr + offset; in sshp_buf_span_from() 141 span->len = buf->len - offset; in sshp_buf_span_from()
|
/linux-6.6.21/drivers/net/wireless/realtek/rtw89/ |
D | sar.c | 58 #define RTW89_SAR_SPAN_VALID(span) ((span)->subband_high) argument 92 const struct rtw89_sar_span *span = NULL; in rtw89_query_sar_config_common() local 104 span = &rtw89_sar_overlapping_6ghz[idx]; in rtw89_query_sar_config_common() 107 if (span && RTW89_SAR_SPAN_VALID(span)) { in rtw89_query_sar_config_common() 108 subband_l = span->subband_low; in rtw89_query_sar_config_common() 109 subband_h = span->subband_high; in rtw89_query_sar_config_common()
|
/linux-6.6.21/arch/sh/mm/ |
D | pmb.c | 146 unsigned long span; in pmb_mapping_exists() local 169 span = pmbe->size; in pmb_mapping_exists() 176 span += iter->size; in pmb_mapping_exists() 181 if (size <= span) { in pmb_mapping_exists() 657 unsigned long span, newsize; in pmb_merge() local 661 span = newsize = head->size; in pmb_merge() 665 span += tail->size; in pmb_merge() 667 if (pmb_size_valid(span)) { in pmb_merge() 668 newsize = span; in pmb_merge()
|
/linux-6.6.21/Documentation/sphinx-static/ |
D | theme_rtd_colors.css | 29 span.menuselection { 33 code.kbd, code.kbd span {
|