/linux-6.6.21/drivers/net/ethernet/huawei/hinic/ |
D | hinic_hw_csr.h | 21 #define HINIC_CSR_DMA_ATTR_ADDR(idx) \ argument 22 (HINIC_DMA_ATTR_BASE + (idx) * HINIC_DMA_ATTR_STRIDE) 26 #define HINIC_CSR_PPF_ELECTION_ADDR(idx) \ argument 27 (HINIC_ELECTION_BASE + (idx) * HINIC_PPF_ELECTION_STRIDE) 34 #define HINIC_CSR_API_CMD_CHAIN_HEAD_HI_ADDR(idx) \ argument 35 (HINIC_CSR_API_CMD_BASE + 0x0 + (idx) * HINIC_CSR_API_CMD_STRIDE) 37 #define HINIC_CSR_API_CMD_CHAIN_HEAD_LO_ADDR(idx) \ argument 38 (HINIC_CSR_API_CMD_BASE + 0x4 + (idx) * HINIC_CSR_API_CMD_STRIDE) 40 #define HINIC_CSR_API_CMD_STATUS_HI_ADDR(idx) \ argument 41 (HINIC_CSR_API_CMD_BASE + 0x8 + (idx) * HINIC_CSR_API_CMD_STRIDE) [all …]
|
/linux-6.6.21/kernel/events/ |
D | hw_breakpoint_test.c | 31 static struct perf_event *register_test_bp(int cpu, struct task_struct *tsk, int idx) in register_test_bp() argument 35 if (WARN_ON(idx < 0 || idx >= MAX_TEST_BREAKPOINTS)) in register_test_bp() 39 attr.bp_addr = (unsigned long)&break_vars[idx]; in register_test_bp() 124 int idx = 0; in test_one_cpu() local 126 fill_bp_slots(test, &idx, get_test_cpu(0), NULL, 0); in test_one_cpu() 127 TEST_EXPECT_NOSPC(register_test_bp(-1, current, idx)); in test_one_cpu() 128 TEST_EXPECT_NOSPC(register_test_bp(get_test_cpu(0), NULL, idx)); in test_one_cpu() 133 int idx = 0; in test_many_cpus() local 138 bool do_continue = fill_bp_slots(test, &idx, cpu, NULL, 0); in test_many_cpus() 140 TEST_EXPECT_NOSPC(register_test_bp(cpu, NULL, idx)); in test_many_cpus() [all …]
|
/linux-6.6.21/tools/testing/selftests/bpf/ |
D | uprobe_multi.c | 10 #define NAME(name, idx) PASTE(name, idx) argument 12 #define DEF(name, idx) int NAME(name, idx)(void) { return 0; } argument 13 #define CALL(name, idx) NAME(name, idx)(); argument 15 #define F(body, name, idx) body(name, idx) argument 17 #define F10(body, name, idx) \ argument 18 F(body, PASTE(name, idx), 0) F(body, PASTE(name, idx), 1) F(body, PASTE(name, idx), 2) \ 19 F(body, PASTE(name, idx), 3) F(body, PASTE(name, idx), 4) F(body, PASTE(name, idx), 5) \ 20 F(body, PASTE(name, idx), 6) F(body, PASTE(name, idx), 7) F(body, PASTE(name, idx), 8) \ 21 F(body, PASTE(name, idx), 9) 23 #define F100(body, name, idx) \ argument [all …]
|
/linux-6.6.21/drivers/net/ethernet/microchip/vcap/ |
D | vcap_api_debugfs_kunit.c | 41 int idx; in test_val_keyset() local 46 for (idx = 0; idx < kslist->cnt; idx++) { in test_val_keyset() 47 if (kslist->keysets[idx] == VCAP_KFS_ETAG) in test_val_keyset() 48 return kslist->keysets[idx]; in test_val_keyset() 49 if (kslist->keysets[idx] == in test_val_keyset() 51 return kslist->keysets[idx]; in test_val_keyset() 52 if (kslist->keysets[idx] == in test_val_keyset() 54 return kslist->keysets[idx]; in test_val_keyset() 55 if (kslist->keysets[idx] == in test_val_keyset() 57 return kslist->keysets[idx]; in test_val_keyset() [all …]
|
/linux-6.6.21/drivers/net/ethernet/microchip/lan966x/ |
D | lan966x_ethtool.c | 300 uint idx = i * lan966x->num_stats; in lan966x_stats_update() local 308 lan966x_add_cnt(&lan966x->stats[idx++], in lan966x_stats_update() 362 u32 idx; in lan966x_get_eth_mac_stats() local 366 idx = port->chip_port * lan966x->num_stats; in lan966x_get_eth_mac_stats() 371 lan966x->stats[idx + SYS_COUNT_TX_UC] + in lan966x_get_eth_mac_stats() 372 lan966x->stats[idx + SYS_COUNT_TX_MC] + in lan966x_get_eth_mac_stats() 373 lan966x->stats[idx + SYS_COUNT_TX_BC] + in lan966x_get_eth_mac_stats() 374 lan966x->stats[idx + SYS_COUNT_TX_PMAC_UC] + in lan966x_get_eth_mac_stats() 375 lan966x->stats[idx + SYS_COUNT_TX_PMAC_MC] + in lan966x_get_eth_mac_stats() 376 lan966x->stats[idx + SYS_COUNT_TX_PMAC_BC]; in lan966x_get_eth_mac_stats() [all …]
|
/linux-6.6.21/drivers/crypto/ccree/ |
D | cc_aead.c | 303 unsigned int idx = 0; in hmac_setkey() local 309 hw_desc_init(&desc[idx]); in hmac_setkey() 310 set_cipher_mode(&desc[idx], hash_mode); in hmac_setkey() 311 set_din_sram(&desc[idx], in hmac_setkey() 315 set_flow_mode(&desc[idx], S_DIN_to_HASH); in hmac_setkey() 316 set_setup_mode(&desc[idx], SETUP_LOAD_STATE0); in hmac_setkey() 317 idx++; in hmac_setkey() 320 hw_desc_init(&desc[idx]); in hmac_setkey() 321 set_cipher_mode(&desc[idx], hash_mode); in hmac_setkey() 322 set_din_const(&desc[idx], 0, ctx->hash_len); in hmac_setkey() [all …]
|
D | cc_hash.c | 342 int idx) in cc_fin_result() argument 350 hw_desc_init(&desc[idx]); in cc_fin_result() 351 set_hash_cipher_mode(&desc[idx], ctx->hw_mode, ctx->hash_mode); in cc_fin_result() 352 set_dout_dlli(&desc[idx], state->digest_result_dma_addr, digestsize, in cc_fin_result() 354 set_queue_last_ind(ctx->drvdata, &desc[idx]); in cc_fin_result() 355 set_flow_mode(&desc[idx], S_HASH_to_DOUT); in cc_fin_result() 356 set_setup_mode(&desc[idx], SETUP_WRITE_STATE0); in cc_fin_result() 357 set_cipher_config1(&desc[idx], HASH_PADDING_DISABLED); in cc_fin_result() 358 cc_set_endianity(ctx->hash_mode, &desc[idx]); in cc_fin_result() 359 idx++; in cc_fin_result() [all …]
|
/linux-6.6.21/drivers/net/can/sja1000/ |
D | sja1000_isa.c | 122 int idx = pdev->id; in sja1000_isa_probe() local 126 idx, port[idx], mem[idx], irq[idx]); in sja1000_isa_probe() 128 if (mem[idx]) { in sja1000_isa_probe() 129 if (!request_mem_region(mem[idx], iosize, DRV_NAME)) { in sja1000_isa_probe() 133 base = ioremap(mem[idx], iosize); in sja1000_isa_probe() 139 if (indirect[idx] > 0 || in sja1000_isa_probe() 140 (indirect[idx] == -1 && indirect[0] > 0)) in sja1000_isa_probe() 142 if (!request_region(port[idx], iosize, DRV_NAME)) { in sja1000_isa_probe() 155 dev->irq = irq[idx]; in sja1000_isa_probe() 157 if (mem[idx]) { in sja1000_isa_probe() [all …]
|
/linux-6.6.21/drivers/net/ethernet/microchip/sparx5/ |
D | sparx5_calendar.c | 162 u32 cal[7], value, idx, portno; in sparx5_config_auto_calendar() local 216 for (idx = 0; idx < ARRAY_SIZE(cal); idx++) in sparx5_config_auto_calendar() 217 spx5_wr(cal[idx], sparx5, QSYS_CAL_AUTO(idx)); in sparx5_config_auto_calendar() 228 for (idx = 2; idx < 5; idx++) in sparx5_config_auto_calendar() 231 HSCH_OUTB_SHARE_ENA(idx)); in sparx5_config_auto_calendar() 256 u32 idx = 0, len = 0; in sparx5_dsm_cal_len() local 258 while (idx < SPX5_DSM_CAL_LEN) { in sparx5_dsm_cal_len() 259 if (cal[idx] != SPX5_DSM_CAL_EMPTY) in sparx5_dsm_cal_len() 261 idx++; in sparx5_dsm_cal_len() 268 u32 idx = 0, tmp; in sparx5_dsm_cp_cal() local [all …]
|
/linux-6.6.21/drivers/net/netdevsim/ |
D | macsec.c | 38 int idx; in nsim_macsec_add_secy() local 43 for (idx = 0; idx < NSIM_MACSEC_MAX_SECY_COUNT; idx++) { in nsim_macsec_add_secy() 44 if (!ns->macsec.nsim_secy[idx].used) in nsim_macsec_add_secy() 48 if (idx == NSIM_MACSEC_MAX_SECY_COUNT) { in nsim_macsec_add_secy() 55 __func__, sci_to_cpu(ctx->secy->sci), idx); in nsim_macsec_add_secy() 56 ns->macsec.nsim_secy[idx].used = true; in nsim_macsec_add_secy() 57 ns->macsec.nsim_secy[idx].nsim_rxsc_count = 0; in nsim_macsec_add_secy() 58 ns->macsec.nsim_secy[idx].sci = ctx->secy->sci; in nsim_macsec_add_secy() 67 int idx; in nsim_macsec_upd_secy() local 69 idx = nsim_macsec_find_secy(ns, ctx->secy->sci); in nsim_macsec_upd_secy() [all …]
|
/linux-6.6.21/drivers/net/can/cc770/ |
D | cc770_isa.c | 167 int idx = pdev->id; in cc770_isa_probe() local 172 idx, port[idx], mem[idx], irq[idx]); in cc770_isa_probe() 173 if (mem[idx]) { in cc770_isa_probe() 174 if (!request_mem_region(mem[idx], iosize, KBUILD_MODNAME)) { in cc770_isa_probe() 178 base = ioremap(mem[idx], iosize); in cc770_isa_probe() 184 if (indirect[idx] > 0 || in cc770_isa_probe() 185 (indirect[idx] == -1 && indirect[0] > 0)) in cc770_isa_probe() 187 if (!request_region(port[idx], iosize, KBUILD_MODNAME)) { in cc770_isa_probe() 200 dev->irq = irq[idx]; in cc770_isa_probe() 202 if (mem[idx]) { in cc770_isa_probe() [all …]
|
/linux-6.6.21/tools/testing/selftests/kvm/lib/ |
D | sparsebit.c | 172 sparsebit_idx_t idx; /* index of least-significant bit in mask */ member 287 root->idx = subtree->idx; in node_copy_subtree() 310 static struct node *node_find(struct sparsebit *s, sparsebit_idx_t idx) in node_find() argument 316 nodep = nodep->idx > idx ? nodep->left : nodep->right) { in node_find() 317 if (idx >= nodep->idx && in node_find() 318 idx <= nodep->idx + MASK_BITS + nodep->num_after - 1) in node_find() 333 static struct node *node_add(struct sparsebit *s, sparsebit_idx_t idx) in node_add() argument 344 nodep->idx = idx & -MASK_BITS; in node_add() 358 if (idx < parentp->idx) { in node_add() 366 assert(idx > parentp->idx + MASK_BITS + parentp->num_after - 1); in node_add() [all …]
|
/linux-6.6.21/lib/ |
D | find_bit.c | 30 unsigned long idx, val, sz = (size); \ 32 for (idx = 0; idx * BITS_PER_LONG < sz; idx++) { \ 35 sz = min(idx * BITS_PER_LONG + __ffs(MUNGE(val)), sz); \ 52 unsigned long mask, idx, tmp, sz = (size), __start = (start); \ 58 idx = __start / BITS_PER_LONG; \ 61 if ((idx + 1) * BITS_PER_LONG >= sz) \ 63 idx++; \ 66 sz = min(idx * BITS_PER_LONG + __ffs(MUNGE(tmp)), sz); \ 73 unsigned long sz = (size), nr = (num), idx, w, tmp; \ 75 for (idx = 0; (idx + 1) * BITS_PER_LONG <= sz; idx++) { \ [all …]
|
/linux-6.6.21/kernel/sched/ |
D | cpudeadline.c | 25 static void cpudl_heapify_down(struct cpudl *cp, int idx) in cpudl_heapify_down() argument 29 int orig_cpu = cp->elements[idx].cpu; in cpudl_heapify_down() 30 u64 orig_dl = cp->elements[idx].dl; in cpudl_heapify_down() 32 if (left_child(idx) >= cp->size) in cpudl_heapify_down() 39 l = left_child(idx); in cpudl_heapify_down() 40 r = right_child(idx); in cpudl_heapify_down() 41 largest = idx; in cpudl_heapify_down() 53 if (largest == idx) in cpudl_heapify_down() 57 cp->elements[idx].cpu = cp->elements[largest].cpu; in cpudl_heapify_down() 58 cp->elements[idx].dl = cp->elements[largest].dl; in cpudl_heapify_down() [all …]
|
/linux-6.6.21/arch/x86/um/ |
D | tls_32.c | 66 int idx; in get_free_idx() local 68 for (idx = 0; idx < GDT_ENTRY_TLS_ENTRIES; idx++) in get_free_idx() 69 if (!t->arch.tls_array[idx].present) in get_free_idx() 70 return idx + GDT_ENTRY_TLS_MIN; in get_free_idx() 92 int idx; in load_TLS() local 94 for (idx = GDT_ENTRY_TLS_MIN; idx < GDT_ENTRY_TLS_MAX; idx++) { in load_TLS() 96 &to->thread.arch.tls_array[idx - GDT_ENTRY_TLS_MIN]; in load_TLS() 105 curr->tls.entry_number = idx; in load_TLS() 201 int idx, int flushed) in set_tls_entry() argument 205 if (idx < GDT_ENTRY_TLS_MIN || idx > GDT_ENTRY_TLS_MAX) in set_tls_entry() [all …]
|
/linux-6.6.21/tools/perf/scripts/perl/Perf-Trace-Util/lib/Perf/Trace/ |
D | Core.pm | 39 foreach my $idx (sort {$a <=> $b} keys %trace_flags) { 40 if (!$value && !$idx) { 45 if ($idx && ($value & $idx) == $idx) { 49 $string .= "$trace_flags{$idx}"; 51 $value &= ~$idx; 69 foreach my $idx (sort {$a <=> $b} keys %{$flag_fields{$event_name}{$field_name}{"values"}}) { 70 if (!$value && !$idx) { 71 $string .= "$flag_fields{$event_name}{$field_name}{'values'}{$idx}"; 74 if ($idx && ($value & $idx) == $idx) { 78 $string .= "$flag_fields{$event_name}{$field_name}{'values'}{$idx}"; [all …]
|
/linux-6.6.21/drivers/clk/uniphier/ |
D | clk-uniphier-sys.c | 27 #define UNIPHIER_LD4_SYS_CLK_NAND(idx) \ argument 29 UNIPHIER_CLK_GATE("nand", (idx), "nand-50m", 0x2104, 2) 31 #define UNIPHIER_PRO5_SYS_CLK_NAND(idx) \ argument 33 UNIPHIER_CLK_GATE("nand", (idx), "nand-50m", 0x2104, 2) 35 #define UNIPHIER_LD11_SYS_CLK_NAND(idx) \ argument 37 UNIPHIER_CLK_GATE("nand", (idx), "nand-50m", 0x210c, 0) 39 #define UNIPHIER_SYS_CLK_NAND_4X(idx) \ argument 40 UNIPHIER_CLK_FACTOR("nand-4x", (idx), "nand", 4, 1) 42 #define UNIPHIER_LD11_SYS_CLK_EMMC(idx) \ argument 43 UNIPHIER_CLK_GATE("emmc", (idx), NULL, 0x210c, 2) [all …]
|
/linux-6.6.21/include/asm-generic/ |
D | fixmap.h | 30 static __always_inline unsigned long fix_to_virt(const unsigned int idx) in fix_to_virt() argument 32 BUILD_BUG_ON(idx >= __end_of_fixed_addresses); in fix_to_virt() 33 return __fix_to_virt(idx); in fix_to_virt() 64 #define set_fixmap(idx, phys) \ argument 65 __set_fixmap(idx, phys, FIXMAP_PAGE_NORMAL) 69 #define clear_fixmap(idx) \ argument 70 __set_fixmap(idx, 0, FIXMAP_PAGE_CLEAR) 74 #define __set_fixmap_offset(idx, phys, flags) \ argument 77 __set_fixmap(idx, phys, flags); \ 78 ________addr = fix_to_virt(idx) + ((phys) & (PAGE_SIZE - 1)); \ [all …]
|
/linux-6.6.21/mm/ |
D | hugetlb_cgroup.c | 33 __hugetlb_cgroup_counter_from_cgroup(struct hugetlb_cgroup *h_cg, int idx, in __hugetlb_cgroup_counter_from_cgroup() argument 37 return &h_cg->rsvd_hugepage[idx]; in __hugetlb_cgroup_counter_from_cgroup() 38 return &h_cg->hugepage[idx]; in __hugetlb_cgroup_counter_from_cgroup() 42 hugetlb_cgroup_counter_from_cgroup(struct hugetlb_cgroup *h_cg, int idx) in hugetlb_cgroup_counter_from_cgroup() argument 44 return __hugetlb_cgroup_counter_from_cgroup(h_cg, idx, false); in hugetlb_cgroup_counter_from_cgroup() 48 hugetlb_cgroup_counter_from_cgroup_rsvd(struct hugetlb_cgroup *h_cg, int idx) in hugetlb_cgroup_counter_from_cgroup_rsvd() argument 50 return __hugetlb_cgroup_counter_from_cgroup(h_cg, idx, true); in hugetlb_cgroup_counter_from_cgroup_rsvd() 91 int idx; in hugetlb_cgroup_init() local 93 for (idx = 0; idx < HUGE_MAX_HSTATE; idx++) { in hugetlb_cgroup_init() 101 parent_h_cgroup, idx); in hugetlb_cgroup_init() [all …]
|
/linux-6.6.21/arch/x86/events/intel/ |
D | uncore_nhmex.c | 249 if (hwc->idx == UNCORE_PMC_IDX_FIXED) in nhmex_uncore_msr_enable_event() 371 reg1->idx = 0; in nhmex_bbox_hw_config() 383 if (reg1->idx != EXTRA_REG_NONE) { in nhmex_bbox_msr_enable_event() 456 reg1->idx = 0; in nhmex_sbox_hw_config() 468 if (reg1->idx != EXTRA_REG_NONE) { in nhmex_sbox_msr_enable_event() 553 static bool nhmex_mbox_get_shared_reg(struct intel_uncore_box *box, int idx, u64 config) in nhmex_mbox_get_shared_reg() argument 560 if (idx < EXTRA_REG_NHMEX_M_ZDP_CTL_FVC) { in nhmex_mbox_get_shared_reg() 561 er = &box->shared_regs[idx]; in nhmex_mbox_get_shared_reg() 577 idx -= EXTRA_REG_NHMEX_M_ZDP_CTL_FVC; in nhmex_mbox_get_shared_reg() 578 if (WARN_ON_ONCE(idx >= 4)) in nhmex_mbox_get_shared_reg() [all …]
|
/linux-6.6.21/drivers/gpu/drm/radeon/ |
D | evergreen_cs.c | 755 unsigned idx) in evergreen_cs_track_validate_texture() argument 763 texdw[0] = radeon_get_ib_value(p, idx + 0); in evergreen_cs_track_validate_texture() 764 texdw[1] = radeon_get_ib_value(p, idx + 1); in evergreen_cs_track_validate_texture() 765 texdw[2] = radeon_get_ib_value(p, idx + 2); in evergreen_cs_track_validate_texture() 766 texdw[3] = radeon_get_ib_value(p, idx + 3); in evergreen_cs_track_validate_texture() 767 texdw[4] = radeon_get_ib_value(p, idx + 4); in evergreen_cs_track_validate_texture() 768 texdw[5] = radeon_get_ib_value(p, idx + 5); in evergreen_cs_track_validate_texture() 769 texdw[6] = radeon_get_ib_value(p, idx + 6); in evergreen_cs_track_validate_texture() 770 texdw[7] = radeon_get_ib_value(p, idx + 7); in evergreen_cs_track_validate_texture() 1049 unsigned idx, unsigned reg) in evergreen_packet0_check() argument [all …]
|
/linux-6.6.21/drivers/net/ethernet/ti/ |
D | cpsw_ale.c | 109 int idx, idx2; in cpsw_ale_get_field() local 112 idx = start / 32; in cpsw_ale_get_field() 115 if (idx != idx2) { in cpsw_ale_get_field() 119 start -= idx * 32; in cpsw_ale_get_field() 120 idx = 2 - idx; /* flip */ in cpsw_ale_get_field() 121 return (hi_val + (ale_entry[idx] >> start)) & BITMASK(bits); in cpsw_ale_get_field() 127 int idx, idx2; in cpsw_ale_set_field() local 130 idx = start / 32; in cpsw_ale_set_field() 133 if (idx != idx2) { in cpsw_ale_set_field() 138 start -= idx * 32; in cpsw_ale_set_field() [all …]
|
/linux-6.6.21/drivers/net/ethernet/chelsio/cxgb/ |
D | fpga_defs.h | 215 #define MAC_REG_ADDR(idx, reg) (MAC_REG_BASE + (idx) * 128 + (reg)) argument 217 #define MAC_REG_IDLO(idx) MAC_REG_ADDR(idx, A_GMAC_MACID_LO) argument 218 #define MAC_REG_IDHI(idx) MAC_REG_ADDR(idx, A_GMAC_MACID_HI) argument 219 #define MAC_REG_CSR(idx) MAC_REG_ADDR(idx, A_GMAC_CSR) argument 220 #define MAC_REG_IFS(idx) MAC_REG_ADDR(idx, A_GMAC_IFS) argument 221 #define MAC_REG_LARGEFRAMELENGTH(idx) MAC_REG_ADDR(idx, A_GMAC_JUMBO_FRAME_LEN) argument 222 #define MAC_REG_LINKDLY(idx) MAC_REG_ADDR(idx, A_GMAC_LNK_DLY) argument 223 #define MAC_REG_PAUSETIME(idx) MAC_REG_ADDR(idx, A_GMAC_PAUSETIME) argument 224 #define MAC_REG_CASTLO(idx) MAC_REG_ADDR(idx, A_GMAC_MCAST_LO) argument 225 #define MAC_REG_MCASTHI(idx) MAC_REG_ADDR(idx, A_GMAC_MCAST_HI) argument [all …]
|
/linux-6.6.21/tools/testing/selftests/kvm/x86_64/ |
D | hyperv_features.c | 25 uint32_t idx; member 47 GUEST_ASSERT(msr->idx); in guest_msr() 50 vector = wrmsr_safe(msr->idx, msr->write_val); in guest_msr() 52 if (!vector && (!msr->write || !is_write_only_msr(msr->idx))) in guest_msr() 53 vector = rdmsr_safe(msr->idx, &msr_val); in guest_msr() 58 msr->idx, msr->write ? "WR" : "RD", vector); in guest_msr() 62 msr->idx, msr->write ? "WR" : "RD", vector); in guest_msr() 64 if (vector || is_write_only_msr(msr->idx)) in guest_msr() 70 msr->idx, msr->write_val, msr_val); in guest_msr() 73 if (msr->idx == HV_X64_MSR_TSC_INVARIANT_CONTROL) { in guest_msr() [all …]
|
/linux-6.6.21/tools/lib/ |
D | find_bit.c | 29 unsigned long idx, val, sz = (size); \ 31 for (idx = 0; idx * BITS_PER_LONG < sz; idx++) { \ 34 sz = min(idx * BITS_PER_LONG + __ffs(MUNGE(val)), sz); \ 51 unsigned long mask, idx, tmp, sz = (size), __start = (start); \ 57 idx = __start / BITS_PER_LONG; \ 60 if ((idx + 1) * BITS_PER_LONG >= sz) \ 62 idx++; \ 65 sz = min(idx * BITS_PER_LONG + __ffs(MUNGE(tmp)), sz); \ 76 return FIND_FIRST_BIT(addr[idx], /* nop */, size); in _find_first_bit() 88 return FIND_FIRST_BIT(addr1[idx] & addr2[idx], /* nop */, size); in _find_first_and_bit() [all …]
|