Home
last modified time | relevance | path

Searched refs:node_mask (Results 1 – 7 of 7) sorted by relevance

/linux-5.19.10/tools/perf/util/
Dmmap.c101 unsigned long *node_mask; in perf_mmap__aio_bind() local
109 node_mask = bitmap_zalloc(node_index + 1); in perf_mmap__aio_bind()
110 if (!node_mask) { in perf_mmap__aio_bind()
114 set_bit(node_index, node_mask); in perf_mmap__aio_bind()
115 if (mbind(data, mmap_len, MPOL_BIND, node_mask, node_index + 1 + 1, 0)) { in perf_mmap__aio_bind()
120 bitmap_free(node_mask); in perf_mmap__aio_bind()
/linux-5.19.10/net/netfilter/
Dxt_cluster.c118 return !!((1 << hash) & info->node_mask) ^ in xt_cluster_mt()
132 if (info->node_mask >= (1ULL << info->total_nodes)) { in xt_cluster_mt_checkentry()
/linux-5.19.10/include/uapi/linux/netfilter/
Dxt_cluster.h13 __u32 node_mask; member
/linux-5.19.10/tools/perf/bench/
Dnuma.c394 struct bitmask *node_mask; in bind_to_memnode() local
400 node_mask = numa_allocate_nodemask(); in bind_to_memnode()
401 BUG_ON(!node_mask); in bind_to_memnode()
403 numa_bitmask_clearall(node_mask); in bind_to_memnode()
404 numa_bitmask_setbit(node_mask, node); in bind_to_memnode()
406 ret = set_mempolicy(MPOL_BIND, node_mask->maskp, node_mask->size + 1); in bind_to_memnode()
407 dprintf("binding to node %d, mask: %016lx => %d\n", node, *node_mask->maskp, ret); in bind_to_memnode()
409 numa_bitmask_free(node_mask); in bind_to_memnode()
/linux-5.19.10/drivers/scsi/
Dstorvsc_drv.c1353 const struct cpumask *node_mask; in get_og_chn() local
1370 node_mask = cpumask_of_node(cpu_to_node(q_num)); in get_og_chn()
1374 if (cpumask_test_cpu(tgt_cpu, node_mask)) in get_og_chn()
1387 if (!cpumask_test_cpu(tgt_cpu, node_mask)) in get_og_chn()
1408 const struct cpumask *node_mask; in storvsc_do_io() local
1430 node_mask = cpumask_of_node(cpu_to_node(q_num)); in storvsc_do_io()
1433 if (!cpumask_test_cpu(tgt_cpu, node_mask)) in storvsc_do_io()
1464 if (cpumask_test_cpu(tgt_cpu, node_mask)) in storvsc_do_io()
/linux-5.19.10/drivers/infiniband/hw/hfi1/
Daffinity.c999 const struct cpumask *node_mask, in hfi1_get_proc_affinity() local
1110 node_mask = cpumask_of_node(node); in hfi1_get_proc_affinity()
1112 cpumask_pr_args(node_mask)); in hfi1_get_proc_affinity()
1115 cpumask_and(available_mask, hw_thread_mask, node_mask); in hfi1_get_proc_affinity()
1144 cpumask_andnot(available_mask, available_mask, node_mask); in hfi1_get_proc_affinity()
/linux-5.19.10/drivers/block/mtip32xx/
Dmtip32xx.c3710 const struct cpumask *node_mask; in get_least_used_cpu_on_node() local
3712 node_mask = cpumask_of_node(node); in get_least_used_cpu_on_node()
3713 least_used_cpu = cpumask_first(node_mask); in get_least_used_cpu_on_node()
3717 for_each_cpu(cpu, node_mask) { in get_least_used_cpu_on_node()
3813 const struct cpumask *node_mask; in mtip_pci_probe() local
3878 node_mask = cpumask_of_node(dd->numa_node); in mtip_pci_probe()
3879 if (!cpumask_empty(node_mask)) { in mtip_pci_probe()
3880 for_each_cpu(cpu, node_mask) in mtip_pci_probe()
3888 topology_physical_package_id(cpumask_first(node_mask)), in mtip_pci_probe()