Searched refs:workers (Results 1 – 17 of 17) sorted by relevance
/linux-6.6.21/Documentation/core-api/ |
D | workqueue.rst | 34 number of workers as the number of CPUs. The kernel grew a lot of MT 118 number of the currently runnable workers. Generally, work items are 122 workers on the CPU, the worker-pool doesn't start execution of a new 125 are pending work items. This allows using a minimal number of workers 128 Keeping idle workers around doesn't cost other than the memory space 140 Forward progress guarantee relies on that workers can be created when 142 through the use of rescue workers. All work items which might be used 169 worker-pools which host workers which are not bound to any 178 of mostly unused workers across different CPUs as the issuer 200 each other. Each maintains its separate pool of workers and [all …]
|
/linux-6.6.21/tools/testing/selftests/bpf/ |
D | test_progs.c | 420 if (verbose() && !env.workers) in test__end_subtest() 830 env->workers = atoi(arg); in parse_arg() 831 if (!env->workers) { in parse_arg() 836 env->workers = get_nprocs(); in parse_arg() 1008 for (i = 0; i < env.workers; i++) in sigint_handler() 1370 dispatcher_threads = calloc(sizeof(pthread_t), env.workers); in server_main() 1371 data = calloc(sizeof(struct dispatch_data), env.workers); in server_main() 1373 env.worker_current_test = calloc(sizeof(int), env.workers); in server_main() 1374 for (i = 0; i < env.workers; i++) { in server_main() 1387 for (i = 0; i < env.workers; i++) { in server_main() [all …]
|
D | test_progs.h | 125 int workers; /* number of worker process */ member
|
/linux-6.6.21/fs/erofs/ |
D | Kconfig | 129 bool "EROFS per-cpu decompression kthread workers" 132 Saying Y here enables per-CPU kthread workers pool to carry out 138 bool "EROFS high priority per-CPU kthread workers" 142 This permits EROFS to configure per-CPU kthread workers to run
|
/linux-6.6.21/drivers/md/ |
D | raid5.h | 518 struct r5worker *workers; member
|
D | raid5.c | 204 group->workers[0].working = true; in raid5_wakeup_stripe_thread() 206 queue_work_on(sh->cpu, raid5_wq, &group->workers[0].work); in raid5_wakeup_stripe_thread() 211 if (group->workers[i].working == false) { in raid5_wakeup_stripe_thread() 212 group->workers[i].working = true; in raid5_wakeup_stripe_thread() 214 &group->workers[i].work); in raid5_wakeup_stripe_thread() 7257 kfree(old_groups[0].workers); in raid5_store_group_thread_cnt() 7294 struct r5worker *workers; in alloc_thread_groups() local 7303 workers = kcalloc(size, *group_cnt, GFP_NOIO); in alloc_thread_groups() 7306 if (!*worker_groups || !workers) { in alloc_thread_groups() 7307 kfree(workers); in alloc_thread_groups() [all …]
|
/linux-6.6.21/net/l2tp/ |
D | Kconfig | 23 with home workers to connect to their offices.
|
/linux-6.6.21/fs/btrfs/ |
D | fs.h | 543 struct btrfs_workqueue *workers; member
|
D | bio.c | 641 btrfs_queue_work(fs_info->workers, &async->work); in btrfs_wq_submit_bio()
|
D | disk-io.c | 1759 btrfs_destroy_workqueue(fs_info->workers); in btrfs_stop_all_workers() 1952 fs_info->workers = in btrfs_init_workqueues() 1991 if (!(fs_info->workers && in btrfs_init_workqueues()
|
D | super.c | 1637 btrfs_workqueue_set_max(fs_info->workers, new_pool_size); in btrfs_resize_thread_pool()
|
/linux-6.6.21/drivers/block/mtip32xx/ |
D | mtip32xx.c | 733 int do_irq_enable = 1, i, workers; in mtip_handle_irq() local 754 for (i = 0, workers = 0; i < MTIP_MAX_SLOT_GROUPS; in mtip_handle_irq() 759 workers++; in mtip_handle_irq() 762 atomic_set(&dd->irq_workers_active, workers); in mtip_handle_irq() 763 if (workers) { in mtip_handle_irq()
|
/linux-6.6.21/kernel/ |
D | workqueue.c | 187 struct list_head workers; /* A: attached workers */ member 490 list_for_each_entry((worker), &(pool)->workers, node) \ 2114 list_add_tail(&worker->node, &pool->workers); in worker_attach_to_pool() 2139 if (list_empty(&pool->workers) && list_empty(&pool->dying_workers)) in worker_detach_from_pool() 3869 INIT_LIST_HEAD(&pool->workers); in init_worker_pool() 4011 if (!list_empty(&pool->workers) || !list_empty(&pool->dying_workers)) in put_unbound_pool()
|
/linux-6.6.21/Documentation/admin-guide/ |
D | workload-tracing.rst | 126 starts specified number (N) of workers that exercise various netdevice 264 The netdev stressor starts N workers that exercise various netdevice ioctl
|
D | kernel-per-CPU-kthreads.rst | 262 d. As of v3.18, Christoph Lameter's on-demand vmstat workers
|
/linux-6.6.21/Documentation/dev-tools/ |
D | kcov.rst | 248 exits (e.g. vhost workers).
|
/linux-6.6.21/Documentation/filesystems/ |
D | xfs-online-fsck-design.rst | 4898 Each inode btree chunk found by the first workqueue's workers are queued to the 4903 first workqueue's workers until the backlog eases. 4950 1. Start a round of repair with a workqueue and enough workers to keep the CPUs
|