1 /* SPDX-License-Identifier: GPL-2.0 */
2
3 #define _GNU_SOURCE
4 #include <linux/limits.h>
5 #include <linux/sched.h>
6 #include <sys/types.h>
7 #include <sys/mman.h>
8 #include <sys/wait.h>
9 #include <unistd.h>
10 #include <fcntl.h>
11 #include <sched.h>
12 #include <stdio.h>
13 #include <errno.h>
14 #include <signal.h>
15 #include <string.h>
16 #include <pthread.h>
17
18 #include "../kselftest.h"
19 #include "cgroup_util.h"
20
touch_anon(char * buf,size_t size)21 static int touch_anon(char *buf, size_t size)
22 {
23 int fd;
24 char *pos = buf;
25
26 fd = open("/dev/urandom", O_RDONLY);
27 if (fd < 0)
28 return -1;
29
30 while (size > 0) {
31 ssize_t ret = read(fd, pos, size);
32
33 if (ret < 0) {
34 if (errno != EINTR) {
35 close(fd);
36 return -1;
37 }
38 } else {
39 pos += ret;
40 size -= ret;
41 }
42 }
43 close(fd);
44
45 return 0;
46 }
47
alloc_and_touch_anon_noexit(const char * cgroup,void * arg)48 static int alloc_and_touch_anon_noexit(const char *cgroup, void *arg)
49 {
50 int ppid = getppid();
51 size_t size = (size_t)arg;
52 void *buf;
53
54 buf = mmap(NULL, size, PROT_READ | PROT_WRITE, MAP_PRIVATE | MAP_ANON,
55 0, 0);
56 if (buf == MAP_FAILED)
57 return -1;
58
59 if (touch_anon((char *)buf, size)) {
60 munmap(buf, size);
61 return -1;
62 }
63
64 while (getppid() == ppid)
65 sleep(1);
66
67 munmap(buf, size);
68 return 0;
69 }
70
71 /*
72 * Create a child process that allocates and touches 100MB, then waits to be
73 * killed. Wait until the child is attached to the cgroup, kill all processes
74 * in that cgroup and wait until "cgroup.procs" is empty. At this point try to
75 * destroy the empty cgroup. The test helps detect race conditions between
76 * dying processes leaving the cgroup and cgroup destruction path.
77 */
test_cgcore_destroy(const char * root)78 static int test_cgcore_destroy(const char *root)
79 {
80 int ret = KSFT_FAIL;
81 char *cg_test = NULL;
82 int child_pid;
83 char buf[PAGE_SIZE];
84
85 cg_test = cg_name(root, "cg_test");
86
87 if (!cg_test)
88 goto cleanup;
89
90 for (int i = 0; i < 10; i++) {
91 if (cg_create(cg_test))
92 goto cleanup;
93
94 child_pid = cg_run_nowait(cg_test, alloc_and_touch_anon_noexit,
95 (void *) MB(100));
96
97 if (child_pid < 0)
98 goto cleanup;
99
100 /* wait for the child to enter cgroup */
101 if (cg_wait_for_proc_count(cg_test, 1))
102 goto cleanup;
103
104 if (cg_killall(cg_test))
105 goto cleanup;
106
107 /* wait for cgroup to be empty */
108 while (1) {
109 if (cg_read(cg_test, "cgroup.procs", buf, sizeof(buf)))
110 goto cleanup;
111 if (buf[0] == '\0')
112 break;
113 usleep(1000);
114 }
115
116 if (rmdir(cg_test))
117 goto cleanup;
118
119 if (waitpid(child_pid, NULL, 0) < 0)
120 goto cleanup;
121 }
122 ret = KSFT_PASS;
123 cleanup:
124 if (cg_test)
125 cg_destroy(cg_test);
126 free(cg_test);
127 return ret;
128 }
129
130 /*
131 * A(0) - B(0) - C(1)
132 * \ D(0)
133 *
134 * A, B and C's "populated" fields would be 1 while D's 0.
135 * test that after the one process in C is moved to root,
136 * A,B and C's "populated" fields would flip to "0" and file
137 * modified events will be generated on the
138 * "cgroup.events" files of both cgroups.
139 */
test_cgcore_populated(const char * root)140 static int test_cgcore_populated(const char *root)
141 {
142 int ret = KSFT_FAIL;
143 int err;
144 char *cg_test_a = NULL, *cg_test_b = NULL;
145 char *cg_test_c = NULL, *cg_test_d = NULL;
146 int cgroup_fd = -EBADF;
147 pid_t pid;
148
149 cg_test_a = cg_name(root, "cg_test_a");
150 cg_test_b = cg_name(root, "cg_test_a/cg_test_b");
151 cg_test_c = cg_name(root, "cg_test_a/cg_test_b/cg_test_c");
152 cg_test_d = cg_name(root, "cg_test_a/cg_test_b/cg_test_d");
153
154 if (!cg_test_a || !cg_test_b || !cg_test_c || !cg_test_d)
155 goto cleanup;
156
157 if (cg_create(cg_test_a))
158 goto cleanup;
159
160 if (cg_create(cg_test_b))
161 goto cleanup;
162
163 if (cg_create(cg_test_c))
164 goto cleanup;
165
166 if (cg_create(cg_test_d))
167 goto cleanup;
168
169 if (cg_enter_current(cg_test_c))
170 goto cleanup;
171
172 if (cg_read_strcmp(cg_test_a, "cgroup.events", "populated 1\n"))
173 goto cleanup;
174
175 if (cg_read_strcmp(cg_test_b, "cgroup.events", "populated 1\n"))
176 goto cleanup;
177
178 if (cg_read_strcmp(cg_test_c, "cgroup.events", "populated 1\n"))
179 goto cleanup;
180
181 if (cg_read_strcmp(cg_test_d, "cgroup.events", "populated 0\n"))
182 goto cleanup;
183
184 if (cg_enter_current(root))
185 goto cleanup;
186
187 if (cg_read_strcmp(cg_test_a, "cgroup.events", "populated 0\n"))
188 goto cleanup;
189
190 if (cg_read_strcmp(cg_test_b, "cgroup.events", "populated 0\n"))
191 goto cleanup;
192
193 if (cg_read_strcmp(cg_test_c, "cgroup.events", "populated 0\n"))
194 goto cleanup;
195
196 if (cg_read_strcmp(cg_test_d, "cgroup.events", "populated 0\n"))
197 goto cleanup;
198
199 /* Test that we can directly clone into a new cgroup. */
200 cgroup_fd = dirfd_open_opath(cg_test_d);
201 if (cgroup_fd < 0)
202 goto cleanup;
203
204 pid = clone_into_cgroup(cgroup_fd);
205 if (pid < 0) {
206 if (errno == ENOSYS)
207 goto cleanup_pass;
208 goto cleanup;
209 }
210
211 if (pid == 0) {
212 if (raise(SIGSTOP))
213 exit(EXIT_FAILURE);
214 exit(EXIT_SUCCESS);
215 }
216
217 err = cg_read_strcmp(cg_test_d, "cgroup.events", "populated 1\n");
218
219 (void)clone_reap(pid, WSTOPPED);
220 (void)kill(pid, SIGCONT);
221 (void)clone_reap(pid, WEXITED);
222
223 if (err)
224 goto cleanup;
225
226 if (cg_read_strcmp(cg_test_d, "cgroup.events", "populated 0\n"))
227 goto cleanup;
228
229 /* Remove cgroup. */
230 if (cg_test_d) {
231 cg_destroy(cg_test_d);
232 free(cg_test_d);
233 cg_test_d = NULL;
234 }
235
236 pid = clone_into_cgroup(cgroup_fd);
237 if (pid < 0)
238 goto cleanup_pass;
239 if (pid == 0)
240 exit(EXIT_SUCCESS);
241 (void)clone_reap(pid, WEXITED);
242 goto cleanup;
243
244 cleanup_pass:
245 ret = KSFT_PASS;
246
247 cleanup:
248 if (cg_test_d)
249 cg_destroy(cg_test_d);
250 if (cg_test_c)
251 cg_destroy(cg_test_c);
252 if (cg_test_b)
253 cg_destroy(cg_test_b);
254 if (cg_test_a)
255 cg_destroy(cg_test_a);
256 free(cg_test_d);
257 free(cg_test_c);
258 free(cg_test_b);
259 free(cg_test_a);
260 if (cgroup_fd >= 0)
261 close(cgroup_fd);
262 return ret;
263 }
264
265 /*
266 * A (domain threaded) - B (threaded) - C (domain)
267 *
268 * test that C can't be used until it is turned into a
269 * threaded cgroup. "cgroup.type" file will report "domain (invalid)" in
270 * these cases. Operations which fail due to invalid topology use
271 * EOPNOTSUPP as the errno.
272 */
test_cgcore_invalid_domain(const char * root)273 static int test_cgcore_invalid_domain(const char *root)
274 {
275 int ret = KSFT_FAIL;
276 char *grandparent = NULL, *parent = NULL, *child = NULL;
277
278 grandparent = cg_name(root, "cg_test_grandparent");
279 parent = cg_name(root, "cg_test_grandparent/cg_test_parent");
280 child = cg_name(root, "cg_test_grandparent/cg_test_parent/cg_test_child");
281 if (!parent || !child || !grandparent)
282 goto cleanup;
283
284 if (cg_create(grandparent))
285 goto cleanup;
286
287 if (cg_create(parent))
288 goto cleanup;
289
290 if (cg_create(child))
291 goto cleanup;
292
293 if (cg_write(parent, "cgroup.type", "threaded"))
294 goto cleanup;
295
296 if (cg_read_strcmp(child, "cgroup.type", "domain invalid\n"))
297 goto cleanup;
298
299 if (!cg_enter_current(child))
300 goto cleanup;
301
302 if (errno != EOPNOTSUPP)
303 goto cleanup;
304
305 if (!clone_into_cgroup_run_wait(child))
306 goto cleanup;
307
308 if (errno == ENOSYS)
309 goto cleanup_pass;
310
311 if (errno != EOPNOTSUPP)
312 goto cleanup;
313
314 cleanup_pass:
315 ret = KSFT_PASS;
316
317 cleanup:
318 cg_enter_current(root);
319 if (child)
320 cg_destroy(child);
321 if (parent)
322 cg_destroy(parent);
323 if (grandparent)
324 cg_destroy(grandparent);
325 free(child);
326 free(parent);
327 free(grandparent);
328 return ret;
329 }
330
331 /*
332 * Test that when a child becomes threaded
333 * the parent type becomes domain threaded.
334 */
test_cgcore_parent_becomes_threaded(const char * root)335 static int test_cgcore_parent_becomes_threaded(const char *root)
336 {
337 int ret = KSFT_FAIL;
338 char *parent = NULL, *child = NULL;
339
340 parent = cg_name(root, "cg_test_parent");
341 child = cg_name(root, "cg_test_parent/cg_test_child");
342 if (!parent || !child)
343 goto cleanup;
344
345 if (cg_create(parent))
346 goto cleanup;
347
348 if (cg_create(child))
349 goto cleanup;
350
351 if (cg_write(child, "cgroup.type", "threaded"))
352 goto cleanup;
353
354 if (cg_read_strcmp(parent, "cgroup.type", "domain threaded\n"))
355 goto cleanup;
356
357 ret = KSFT_PASS;
358
359 cleanup:
360 if (child)
361 cg_destroy(child);
362 if (parent)
363 cg_destroy(parent);
364 free(child);
365 free(parent);
366 return ret;
367
368 }
369
370 /*
371 * Test that there's no internal process constrain on threaded cgroups.
372 * You can add threads/processes on a parent with a controller enabled.
373 */
test_cgcore_no_internal_process_constraint_on_threads(const char * root)374 static int test_cgcore_no_internal_process_constraint_on_threads(const char *root)
375 {
376 int ret = KSFT_FAIL;
377 char *parent = NULL, *child = NULL;
378
379 if (cg_read_strstr(root, "cgroup.controllers", "cpu") ||
380 cg_write(root, "cgroup.subtree_control", "+cpu")) {
381 ret = KSFT_SKIP;
382 goto cleanup;
383 }
384
385 parent = cg_name(root, "cg_test_parent");
386 child = cg_name(root, "cg_test_parent/cg_test_child");
387 if (!parent || !child)
388 goto cleanup;
389
390 if (cg_create(parent))
391 goto cleanup;
392
393 if (cg_create(child))
394 goto cleanup;
395
396 if (cg_write(parent, "cgroup.type", "threaded"))
397 goto cleanup;
398
399 if (cg_write(child, "cgroup.type", "threaded"))
400 goto cleanup;
401
402 if (cg_write(parent, "cgroup.subtree_control", "+cpu"))
403 goto cleanup;
404
405 if (cg_enter_current(parent))
406 goto cleanup;
407
408 ret = KSFT_PASS;
409
410 cleanup:
411 cg_enter_current(root);
412 cg_enter_current(root);
413 if (child)
414 cg_destroy(child);
415 if (parent)
416 cg_destroy(parent);
417 free(child);
418 free(parent);
419 return ret;
420 }
421
422 /*
423 * Test that you can't enable a controller on a child if it's not enabled
424 * on the parent.
425 */
test_cgcore_top_down_constraint_enable(const char * root)426 static int test_cgcore_top_down_constraint_enable(const char *root)
427 {
428 int ret = KSFT_FAIL;
429 char *parent = NULL, *child = NULL;
430
431 parent = cg_name(root, "cg_test_parent");
432 child = cg_name(root, "cg_test_parent/cg_test_child");
433 if (!parent || !child)
434 goto cleanup;
435
436 if (cg_create(parent))
437 goto cleanup;
438
439 if (cg_create(child))
440 goto cleanup;
441
442 if (!cg_write(child, "cgroup.subtree_control", "+memory"))
443 goto cleanup;
444
445 ret = KSFT_PASS;
446
447 cleanup:
448 if (child)
449 cg_destroy(child);
450 if (parent)
451 cg_destroy(parent);
452 free(child);
453 free(parent);
454 return ret;
455 }
456
457 /*
458 * Test that you can't disable a controller on a parent
459 * if it's enabled in a child.
460 */
test_cgcore_top_down_constraint_disable(const char * root)461 static int test_cgcore_top_down_constraint_disable(const char *root)
462 {
463 int ret = KSFT_FAIL;
464 char *parent = NULL, *child = NULL;
465
466 parent = cg_name(root, "cg_test_parent");
467 child = cg_name(root, "cg_test_parent/cg_test_child");
468 if (!parent || !child)
469 goto cleanup;
470
471 if (cg_create(parent))
472 goto cleanup;
473
474 if (cg_create(child))
475 goto cleanup;
476
477 if (cg_write(parent, "cgroup.subtree_control", "+memory"))
478 goto cleanup;
479
480 if (cg_write(child, "cgroup.subtree_control", "+memory"))
481 goto cleanup;
482
483 if (!cg_write(parent, "cgroup.subtree_control", "-memory"))
484 goto cleanup;
485
486 ret = KSFT_PASS;
487
488 cleanup:
489 if (child)
490 cg_destroy(child);
491 if (parent)
492 cg_destroy(parent);
493 free(child);
494 free(parent);
495 return ret;
496 }
497
498 /*
499 * Test internal process constraint.
500 * You can't add a pid to a domain parent if a controller is enabled.
501 */
test_cgcore_internal_process_constraint(const char * root)502 static int test_cgcore_internal_process_constraint(const char *root)
503 {
504 int ret = KSFT_FAIL;
505 char *parent = NULL, *child = NULL;
506
507 parent = cg_name(root, "cg_test_parent");
508 child = cg_name(root, "cg_test_parent/cg_test_child");
509 if (!parent || !child)
510 goto cleanup;
511
512 if (cg_create(parent))
513 goto cleanup;
514
515 if (cg_create(child))
516 goto cleanup;
517
518 if (cg_write(parent, "cgroup.subtree_control", "+memory"))
519 goto cleanup;
520
521 if (!cg_enter_current(parent))
522 goto cleanup;
523
524 if (!clone_into_cgroup_run_wait(parent))
525 goto cleanup;
526
527 ret = KSFT_PASS;
528
529 cleanup:
530 if (child)
531 cg_destroy(child);
532 if (parent)
533 cg_destroy(parent);
534 free(child);
535 free(parent);
536 return ret;
537 }
538
dummy_thread_fn(void * arg)539 static void *dummy_thread_fn(void *arg)
540 {
541 return (void *)(size_t)pause();
542 }
543
544 /*
545 * Test threadgroup migration.
546 * All threads of a process are migrated together.
547 */
test_cgcore_proc_migration(const char * root)548 static int test_cgcore_proc_migration(const char *root)
549 {
550 int ret = KSFT_FAIL;
551 int t, c_threads = 0, n_threads = 13;
552 char *src = NULL, *dst = NULL;
553 pthread_t threads[n_threads];
554
555 src = cg_name(root, "cg_src");
556 dst = cg_name(root, "cg_dst");
557 if (!src || !dst)
558 goto cleanup;
559
560 if (cg_create(src))
561 goto cleanup;
562 if (cg_create(dst))
563 goto cleanup;
564
565 if (cg_enter_current(src))
566 goto cleanup;
567
568 for (c_threads = 0; c_threads < n_threads; ++c_threads) {
569 if (pthread_create(&threads[c_threads], NULL, dummy_thread_fn, NULL))
570 goto cleanup;
571 }
572
573 cg_enter_current(dst);
574 if (cg_read_lc(dst, "cgroup.threads") != n_threads + 1)
575 goto cleanup;
576
577 ret = KSFT_PASS;
578
579 cleanup:
580 for (t = 0; t < c_threads; ++t) {
581 pthread_cancel(threads[t]);
582 }
583
584 for (t = 0; t < c_threads; ++t) {
585 pthread_join(threads[t], NULL);
586 }
587
588 cg_enter_current(root);
589
590 if (dst)
591 cg_destroy(dst);
592 if (src)
593 cg_destroy(src);
594 free(dst);
595 free(src);
596 return ret;
597 }
598
migrating_thread_fn(void * arg)599 static void *migrating_thread_fn(void *arg)
600 {
601 int g, i, n_iterations = 1000;
602 char **grps = arg;
603 char lines[3][PATH_MAX];
604
605 for (g = 1; g < 3; ++g)
606 snprintf(lines[g], sizeof(lines[g]), "0::%s", grps[g] + strlen(grps[0]));
607
608 for (i = 0; i < n_iterations; ++i) {
609 cg_enter_current_thread(grps[(i % 2) + 1]);
610
611 if (proc_read_strstr(0, 1, "cgroup", lines[(i % 2) + 1]))
612 return (void *)-1;
613 }
614 return NULL;
615 }
616
617 /*
618 * Test single thread migration.
619 * Threaded cgroups allow successful migration of a thread.
620 */
test_cgcore_thread_migration(const char * root)621 static int test_cgcore_thread_migration(const char *root)
622 {
623 int ret = KSFT_FAIL;
624 char *dom = NULL;
625 char line[PATH_MAX];
626 char *grps[3] = { (char *)root, NULL, NULL };
627 pthread_t thr;
628 void *retval;
629
630 dom = cg_name(root, "cg_dom");
631 grps[1] = cg_name(root, "cg_dom/cg_src");
632 grps[2] = cg_name(root, "cg_dom/cg_dst");
633 if (!grps[1] || !grps[2] || !dom)
634 goto cleanup;
635
636 if (cg_create(dom))
637 goto cleanup;
638 if (cg_create(grps[1]))
639 goto cleanup;
640 if (cg_create(grps[2]))
641 goto cleanup;
642
643 if (cg_write(grps[1], "cgroup.type", "threaded"))
644 goto cleanup;
645 if (cg_write(grps[2], "cgroup.type", "threaded"))
646 goto cleanup;
647
648 if (cg_enter_current(grps[1]))
649 goto cleanup;
650
651 if (pthread_create(&thr, NULL, migrating_thread_fn, grps))
652 goto cleanup;
653
654 if (pthread_join(thr, &retval))
655 goto cleanup;
656
657 if (retval)
658 goto cleanup;
659
660 snprintf(line, sizeof(line), "0::%s", grps[1] + strlen(grps[0]));
661 if (proc_read_strstr(0, 1, "cgroup", line))
662 goto cleanup;
663
664 ret = KSFT_PASS;
665
666 cleanup:
667 cg_enter_current(root);
668 if (grps[2])
669 cg_destroy(grps[2]);
670 if (grps[1])
671 cg_destroy(grps[1]);
672 if (dom)
673 cg_destroy(dom);
674 free(grps[2]);
675 free(grps[1]);
676 free(dom);
677 return ret;
678 }
679
680 /*
681 * cgroup migration permission check should be performed based on the
682 * credentials at the time of open instead of write.
683 */
test_cgcore_lesser_euid_open(const char * root)684 static int test_cgcore_lesser_euid_open(const char *root)
685 {
686 const uid_t test_euid = TEST_UID;
687 int ret = KSFT_FAIL;
688 char *cg_test_a = NULL, *cg_test_b = NULL;
689 char *cg_test_a_procs = NULL, *cg_test_b_procs = NULL;
690 int cg_test_b_procs_fd = -1;
691 uid_t saved_uid;
692
693 cg_test_a = cg_name(root, "cg_test_a");
694 cg_test_b = cg_name(root, "cg_test_b");
695
696 if (!cg_test_a || !cg_test_b)
697 goto cleanup;
698
699 cg_test_a_procs = cg_name(cg_test_a, "cgroup.procs");
700 cg_test_b_procs = cg_name(cg_test_b, "cgroup.procs");
701
702 if (!cg_test_a_procs || !cg_test_b_procs)
703 goto cleanup;
704
705 if (cg_create(cg_test_a) || cg_create(cg_test_b))
706 goto cleanup;
707
708 if (cg_enter_current(cg_test_a))
709 goto cleanup;
710
711 if (chown(cg_test_a_procs, test_euid, -1) ||
712 chown(cg_test_b_procs, test_euid, -1))
713 goto cleanup;
714
715 saved_uid = geteuid();
716 if (seteuid(test_euid))
717 goto cleanup;
718
719 cg_test_b_procs_fd = open(cg_test_b_procs, O_RDWR);
720
721 if (seteuid(saved_uid))
722 goto cleanup;
723
724 if (cg_test_b_procs_fd < 0)
725 goto cleanup;
726
727 if (write(cg_test_b_procs_fd, "0", 1) >= 0 || errno != EACCES)
728 goto cleanup;
729
730 ret = KSFT_PASS;
731
732 cleanup:
733 cg_enter_current(root);
734 if (cg_test_b_procs_fd >= 0)
735 close(cg_test_b_procs_fd);
736 if (cg_test_b)
737 cg_destroy(cg_test_b);
738 if (cg_test_a)
739 cg_destroy(cg_test_a);
740 free(cg_test_b_procs);
741 free(cg_test_a_procs);
742 free(cg_test_b);
743 free(cg_test_a);
744 return ret;
745 }
746
747 struct lesser_ns_open_thread_arg {
748 const char *path;
749 int fd;
750 int err;
751 };
752
lesser_ns_open_thread_fn(void * arg)753 static int lesser_ns_open_thread_fn(void *arg)
754 {
755 struct lesser_ns_open_thread_arg *targ = arg;
756
757 targ->fd = open(targ->path, O_RDWR);
758 targ->err = errno;
759 return 0;
760 }
761
762 /*
763 * cgroup migration permission check should be performed based on the cgroup
764 * namespace at the time of open instead of write.
765 */
test_cgcore_lesser_ns_open(const char * root)766 static int test_cgcore_lesser_ns_open(const char *root)
767 {
768 static char stack[65536];
769 const uid_t test_euid = 65534; /* usually nobody, any !root is fine */
770 int ret = KSFT_FAIL;
771 char *cg_test_a = NULL, *cg_test_b = NULL;
772 char *cg_test_a_procs = NULL, *cg_test_b_procs = NULL;
773 int cg_test_b_procs_fd = -1;
774 struct lesser_ns_open_thread_arg targ = { .fd = -1 };
775 pid_t pid;
776 int status;
777
778 cg_test_a = cg_name(root, "cg_test_a");
779 cg_test_b = cg_name(root, "cg_test_b");
780
781 if (!cg_test_a || !cg_test_b)
782 goto cleanup;
783
784 cg_test_a_procs = cg_name(cg_test_a, "cgroup.procs");
785 cg_test_b_procs = cg_name(cg_test_b, "cgroup.procs");
786
787 if (!cg_test_a_procs || !cg_test_b_procs)
788 goto cleanup;
789
790 if (cg_create(cg_test_a) || cg_create(cg_test_b))
791 goto cleanup;
792
793 if (cg_enter_current(cg_test_b))
794 goto cleanup;
795
796 if (chown(cg_test_a_procs, test_euid, -1) ||
797 chown(cg_test_b_procs, test_euid, -1))
798 goto cleanup;
799
800 targ.path = cg_test_b_procs;
801 pid = clone(lesser_ns_open_thread_fn, stack + sizeof(stack),
802 CLONE_NEWCGROUP | CLONE_FILES | CLONE_VM | SIGCHLD,
803 &targ);
804 if (pid < 0)
805 goto cleanup;
806
807 if (waitpid(pid, &status, 0) < 0)
808 goto cleanup;
809
810 if (!WIFEXITED(status))
811 goto cleanup;
812
813 cg_test_b_procs_fd = targ.fd;
814 if (cg_test_b_procs_fd < 0)
815 goto cleanup;
816
817 if (cg_enter_current(cg_test_a))
818 goto cleanup;
819
820 if ((status = write(cg_test_b_procs_fd, "0", 1)) >= 0 || errno != ENOENT)
821 goto cleanup;
822
823 ret = KSFT_PASS;
824
825 cleanup:
826 cg_enter_current(root);
827 if (cg_test_b_procs_fd >= 0)
828 close(cg_test_b_procs_fd);
829 if (cg_test_b)
830 cg_destroy(cg_test_b);
831 if (cg_test_a)
832 cg_destroy(cg_test_a);
833 free(cg_test_b_procs);
834 free(cg_test_a_procs);
835 free(cg_test_b);
836 free(cg_test_a);
837 return ret;
838 }
839
840 #define T(x) { x, #x }
841 struct corecg_test {
842 int (*fn)(const char *root);
843 const char *name;
844 } tests[] = {
845 T(test_cgcore_internal_process_constraint),
846 T(test_cgcore_top_down_constraint_enable),
847 T(test_cgcore_top_down_constraint_disable),
848 T(test_cgcore_no_internal_process_constraint_on_threads),
849 T(test_cgcore_parent_becomes_threaded),
850 T(test_cgcore_invalid_domain),
851 T(test_cgcore_populated),
852 T(test_cgcore_proc_migration),
853 T(test_cgcore_thread_migration),
854 T(test_cgcore_destroy),
855 T(test_cgcore_lesser_euid_open),
856 T(test_cgcore_lesser_ns_open),
857 };
858 #undef T
859
main(int argc,char * argv[])860 int main(int argc, char *argv[])
861 {
862 char root[PATH_MAX];
863 int i, ret = EXIT_SUCCESS;
864
865 if (cg_find_unified_root(root, sizeof(root)))
866 ksft_exit_skip("cgroup v2 isn't mounted\n");
867
868 if (cg_read_strstr(root, "cgroup.subtree_control", "memory"))
869 if (cg_write(root, "cgroup.subtree_control", "+memory"))
870 ksft_exit_skip("Failed to set memory controller\n");
871
872 for (i = 0; i < ARRAY_SIZE(tests); i++) {
873 switch (tests[i].fn(root)) {
874 case KSFT_PASS:
875 ksft_test_result_pass("%s\n", tests[i].name);
876 break;
877 case KSFT_SKIP:
878 ksft_test_result_skip("%s\n", tests[i].name);
879 break;
880 default:
881 ret = EXIT_FAILURE;
882 ksft_test_result_fail("%s\n", tests[i].name);
883 break;
884 }
885 }
886
887 return ret;
888 }
889