R Under development (unstable) (2025-05-30 r88253 ucrt) -- "Unsuffered Consequences" Copyright (C) 2025 The R Foundation for Statistical Computing Platform: x86_64-w64-mingw32/x64 R is free software and comes with ABSOLUTELY NO WARRANTY. You are welcome to redistribute it under certain conditions. Type 'license()' or 'licence()' for distribution details. R is a collaborative project with many contributors. Type 'contributors()' for more information and 'citation()' on how to cite R or R packages in publications. Type 'demo()' for some demos, 'help()' for on-line help, or 'help.start()' for an HTML browser interface to help. Type 'q()' to quit R. > ## This runs testme test script inst/testme/test-cgroups.R > ## Don't edit - it was autogenerated by inst/testme/deploy.R > parallelly:::testme("cgroups") Test 'cgroups' ... Sourcing 7 prologue scripts ... 01/07 prologue script 'D:/RCompile/CRANincoming/R-devel/lib/parallelly/testme/_prologue/001.load.R' 02/07 prologue script 'D:/RCompile/CRANincoming/R-devel/lib/parallelly/testme/_prologue/002.record-state.R' 03/07 prologue script 'D:/RCompile/CRANincoming/R-devel/lib/parallelly/testme/_prologue/030.imports.R' 04/07 prologue script 'D:/RCompile/CRANincoming/R-devel/lib/parallelly/testme/_prologue/050.utils.R' 05/07 prologue script 'D:/RCompile/CRANincoming/R-devel/lib/parallelly/testme/_prologue/090.context.R' 06/07 prologue script 'D:/RCompile/CRANincoming/R-devel/lib/parallelly/testme/_prologue/090.options.R' 07/07 prologue script 'D:/RCompile/CRANincoming/R-devel/lib/parallelly/testme/_prologue/091.envvars.R' Sourcing 7 prologue scripts ... done Running test script: 'D:/RCompile/CRANincoming/R-devel/lib/parallelly/testme/test-cgroups.R' > library(parallelly) > message("*** cgroups ...") *** cgroups ... > message("- getCGroups()") - getCGroups() > cgroups <- parallelly:::getCGroups() > print(cgroups) [1] hierarchy_id controller path <0 rows> (or 0-length row.names) > stopifnot(is.data.frame(cgroups), identical(colnames(cgroups), + c("hierarchy_id", "controller", "path")), nrow(cgroups) == + 0 || !is.nul .... [TRUNCATED] > message("- getCGroupsRoot()") - getCGroupsRoot() > root <- parallelly:::getCGroupsRoot() > cat(sprintf("cgroups root path: %s\n", sQuote(root))) cgroups root path: 'NA' > stopifnot(length(root) == 1, is.character(root)) > message("- getCGroups()") - getCGroups() > cgroups <- parallelly:::getCGroups() > print(cgroups) [1] hierarchy_id controller path <0 rows> (or 0-length row.names) > stopifnot(is.data.frame(cgroups)) > message("- getCGroupsPath()") - getCGroupsPath() > path <- parallelly:::getCGroupsPath("cpu") > cat(sprintf("cgroups 'cpu' path: %s\n", sQuote(path))) cgroups 'cpu' path: 'NA' > stopifnot(length(path) == 1, is.character(path)) > path <- parallelly:::getCGroupsPath("cpuset") > cat(sprintf("cgroups 'cpuset' path: %s\n", sQuote(path))) cgroups 'cpuset' path: 'NA' > stopifnot(length(path) == 1, is.character(path)) > message("- getCGroups1Value()") - getCGroups1Value() > value <- parallelly:::getCGroups1Value("cpu", "cpu.cfs_quota_us") > cat(sprintf("cgroups v1 'cpu.cfs_quota_us' value: %s\n", + sQuote(value))) cgroups v1 'cpu.cfs_quota_us' value: 'NA' > stopifnot(length(value) == 1, is.character(value)) > value <- parallelly:::getCGroups1Value("cpu", "cpu.cfs_total_us") > cat(sprintf("cgroups v1 'cpu.cfs_total_us' value: %s\n", + sQuote(value))) cgroups v1 'cpu.cfs_total_us' value: 'NA' > stopifnot(length(value) == 1, is.character(value)) > value <- parallelly:::getCGroups1Value("cpuset", "cpuset.cpus") > cat(sprintf("cgroups v1 'cpuset.cpus' value: %s\n", + sQuote(value))) cgroups v1 'cpuset.cpus' value: 'NA' > stopifnot(length(value) == 1, is.character(value)) > message("- getCGroups1CpuSet()") - getCGroups1CpuSet() > value <- parallelly:::getCGroups1CpuSet() > cat(sprintf("CPU set: [n=%d] %s\n", length(value), + paste(sQuote(value), collapse = ", "))) CPU set: [n=0] > stopifnot(length(value) >= 0, is.integer(value), !any(is.na(value))) > message("- getCGroups1CpuQuotaMicroseconds()") - getCGroups1CpuQuotaMicroseconds() > value <- parallelly:::getCGroups1CpuQuotaMicroseconds() > cat(sprintf("CPU quota (ms): %d\n", value)) CPU quota (ms): NA > stopifnot(length(value) == 1, is.integer(value), is.na(value) || + value == -1 || value > 0) > message("- getCGroups1CpuPeriodMicroseconds()") - getCGroups1CpuPeriodMicroseconds() > value <- parallelly:::getCGroups1CpuPeriodMicroseconds() > cat(sprintf("CPU total (ms): %d\n", value)) CPU total (ms): NA > stopifnot(length(value) == 1, is.integer(value), is.na(value) || + value > 0) > message("- getCGroups1CpuQuota()") - getCGroups1CpuQuota() > value <- parallelly:::getCGroups1CpuQuota() > cat(sprintf("CPU quota (ratio): %g\n", value)) CPU quota (ratio): NA > stopifnot(length(value) == 1, is.numeric(value), !is.infinite(value), + is.na(value) || value > 0) > message("- getCGroups2CpuMax()") - getCGroups2CpuMax() > value <- parallelly:::getCGroups2CpuMax() > cat(sprintf("CPU quota (ratio): %g\n", value)) CPU quota (ratio): NA > stopifnot(length(value) == 1, is.numeric(value), !is.infinite(value), + is.na(value) || value > 0) > message("*** cgroups ... DONE") *** cgroups ... DONE > root <- system.file(package = "parallelly", "test-data", + mustWork = TRUE) > for (dir in c("no-cgroups", "mixed-cgroups", "cgroups1", + "cgroups2")) { + message(sprintf("%s - real-world ...", dir)) + path <- file .... [TRUNCATED] no-cgroups - real-world ... CGroups for system 'no-cgroups' ... - Using temporary folder: 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed8539e7251' - UID: 1000 - procPath(): 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed8539e7251/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: tmpfs /run tmpfs rw,nosuid,nodev,noexec,relatime,size=3256456k,mode=755,inode64 0 0 - getCGroupsVersion(): -1 - getCGroupsMounts(): [1] device mountpoint type options dump pass <0 rows> (or 0-length row.names) - getCGroups(): [1] hierarchy_id controller path <0 rows> (or 0-length row.names) - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 Running custom test script: 'D:/RCompile/CRANincoming/R-devel/lib/parallelly/test-data/no-cgroups/no-cgroups.R' system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 CGroups for system 'no-cgroups' ... done no-cgroups - real-world ... done mixed-cgroups - real-world ... CGroups for system 'piavpn' ... - Using temporary folder: 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed82080218b' - UID: 1000 - procPath(): 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed82080218b/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup2 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed82080218b/sys/fs/cgroup cgroup2 rw,nosuid,nodev,noexec,relatime,nsdelegate,memory_recursiveprot 0 0 02: none D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed82080218b/opt/piavpn/etc/cgroup/net_cls cgroup rw,relatime,net_cls 0 0 - getCGroupsVersion(): 1 - getCGroupsMounts(): device 1 cgroup2 2 none mountpoint 1 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed82080218b/sys/fs/cgroup 2 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed82080218b/opt/piavpn/etc/cgroup/net_cls type options dump 1 cgroup2 rw,nosuid,nodev,noexec,relatime,nsdelegate,memory_recursiveprot 0 2 cgroup rw,relatime,net_cls 0 pass 1 0 2 0 - getCGroups(): hierarchy_id controller 2 0 1 1 net_cls path 2 /user.slice/user-1000.slice/user@1000.service/app.slice/app-org.kde.konsole-4415.scope/tab(4427).scope 1 / - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 Running custom test script: 'D:/RCompile/CRANincoming/R-devel/lib/parallelly/test-data/mixed-cgroups/piavpn.R' system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 CGroups for system 'piavpn' ... done CGroups for system 'slurm' ... - Using temporary folder: 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed83c521038' - UID: 632771201 - procPath(): 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed83c521038/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup2 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/unified cgroup2 rw,nosuid,nodev,noexec,relatime,nsdelegate 0 0 02: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/systemd cgroup rw,nosuid,nodev,noexec,relatime,xattr,name=systemd 0 0 03: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/memory cgroup rw,nosuid,nodev,noexec,relatime,memory 0 0 04: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/cpu,cpuacct cgroup rw,nosuid,nodev,noexec,relatime,cpu,cpuacct 0 0 05: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/perf_event cgroup rw,nosuid,nodev,noexec,relatime,perf_event 0 0 06: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/hugetlb cgroup rw,nosuid,nodev,noexec,relatime,hugetlb 0 0 07: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/cpuset cgroup rw,nosuid,nodev,noexec,relatime,cpuset 0 0 08: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/rdma cgroup rw,nosuid,nodev,noexec,relatime,rdma 0 0 09: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/misc cgroup rw,nosuid,nodev,noexec,relatime,misc 0 0 10: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/net_cls,net_prio cgroup rw,nosuid,nodev,noexec,relatime,net_cls,net_prio 0 0 11: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/freezer cgroup rw,nosuid,nodev,noexec,relatime,freezer 0 0 12: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/pids cgroup rw,nosuid,nodev,noexec,relatime,pids 0 0 13: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/blkio cgroup rw,nosuid,nodev,noexec,relatime,blkio 0 0 14: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/devices cgroup rw,nosuid,nodev,noexec,relatime,devices 0 0 - getCGroupsVersion(): 1 - getCGroupsMounts(): device 1 cgroup2 2 cgroup 3 cgroup 4 cgroup 5 cgroup 6 cgroup 7 cgroup 8 cgroup 9 cgroup 10 cgroup 11 cgroup 12 cgroup 13 cgroup 14 cgroup mountpoint 1 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/unified 2 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/systemd 3 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/memory 4 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/cpu,cpuacct 5 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/perf_event 6 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/hugetlb 7 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/cpuset 8 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/rdma 9 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/misc 10 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/net_cls,net_prio 11 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/freezer 12 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/pids 13 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/blkio 14 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83c521038/sys/fs/cgroup/devices type options dump pass 1 cgroup2 rw,nosuid,nodev,noexec,relatime,nsdelegate 0 0 2 cgroup rw,nosuid,nodev,noexec,relatime,xattr,name=systemd 0 0 3 cgroup rw,nosuid,nodev,noexec,relatime,memory 0 0 4 cgroup rw,nosuid,nodev,noexec,relatime,cpu,cpuacct 0 0 5 cgroup rw,nosuid,nodev,noexec,relatime,perf_event 0 0 6 cgroup rw,nosuid,nodev,noexec,relatime,hugetlb 0 0 7 cgroup rw,nosuid,nodev,noexec,relatime,cpuset 0 0 8 cgroup rw,nosuid,nodev,noexec,relatime,rdma 0 0 9 cgroup rw,nosuid,nodev,noexec,relatime,misc 0 0 10 cgroup rw,nosuid,nodev,noexec,relatime,net_cls,net_prio 0 0 11 cgroup rw,nosuid,nodev,noexec,relatime,freezer 0 0 12 cgroup rw,nosuid,nodev,noexec,relatime,pids 0 0 13 cgroup rw,nosuid,nodev,noexec,relatime,blkio 0 0 14 cgroup rw,nosuid,nodev,noexec,relatime,devices 0 0 - getCGroups(): hierarchy_id controller path 14 0 /system.slice/slurmd.service 13 1 name=systemd /system.slice/slurmd.service 12 2 memory /slurm/uid_632771201/job_1140459 11 3 cpu /slurm/uid_632771201/job_1140459 111 3 cpuacct /slurm/uid_632771201/job_1140459 10 4 perf_event / 9 5 hugetlb / 8 6 cpuset /slurm/uid_632771201/job_1140459/step_0 7 7 rdma / 6 8 misc / 5 9 net_cls / 51 9 net_prio / 4 10 freezer /slurm/uid_632771201/job_1140459/step_0 3 11 pids /system.slice/slurmd.service 2 12 blkio /slurm/uid_632771201/job_1140459 1 13 devices /slurm/uid_632771201/job_1140459/step_0/task_0 - length(getCGroups1CpuSet()): 1 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status cgroups.cpuset 96 96 1 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 Running custom test script: 'D:/RCompile/CRANincoming/R-devel/lib/parallelly/test-data/mixed-cgroups/slurm.R' system /proc/self/status cgroups.cpuset 96 96 1 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 CGroups for system 'slurm' ... done CGroups for system 'wsl2-invalid-mounts' ... - Using temporary folder: 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed86e04359b' - procPath(): 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed86e04359b/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: none /usr/lib/modules/5.15.167.4-microsoft-standard-WSL2 overlay rw,nosuid,nodev,noatime,lowerdir=/modules,upperdir=/lib/modules/5.15.167.4-microsoft-standard-WSL2/rw/upper,workdir=/lib/modules/5.15.167.4-microsoft-standard-WSL2/rw/work 0 0 02: none /mnt/wsl tmpfs rw,relatime 0 0 03: drivers /usr/lib/wsl/drivers 9p ro,dirsync,nosuid,nodev,noatime,aname=drivers;fmask=222;dmask=222,mmap,access=client,msize=65536,trans=fd,rfd=8,wfd=8 0 0 04: /dev/sde / ext4 rw,relatime,discard,errors=remount-ro,data=ordered 0 0 05: none /mnt/wslg tmpfs rw,relatime 0 0 06: /dev/sde /mnt/wslg/distro ext4 ro,relatime,discard,errors=remount-ro,data=ordered 0 0 07: none /usr/lib/wsl/lib overlay rw,nosuid,nodev,noatime,lowerdir=/gpu_lib_packaged:/gpu_lib_inbox,upperdir=/gpu_lib/rw/upper,workdir=/gpu_lib/rw/work 0 0 08: rootfs /init rootfs ro,size=56660288k,nr_inodes=14165072 0 0 09: none /dev devtmpfs rw,nosuid,relatime,size=56660288k,nr_inodes=14165072,mode=755 0 0 10: sysfs /sys sysfs rw,nosuid,nodev,noexec,noatime 0 0 11: proc /proc proc rw,nosuid,nodev,noexec,noatime 0 0 12: devpts /dev/pts devpts rw,nosuid,noexec,noatime,gid=5,mode=620,ptmxmode=000 0 0 13: none /run tmpfs rw,nosuid,nodev,mode=755 0 0 14: none /run/lock tmpfs rw,nosuid,nodev,noexec,noatime 0 0 15: none /run/shm tmpfs rw,nosuid,nodev,noatime 0 0 16: none /dev/shm tmpfs rw,nosuid,nodev,noatime 0 0 17: none /run/user tmpfs rw,nosuid,nodev,noexec,noatime,mode=755 0 0 18: binfmt_misc /proc/sys/fs/binfmt_misc binfmt_misc rw,relatime 0 0 19: tmpfs /sys/fs/cgroup tmpfs ro,nosuid,nodev,noexec,mode=755 0 0 20: cgroup2 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/unified cgroup2 rw,nosuid,nodev,noexec,relatime 0 0 21: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/cpuset cgroup rw,nosuid,nodev,noexec,relatime,cpuset 0 0 22: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/cpu cgroup rw,nosuid,nodev,noexec,relatime,cpu 0 0 23: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/cpuacct cgroup rw,nosuid,nodev,noexec,relatime,cpuacct 0 0 24: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/blkio cgroup rw,nosuid,nodev,noexec,relatime,blkio 0 0 25: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/memory cgroup rw,nosuid,nodev,noexec,relatime,memory 0 0 26: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/devices cgroup rw,nosuid,nodev,noexec,relatime,devices 0 0 27: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/freezer cgroup rw,nosuid,nodev,noexec,relatime,freezer 0 0 28: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/net_cls cgroup rw,nosuid,nodev,noexec,relatime,net_cls 0 0 29: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/perf_event cgroup rw,nosuid,nodev,noexec,relatime,perf_event 0 0 30: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/net_prio cgroup rw,nosuid,nodev,noexec,relatime,net_prio 0 0 31: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/hugetlb cgroup rw,nosuid,nodev,noexec,relatime,hugetlb 0 0 32: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/pids cgroup rw,nosuid,nodev,noexec,relatime,pids 0 0 33: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/rdma cgroup rw,nosuid,nodev,noexec,relatime,rdma 0 0 34: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/misc cgroup rw,nosuid,nodev,noexec,relatime,misc 0 0 35: none /mnt/wslg/versions.txt overlay rw,relatime,lowerdir=/systemvhd,upperdir=/system/rw/upper,workdir=/system/rw/work 0 0 36: none /mnt/wslg/doc overlay rw,relatime,lowerdir=/systemvhd,upperdir=/system/rw/upper,workdir=/system/rw/work 0 0 37: none /tmp/.X11-unix tmpfs ro,relatime 0 0 38: C:\\134 /mnt/c 9p rw,dirsync,noatime,aname=drvfs;path=C:\\;uid=1000;gid=1000;symlinkroot=/mnt/,mmap,access=client,msize=65536,trans=fd,rfd=6,wfd=6 0 0 39: D:\\134 /mnt/d 9p rw,dirsync,noatime,aname=drvfs;path=D:\\;uid=1000;gid=1000;symlinkroot=/mnt/,mmap,access=client,msize=65536,trans=fd,rfd=6,wfd=6 0 0 40: E:\\134 /mnt/e 9p rw,dirsync,noatime,aname=drvfs;path=E:\\;uid=1000;gid=1000;symlinkroot=/mnt/,mmap,access=client,msize=65536,trans=fd,rfd=6,wfd=6 0 0 41: F:\\134 /mnt/f 9p rw,dirsync,noatime,aname=drvfs;path=F:\\;uid=1000;gid=1000;symlinkroot=/mnt/,mmap,access=client,msize=65536,trans=fd,rfd=6,wfd=6 0 0 42: G:\\134 /mnt/g 9p rw,dirsync,noatime,aname=drvfs;path=G:\\;uid=1000;gid=1000;symlinkroot=/mnt/,mmap,access=client,msize=65536,trans=fd,rfd=6,wfd=6 0 0 43: none /run/user tmpfs rw,relatime 0 0 44: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/systemd cgroup rw,nosuid,nodev,noexec,relatime,xattr,name=systemd 0 0 45: hugetlbfs /dev/hugepages hugetlbfs rw,relatime,pagesize=2M 0 0 46: mqueue /dev/mqueue mqueue rw,nosuid,nodev,noexec,relatime 0 0 47: debugfs /sys/kernel/debug debugfs rw,nosuid,nodev,noexec,relatime 0 0 48: tracefs /sys/kernel/tracing tracefs rw,nosuid,nodev,noexec,relatime 0 0 49: fusectl /sys/fs/fuse/connections fusectl rw,nosuid,nodev,noexec,relatime 0 0 50: /dev/sde /snap ext4 rw,relatime,discard,errors=remount-ro,data=ordered 0 0 51: snapfuse /snap/lxd/29619 fuse.snapfuse ro,nodev,relatime,user_id=0,group_id=0,allow_other 0 0 52: snapfuse /snap/core20/2501 fuse.snapfuse ro,nodev,relatime,user_id=0,group_id=0,allow_other 0 0 53: snapfuse /snap/core20/2571 fuse.snapfuse ro,nodev,relatime,user_id=0,group_id=0,allow_other 0 0 54: snapfuse /snap/lxd/32662 fuse.snapfuse ro,nodev,relatime,user_id=0,group_id=0,allow_other 0 0 55: snapfuse /snap/snapd/23771 fuse.snapfuse ro,nodev,relatime,user_id=0,group_id=0,allow_other 0 0 56: snapfuse /snap/snapd/24505 fuse.snapfuse ro,nodev,relatime,user_id=0,group_id=0,allow_other 0 0 57: /dev/sde /var/lib/docker ext4 rw,relatime,discard,errors=remount-ro,data=ordered 0 0 58: none /run/snapd/ns tmpfs rw,nosuid,nodev,mode=755 0 0 59: nsfs /run/snapd/ns/lxd.mnt nsfs rw 0 0 60: none /mnt/wsl/docker-desktop-bind-mounts/Ubuntu-20.04/71329c4cc6e32171553fa81d044eb31d1a3aac52ba9376c4a99f4505c494cf5b tmpfs rw,nosuid,nodev,mode=755 0 0 61: none /mnt/wsl/docker-desktop/shared-sockets/guest-services tmpfs rw,nosuid,nodev,mode=755 0 0 62: none /mnt/wsl/docker-desktop/shared-sockets/host-services tmpfs rw,nosuid,nodev,mode=755 0 0 63: /dev/sdf /mnt/wsl/docker-desktop/docker-desktop-user-distro ext4 rw,relatime,discard,errors=remount-ro,data=ordered 0 0 64: none /mnt/wsl/docker-desktop-bind-mounts/Ubuntu-20.04/docker.sock tmpfs rw,nosuid,nodev,mode=755 0 0 65: /dev/loop0 /mnt/wsl/docker-desktop/cli-tools iso9660 ro,relatime,nojoliet,check=s,map=n,blocksize=2048,iocharset=iso8859-1 0 0 - getCGroupsVersion(): -1 - getCGroupsMounts(): device 20 cgroup2 21 cgroup 22 cgroup 23 cgroup 24 cgroup 25 cgroup 26 cgroup 27 cgroup 28 cgroup 29 cgroup 30 cgroup 31 cgroup 32 cgroup 33 cgroup 34 cgroup 44 cgroup mountpoint 20 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/unified 21 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/cpuset 22 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/cpu 23 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/cpuacct 24 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/blkio 25 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/memory 26 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/devices 27 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/freezer 28 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/net_cls 29 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/perf_event 30 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/net_prio 31 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/hugetlb 32 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/pids 33 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/rdma 34 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/misc 44 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/systemd type options dump pass 20 cgroup2 rw,nosuid,nodev,noexec,relatime 0 0 21 cgroup rw,nosuid,nodev,noexec,relatime,cpuset 0 0 22 cgroup rw,nosuid,nodev,noexec,relatime,cpu 0 0 23 cgroup rw,nosuid,nodev,noexec,relatime,cpuacct 0 0 24 cgroup rw,nosuid,nodev,noexec,relatime,blkio 0 0 25 cgroup rw,nosuid,nodev,noexec,relatime,memory 0 0 26 cgroup rw,nosuid,nodev,noexec,relatime,devices 0 0 27 cgroup rw,nosuid,nodev,noexec,relatime,freezer 0 0 28 cgroup rw,nosuid,nodev,noexec,relatime,net_cls 0 0 29 cgroup rw,nosuid,nodev,noexec,relatime,perf_event 0 0 30 cgroup rw,nosuid,nodev,noexec,relatime,net_prio 0 0 31 cgroup rw,nosuid,nodev,noexec,relatime,hugetlb 0 0 32 cgroup rw,nosuid,nodev,noexec,relatime,pids 0 0 33 cgroup rw,nosuid,nodev,noexec,relatime,rdma 0 0 34 cgroup rw,nosuid,nodev,noexec,relatime,misc 0 0 44 cgroup rw,nosuid,nodev,noexec,relatime,xattr,name=systemd 0 0 - getCGroups(): [1] hierarchy_id controller path <0 rows> (or 0-length row.names) - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 Running custom test script: 'D:/RCompile/CRANincoming/R-devel/lib/parallelly/test-data/mixed-cgroups/wsl2-invalid-mounts.R' device 20 cgroup2 21 cgroup 22 cgroup 23 cgroup 24 cgroup 25 cgroup 26 cgroup 27 cgroup 28 cgroup 29 cgroup 30 cgroup 31 cgroup 32 cgroup 33 cgroup 34 cgroup 44 cgroup mountpoint 20 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/unified 21 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/cpuset 22 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/cpu 23 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/cpuacct 24 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/blkio 25 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/memory 26 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/devices 27 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/freezer 28 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/net_cls 29 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/perf_event 30 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/net_prio 31 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/hugetlb 32 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/pids 33 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/rdma 34 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/misc 44 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed86e04359b/sys/fs/cgroup/systemd type options dump pass 20 cgroup2 rw,nosuid,nodev,noexec,relatime 0 0 21 cgroup rw,nosuid,nodev,noexec,relatime,cpuset 0 0 22 cgroup rw,nosuid,nodev,noexec,relatime,cpu 0 0 23 cgroup rw,nosuid,nodev,noexec,relatime,cpuacct 0 0 24 cgroup rw,nosuid,nodev,noexec,relatime,blkio 0 0 25 cgroup rw,nosuid,nodev,noexec,relatime,memory 0 0 26 cgroup rw,nosuid,nodev,noexec,relatime,devices 0 0 27 cgroup rw,nosuid,nodev,noexec,relatime,freezer 0 0 28 cgroup rw,nosuid,nodev,noexec,relatime,net_cls 0 0 29 cgroup rw,nosuid,nodev,noexec,relatime,perf_event 0 0 30 cgroup rw,nosuid,nodev,noexec,relatime,net_prio 0 0 31 cgroup rw,nosuid,nodev,noexec,relatime,hugetlb 0 0 32 cgroup rw,nosuid,nodev,noexec,relatime,pids 0 0 33 cgroup rw,nosuid,nodev,noexec,relatime,rdma 0 0 34 cgroup rw,nosuid,nodev,noexec,relatime,misc 0 0 44 cgroup rw,nosuid,nodev,noexec,relatime,xattr,name=systemd 0 0 CGroups for system 'wsl2-invalid-mounts' ... done mixed-cgroups - real-world ... done cgroups1 - real-world ... CGroups for system 'c4-dev3' ... - Using temporary folder: 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed83397144d' - UID: 581 - procPath(): 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed83397144d/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/systemd cgroup rw,nosuid,nodev,noexec,relatime,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd 0 0 02: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/rdma cgroup rw,nosuid,nodev,noexec,relatime,rdma 0 0 03: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/cpu,cpuacct cgroup rw,nosuid,nodev,noexec,relatime,cpu,cpuacct 0 0 04: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/freezer cgroup rw,nosuid,nodev,noexec,relatime,freezer 0 0 05: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/devices cgroup rw,nosuid,nodev,noexec,relatime,devices 0 0 06: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/net_cls,net_prio cgroup rw,nosuid,nodev,noexec,relatime,net_cls,net_prio 0 0 07: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/memory cgroup rw,nosuid,nodev,noexec,relatime,memory 0 0 08: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/pids cgroup rw,nosuid,nodev,noexec,relatime,pids 0 0 09: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/hugetlb cgroup rw,nosuid,nodev,noexec,relatime,hugetlb 0 0 10: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/cpuset cgroup rw,nosuid,nodev,noexec,relatime,cpuset 0 0 11: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/perf_event cgroup rw,nosuid,nodev,noexec,relatime,perf_event 0 0 12: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/blkio cgroup rw,nosuid,nodev,noexec,relatime,blkio 0 0 - getCGroupsVersion(): 1 - getCGroupsMounts(): device 1 cgroup 2 cgroup 3 cgroup 4 cgroup 5 cgroup 6 cgroup 7 cgroup 8 cgroup 9 cgroup 10 cgroup 11 cgroup 12 cgroup mountpoint 1 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/systemd 2 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/rdma 3 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/cpu,cpuacct 4 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/freezer 5 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/devices 6 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/net_cls,net_prio 7 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/memory 8 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/pids 9 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/hugetlb 10 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/cpuset 11 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/perf_event 12 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed83397144d/sys/fs/cgroup/blkio type 1 cgroup 2 cgroup 3 cgroup 4 cgroup 5 cgroup 6 cgroup 7 cgroup 8 cgroup 9 cgroup 10 cgroup 11 cgroup 12 cgroup options 1 rw,nosuid,nodev,noexec,relatime,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd 2 rw,nosuid,nodev,noexec,relatime,rdma 3 rw,nosuid,nodev,noexec,relatime,cpu,cpuacct 4 rw,nosuid,nodev,noexec,relatime,freezer 5 rw,nosuid,nodev,noexec,relatime,devices 6 rw,nosuid,nodev,noexec,relatime,net_cls,net_prio 7 rw,nosuid,nodev,noexec,relatime,memory 8 rw,nosuid,nodev,noexec,relatime,pids 9 rw,nosuid,nodev,noexec,relatime,hugetlb 10 rw,nosuid,nodev,noexec,relatime,cpuset 11 rw,nosuid,nodev,noexec,relatime,perf_event 12 rw,nosuid,nodev,noexec,relatime,blkio dump pass 1 0 0 2 0 0 3 0 0 4 0 0 5 0 0 6 0 0 7 0 0 8 0 0 9 0 0 10 0 0 11 0 0 12 0 0 - getCGroups(): hierarchy_id controller path 12 1 name=systemd /user.slice/user-581.slice/session-9171.scope 11 2 rdma / 10 3 cpu / 101 3 cpuacct / 9 4 freezer / 8 5 devices /user.slice 7 6 net_cls / 71 6 net_prio / 6 7 memory /user.slice/user-581.slice/session-9171.scope 5 8 pids /user.slice/user-581.slice/session-9171.scope 4 9 hugetlb / 3 10 cpuset / 2 11 perf_event / 1 12 blkio /system.slice/sshd.service - length(getCGroups1CpuSet()): 40 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status cgroups.cpuset 96 96 40 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 Running custom test script: 'D:/RCompile/CRANincoming/R-devel/lib/parallelly/test-data/cgroups1/c4-dev3.R' system /proc/self/status cgroups.cpuset 96 96 40 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 CGroups for system 'c4-dev3' ... done CGroups for system 'c4-n39-4cores' ... - Using temporary folder: 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed8175a3f1a' - UID: 581 - procPath(): 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed8175a3f1a/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/systemd cgroup rw,nosuid,nodev,noexec,relatime,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd 0 0 02: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/net_cls,net_prio cgroup rw,nosuid,nodev,noexec,relatime,net_cls,net_prio 0 0 03: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/hugetlb cgroup rw,nosuid,nodev,noexec,relatime,hugetlb 0 0 04: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/blkio cgroup rw,nosuid,nodev,noexec,relatime,blkio 0 0 05: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/freezer cgroup rw,nosuid,nodev,noexec,relatime,freezer 0 0 06: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/pids cgroup rw,nosuid,nodev,noexec,relatime,pids 0 0 07: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/memory cgroup rw,nosuid,nodev,noexec,relatime,memory 0 0 08: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/perf_event cgroup rw,nosuid,nodev,noexec,relatime,perf_event 0 0 09: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/cpu,cpuacct cgroup rw,nosuid,nodev,noexec,relatime,cpu,cpuacct 0 0 10: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/cpuset cgroup rw,nosuid,nodev,noexec,relatime,cpuset 0 0 11: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/devices cgroup rw,nosuid,nodev,noexec,relatime,devices 0 0 12: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/rdma cgroup rw,nosuid,nodev,noexec,relatime,rdma 0 0 - getCGroupsVersion(): 1 - getCGroupsMounts(): device 1 cgroup 2 cgroup 3 cgroup 4 cgroup 5 cgroup 6 cgroup 7 cgroup 8 cgroup 9 cgroup 10 cgroup 11 cgroup 12 cgroup mountpoint 1 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/systemd 2 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/net_cls,net_prio 3 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/hugetlb 4 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/blkio 5 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/freezer 6 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/pids 7 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/memory 8 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/perf_event 9 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/cpu,cpuacct 10 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/cpuset 11 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/devices 12 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8175a3f1a/sys/fs/cgroup/rdma type 1 cgroup 2 cgroup 3 cgroup 4 cgroup 5 cgroup 6 cgroup 7 cgroup 8 cgroup 9 cgroup 10 cgroup 11 cgroup 12 cgroup options 1 rw,nosuid,nodev,noexec,relatime,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd 2 rw,nosuid,nodev,noexec,relatime,net_cls,net_prio 3 rw,nosuid,nodev,noexec,relatime,hugetlb 4 rw,nosuid,nodev,noexec,relatime,blkio 5 rw,nosuid,nodev,noexec,relatime,freezer 6 rw,nosuid,nodev,noexec,relatime,pids 7 rw,nosuid,nodev,noexec,relatime,memory 8 rw,nosuid,nodev,noexec,relatime,perf_event 9 rw,nosuid,nodev,noexec,relatime,cpu,cpuacct 10 rw,nosuid,nodev,noexec,relatime,cpuset 11 rw,nosuid,nodev,noexec,relatime,devices 12 rw,nosuid,nodev,noexec,relatime,rdma dump pass 1 0 0 2 0 0 3 0 0 4 0 0 5 0 0 6 0 0 7 0 0 8 0 0 9 0 0 10 0 0 11 0 0 12 0 0 - getCGroups(): hierarchy_id controller path 12 1 name=systemd /system.slice/slurmd.service 11 2 net_cls / 111 2 net_prio / 10 3 hugetlb / 9 4 blkio /system.slice/slurmd.service 8 5 freezer /slurm/uid_581/job_255208/step_0 7 6 pids /system.slice/slurmd.service 6 7 memory /slurm/uid_581/job_255208/step_0 5 8 perf_event / 4 9 cpu /system.slice/slurmd.service 41 9 cpuacct /system.slice/slurmd.service 3 10 cpuset /slurm/uid_581/job_255208/step_0 2 11 devices /slurm/uid_581/job_255208/step_0/task_0 1 12 rdma / - length(getCGroups1CpuSet()): 4 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status cgroups.cpuset 96 96 4 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 CGroups for system 'c4-n39-4cores' ... done CGroups for system 'posit-workbench' ... - Using temporary folder: 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed813d355af' - UID: 6776184 - procPath(): 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed813d355af/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/systemd cgroup ro,nosuid,nodev,noexec,relatime,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd 0 0 02: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/hugetlb cgroup ro,nosuid,nodev,noexec,relatime,hugetlb 0 0 03: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/cpu,cpuacct cgroup ro,nosuid,nodev,noexec,relatime,cpu,cpuacct 0 0 04: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/cpuset cgroup ro,nosuid,nodev,noexec,relatime,cpuset 0 0 05: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/perf_event cgroup ro,nosuid,nodev,noexec,relatime,perf_event 0 0 06: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/net_cls,net_prio cgroup ro,nosuid,nodev,noexec,relatime,net_cls,net_prio 0 0 07: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/memory cgroup ro,nosuid,nodev,noexec,relatime,memory 0 0 08: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/blkio cgroup ro,nosuid,nodev,noexec,relatime,blkio 0 0 09: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/freezer cgroup ro,nosuid,nodev,noexec,relatime,freezer 0 0 10: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/devices cgroup ro,nosuid,nodev,noexec,relatime,devices 0 0 11: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/pids cgroup ro,nosuid,nodev,noexec,relatime,pids 0 0 - getCGroupsVersion(): 1 - getCGroupsMounts(): device 1 cgroup 2 cgroup 3 cgroup 4 cgroup 5 cgroup 6 cgroup 7 cgroup 8 cgroup 9 cgroup 10 cgroup 11 cgroup mountpoint 1 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/systemd 2 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/hugetlb 3 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/cpu,cpuacct 4 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/cpuset 5 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/perf_event 6 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/net_cls,net_prio 7 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/memory 8 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/blkio 9 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/freezer 10 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/devices 11 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed813d355af/sys/fs/cgroup/pids type 1 cgroup 2 cgroup 3 cgroup 4 cgroup 5 cgroup 6 cgroup 7 cgroup 8 cgroup 9 cgroup 10 cgroup 11 cgroup options 1 ro,nosuid,nodev,noexec,relatime,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd 2 ro,nosuid,nodev,noexec,relatime,hugetlb 3 ro,nosuid,nodev,noexec,relatime,cpu,cpuacct 4 ro,nosuid,nodev,noexec,relatime,cpuset 5 ro,nosuid,nodev,noexec,relatime,perf_event 6 ro,nosuid,nodev,noexec,relatime,net_cls,net_prio 7 ro,nosuid,nodev,noexec,relatime,memory 8 ro,nosuid,nodev,noexec,relatime,blkio 9 ro,nosuid,nodev,noexec,relatime,freezer 10 ro,nosuid,nodev,noexec,relatime,devices 11 ro,nosuid,nodev,noexec,relatime,pids dump pass 1 0 0 2 0 0 3 0 0 4 0 0 5 0 0 6 0 0 7 0 0 8 0 0 9 0 0 10 0 0 11 0 0 - getCGroups(): hierarchy_id controller 11 1 name=systemd 10 2 hugetlb 9 3 cpu 91 3 cpuacct 8 4 cpuset 7 5 perf_event 6 6 net_cls 61 6 net_prio 5 7 memory 4 8 blkio 3 9 freezer 2 10 devices 1 11 pids path 11 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 10 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 9 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 91 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 8 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 7 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 6 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 61 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 5 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 4 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 3 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 2 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 1 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope - length(getCGroups1CpuSet()): 32 - getCGroups1CpuQuota(): 2 - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status cgroups.cpuset 96 96 32 cgroups.cpuquota mc.cores _R_CHECK_LIMIT_CORES_ 2 2 2 Running custom test script: 'D:/RCompile/CRANincoming/R-devel/lib/parallelly/test-data/cgroups1/posit-workbench.R' system /proc/self/status cgroups.cpuset 96 96 32 cgroups.cpuquota mc.cores _R_CHECK_LIMIT_CORES_ 2 2 2 CGroups for system 'posit-workbench' ... done cgroups1 - real-world ... done cgroups2 - real-world ... CGroups for system 'dev1' ... - Using temporary folder: 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed824333fc9' - UID: 34002 - procPath(): 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed824333fc9/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup2 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed824333fc9/sys/fs/cgroup cgroup2 rw,seclabel,nosuid,nodev,noexec,relatime,nsdelegate 0 0 - getCGroupsVersion(): 2 - getCGroupsMounts(): device 1 cgroup2 mountpoint 1 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed824333fc9/sys/fs/cgroup type options dump pass 1 cgroup2 rw,seclabel,nosuid,nodev,noexec,relatime,nsdelegate 0 0 - getCGroups(): hierarchy_id controller path 1 0 /user.slice/user-34002.slice/session-139474.scope - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 Running custom test script: 'D:/RCompile/CRANincoming/R-devel/lib/parallelly/test-data/cgroups2/dev1.R' system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 CGroups for system 'dev1' ... done CGroups for system 'dt1' ... - Using temporary folder: 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed84d5462c1' - UID: 34002 - procPath(): 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed84d5462c1/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup2 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed84d5462c1/sys/fs/cgroup cgroup2 rw,seclabel,nosuid,nodev,noexec,relatime,nsdelegate 0 0 - getCGroupsVersion(): 2 - getCGroupsMounts(): device 1 cgroup2 mountpoint 1 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed84d5462c1/sys/fs/cgroup type options dump pass 1 cgroup2 rw,seclabel,nosuid,nodev,noexec,relatime,nsdelegate 0 0 - getCGroups(): hierarchy_id controller path 1 0 /user.slice/user-34002.slice - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): 2 - availableCores(which = 'all'): system /proc/self/status cgroups2.cpu.max 96 96 2 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 Running custom test script: 'D:/RCompile/CRANincoming/R-devel/lib/parallelly/test-data/cgroups2/dt1.R' system /proc/self/status cgroups2.cpu.max 96 96 2 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 CGroups for system 'dt1' ... done CGroups for system 'log1' ... - Using temporary folder: 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed8b7c2cb3' - UID: 34002 - procPath(): 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed8b7c2cb3/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup2 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8b7c2cb3/sys/fs/cgroup cgroup2 rw,seclabel,nosuid,nodev,noexec,relatime,nsdelegate 0 0 - getCGroupsVersion(): 2 - getCGroupsMounts(): device 1 cgroup2 mountpoint 1 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8b7c2cb3/sys/fs/cgroup type options dump pass 1 cgroup2 rw,seclabel,nosuid,nodev,noexec,relatime,nsdelegate 0 0 - getCGroups(): hierarchy_id controller path 1 0 /user.slice/user-34002.slice - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): 1 - availableCores(which = 'all'): system /proc/self/status cgroups2.cpu.max 96 96 1 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 Running custom test script: 'D:/RCompile/CRANincoming/R-devel/lib/parallelly/test-data/cgroups2/log1.R' system /proc/self/status cgroups2.cpu.max 96 96 1 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 CGroups for system 'log1' ... done CGroups for system 'pi-2021' ... - Using temporary folder: 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed833896989' - UID: 1000 - procPath(): 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed833896989/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup2 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed833896989/sys/fs/cgroup cgroup2 rw,nosuid,nodev,noexec,relatime,nsdelegate,memory_recursiveprot 0 0 - getCGroupsVersion(): 2 - getCGroupsMounts(): device 1 cgroup2 mountpoint 1 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed833896989/sys/fs/cgroup type options dump 1 cgroup2 rw,nosuid,nodev,noexec,relatime,nsdelegate,memory_recursiveprot 0 pass 1 0 - getCGroups(): hierarchy_id controller path 1 0 /user.slice/user-1000.slice/session-1498.scope - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 Running custom test script: 'D:/RCompile/CRANincoming/R-devel/lib/parallelly/test-data/cgroups2/pi-2021.R' system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 CGroups for system 'pi-2021' ... done CGroups for system 'posit.cloud-free' ... - Using temporary folder: 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed8a3139b2' - UID: 2025988 - procPath(): 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed8a3139b2/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8a3139b2/sys/fs/cgroup cgroup2 ro,nosuid,nodev,noexec,relatime 0 0 - getCGroupsVersion(): 2 - getCGroupsMounts(): device 1 cgroup mountpoint 1 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed8a3139b2/sys/fs/cgroup type options dump pass 1 cgroup2 ro,nosuid,nodev,noexec,relatime 0 0 - getCGroups(): hierarchy_id controller path 1 0 / - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): 1 - availableCores(which = 'all'): system /proc/self/status cgroups2.cpu.max 96 96 1 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 Running custom test script: 'D:/RCompile/CRANincoming/R-devel/lib/parallelly/test-data/cgroups2/posit.cloud-free.R' system /proc/self/status cgroups2.cpu.max 96 96 1 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 CGroups for system 'posit.cloud-free' ... done CGroups for system 'x1' ... - Using temporary folder: 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed82a727f56' - UID: 1000 - procPath(): 'D:\temp\2025_06_02_21_45_16_15648\RtmpSuLAF3\fileed82a727f56/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup2 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed82a727f56/sys/fs/cgroup cgroup2 rw,nosuid,nodev,noexec,relatime 0 0 - getCGroupsVersion(): 2 - getCGroupsMounts(): device 1 cgroup2 mountpoint 1 D:/temp/2025_06_02_21_45_16_15648/RtmpSuLAF3/fileed82a727f56/sys/fs/cgroup type options dump pass 1 cgroup2 rw,nosuid,nodev,noexec,relatime 0 0 - getCGroups(): hierarchy_id controller 1 0 path 1 /user.slice/user-1000.slice/user@1000.service/app.slice/app-org.gnome.Terminal.slice/vte-spawn-2b165c43-93a4-4257-bef3-0c9d5a3b7f97.scope - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 Running custom test script: 'D:/RCompile/CRANincoming/R-devel/lib/parallelly/test-data/cgroups2/x1.R' system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 CGroups for system 'x1' ... done cgroups2 - real-world ... done > message("cgroups - real-world ... DONE") cgroups - real-world ... DONE Test time: user.self=0.2s, sys.self=0.1s, elapsed=1s, user.child=NAs, sys.child=NAs Test 'cgroups' ... success > > proc.time() user system elapsed 0.35 0.17 1.65