R Under development (unstable) (2025-01-29 r87668 ucrt) -- "Unsuffered Consequences" Copyright (C) 2025 The R Foundation for Statistical Computing Platform: x86_64-w64-mingw32/x64 R is free software and comes with ABSOLUTELY NO WARRANTY. You are welcome to redistribute it under certain conditions. Type 'license()' or 'licence()' for distribution details. R is a collaborative project with many contributors. Type 'contributors()' for more information and 'citation()' on how to cite R or R packages in publications. Type 'demo()' for some demos, 'help()' for on-line help, or 'help.start()' for an HTML browser interface to help. Type 'q()' to quit R. > source("incl/start.R") > > message("*** cgroups ...") *** cgroups ... > > message("- getCGroups()") - getCGroups() > cgroups <- parallelly:::getCGroups() > print(cgroups) [1] hierarchy_id controller path <0 rows> (or 0-length row.names) > stopifnot( + is.data.frame(cgroups), + identical(colnames(cgroups), c("hierarchy_id", "controller", "path")), + nrow(cgroups) == 0L || !is.null(cgroups$controller) + ) > > message("- getCGroupsRoot()") - getCGroupsRoot() > root <- parallelly:::getCGroupsRoot() > cat(sprintf("cgroups root path: %s\n", sQuote(root))) cgroups root path: 'NA' > stopifnot(length(root) == 1L, is.character(root)) > > message("- getCGroups()") - getCGroups() > cgroups <- parallelly:::getCGroups() > print(cgroups) [1] hierarchy_id controller path <0 rows> (or 0-length row.names) > stopifnot(is.data.frame(cgroups)) > > message("- getCGroupsPath()") - getCGroupsPath() > path <- parallelly:::getCGroupsPath("cpu") > cat(sprintf("cgroups 'cpu' path: %s\n", sQuote(path))) cgroups 'cpu' path: 'NA' > stopifnot(length(path) == 1L, is.character(path)) > > path <- parallelly:::getCGroupsPath("cpuset") > cat(sprintf("cgroups 'cpuset' path: %s\n", sQuote(path))) cgroups 'cpuset' path: 'NA' > stopifnot(length(path) == 1L, is.character(path)) > > > message("- getCGroups1Value()") - getCGroups1Value() > value <- parallelly:::getCGroups1Value("cpu", "cpu.cfs_quota_us") > cat(sprintf("cgroups v1 'cpu.cfs_quota_us' value: %s\n", sQuote(value))) cgroups v1 'cpu.cfs_quota_us' value: 'NA' > stopifnot(length(value) == 1L, is.character(value)) > > value <- parallelly:::getCGroups1Value("cpu", "cpu.cfs_total_us") > cat(sprintf("cgroups v1 'cpu.cfs_total_us' value: %s\n", sQuote(value))) cgroups v1 'cpu.cfs_total_us' value: 'NA' > stopifnot(length(value) == 1L, is.character(value)) > > value <- parallelly:::getCGroups1Value("cpuset", "cpuset.cpus") > cat(sprintf("cgroups v1 'cpuset.cpus' value: %s\n", sQuote(value))) cgroups v1 'cpuset.cpus' value: 'NA' > stopifnot(length(value) == 1L, is.character(value)) > > > message("- getCGroups1CpuSet()") - getCGroups1CpuSet() > value <- parallelly:::getCGroups1CpuSet() > cat(sprintf("CPU set: [n=%d] %s\n", length(value), paste(sQuote(value), collapse = ", "))) CPU set: [n=0] > stopifnot(length(value) >= 0L, is.integer(value), !any(is.na(value))) > > > message("- getCGroups1CpuQuotaMicroseconds()") - getCGroups1CpuQuotaMicroseconds() > value <- parallelly:::getCGroups1CpuQuotaMicroseconds() > cat(sprintf("CPU quota (ms): %d\n", value)) CPU quota (ms): NA > stopifnot( + length(value) == 1L, + is.integer(value), + is.na(value) || value == -1 || value > 0 + ) > > message("- getCGroups1CpuPeriodMicroseconds()") - getCGroups1CpuPeriodMicroseconds() > value <- parallelly:::getCGroups1CpuPeriodMicroseconds() > cat(sprintf("CPU total (ms): %d\n", value)) CPU total (ms): NA > stopifnot( + length(value) == 1L, + is.integer(value), + is.na(value) || value > 0 + ) > > message("- getCGroups1CpuQuota()") - getCGroups1CpuQuota() > value <- parallelly:::getCGroups1CpuQuota() > cat(sprintf("CPU quota (ratio): %g\n", value)) CPU quota (ratio): NA > stopifnot( + length(value) == 1L, + is.numeric(value), + !is.infinite(value), + is.na(value) || value > 0 + ) > > message("- getCGroups2CpuMax()") - getCGroups2CpuMax() > value <- parallelly:::getCGroups2CpuMax() > cat(sprintf("CPU quota (ratio): %g\n", value)) CPU quota (ratio): NA > stopifnot( + length(value) == 1L, + is.numeric(value), + !is.infinite(value), + is.na(value) || value > 0 + ) > > message("*** cgroups ... DONE") *** cgroups ... DONE > > root <- system.file(package = "parallelly", "test-data", mustWork = TRUE) > for (dir in c("no-cgroups", "mixed-cgroups", "cgroups1", "cgroups2")) { + message(sprintf("%s - real-world ...", dir)) + path <- file.path(root, dir) + stopifnot(file_test("-d", path)) + tarballs <- dir(path = path, pattern = ".*[.]tar[.]gz", full.names = TRUE) + names(tarballs) <- sub("[.]tar[.]gz$", "", basename(tarballs)) + + for (name in names(tarballs)) { + parallelly:::withCGroups(tarballs[name], { + file <- file.path(path, sprintf("%s.R", name)) + if (file_test("-f", file)) { + source(file, local = FALSE) + } + }) + } + message(sprintf("%s - real-world ... done", dir)) + } no-cgroups - real-world ... CGroups for system 'no-cgroups' ... - Using temporary folder: 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f4619f1854' - UID: 1000 - procPath(): 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f4619f1854/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: tmpfs /run tmpfs rw,nosuid,nodev,noexec,relatime,size=3256456k,mode=755,inode64 0 0 - getCGroupsVersion(): -1 - getCGroupsMounts(): [1] device mountpoint type options dump pass <0 rows> (or 0-length row.names) - getCGroups(): [1] hierarchy_id controller path <0 rows> (or 0-length row.names) - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 CGroups for system 'no-cgroups' ... done no-cgroups - real-world ... done mixed-cgroups - real-world ... CGroups for system 'piavpn-cgroups' ... - Using temporary folder: 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f43ebd5517' - UID: 1000 - procPath(): 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f43ebd5517/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup2 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ebd5517/sys/fs/cgroup cgroup2 rw,nosuid,nodev,noexec,relatime,nsdelegate,memory_recursiveprot 0 0 02: none D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ebd5517/opt/piavpn/etc/cgroup/net_cls cgroup rw,relatime,net_cls 0 0 - getCGroupsVersion(): 1 - getCGroupsMounts(): device 1 cgroup2 2 none mountpoint 1 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ebd5517/sys/fs/cgroup 2 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ebd5517/opt/piavpn/etc/cgroup/net_cls type options dump 1 cgroup2 rw,nosuid,nodev,noexec,relatime,nsdelegate,memory_recursiveprot 0 2 cgroup rw,relatime,net_cls 0 pass 1 0 2 0 - getCGroups(): hierarchy_id controller 2 0 1 1 net_cls path 2 /user.slice/user-1000.slice/user@1000.service/app.slice/app-org.kde.konsole-4415.scope/tab(4427).scope 1 / - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 CGroups for system 'piavpn-cgroups' ... done CGroups for system 'slurm-cgroups' ... - Using temporary folder: 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f479b5785' - UID: 632771201 - procPath(): 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f479b5785/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup2 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/unified cgroup2 rw,nosuid,nodev,noexec,relatime,nsdelegate 0 0 02: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/systemd cgroup rw,nosuid,nodev,noexec,relatime,xattr,name=systemd 0 0 03: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/memory cgroup rw,nosuid,nodev,noexec,relatime,memory 0 0 04: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/cpu,cpuacct cgroup rw,nosuid,nodev,noexec,relatime,cpu,cpuacct 0 0 05: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/perf_event cgroup rw,nosuid,nodev,noexec,relatime,perf_event 0 0 06: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/hugetlb cgroup rw,nosuid,nodev,noexec,relatime,hugetlb 0 0 07: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/cpuset cgroup rw,nosuid,nodev,noexec,relatime,cpuset 0 0 08: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/rdma cgroup rw,nosuid,nodev,noexec,relatime,rdma 0 0 09: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/misc cgroup rw,nosuid,nodev,noexec,relatime,misc 0 0 10: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/net_cls,net_prio cgroup rw,nosuid,nodev,noexec,relatime,net_cls,net_prio 0 0 11: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/freezer cgroup rw,nosuid,nodev,noexec,relatime,freezer 0 0 12: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/pids cgroup rw,nosuid,nodev,noexec,relatime,pids 0 0 13: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/blkio cgroup rw,nosuid,nodev,noexec,relatime,blkio 0 0 14: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/devices cgroup rw,nosuid,nodev,noexec,relatime,devices 0 0 - getCGroupsVersion(): 1 - getCGroupsMounts(): device 1 cgroup2 2 cgroup 3 cgroup 4 cgroup 5 cgroup 6 cgroup 7 cgroup 8 cgroup 9 cgroup 10 cgroup 11 cgroup 12 cgroup 13 cgroup 14 cgroup mountpoint 1 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/unified 2 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/systemd 3 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/memory 4 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/cpu,cpuacct 5 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/perf_event 6 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/hugetlb 7 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/cpuset 8 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/rdma 9 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/misc 10 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/net_cls,net_prio 11 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/freezer 12 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/pids 13 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/blkio 14 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f479b5785/sys/fs/cgroup/devices type options dump pass 1 cgroup2 rw,nosuid,nodev,noexec,relatime,nsdelegate 0 0 2 cgroup rw,nosuid,nodev,noexec,relatime,xattr,name=systemd 0 0 3 cgroup rw,nosuid,nodev,noexec,relatime,memory 0 0 4 cgroup rw,nosuid,nodev,noexec,relatime,cpu,cpuacct 0 0 5 cgroup rw,nosuid,nodev,noexec,relatime,perf_event 0 0 6 cgroup rw,nosuid,nodev,noexec,relatime,hugetlb 0 0 7 cgroup rw,nosuid,nodev,noexec,relatime,cpuset 0 0 8 cgroup rw,nosuid,nodev,noexec,relatime,rdma 0 0 9 cgroup rw,nosuid,nodev,noexec,relatime,misc 0 0 10 cgroup rw,nosuid,nodev,noexec,relatime,net_cls,net_prio 0 0 11 cgroup rw,nosuid,nodev,noexec,relatime,freezer 0 0 12 cgroup rw,nosuid,nodev,noexec,relatime,pids 0 0 13 cgroup rw,nosuid,nodev,noexec,relatime,blkio 0 0 14 cgroup rw,nosuid,nodev,noexec,relatime,devices 0 0 - getCGroups(): hierarchy_id controller path 14 0 /system.slice/slurmd.service 13 1 name=systemd /system.slice/slurmd.service 12 2 memory /slurm/uid_632771201/job_1140459 11 3 cpu /slurm/uid_632771201/job_1140459 111 3 cpuacct /slurm/uid_632771201/job_1140459 10 4 perf_event / 9 5 hugetlb / 8 6 cpuset /slurm/uid_632771201/job_1140459/step_0 7 7 rdma / 6 8 misc / 5 9 net_cls / 51 9 net_prio / 4 10 freezer /slurm/uid_632771201/job_1140459/step_0 3 11 pids /system.slice/slurmd.service 2 12 blkio /slurm/uid_632771201/job_1140459 1 13 devices /slurm/uid_632771201/job_1140459/step_0/task_0 - length(getCGroups1CpuSet()): 1 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status cgroups.cpuset 96 96 1 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 CGroups for system 'slurm-cgroups' ... done mixed-cgroups - real-world ... done cgroups1 - real-world ... CGroups for system 'c4-dev3' ... - Using temporary folder: 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f4130b19ca' - UID: 581 - procPath(): 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f4130b19ca/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/systemd cgroup rw,nosuid,nodev,noexec,relatime,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd 0 0 02: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/rdma cgroup rw,nosuid,nodev,noexec,relatime,rdma 0 0 03: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/cpu,cpuacct cgroup rw,nosuid,nodev,noexec,relatime,cpu,cpuacct 0 0 04: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/freezer cgroup rw,nosuid,nodev,noexec,relatime,freezer 0 0 05: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/devices cgroup rw,nosuid,nodev,noexec,relatime,devices 0 0 06: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/net_cls,net_prio cgroup rw,nosuid,nodev,noexec,relatime,net_cls,net_prio 0 0 07: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/memory cgroup rw,nosuid,nodev,noexec,relatime,memory 0 0 08: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/pids cgroup rw,nosuid,nodev,noexec,relatime,pids 0 0 09: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/hugetlb cgroup rw,nosuid,nodev,noexec,relatime,hugetlb 0 0 10: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/cpuset cgroup rw,nosuid,nodev,noexec,relatime,cpuset 0 0 11: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/perf_event cgroup rw,nosuid,nodev,noexec,relatime,perf_event 0 0 12: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/blkio cgroup rw,nosuid,nodev,noexec,relatime,blkio 0 0 - getCGroupsVersion(): 1 - getCGroupsMounts(): device 1 cgroup 2 cgroup 3 cgroup 4 cgroup 5 cgroup 6 cgroup 7 cgroup 8 cgroup 9 cgroup 10 cgroup 11 cgroup 12 cgroup mountpoint 1 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/systemd 2 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/rdma 3 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/cpu,cpuacct 4 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/freezer 5 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/devices 6 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/net_cls,net_prio 7 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/memory 8 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/pids 9 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/hugetlb 10 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/cpuset 11 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/perf_event 12 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4130b19ca/sys/fs/cgroup/blkio type 1 cgroup 2 cgroup 3 cgroup 4 cgroup 5 cgroup 6 cgroup 7 cgroup 8 cgroup 9 cgroup 10 cgroup 11 cgroup 12 cgroup options 1 rw,nosuid,nodev,noexec,relatime,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd 2 rw,nosuid,nodev,noexec,relatime,rdma 3 rw,nosuid,nodev,noexec,relatime,cpu,cpuacct 4 rw,nosuid,nodev,noexec,relatime,freezer 5 rw,nosuid,nodev,noexec,relatime,devices 6 rw,nosuid,nodev,noexec,relatime,net_cls,net_prio 7 rw,nosuid,nodev,noexec,relatime,memory 8 rw,nosuid,nodev,noexec,relatime,pids 9 rw,nosuid,nodev,noexec,relatime,hugetlb 10 rw,nosuid,nodev,noexec,relatime,cpuset 11 rw,nosuid,nodev,noexec,relatime,perf_event 12 rw,nosuid,nodev,noexec,relatime,blkio dump pass 1 0 0 2 0 0 3 0 0 4 0 0 5 0 0 6 0 0 7 0 0 8 0 0 9 0 0 10 0 0 11 0 0 12 0 0 - getCGroups(): hierarchy_id controller path 12 1 name=systemd /user.slice/user-581.slice/session-9171.scope 11 2 rdma / 10 3 cpu / 101 3 cpuacct / 9 4 freezer / 8 5 devices /user.slice 7 6 net_cls / 71 6 net_prio / 6 7 memory /user.slice/user-581.slice/session-9171.scope 5 8 pids /user.slice/user-581.slice/session-9171.scope 4 9 hugetlb / 3 10 cpuset / 2 11 perf_event / 1 12 blkio /system.slice/sshd.service - length(getCGroups1CpuSet()): 40 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status cgroups.cpuset 96 96 40 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 system /proc/self/status cgroups.cpuset 96 96 40 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 CGroups for system 'c4-dev3' ... done CGroups for system 'c4-n39-4cores' ... - Using temporary folder: 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f4710332a4' - UID: 581 - procPath(): 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f4710332a4/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/systemd cgroup rw,nosuid,nodev,noexec,relatime,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd 0 0 02: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/net_cls,net_prio cgroup rw,nosuid,nodev,noexec,relatime,net_cls,net_prio 0 0 03: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/hugetlb cgroup rw,nosuid,nodev,noexec,relatime,hugetlb 0 0 04: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/blkio cgroup rw,nosuid,nodev,noexec,relatime,blkio 0 0 05: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/freezer cgroup rw,nosuid,nodev,noexec,relatime,freezer 0 0 06: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/pids cgroup rw,nosuid,nodev,noexec,relatime,pids 0 0 07: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/memory cgroup rw,nosuid,nodev,noexec,relatime,memory 0 0 08: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/perf_event cgroup rw,nosuid,nodev,noexec,relatime,perf_event 0 0 09: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/cpu,cpuacct cgroup rw,nosuid,nodev,noexec,relatime,cpu,cpuacct 0 0 10: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/cpuset cgroup rw,nosuid,nodev,noexec,relatime,cpuset 0 0 11: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/devices cgroup rw,nosuid,nodev,noexec,relatime,devices 0 0 12: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/rdma cgroup rw,nosuid,nodev,noexec,relatime,rdma 0 0 - getCGroupsVersion(): 1 - getCGroupsMounts(): device 1 cgroup 2 cgroup 3 cgroup 4 cgroup 5 cgroup 6 cgroup 7 cgroup 8 cgroup 9 cgroup 10 cgroup 11 cgroup 12 cgroup mountpoint 1 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/systemd 2 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/net_cls,net_prio 3 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/hugetlb 4 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/blkio 5 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/freezer 6 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/pids 7 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/memory 8 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/perf_event 9 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/cpu,cpuacct 10 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/cpuset 11 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/devices 12 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4710332a4/sys/fs/cgroup/rdma type 1 cgroup 2 cgroup 3 cgroup 4 cgroup 5 cgroup 6 cgroup 7 cgroup 8 cgroup 9 cgroup 10 cgroup 11 cgroup 12 cgroup options 1 rw,nosuid,nodev,noexec,relatime,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd 2 rw,nosuid,nodev,noexec,relatime,net_cls,net_prio 3 rw,nosuid,nodev,noexec,relatime,hugetlb 4 rw,nosuid,nodev,noexec,relatime,blkio 5 rw,nosuid,nodev,noexec,relatime,freezer 6 rw,nosuid,nodev,noexec,relatime,pids 7 rw,nosuid,nodev,noexec,relatime,memory 8 rw,nosuid,nodev,noexec,relatime,perf_event 9 rw,nosuid,nodev,noexec,relatime,cpu,cpuacct 10 rw,nosuid,nodev,noexec,relatime,cpuset 11 rw,nosuid,nodev,noexec,relatime,devices 12 rw,nosuid,nodev,noexec,relatime,rdma dump pass 1 0 0 2 0 0 3 0 0 4 0 0 5 0 0 6 0 0 7 0 0 8 0 0 9 0 0 10 0 0 11 0 0 12 0 0 - getCGroups(): hierarchy_id controller path 12 1 name=systemd /system.slice/slurmd.service 11 2 net_cls / 111 2 net_prio / 10 3 hugetlb / 9 4 blkio /system.slice/slurmd.service 8 5 freezer /slurm/uid_581/job_255208/step_0 7 6 pids /system.slice/slurmd.service 6 7 memory /slurm/uid_581/job_255208/step_0 5 8 perf_event / 4 9 cpu /system.slice/slurmd.service 41 9 cpuacct /system.slice/slurmd.service 3 10 cpuset /slurm/uid_581/job_255208/step_0 2 11 devices /slurm/uid_581/job_255208/step_0/task_0 1 12 rdma / - length(getCGroups1CpuSet()): 4 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status cgroups.cpuset 96 96 4 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 CGroups for system 'c4-n39-4cores' ... done CGroups for system 'posit-workbench' ... - Using temporary folder: 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f43ace125d' - UID: 6776184 - procPath(): 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f43ace125d/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/systemd cgroup ro,nosuid,nodev,noexec,relatime,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd 0 0 02: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/hugetlb cgroup ro,nosuid,nodev,noexec,relatime,hugetlb 0 0 03: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/cpu,cpuacct cgroup ro,nosuid,nodev,noexec,relatime,cpu,cpuacct 0 0 04: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/cpuset cgroup ro,nosuid,nodev,noexec,relatime,cpuset 0 0 05: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/perf_event cgroup ro,nosuid,nodev,noexec,relatime,perf_event 0 0 06: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/net_cls,net_prio cgroup ro,nosuid,nodev,noexec,relatime,net_cls,net_prio 0 0 07: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/memory cgroup ro,nosuid,nodev,noexec,relatime,memory 0 0 08: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/blkio cgroup ro,nosuid,nodev,noexec,relatime,blkio 0 0 09: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/freezer cgroup ro,nosuid,nodev,noexec,relatime,freezer 0 0 10: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/devices cgroup ro,nosuid,nodev,noexec,relatime,devices 0 0 11: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/pids cgroup ro,nosuid,nodev,noexec,relatime,pids 0 0 - getCGroupsVersion(): 1 - getCGroupsMounts(): device 1 cgroup 2 cgroup 3 cgroup 4 cgroup 5 cgroup 6 cgroup 7 cgroup 8 cgroup 9 cgroup 10 cgroup 11 cgroup mountpoint 1 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/systemd 2 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/hugetlb 3 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/cpu,cpuacct 4 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/cpuset 5 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/perf_event 6 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/net_cls,net_prio 7 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/memory 8 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/blkio 9 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/freezer 10 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/devices 11 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43ace125d/sys/fs/cgroup/pids type 1 cgroup 2 cgroup 3 cgroup 4 cgroup 5 cgroup 6 cgroup 7 cgroup 8 cgroup 9 cgroup 10 cgroup 11 cgroup options 1 ro,nosuid,nodev,noexec,relatime,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd 2 ro,nosuid,nodev,noexec,relatime,hugetlb 3 ro,nosuid,nodev,noexec,relatime,cpu,cpuacct 4 ro,nosuid,nodev,noexec,relatime,cpuset 5 ro,nosuid,nodev,noexec,relatime,perf_event 6 ro,nosuid,nodev,noexec,relatime,net_cls,net_prio 7 ro,nosuid,nodev,noexec,relatime,memory 8 ro,nosuid,nodev,noexec,relatime,blkio 9 ro,nosuid,nodev,noexec,relatime,freezer 10 ro,nosuid,nodev,noexec,relatime,devices 11 ro,nosuid,nodev,noexec,relatime,pids dump pass 1 0 0 2 0 0 3 0 0 4 0 0 5 0 0 6 0 0 7 0 0 8 0 0 9 0 0 10 0 0 11 0 0 - getCGroups(): hierarchy_id controller 11 1 name=systemd 10 2 hugetlb 9 3 cpu 91 3 cpuacct 8 4 cpuset 7 5 perf_event 6 6 net_cls 61 6 net_prio 5 7 memory 4 8 blkio 3 9 freezer 2 10 devices 1 11 pids path 11 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 10 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 9 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 91 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 8 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 7 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 6 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 61 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 5 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 4 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 3 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 2 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 1 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope - length(getCGroups1CpuSet()): 32 - getCGroups1CpuQuota(): 2 - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status cgroups.cpuset 96 96 32 cgroups.cpuquota mc.cores _R_CHECK_LIMIT_CORES_ 2 2 2 system /proc/self/status cgroups.cpuset 96 96 32 cgroups.cpuquota mc.cores _R_CHECK_LIMIT_CORES_ 2 2 2 CGroups for system 'posit-workbench' ... done cgroups1 - real-world ... done cgroups2 - real-world ... CGroups for system 'dev1' ... - Using temporary folder: 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f475fb1cec' - UID: 34002 - procPath(): 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f475fb1cec/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup2 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f475fb1cec/sys/fs/cgroup cgroup2 rw,seclabel,nosuid,nodev,noexec,relatime,nsdelegate 0 0 - getCGroupsVersion(): 2 - getCGroupsMounts(): device 1 cgroup2 mountpoint 1 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f475fb1cec/sys/fs/cgroup type options dump pass 1 cgroup2 rw,seclabel,nosuid,nodev,noexec,relatime,nsdelegate 0 0 - getCGroups(): hierarchy_id controller path 1 0 /user.slice/user-34002.slice/session-139474.scope - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 CGroups for system 'dev1' ... done CGroups for system 'dt1' ... - Using temporary folder: 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f440aa21ef' - UID: 34002 - procPath(): 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f440aa21ef/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup2 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f440aa21ef/sys/fs/cgroup cgroup2 rw,seclabel,nosuid,nodev,noexec,relatime,nsdelegate 0 0 - getCGroupsVersion(): 2 - getCGroupsMounts(): device 1 cgroup2 mountpoint 1 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f440aa21ef/sys/fs/cgroup type options dump pass 1 cgroup2 rw,seclabel,nosuid,nodev,noexec,relatime,nsdelegate 0 0 - getCGroups(): hierarchy_id controller path 1 0 /user.slice/user-34002.slice - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): 2 - availableCores(which = 'all'): system /proc/self/status cgroups2.cpu.max 96 96 2 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 system /proc/self/status cgroups2.cpu.max 96 96 2 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 CGroups for system 'dt1' ... done CGroups for system 'log1' ... - Using temporary folder: 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f4160f3996' - UID: 34002 - procPath(): 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f4160f3996/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup2 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4160f3996/sys/fs/cgroup cgroup2 rw,seclabel,nosuid,nodev,noexec,relatime,nsdelegate 0 0 - getCGroupsVersion(): 2 - getCGroupsMounts(): device 1 cgroup2 mountpoint 1 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f4160f3996/sys/fs/cgroup type options dump pass 1 cgroup2 rw,seclabel,nosuid,nodev,noexec,relatime,nsdelegate 0 0 - getCGroups(): hierarchy_id controller path 1 0 /user.slice/user-34002.slice - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): 1 - availableCores(which = 'all'): system /proc/self/status cgroups2.cpu.max 96 96 1 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 system /proc/self/status cgroups2.cpu.max 96 96 1 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 CGroups for system 'log1' ... done CGroups for system 'pi-2021' ... - Using temporary folder: 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f438405b4' - UID: 1000 - procPath(): 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f438405b4/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup2 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f438405b4/sys/fs/cgroup cgroup2 rw,nosuid,nodev,noexec,relatime,nsdelegate,memory_recursiveprot 0 0 - getCGroupsVersion(): 2 - getCGroupsMounts(): device 1 cgroup2 mountpoint 1 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f438405b4/sys/fs/cgroup type options dump 1 cgroup2 rw,nosuid,nodev,noexec,relatime,nsdelegate,memory_recursiveprot 0 pass 1 0 - getCGroups(): hierarchy_id controller path 1 0 /user.slice/user-1000.slice/session-1498.scope - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 CGroups for system 'pi-2021' ... done CGroups for system 'posit.cloud-free' ... - Using temporary folder: 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f471726094' - UID: 2025988 - procPath(): 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f471726094/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f471726094/sys/fs/cgroup cgroup2 ro,nosuid,nodev,noexec,relatime 0 0 - getCGroupsVersion(): 2 - getCGroupsMounts(): device 1 cgroup mountpoint 1 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f471726094/sys/fs/cgroup type options dump pass 1 cgroup2 ro,nosuid,nodev,noexec,relatime 0 0 - getCGroups(): hierarchy_id controller path 1 0 / - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): 1 - availableCores(which = 'all'): system /proc/self/status cgroups2.cpu.max 96 96 1 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 system /proc/self/status cgroups2.cpu.max 96 96 1 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 CGroups for system 'posit.cloud-free' ... done CGroups for system 'x1' ... - Using temporary folder: 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f43009f48' - UID: 1000 - procPath(): 'D:\temp\2025_01_30_16_40_16_30529\RtmpmAbHYA\file141f43009f48/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup2 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43009f48/sys/fs/cgroup cgroup2 rw,nosuid,nodev,noexec,relatime 0 0 - getCGroupsVersion(): 2 - getCGroupsMounts(): device 1 cgroup2 mountpoint 1 D:/temp/2025_01_30_16_40_16_30529/RtmpmAbHYA/file141f43009f48/sys/fs/cgroup type options dump pass 1 cgroup2 rw,nosuid,nodev,noexec,relatime 0 0 - getCGroups(): hierarchy_id controller 1 0 path 1 /user.slice/user-1000.slice/user@1000.service/app.slice/app-org.gnome.Terminal.slice/vte-spawn-2b165c43-93a4-4257-bef3-0c9d5a3b7f97.scope - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 CGroups for system 'x1' ... done cgroups2 - real-world ... done > > message("cgroups - real-world ... DONE") cgroups - real-world ... DONE > > source("incl/end.R") > > proc.time() user system elapsed 0.62 0.45 2.84