R Under development (unstable) (2024-12-17 r87446 ucrt) -- "Unsuffered Consequences" Copyright (C) 2024 The R Foundation for Statistical Computing Platform: x86_64-w64-mingw32/x64 R is free software and comes with ABSOLUTELY NO WARRANTY. You are welcome to redistribute it under certain conditions. Type 'license()' or 'licence()' for distribution details. R is a collaborative project with many contributors. Type 'contributors()' for more information and 'citation()' on how to cite R or R packages in publications. Type 'demo()' for some demos, 'help()' for on-line help, or 'help.start()' for an HTML browser interface to help. Type 'q()' to quit R. > source("incl/start.R") > > message("*** cgroups ...") *** cgroups ... > > message("- getCGroups()") - getCGroups() > cgroups <- parallelly:::getCGroups() > print(cgroups) [1] hierarchy_id controller path <0 rows> (or 0-length row.names) > stopifnot( + is.data.frame(cgroups), + identical(colnames(cgroups), c("hierarchy_id", "controller", "path")), + nrow(cgroups) == 0L || !is.null(cgroups$controller) + ) > > message("- getCGroupsRoot()") - getCGroupsRoot() > root <- parallelly:::getCGroupsRoot() > cat(sprintf("cgroups root path: %s\n", sQuote(root))) cgroups root path: 'NA' > stopifnot(length(root) == 1L, is.character(root)) > > message("- getCGroups()") - getCGroups() > cgroups <- parallelly:::getCGroups() > print(cgroups) [1] hierarchy_id controller path <0 rows> (or 0-length row.names) > stopifnot(is.data.frame(cgroups)) > > message("- getCGroupsPath()") - getCGroupsPath() > path <- parallelly:::getCGroupsPath("cpu") > cat(sprintf("cgroups 'cpu' path: %s\n", sQuote(path))) cgroups 'cpu' path: 'NA' > stopifnot(length(path) == 1L, is.character(path)) > > path <- parallelly:::getCGroupsPath("cpuset") > cat(sprintf("cgroups 'cpuset' path: %s\n", sQuote(path))) cgroups 'cpuset' path: 'NA' > stopifnot(length(path) == 1L, is.character(path)) > > > message("- getCGroups1Value()") - getCGroups1Value() > value <- parallelly:::getCGroups1Value("cpu", "cpu.cfs_quota_us") > cat(sprintf("cgroups v1 'cpu.cfs_quota_us' value: %s\n", sQuote(value))) cgroups v1 'cpu.cfs_quota_us' value: 'NA' > stopifnot(length(value) == 1L, is.character(value)) > > value <- parallelly:::getCGroups1Value("cpu", "cpu.cfs_total_us") > cat(sprintf("cgroups v1 'cpu.cfs_total_us' value: %s\n", sQuote(value))) cgroups v1 'cpu.cfs_total_us' value: 'NA' > stopifnot(length(value) == 1L, is.character(value)) > > value <- parallelly:::getCGroups1Value("cpuset", "cpuset.cpus") > cat(sprintf("cgroups v1 'cpuset.cpus' value: %s\n", sQuote(value))) cgroups v1 'cpuset.cpus' value: 'NA' > stopifnot(length(value) == 1L, is.character(value)) > > > message("- getCGroups1CpuSet()") - getCGroups1CpuSet() > value <- parallelly:::getCGroups1CpuSet() > cat(sprintf("CPU set: [n=%d] %s\n", length(value), paste(sQuote(value), collapse = ", "))) CPU set: [n=0] > stopifnot(length(value) >= 0L, is.integer(value), !any(is.na(value))) > > > message("- getCGroups1CpuQuotaMicroseconds()") - getCGroups1CpuQuotaMicroseconds() > value <- parallelly:::getCGroups1CpuQuotaMicroseconds() > cat(sprintf("CPU quota (ms): %d\n", value)) CPU quota (ms): NA > stopifnot( + length(value) == 1L, + is.integer(value), + is.na(value) || value == -1 || value > 0 + ) > > message("- getCGroups1CpuPeriodMicroseconds()") - getCGroups1CpuPeriodMicroseconds() > value <- parallelly:::getCGroups1CpuPeriodMicroseconds() > cat(sprintf("CPU total (ms): %d\n", value)) CPU total (ms): NA > stopifnot( + length(value) == 1L, + is.integer(value), + is.na(value) || value > 0 + ) > > message("- getCGroups1CpuQuota()") - getCGroups1CpuQuota() > value <- parallelly:::getCGroups1CpuQuota() > cat(sprintf("CPU quota (ratio): %g\n", value)) CPU quota (ratio): NA > stopifnot( + length(value) == 1L, + is.numeric(value), + !is.infinite(value), + is.na(value) || value > 0 + ) > > message("- getCGroups2CpuMax()") - getCGroups2CpuMax() > value <- parallelly:::getCGroups2CpuMax() > cat(sprintf("CPU quota (ratio): %g\n", value)) CPU quota (ratio): NA > stopifnot( + length(value) == 1L, + is.numeric(value), + !is.infinite(value), + is.na(value) || value > 0 + ) > > message("*** cgroups ... DONE") *** cgroups ... DONE > > root <- system.file(package = "parallelly", "test-data", mustWork = TRUE) > for (dir in c("no-cgroups", "cgroups1", "cgroups2")) { + message(sprintf("%s - real-world ...", dir)) + path <- file.path(root, dir) + stopifnot(file_test("-d", path)) + tarballs <- dir(path = path, pattern = ".*[.]tar[.]gz", full.names = TRUE) + names(tarballs) <- sub("[.]tar[.]gz$", "", basename(tarballs)) + + for (name in names(tarballs)) { + parallelly:::withCGroups(tarballs[name], { + file <- file.path(path, sprintf("%s.R", name)) + if (file_test("-f", file)) { + source(file, local = FALSE) + } + }) + } + message(sprintf("%s - real-world ... done", dir)) + } no-cgroups - real-world ... CGroups for system 'no-cgroups' ... - Using temporary folder: 'D:\temp\2024_12_18_18_30_17_11197\RtmpK4AEGr\file828c1d924a9' - UID: 1000 - procPath(): 'D:\temp\2024_12_18_18_30_17_11197\RtmpK4AEGr\file828c1d924a9/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: tmpfs /run tmpfs rw,nosuid,nodev,noexec,relatime,size=3256456k,mode=755,inode64 0 0 - getCGroupsVersion(): -1 - getCGroupsMounts(): [1] device mountpoint type options dump pass <0 rows> (or 0-length row.names) - getCGroups(): [1] hierarchy_id controller path <0 rows> (or 0-length row.names) - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 CGroups for system 'no-cgroups' ... done no-cgroups - real-world ... done cgroups1 - real-world ... CGroups for system 'c4-dev3' ... - Using temporary folder: 'D:\temp\2024_12_18_18_30_17_11197\RtmpK4AEGr\file828c4b02c66' - UID: 581 - procPath(): 'D:\temp\2024_12_18_18_30_17_11197\RtmpK4AEGr\file828c4b02c66/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/systemd cgroup rw,nosuid,nodev,noexec,relatime,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd 0 0 02: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/rdma cgroup rw,nosuid,nodev,noexec,relatime,rdma 0 0 03: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/cpu,cpuacct cgroup rw,nosuid,nodev,noexec,relatime,cpu,cpuacct 0 0 04: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/freezer cgroup rw,nosuid,nodev,noexec,relatime,freezer 0 0 05: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/devices cgroup rw,nosuid,nodev,noexec,relatime,devices 0 0 06: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/net_cls,net_prio cgroup rw,nosuid,nodev,noexec,relatime,net_cls,net_prio 0 0 07: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/memory cgroup rw,nosuid,nodev,noexec,relatime,memory 0 0 08: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/pids cgroup rw,nosuid,nodev,noexec,relatime,pids 0 0 09: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/hugetlb cgroup rw,nosuid,nodev,noexec,relatime,hugetlb 0 0 10: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/cpuset cgroup rw,nosuid,nodev,noexec,relatime,cpuset 0 0 11: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/perf_event cgroup rw,nosuid,nodev,noexec,relatime,perf_event 0 0 12: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/blkio cgroup rw,nosuid,nodev,noexec,relatime,blkio 0 0 - getCGroupsVersion(): 1 - getCGroupsMounts(): device 1 cgroup 2 cgroup 3 cgroup 4 cgroup 5 cgroup 6 cgroup 7 cgroup 8 cgroup 9 cgroup 10 cgroup 11 cgroup 12 cgroup mountpoint 1 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/systemd 2 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/rdma 3 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/cpu,cpuacct 4 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/freezer 5 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/devices 6 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/net_cls,net_prio 7 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/memory 8 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/pids 9 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/hugetlb 10 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/cpuset 11 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/perf_event 12 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c4b02c66/sys/fs/cgroup/blkio type 1 cgroup 2 cgroup 3 cgroup 4 cgroup 5 cgroup 6 cgroup 7 cgroup 8 cgroup 9 cgroup 10 cgroup 11 cgroup 12 cgroup options 1 rw,nosuid,nodev,noexec,relatime,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd 2 rw,nosuid,nodev,noexec,relatime,rdma 3 rw,nosuid,nodev,noexec,relatime,cpu,cpuacct 4 rw,nosuid,nodev,noexec,relatime,freezer 5 rw,nosuid,nodev,noexec,relatime,devices 6 rw,nosuid,nodev,noexec,relatime,net_cls,net_prio 7 rw,nosuid,nodev,noexec,relatime,memory 8 rw,nosuid,nodev,noexec,relatime,pids 9 rw,nosuid,nodev,noexec,relatime,hugetlb 10 rw,nosuid,nodev,noexec,relatime,cpuset 11 rw,nosuid,nodev,noexec,relatime,perf_event 12 rw,nosuid,nodev,noexec,relatime,blkio dump pass 1 0 0 2 0 0 3 0 0 4 0 0 5 0 0 6 0 0 7 0 0 8 0 0 9 0 0 10 0 0 11 0 0 12 0 0 - getCGroups(): hierarchy_id controller path 12 1 name=systemd /user.slice/user-581.slice/session-9171.scope 11 2 rdma / 10 3 cpu / 101 3 cpuacct / 9 4 freezer / 8 5 devices /user.slice 7 6 net_cls / 71 6 net_prio / 6 7 memory /user.slice/user-581.slice/session-9171.scope 5 8 pids /user.slice/user-581.slice/session-9171.scope 4 9 hugetlb / 3 10 cpuset / 2 11 perf_event / 1 12 blkio /system.slice/sshd.service - length(getCGroups1CpuSet()): 40 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status cgroups.cpuset 96 96 40 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 system /proc/self/status cgroups.cpuset 96 96 40 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 CGroups for system 'c4-dev3' ... done CGroups for system 'c4-n39-4cores' ... - Using temporary folder: 'D:\temp\2024_12_18_18_30_17_11197\RtmpK4AEGr\file828c3a4f2c24' - UID: 581 - procPath(): 'D:\temp\2024_12_18_18_30_17_11197\RtmpK4AEGr\file828c3a4f2c24/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/systemd cgroup rw,nosuid,nodev,noexec,relatime,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd 0 0 02: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/net_cls,net_prio cgroup rw,nosuid,nodev,noexec,relatime,net_cls,net_prio 0 0 03: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/hugetlb cgroup rw,nosuid,nodev,noexec,relatime,hugetlb 0 0 04: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/blkio cgroup rw,nosuid,nodev,noexec,relatime,blkio 0 0 05: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/freezer cgroup rw,nosuid,nodev,noexec,relatime,freezer 0 0 06: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/pids cgroup rw,nosuid,nodev,noexec,relatime,pids 0 0 07: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/memory cgroup rw,nosuid,nodev,noexec,relatime,memory 0 0 08: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/perf_event cgroup rw,nosuid,nodev,noexec,relatime,perf_event 0 0 09: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/cpu,cpuacct cgroup rw,nosuid,nodev,noexec,relatime,cpu,cpuacct 0 0 10: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/cpuset cgroup rw,nosuid,nodev,noexec,relatime,cpuset 0 0 11: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/devices cgroup rw,nosuid,nodev,noexec,relatime,devices 0 0 12: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/rdma cgroup rw,nosuid,nodev,noexec,relatime,rdma 0 0 - getCGroupsVersion(): 1 - getCGroupsMounts(): device 1 cgroup 2 cgroup 3 cgroup 4 cgroup 5 cgroup 6 cgroup 7 cgroup 8 cgroup 9 cgroup 10 cgroup 11 cgroup 12 cgroup mountpoint 1 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/systemd 2 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/net_cls,net_prio 3 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/hugetlb 4 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/blkio 5 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/freezer 6 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/pids 7 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/memory 8 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/perf_event 9 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/cpu,cpuacct 10 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/cpuset 11 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/devices 12 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c3a4f2c24/sys/fs/cgroup/rdma type 1 cgroup 2 cgroup 3 cgroup 4 cgroup 5 cgroup 6 cgroup 7 cgroup 8 cgroup 9 cgroup 10 cgroup 11 cgroup 12 cgroup options 1 rw,nosuid,nodev,noexec,relatime,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd 2 rw,nosuid,nodev,noexec,relatime,net_cls,net_prio 3 rw,nosuid,nodev,noexec,relatime,hugetlb 4 rw,nosuid,nodev,noexec,relatime,blkio 5 rw,nosuid,nodev,noexec,relatime,freezer 6 rw,nosuid,nodev,noexec,relatime,pids 7 rw,nosuid,nodev,noexec,relatime,memory 8 rw,nosuid,nodev,noexec,relatime,perf_event 9 rw,nosuid,nodev,noexec,relatime,cpu,cpuacct 10 rw,nosuid,nodev,noexec,relatime,cpuset 11 rw,nosuid,nodev,noexec,relatime,devices 12 rw,nosuid,nodev,noexec,relatime,rdma dump pass 1 0 0 2 0 0 3 0 0 4 0 0 5 0 0 6 0 0 7 0 0 8 0 0 9 0 0 10 0 0 11 0 0 12 0 0 - getCGroups(): hierarchy_id controller path 12 1 name=systemd /system.slice/slurmd.service 11 2 net_cls / 111 2 net_prio / 10 3 hugetlb / 9 4 blkio /system.slice/slurmd.service 8 5 freezer /slurm/uid_581/job_255208/step_0 7 6 pids /system.slice/slurmd.service 6 7 memory /slurm/uid_581/job_255208/step_0 5 8 perf_event / 4 9 cpu /system.slice/slurmd.service 41 9 cpuacct /system.slice/slurmd.service 3 10 cpuset /slurm/uid_581/job_255208/step_0 2 11 devices /slurm/uid_581/job_255208/step_0/task_0 1 12 rdma / - length(getCGroups1CpuSet()): 4 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status cgroups.cpuset 96 96 4 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 CGroups for system 'c4-n39-4cores' ... done CGroups for system 'posit-workbench' ... - Using temporary folder: 'D:\temp\2024_12_18_18_30_17_11197\RtmpK4AEGr\file828c43e73506' - UID: 6776184 - procPath(): 'D:\temp\2024_12_18_18_30_17_11197\RtmpK4AEGr\file828c43e73506/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/systemd cgroup ro,nosuid,nodev,noexec,relatime,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd 0 0 02: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/hugetlb cgroup ro,nosuid,nodev,noexec,relatime,hugetlb 0 0 03: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/cpu,cpuacct cgroup ro,nosuid,nodev,noexec,relatime,cpu,cpuacct 0 0 04: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/cpuset cgroup ro,nosuid,nodev,noexec,relatime,cpuset 0 0 05: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/perf_event cgroup ro,nosuid,nodev,noexec,relatime,perf_event 0 0 06: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/net_cls,net_prio cgroup ro,nosuid,nodev,noexec,relatime,net_cls,net_prio 0 0 07: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/memory cgroup ro,nosuid,nodev,noexec,relatime,memory 0 0 08: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/blkio cgroup ro,nosuid,nodev,noexec,relatime,blkio 0 0 09: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/freezer cgroup ro,nosuid,nodev,noexec,relatime,freezer 0 0 10: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/devices cgroup ro,nosuid,nodev,noexec,relatime,devices 0 0 11: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/pids cgroup ro,nosuid,nodev,noexec,relatime,pids 0 0 - getCGroupsVersion(): 1 - getCGroupsMounts(): device 1 cgroup 2 cgroup 3 cgroup 4 cgroup 5 cgroup 6 cgroup 7 cgroup 8 cgroup 9 cgroup 10 cgroup 11 cgroup mountpoint 1 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/systemd 2 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/hugetlb 3 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/cpu,cpuacct 4 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/cpuset 5 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/perf_event 6 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/net_cls,net_prio 7 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/memory 8 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/blkio 9 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/freezer 10 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/devices 11 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c43e73506/sys/fs/cgroup/pids type 1 cgroup 2 cgroup 3 cgroup 4 cgroup 5 cgroup 6 cgroup 7 cgroup 8 cgroup 9 cgroup 10 cgroup 11 cgroup options 1 ro,nosuid,nodev,noexec,relatime,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd 2 ro,nosuid,nodev,noexec,relatime,hugetlb 3 ro,nosuid,nodev,noexec,relatime,cpu,cpuacct 4 ro,nosuid,nodev,noexec,relatime,cpuset 5 ro,nosuid,nodev,noexec,relatime,perf_event 6 ro,nosuid,nodev,noexec,relatime,net_cls,net_prio 7 ro,nosuid,nodev,noexec,relatime,memory 8 ro,nosuid,nodev,noexec,relatime,blkio 9 ro,nosuid,nodev,noexec,relatime,freezer 10 ro,nosuid,nodev,noexec,relatime,devices 11 ro,nosuid,nodev,noexec,relatime,pids dump pass 1 0 0 2 0 0 3 0 0 4 0 0 5 0 0 6 0 0 7 0 0 8 0 0 9 0 0 10 0 0 11 0 0 - getCGroups(): hierarchy_id controller 11 1 name=systemd 10 2 hugetlb 9 3 cpu 91 3 cpuacct 8 4 cpuset 7 5 perf_event 6 6 net_cls 61 6 net_prio 5 7 memory 4 8 blkio 3 9 freezer 2 10 devices 1 11 pids path 11 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 10 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 9 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 91 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 8 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 7 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 6 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 61 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 5 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 4 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 3 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 2 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope 1 /kubepods.slice/kubepods-pod0f2ea08d_d925_452c_b489_a089fbd80d96.slice/cri-containerd-f7f17ada6ed1935b2590745f64ccd612de2ef85078ac37bc962a9669e1280c66.scope - length(getCGroups1CpuSet()): 32 - getCGroups1CpuQuota(): 2 - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status cgroups.cpuset 96 96 32 cgroups.cpuquota mc.cores _R_CHECK_LIMIT_CORES_ 2 2 2 system /proc/self/status cgroups.cpuset 96 96 32 cgroups.cpuquota mc.cores _R_CHECK_LIMIT_CORES_ 2 2 2 CGroups for system 'posit-workbench' ... done cgroups1 - real-world ... done cgroups2 - real-world ... CGroups for system 'dev1' ... - Using temporary folder: 'D:\temp\2024_12_18_18_30_17_11197\RtmpK4AEGr\file828c77df6e68' - UID: 34002 - procPath(): 'D:\temp\2024_12_18_18_30_17_11197\RtmpK4AEGr\file828c77df6e68/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup2 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c77df6e68/sys/fs/cgroup cgroup2 rw,seclabel,nosuid,nodev,noexec,relatime,nsdelegate 0 0 - getCGroupsVersion(): 2 - getCGroupsMounts(): device 1 cgroup2 mountpoint 1 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c77df6e68/sys/fs/cgroup type options dump pass 1 cgroup2 rw,seclabel,nosuid,nodev,noexec,relatime,nsdelegate 0 0 - getCGroups(): hierarchy_id controller path 1 0 /user.slice/user-34002.slice/session-139474.scope - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 CGroups for system 'dev1' ... done CGroups for system 'dt1' ... - Using temporary folder: 'D:\temp\2024_12_18_18_30_17_11197\RtmpK4AEGr\file828c886502d' - UID: 34002 - procPath(): 'D:\temp\2024_12_18_18_30_17_11197\RtmpK4AEGr\file828c886502d/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup2 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c886502d/sys/fs/cgroup cgroup2 rw,seclabel,nosuid,nodev,noexec,relatime,nsdelegate 0 0 - getCGroupsVersion(): 2 - getCGroupsMounts(): device 1 cgroup2 mountpoint 1 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c886502d/sys/fs/cgroup type options dump pass 1 cgroup2 rw,seclabel,nosuid,nodev,noexec,relatime,nsdelegate 0 0 - getCGroups(): hierarchy_id controller path 1 0 /user.slice/user-34002.slice - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): 2 - availableCores(which = 'all'): system /proc/self/status cgroups2.cpu.max 96 96 2 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 system /proc/self/status cgroups2.cpu.max 96 96 2 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 CGroups for system 'dt1' ... done CGroups for system 'log1' ... - Using temporary folder: 'D:\temp\2024_12_18_18_30_17_11197\RtmpK4AEGr\file828c49a474af' - UID: 34002 - procPath(): 'D:\temp\2024_12_18_18_30_17_11197\RtmpK4AEGr\file828c49a474af/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup2 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c49a474af/sys/fs/cgroup cgroup2 rw,seclabel,nosuid,nodev,noexec,relatime,nsdelegate 0 0 - getCGroupsVersion(): 2 - getCGroupsMounts(): device 1 cgroup2 mountpoint 1 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c49a474af/sys/fs/cgroup type options dump pass 1 cgroup2 rw,seclabel,nosuid,nodev,noexec,relatime,nsdelegate 0 0 - getCGroups(): hierarchy_id controller path 1 0 /user.slice/user-34002.slice - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): 1 - availableCores(which = 'all'): system /proc/self/status cgroups2.cpu.max 96 96 1 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 system /proc/self/status cgroups2.cpu.max 96 96 1 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 CGroups for system 'log1' ... done CGroups for system 'pi-2021' ... - Using temporary folder: 'D:\temp\2024_12_18_18_30_17_11197\RtmpK4AEGr\file828c532c670b' - UID: 1000 - procPath(): 'D:\temp\2024_12_18_18_30_17_11197\RtmpK4AEGr\file828c532c670b/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup2 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c532c670b/sys/fs/cgroup cgroup2 rw,nosuid,nodev,noexec,relatime,nsdelegate,memory_recursiveprot 0 0 - getCGroupsVersion(): 2 - getCGroupsMounts(): device 1 cgroup2 mountpoint 1 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c532c670b/sys/fs/cgroup type options dump 1 cgroup2 rw,nosuid,nodev,noexec,relatime,nsdelegate,memory_recursiveprot 0 pass 1 0 - getCGroups(): hierarchy_id controller path 1 0 /user.slice/user-1000.slice/session-1498.scope - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 CGroups for system 'pi-2021' ... done CGroups for system 'posit.cloud-free' ... - Using temporary folder: 'D:\temp\2024_12_18_18_30_17_11197\RtmpK4AEGr\file828c72db38d6' - UID: 2025988 - procPath(): 'D:\temp\2024_12_18_18_30_17_11197\RtmpK4AEGr\file828c72db38d6/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c72db38d6/sys/fs/cgroup cgroup2 ro,nosuid,nodev,noexec,relatime 0 0 - getCGroupsVersion(): 2 - getCGroupsMounts(): device 1 cgroup mountpoint 1 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c72db38d6/sys/fs/cgroup type options dump pass 1 cgroup2 ro,nosuid,nodev,noexec,relatime 0 0 - getCGroups(): hierarchy_id controller path 1 0 / - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): 1 - availableCores(which = 'all'): system /proc/self/status cgroups2.cpu.max 96 96 1 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 system /proc/self/status cgroups2.cpu.max 96 96 1 mc.cores _R_CHECK_LIMIT_CORES_ 2 2 CGroups for system 'posit.cloud-free' ... done CGroups for system 'x1' ... - Using temporary folder: 'D:\temp\2024_12_18_18_30_17_11197\RtmpK4AEGr\file828c2b924286' - UID: 1000 - procPath(): 'D:\temp\2024_12_18_18_30_17_11197\RtmpK4AEGr\file828c2b924286/proc' - maxCores(): Inf - Adjust /proc/self/mounts accordingly: 01: cgroup2 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c2b924286/sys/fs/cgroup cgroup2 rw,nosuid,nodev,noexec,relatime 0 0 - getCGroupsVersion(): 2 - getCGroupsMounts(): device 1 cgroup2 mountpoint 1 D:/temp/2024_12_18_18_30_17_11197/RtmpK4AEGr/file828c2b924286/sys/fs/cgroup type options dump pass 1 cgroup2 rw,nosuid,nodev,noexec,relatime 0 0 - getCGroups(): hierarchy_id controller 1 0 path 1 /user.slice/user-1000.slice/user@1000.service/app.slice/app-org.gnome.Terminal.slice/vte-spawn-2b165c43-93a4-4257-bef3-0c9d5a3b7f97.scope - length(getCGroups1CpuSet()): 0 - getCGroups1CpuQuota(): NA - getCGroups2CpuMax(): NA - availableCores(which = 'all'): system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 system /proc/self/status mc.cores 96 96 2 _R_CHECK_LIMIT_CORES_ 2 CGroups for system 'x1' ... done cgroups2 - real-world ... done > > message("cgroups - real-world ... DONE") cgroups - real-world ... DONE > > source("incl/end.R") > > proc.time() user system elapsed 0.50 0.37 2.25