# Metrics¶

This section details how the resource usage from the Execution report are computed.

## CPU Usage¶

The plot reports how much CPU resources were used by each process.

Let's illustrate how this plot behaves with several examples.

In the first example, let's consider the simple use case in which a process performs one task of pure computation using one CPU. Then, you expect the Raw Usage tab to report 100%. If the task is distributed over, 2, 3, 4, etc. CPUs, then the Raw Usage will be 200%, 300%, 400%, etc. respectively. The % Allocated tab just rescales the raw value usage with respect to the number of CPUs set with the cpus directive (if not set with the directive, the number of CPUs is set to 1, thus showing the same values as in the Raw Usage tab). Using the program stress as follows would report 100% in the Raw Usage tab and 50% in the % Allocated tab since the process asked twice the number of CPUs needed by the process:

#!/usr/bin/env nextflow

process CpuUsageEx1 {

cpus 2

"""
stress -c 1 -t 10 # compute square-root of random numbers during 10s using 1 CPU
"""
}


In the second example, some time will be spent performing pure computation and some time just waiting. Using the program stress and sleep as follows would report 75% in the Raw Usage tab:

#!/usr/bin/env nextflow

process CpuUsageEx2 {

cpus 1

"""
stress -c 1 -t 10 # compute square-root of random numbers during 10s using 1 CPU
stress -c 1 -t 5 # compute square-root of random numbers during 5s using 1 CPU
sleep 5 # use no CPU during 5s
"""
}


Indeed, the percentage of the CPU that this process got is a weighted average taking into account the percentage of the CPU and duration of each individual program over the job duration (a.k.a. elapsed real time, real time or wall time ) as follows:

$\frac{ 100\% \times 10s + 100\% \times 5s + 0\% \times 5s }{10s+5s+5s} = 75\%$

The third example is similar to the second one except that the pure computation stage is performed in a single step forked on 2 CPUs:

#!/usr/bin/env nextflow

process CpuUsageEx3 {

cpus 2

"""
stress -c 2 -t 10 # compute square-root of random numbers during 10s using 2 CPUs
sleep 10 # use no CPU during 10s
"""
}


The Raw Usage tab would report 100% in the Raw Usage tab:

$\frac{ 200\% \times 10s }{10s+10s} = 100\%$

The % Allocated tab would report 50%, however, it would not be relevant to change the cpus directive from 2 to 1 as the process really uses 2 CPUs at it peak load.

Hint

The stress program can be installed with sudo apt-get install stress or sudo yum install stress depending on your linux distribution.

## Memory Usage¶

The plot has three tabs showing the usage of the physical memory (RAM), the virtual memory (vmem) and the percentage of RAM used by the process with respect to what was set in the memory directive. The peak usage during the execution of the process is reported for both physical and virtual memories.

Hint

To better understand the memory usage plot, it is important to know that:

• the total amount of memory used be a processs is the virtual memory (vmem). The vmem contains all memory areas whether they are in the physical memory (RAM), in the Swap space, on the disk or shared with other processes,
• the resident set size (RSS) is the amount of space of physical memory (RAM) held by a process,
• the relationship is: vmem $$\geq$$ RSS + Swap,
• the memory directive sets the RAM requested by the process.

Let's illustrate how this plot behaves with one example which relies on two C programs.

The first program just allocates a variable of 1 GiB:

  1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54  #include #include #include #include #include #include #include /* Get vmem and rss usage from /proc//statm */ static int mem_used(pid_t pid, unsigned long* vmem, unsigned long* rss) { FILE* file; char path[40]; unsigned int page_size; snprintf(path, 40, "/proc/%ld/statm", (long) pid); file = fopen(path, "r"); // vmem and rss are the first values in the file fscanf(file, "%lu %lu", vmem, rss); // values in statm are in pages so to get bytes we need to know page size page_size = (unsigned) getpagesize(); *vmem = *vmem * page_size; *rss = *rss * page_size; fclose(file); return 0; } int main(int argc, char **argv) { unsigned char *address; char input; size_t size = 1024*1024*1024; // 1 GiB unsigned long i; unsigned long vmem = 0; unsigned long rss = 0; pid_t pid; pid = getpid(); printf("Pid: %ld\n", (long) pid); mem_used(pid, &vmem, &rss); printf("VMEM: %lu RSS: %lu\n", vmem, rss); address = malloc(size); printf("Allocated %d Bytes of memory\n", (int) size); mem_used(pid, &vmem, &rss); printf("VMEM: %lu RSS: %lu\n", vmem, rss); // Leave time for nextflow to get information sleep(15); free(address); return 0; } 

The second program allocates a variable of 1 GiB and fills it with data:

  1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63  #include #include #include #include #include #include #include /* Get vmem and rss usage from /proc//statm */ static int mem_used(pid_t pid, unsigned long* vmem, unsigned long* rss) { FILE* file; char path[40]; unsigned int page_size; snprintf(path, 40, "/proc/%ld/statm", (long) pid); file = fopen(path, "r"); // vmem and rss are the first values in the file fscanf(file, "%lu %lu", vmem, rss); // values in statm are in pages so to get bytes we need to know page size page_size = (unsigned) getpagesize(); *vmem = *vmem * page_size; *rss = *rss * page_size; fclose(file); return 0; } int main(int argc, char **argv) { unsigned char *address; char input; size_t size = 1024*1024*1024; // 1 GiB unsigned long i; unsigned long vmem = 0; unsigned long rss = 0; pid_t pid; pid = getpid(); printf("Pid: %ld\n", (long) pid); mem_used(pid, &vmem, &rss); printf("VMEM: %lu RSS: %lu\n", vmem, rss); address = malloc(size); printf("Allocated %d Bytes of memory\n", (int) size); mem_used(pid, &vmem, &rss); printf("VMEM: %lu RSS: %lu\n", vmem, rss); printf("Filling memory with data..."); fflush(stdout); for (i = 0; i < size; i++) { *(address + i) = 123; } mem_used(pid, &vmem, &rss); printf("\nVMEM: %lu RSS: %lu\n", vmem, rss); // Leave time for nextflow to get information sleep(15); free(address); return 0; } 

The first and second programs are executed in foo and bar processes respectively as follows:

#!/usr/bin/env nextflow

process foo {

memory '1.5 GB'

"""
memory_vmem_1GiB_ram_0Gib
"""

}

process bar {

memory '1.5 GB'

"""
memory_vmem_1GiB_ram_1Gib
"""

}


The Virtual (RAM + Disk swap) tab shows that both foo and bar processes use the same amount of virtual memory (~1 GiB):

However, the Physical (RAM) tab shows that only the bar process uses ~1 GiB of RAM while foo process uses ~0 GiB:

As expected, the % RAM Allocated tab shows that 0% of the resource set in the memory directive was used for foo process while 67% (= 1 / 1.5) of the resource were used for bar process:

Warning

Binary unit are used to report memory raw values. This means that 1KB = $$1024$$ bytes, 1 MB = $$1024^2$$ bytes, 1 GB = $$1024^3$$ bytes, etc.

## Job Duration¶

The plot has two tabs the job duration (a.k.a. elapsed real time, real time or wall time ) in the Raw Usage tag and the percentage of requested time used in the % Allocated tab with respect to the duration set in the time directive of the process.

## I/O Usage¶

The plot has two tabs showing how many data were read and/or written each process. For example, the following processes read and write 1GB and 256MB of data respectively:

#!/usr/bin/env nextflow

"""
dd if=/dev/zero of=/dev/null bs=1G count=1
"""
}