Is my ulimit exceeded

From Linuxintro
Revision as of 19:49, 16 April 2020 by imported>ThorstenStaerk

Under high load it happens from time to time that network connections fail although everything they work perfectly under low load. The reason be the ulimit set on a process. A ulimit restricts the process from opening up more files (and network connections) than a certain number.

To display you ulimit settings use the command ulimit: <source>

# ulimit -a
core file size          (blocks, -c) unlimited
data seg size           (kbytes, -d) unlimited
file size               (blocks, -f) unlimited
pending signals                 (-i) 32768
max locked memory       (kbytes, -l) 32
max memory size         (kbytes, -m) unlimited
open files                      (-n) 1024
pipe size            (512 bytes, -p) 8
POSIX message queues     (bytes, -q) 819200
stack size              (kbytes, -s) 8192
cpu time               (seconds, -t) unlimited
max user processes              (-u) 32768
virtual memory          (kbytes, -v) unlimited
file locks                      (-x) unlimited

</source>

You can permanently set the limits in /etc/security/limits.conf. You will have to re-login afterwards. To set the number of file descriptors for all users, the syntax is: <source>

*                hard    nofile           10000
*		 soft    nofile           10000

</source>

From my practice I can tell that the most prominent ulimit is -n, the count of open files allowed per process. I can also tell that the most interesting questions that are never answered in man pages are:

  • how is my ulimit for a given process?
  • how much of the ulimit is already used up?
  • have there been problem with the limit beeing set too small?

How is my ulimit for a given process?

Let's take firefox as an example: <source>

  1. ps -A | grep firefox

10975 ? 00:00:01 firefox

  1. cd /proc/10975/
  2. cat limits

Limit Soft Limit Hard Limit Units Max cpu time unlimited unlimited seconds Max file size unlimited unlimited bytes Max data size unlimited unlimited bytes Max stack size 8388608 unlimited bytes Max core file size 0 unlimited bytes Max resident set unlimited unlimited bytes Max processes 11848 11848 processes Max open files 1024 4096 files Max locked memory 65536 65536 bytes Max address space unlimited unlimited bytes Max file locks unlimited unlimited locks Max pending signals 11848 11848 signals Max msgqueue size 819200 819200 bytes Max nice priority 0 0 Max realtime priority 0 0 Max realtime timeout unlimited unlimited us </source>

How much of the ulimit is already used up?

Let's see for firefox: <source>

# ps -A | grep firefox
10975 ?        00:00:03 firefox
# cd /proc/10975/fd
# ls -1|wc -l
55

</source> ok, so firefox is consuming 55 of 1024 open file descriptors, much headroom left.

Test case

I wrote a simple C program that does nothing but open files:

main.c <source>

  1. include <stdio.h>

int main() {

 FILE *handle;
 for (int i=0;i<=2048;i++)
 {
   printf("%d",(fopen("testfile", "wb")==0));
 }
 while (true){};

} </source>

Compile this file with the command <source>

g++ main.c 

</source>

Run this program (and send it to the background) then with the command <source>

./a.out &

</source>

Now find out the process ID: <source>

# ps -A|grep a.out
29232 pts/3    00:04:15 a.out

</source>

Now go into the process' file descriptor directory: <source>

cd /proc/29232/fd

</source>

And count the number of files: <source>

# ls -1 | wc -l
1024

</source>

strace'ing it gives me: <source>

open("testfile", O_WRONLY|O_CREAT|O_TRUNC, 0666) = 1023
open("testfile", O_WRONLY|O_CREAT|O_TRUNC, 0666) = -1 EMFILE (Too many open files)

</source>