[LU-17323] fork() leaks ERESTARTNOINTR (errno 513) to user application Created: 29/Nov/23 Updated: 31/Jan/24 |
|
| Status: | Open |
| Project: | Lustre |
| Component/s: | None |
| Affects Version/s: | Lustre 2.11.0, Lustre 2.12.5, Lustre 2.12.6, Lustre 2.12.9 |
| Fix Version/s: | None |
| Type: | Bug | Priority: | Critical |
| Reporter: | Mike D | Assignee: | WC Triage |
| Resolution: | Unresolved | Votes: | 0 |
| Labels: | None | ||
| Environment: |
RHEL6, RHEL7/CentOS7 (various kernels) |
||
| Attachments: |
|
||||||||
| Issue Links: |
|
||||||||
| Severity: | 3 | ||||||||
| Rank (Obsolete): | 9223372036854775807 | ||||||||
| Description |
|
When using file locks on a Lustre mount with the 'flock' mount option, fork() The fork() cleanup code calls exit_files() which calls Lustre code. I'm not It seems there has to be multiple threads involved. My reproducer has two The problem doesn't reproduce with the 'localflock' mount option, so we We've seen this on RHEL6, RHEL7/CentOS7 kernels, Steps to reproduce: 1) Lustre mount must be using 'flock' mount option. Problem usually reproduces within 5-60 seconds. > touch /lustre_mnt/testfile.txt Use POSIX style read lock Use BSD style read lock Don't lock at all (this won't reproduce and will run indefinitely) NOTE: be aware the reproducer can exhaust your maxprocs limit |
| Comments |
| Comment by Patrick Farrell [ 29/Nov/23 ] |
|
Hi Mike, I think I know the issue you're hitting here and its root cause. Historically, Lustre had to do some nasty things with signal handling due to the lack of the ability to express "waiting" without contributing to load, and we also rolled our own for some waiting primitives that didn't exist at the time. This results in some weird behavior with certain signals in some cases. I saw this with ptrace, but this problem has a very similar feel to it. Neil Brown did a thorough rework of signal handling and task waiting in Lustre, spread over a number of patches (If it were just one, I would link it), which I believe landed for 2.13 but I don't think was ported to 2.12 (it was seen as code cleanup rather than fixing specific bugs). (I think your not hitting the problem with 2.12 is probably a coincidence/timing change.) 2.15 is the current maintenance release, so it would be good to see if you can reproduce this with 2.15, which has the full bevy of wait and signal handling changes. |
| Comment by Andreas Dilger [ 29/Nov/23 ] |
|
Mike, thank you for the detailed analysis (including a reproducer!). Since RHEL6/7 are basically EOL at this point, this issue would only be of interest if the problem persists in RHEL8/9 since we've run the full lifetime of EL6/7 without hitting this problem in actual production usage (or at least nothing has been reported to us up to this point). I don't see ERESTARTNOINTR used or returned anywhere in the Lustre code, so this error code is definitely coming from the kernel fork() handling. There is indeed code in libcfs/include/libcfs/linux/linux-wait.h that is clearing TIF_SIGPENDING in the RPC completion wait routines (__wait_event_idle() or __wait_event_lifo(), which are conditionally used depending on the kernel version in use. I suspect these routines are clones of similar code from newer kernels just for compatibility use with older kernels, so there may be some variations. If the problem still persists with newer kernels and Lustre releases then it would be useful to continue investigation and add the repro.c test case into our regression test suite. |
| Comment by Patrick Farrell [ 29/Nov/23 ] |
|
Interesting that the signal clearing is in those macros. Those are the ones neilb ported in to Lustre to replace our hand rolled stuff, so maybe the issue isn't fixed. Or perhaps Neil knows - should've tagged him earlier. |
| Comment by Mike D [ 04/Dec/23 ] |
|
I plan to try Lustre 2.15 client (assuming that will connect to the 2.12.x server) but it will probably be several weeks before I can try it and report back. I don't know if the problem occurs with RHEL8/9 as I don't have an easy way to test that. |
| Comment by Mike D [ 15/Dec/23 ] |
|
I tried the latest Lustre 2.15 client and have not been able to reproduce the issue on CentOS7. However, I did notice a problem (I haven't investigated it much yet): > gcc hello.c > ./a.out ./a.out: Command not found. > /bin/ls a.out a.out > ./a.out hello world
The file isn't there until I do the ls. This is reproducible every time.
Is it recommended to use a Lustre 2.15 client with 2.12.x servers? |
| Comment by Andreas Dilger [ 15/Dec/23 ] |
|
Mike, please file your gcc issue in a separate Jira ticket, or it will be lost here. There should be proper interop between 2.15 clients and 2.12 servers. |
| Comment by Patrick Farrell [ 15/Dec/23 ] |
|
Also, since it's easy to reproduce, but doesn't happen in our usual test environments, could you grab client debug logs and attach them to the new LU? Something like this would do the trick: DEBUGMB=`lctl get_param -n debug_mb` lctl clear lctl set_param *debug=-1 debug_mb=10000 lctl mark "running gcc" gcc hello.c lctl mark "running a.out" ./a.out lctl mark "running ls" /bin/ls a.out lctl mark "running a.out again" ./a.out # Write out logs lctl dk > /tmp/log # Set debug to minimum, this leaves on error, warning, etc lctl set_param debug=0 lctl set_param debug_mb=$DEBUGMB (Will probably want to compress that before attaching it)
|
| Comment by Mike D [ 08/Jan/24 ] |
|
I'll note that csh vs bash behavior is different.
$ gcc hello.c $ ./a.out ./a.out: Command not found $ ./a.out $ ./a.out ./a.out: Command not found $ ls a.out a.out $ ./a.out hello world
$ gcc hello.c $ ./a.out bash: ./a.out: /lib64/ld-linux-x86-64.so.2: bad ELF interpreter: No such file or directory $ ./a.out hello world
The main thing I've noticed in the client debug log: running a.out file.c:2012:ll_file_read_iter() file a.out:[0x2000XXXXX:0x9:0x0], ppos: 0, count: 80 file.c:2012:ll_file_read_iter() file a.out:[0x2000XXXXX:0x9:0x0], ppos: 6456, count: 1984 file.c:2012:ll_file_read_iter() file a.out:[0x2000XXXXX:0x9:0x0], ppos: 6186, count: 268 file.c:2012:ll_file_read_iter() file a.out:[0x2000XXXXX:0x9:0x0], ppos: 64, count: 504 file.c:2012:ll_file_read_iter() file a.out:[0x2000XXXXX:0x9:0x0], ppos: 568, count: 200 running a.out again file.c:2012:ll_file_read_iter() file a.out:[0x2000XXXXX:0x9:0x0], ppos: 0, count: 128 file.c:2012:ll_file_read_iter() file a.out:[0x2000XXXXX:0x9:0x0], ppos: 64, count: 504 file.c:2012:ll_file_read_iter() file a.out:[0x2000XXXXX:0x9:0x0], ppos: 568, count: 28
I may be able to get a full log if needed but I don't know when I'll have time to get back to this. I'll look at entering a new ticket with some of this info. Thanks for the help so far. |
| Comment by Mike D [ 31/Jan/24 ] |
|
Linked LU-17405 which tracks the side issue uncovered with gcc when testing the Lustre 2.15 client with 2.12.x servers. |