Package: nfs-common Version: 1:1.3.4-2.5 OS Release: Buster Dear all:
Since september 2019, the rpc.gssd user space daemon on the NFSv4 file servers (VMware ESXi virtualized hosts) of my department provokes random segmentation violations. Security flavour of NFS exports is set to sec=krb5p. Some monthes back, all NFS server were still running on Debian Stretch. I am about to migrate all "my" NFS file servers to Debian Buster. Unfortunately, the problem persists with Debian Buster using the most recent versions of nfs-common package and Linux Kernel. I now managed to get a backtrace of a recent segfault incident on Debian Buster. Here is the full backtrace: root@server:~# coredumpctl debug PID: 6356 (rpc.gssd) UID: 0 (root) GID: 0 (root) Signal: 11 (SEGV) Timestamp: Thu 2020-06-25 11:46:08 CEST (21h ago) Command Line: /usr/sbin/rpc.gssd -vvvvvvv -rrrrrrr -t 3600 -T 10 Executable: /usr/sbin/rpc.gssd Control Group: /system.slice/rpc-gssd.service Unit: rpc-gssd.service Slice: system.slice Boot ID: (obfuscated) Machine ID: (obfuscated) Hostname: all Storage: /var/lib/systemd/coredump/core.rpc\x2egssd.0.7f31136228274af0a1a855b91ad1e75c.6356.1593078368000000.lz4 Message: Process 6356 (rpc.gssd) of user 0 dumped core. Stack trace of thread 14174: #0 0x000056233fff038e n/a (rpc.gssd) #1 0x000056233fff09f8 n/a (rpc.gssd) #2 0x000056233fff0b92 n/a (rpc.gssd) #3 0x000056233fff13b3 n/a (rpc.gssd) #4 0x00007fb2eb8dbfa3 start_thread (libpthread.so.0) #5 0x00007fb2eb80c4cf __clone (libc.so.6) Stack trace of thread 6356: #0 0x00007fb2eb801819 __GI___poll (libc.so.6) #1 0x00007fb2eb6e7207 send_dg (libresolv.so.2) #2 0x00007fb2eb6e4c43 __GI___res_context_query (libresolv.so.2) #3 0x00007fb2eb6bf536 __GI__nss_dns_gethostbyaddr2_r (libnss_dns.so.2) #4 0x00007fb2eb6bf823 _nss_dns_gethostbyaddr_r (libnss_dns.so.2) #5 0x00007fb2eb81dee2 __gethostbyaddr_r (libc.so.6) #6 0x00007fb2eb8267d5 gni_host_inet_name (libc.so.6) #7 0x000056233ffef455 n/a (rpc.gssd) #8 0x000056233ffef82c n/a (rpc.gssd) #9 0x000056233fff01d0 n/a (rpc.gssd) #10 0x00007fb2ebab49ba n/a (libevent-2.1.so.6) #11 0x00007fb2ebab5537 event_base_loop (libevent-2.1.so.6) #12 0x000056233ffedeaa n/a (rpc.gssd) #13 0x00007fb2eb73709b __libc_start_main (libc.so.6) #14 0x000056233ffee03a n/a (rpc.gssd) GNU gdb (Debian 8.2.1-2+b3) 8.2.1 Copyright (C) 2018 Free Software Foundation, Inc. License GPLv3+: GNU GPL version 3 or later <http://gnu.org/licenses/gpl.html> This is free software: you are free to change and redistribute it. There is NO WARRANTY, to the extent permitted by law. Type "show copying" and "show warranty" for details. This GDB was configured as "x86_64-linux-gnu". Type "show configuration" for configuration details. For bug reporting instructions, please see: <http://www.gnu.org/software/gdb/bugs/>. Find the GDB manual and other documentation resources online at: <http://www.gnu.org/software/gdb/documentation/>. For help, type "help". Type "apropos word" to search for commands related to "word"... Reading symbols from /usr/sbin/rpc.gssd...Reading symbols from /usr/lib/debug/.build-id/08/a9957ac98e4e5a68f9238c4d763a95e9b4d492.debug...done. done. [New LWP 14174] [New LWP 6356] [Thread debugging using libthread_db enabled] Using host libthread_db library "/lib/x86_64-linux-gnu/libthread_db.so.1". Core was generated by `/usr/sbin/rpc.gssd -vvvvvvv -rrrrrrr -t 3600 -T 10'. Program terminated with signal SIGSEGV, Segmentation fault. #0 0x000056233fff038e in create_auth_rpc_client (clp=clp@entry=0x562341008fa0, tgtname=tgtname@entry=0x562341011c8f "h...@client.domain.tu-berlin.de", clnt_return=clnt_return@entry=0x7fb2eaeb9de8, auth_return=auth_return@entry=0x7fb2eaeb9d50, uid=uid@entry=0, cred=cred@entry=0x0, authtype=0) at gssd_proc.c:352 352 gssd_proc.c: No such file or directory. [Current thread is 1 (Thread 0x7fb2eaeba700 (LWP 14174))] (gdb) bt full #0 0x000056233fff038e in create_auth_rpc_client (clp=clp@entry=0x562341008fa0, tgtname=tgtname@entry=0x562341011c8f "h...@client.domain.tu-berlin.de", clnt_return=clnt_return@entry=0x7fb2eaeb9de8, auth_return=auth_return@entry=0x7fb2eaeb9d50, uid=uid@entry=0, cred=cred@entry=0x0, authtype=0) at gssd_proc.c:352 rpc_clnt = 0x0 sec = {mech = 0x56233fffc590 <krb5oid>, qop = 0, svc = RPCSEC_GSS_SVC_NONE, cred = 0x7fb2dc000d60, req_flags = 2} auth = 0x0 retval = -1 min_stat = 256 rpc_errmsg = '\000' <repeats 13 times>, "\001\000\000c\343\000\333@'\243\001\214\002\236ROO\257\303\351\267]<\033\236\250̺ \215\210o=o\316_]\v\266\301\207\307\313\377S\236\353\262\177\000\000c\301\207-\204\\\223\226\024\320E\347\025\307\021`0\314\340B\020'B\370\252\300\327Ɣ\000b3\341\064\r\320\315ۻ\026z\261\035@|⌁\023s\322&e\365\377S\236\353\262\177\000\000\347\351\233\022F\210[\274}!\225\252\213\260\377\343\016\267>$(\261:\313<ޖ\322F\314ᯨi\367\032\255\306\306Ja\203\355&\253\000\064\271M\023\026\243\266c\334\313\000\242\066p\361K\361$\320\357-\366\026)\240\063\320\f\000ܲ\177\000\000\260k"... protocol = 6 timeout = {tv_sec = -418240567167153622, tv_usec = 6354052107285749079} addr = 0x5623410090f8 salen = <optimized out> #1 0x000056233fff09f8 in krb5_use_machine_creds (clp=clp@entry=0x562341008fa0, uid=uid@entry=0, tgtname=tgtname@entry=0x562341011c8f "h...@client.domain.tu-berlin.de", service=service@entry=0x562341011cb5 "nfs", rpc_clnt=rpc_clnt@entry=0x7fb2eaeb9de8) at gssd_proc.c:558 min_stat = 0 auth = 0x0 credlist = 0x7fb2dc02a910 ccname = 0x7fb2dc02a910 nocache = <optimized out> success = 0 #2 0x000056233fff0b92 in process_krb5_upcall (clp=clp@entry=0x562341008fa0, uid=uid@entry=0, fd=11, tgtname=tgtname@entry=0x562341011c8f "h...@client.domain.tu-berlin.de", service=service@entry=0x562341011cb5 "nfs") at gssd_proc.c:646 rpc_clnt = 0x0 auth = <optimized out> pd = {pd_ctx = 0x0, pd_ctx_hndl = {length = 0, value = 0x0}, pd_seq_win = 0} token = {length = 0, value = 0x0} err = 1090591864 downcall_err = -13 maj_stat = <optimized out> min_stat = 1090560944 lifetime_rec = 22051 gacceptor = 0x0 mech = 0x7fb2dc000020 acceptor = {length = 0, value = 0x0} #3 0x000056233fff13b3 in handle_gssd_upcall (info=0x562341011c70) at gssd_proc.c:805 clp = 0x562341008fa0 uid = 0 p = 0x562341011c87 "" mech = <optimized out> uidstr = <optimized out> target = 0x562341011c8f "h...@client.domain.tu-berlin.de" service = 0x562341011cb5 "nfs" enctypes = 0x562341011cc2 "18,17,16,23,3,1,2" upcall_str = 0x7fb2dc001000 "mech=krb5 uid=0 target=h...@client.domain.tu-berlin.de service=nfs enctypes=18,17,16,23,3,1,2 " pbuf = 0x0 __func__ = "handle_gssd_upcall" #4 0x00007fb2eb8dbfa3 in start_thread (arg=<optimized out>) at pthread_create.c:486 ret = <optimized out> pd = <optimized out> now = <optimized out> unwind_buf = {cancel_jmp_buf = {{jmp_buf = {140406422218496, 1116452640567996153, 140736109338078, 140736109338079, 140406422218496, 94709414435248, -1145117077261373703, -1145118435554349319}, mask_was_saved = 0}}, priv = {pad = {0x0, 0x0, 0x0, 0x0}, data = {prev = 0x0, cleanup = 0x0, canceltype = 0}}} not_first_call = <optimized out> #5 0x00007fb2eb80c4cf in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:95 No locals. (gdb) thread apply all bt full Thread 2 (Thread 0x7fb2eb6d9740 (LWP 6356)): #0 0x00007fb2eb801819 in __GI___poll (fds=fds@entry=0x7fffadcdb658, nfds=nfds@entry=1, timeout=5000, timeout@entry=<error reading variable: That operation is not available on integers of more than 8 bytes.>) at ../sysdeps/unix/sysv/linux/poll.c:29 resultvar = 18446744073709551100 sc_cancel_oldtype = 0 sc_ret = <optimized out> #1 0x00007fb2eb6e7207 in send_dg (ansp2_malloced=<optimized out>, resplen2=<optimized out>, anssizp2=<optimized out>, ansp2=<optimized out>, anscp=<optimized out>, gotsomewhere=<synthetic pointer>, v_circuit=<synthetic pointer>, ns=<optimized out>, terrno=0x7fffadcdb618, anssizp=0x7fffadcdb750, ansp=0x7fffadcdb608, buflen2=<optimized out>, buf2=<optimized out>, buflen=<optimized out>, buf=<optimized out>, statp=<optimized out>) at res_send.c:1112 hp = <optimized out> from = {sin6_family = 6512, sin6_port = 16641, sin6_flowinfo = 22051, sin6_addr = {__in6_u = {__u6_addr8 = "\353Hn\353\262\177\000\000h\r\000\000\000\000\000", __u6_addr16 = {18667, 60270, 32690, 0, 3432, 0, 0, 0}, __u6_addr32 = {3949873387, 32690, 3432, 0}}}, sin6_scope_id = 3951866528} seconds = <optimized out> retval = <optimized out> recvresp1 = <optimized out> n = <optimized out> single_request_reopen = <optimized out> recvresp2 = <optimized out> timeout = {tv_sec = 5, tv_nsec = 0} save_gotsomewhere = <optimized out> hp2 = <optimized out> finish = {tv_sec = <optimized out>, tv_nsec = 56066000} pfd = {{fd = 25, events = 1, revents = 0}} resplen = 0 single_request = <optimized out> now = {tv_sec = 1593078368, tv_nsec = 56066000} ptimeout = <optimized out> need_recompute = 0 nwritten = <optimized out> hp = <optimized out> hp2 = <optimized out> now = <optimized out> timeout = <optimized out> finish = <optimized out> pfd = <optimized out> ptimeout = <optimized out> from = <optimized out> resplen = <optimized out> n = <optimized out> seconds = <optimized out> single_request_reopen = <optimized out> single_request = <optimized out> save_gotsomewhere = <optimized out> retval = <optimized out> need_recompute = <optimized out> nwritten = <optimized out> recvresp1 = <optimized out> recvresp2 = <optimized out> __PRETTY_FUNCTION__ = "send_dg" iov = <optimized out> iov2 = <optimized out> reqs = <optimized out> ndg = <optimized out> sr = <optimized out> thisanssizp = <optimized out> thisansp = <optimized out> thisresplenp = <optimized out> anhp = <optimized out> fromlen = <optimized out> newp = <optimized out> #2 __res_context_send (ctx=ctx@entry=0x562341011970, buf=buf@entry=0x7fffadcdb780 "\222M\001", buflen=buflen@entry=46, buf2=buf2@entry=0x0, buflen2=buflen2@entry=0, ans=<optimized out>, ans@entry=0x7fffadcdb980 "d\253\205\200", anssiz=<optimized out>, ansp=<optimized out>, ansp2=<optimized out>, nansp2=<optimized out>, resplen2=<optimized out>, ansp2_malloced=<optimized out>) at res_send.c:519 ns = <optimized out> ns_shift = <optimized out> statp = <optimized out> gotsomewhere = <optimized out> terrno = 110 try = <optimized out> v_circuit = 0 resplen = <optimized out> n = <optimized out> ns_offset = <optimized out> #3 0x00007fb2eb6e4c43 in __GI___res_context_query (ctx=ctx@entry=0x562341011970, name=name@entry=0x7fffadcdbdf0 "135.115.149.130.in-addr.arpa", class=class@entry=1, type=type@entry=12, answer=0x7fffadcdb980 "d\253\205\200", anslen=anslen@entry=1024, answerp=0x7fffadcdbde8, answerp2=0x0, nanswerp2=0x0, resplen2=0x0, answerp2_malloced=0x0) at res_query.c:216 statp = 0x7fb2eb8d2900 <_res> hp = 0x7fffadcdb980 hp2 = <optimized out> n = <optimized out> use_malloc = 0 bufsize = <optimized out> buf = 0x7fffadcdb780 "\222M\001" query1 = 0x7fffadcdb780 "\222M\001" --Type <RET> for more, q to quit, c to continue without paging-- nquery1 = <optimized out> query2 = 0x0 nquery2 = 0 __PRETTY_FUNCTION__ = "__res_context_query" #4 0x00007fb2eb6bf536 in __GI__nss_dns_gethostbyaddr2_r (addr=<optimized out>, len=<optimized out>, af=<optimized out>, result=0x7fffadcdc390, buffer=0x7fffadcdc550 "\377\002", buflen=1024, errnop=0x7fb2eb6d96a0, h_errnop=0x7fffadcdc37c, ttlp=0x0) at nss_dns/dns-host.c:536 mapped = "\000\000\000\000\000\000\000\000\000\000\377\377" tunnelled = '\000' <repeats 11 times> v6local = "\000\000\000\001" uaddr = <optimized out> host_data = 0x7fffadcdc550 host_buffer = {buf = 0x7fffadcdb980, ptr = 0x7fffadcdb980 "d\253\205\200"} orig_host_buffer = 0x7fffadcdb980 qbuf = "135.115.149.130.in-addr.arpa", '\000' <repeats 188 times>... qp = <optimized out> size = <optimized out> n = <optimized out> status = <optimized out> olderr = 22 pad = 0 ctx = 0x562341011970 #5 0x00007fb2eb6bf823 in _nss_dns_gethostbyaddr_r (addr=<optimized out>, len=<optimized out>, af=<optimized out>, result=<optimized out>, buffer=<optimized out>, buflen=<optimized out>, errnop=0x7fb2eb6d96a0, h_errnop=0x7fffadcdc37c) at nss_dns/dns-host.c:576 No locals. #6 0x00007fb2eb81dee2 in __gethostbyaddr_r (addr=addr@entry=0x5623410090fc, len=len@entry=4, type=type@entry=2, resbuf=resbuf@entry=0x7fffadcdc390, buffer=<optimized out>, buflen=<optimized out>, result=<optimized out>, h_errnop=<optimized out>) at ../nss/getXXbyYY_r.c:315 startp_initialized = true startp = 0xa338ee3dbe5a2ef9 start_fct = 0xf01bb91bcc7a2ef9 nip = 0x562340ffdef0 do_merge = 0 mergegrp = <optimized out> mergebuf = 0x0 endptr = 0x0 fct = {l = 0x7fb2eb6bf810 <_nss_dns_gethostbyaddr_r>, ptr = 0x7fb2eb6bf810 <_nss_dns_gethostbyaddr_r>} no_more = 0 err = <optimized out> status = NSS_STATUS_UNAVAIL nscd_status = <optimized out> any_service = true res_ctx = <optimized out> res = <optimized out> #7 0x00007fb2eb8267d5 in gni_host_inet_name (addrlen=<optimized out>, flags=8, hostlen=1025, host=0x7fffadcdc9d0 "client.domain.tu-berlin.de", sa=0x5623410090f8, tmpbuf=0x7fffadcdc540) at ../include/scratch_buffer.h:101 sinp = <optimized out> herrno = 1 th = {h_name = 0x7fffadcdc578 "ip6-allrouters", h_aliases = 0x7fffadcdc5a0, h_addrtype = 2, h_length = 4, h_addr_list = 0x7fffadcdc560} h = 0x0 herrno = <optimized out> th = <optimized out> h = <optimized out> sin6p = <optimized out> sinp = <optimized out> c = <optimized out> do_idn = <optimized out> h_name = <optimized out> len = <optimized out> rc = <optimized out> #8 gni_host_inet (addrlen=<optimized out>, flags=8, hostlen=1025, host=0x7fffadcdc9d0 "client.domain.tu-berlin.de", sa=0x5623410090f8, tmpbuf=0x7fffadcdc540) at getnameinfo.c:367 result = <optimized out> #9 gni_host (addrlen=<optimized out>, flags=8, hostlen=1025, host=0x7fffadcdc9d0 "client.domain.tu-berlin.de", sa=0x5623410090f8, tmpbuf=0x7fffadcdc540) at getnameinfo.c:409 No locals. #10 __GI_getnameinfo (sa=0x5623410090f8, addrlen=<optimized out>, host=0x7fffadcdc9d0 "client.domain.tu-berlin.de", hostlen=1025, serv=0x0, servlen=0, flags=8) at getnameinfo.c:523 result = <optimized out> tmpbuf = {data = 0x7fffadcdc550, length = 1024, __space = {__align = {__max_align_ll = 767, __max_align_ld = 5.13011199855208963934e-4937}, __c = "\377\002", '\000' <repeats 13 times>, "\002P\305ͭ\377\177", '\000' <repeats 11 times>, "f02::2\000ip6-allrouters\000\000calhost\000ip6-loopback\000\000ap\202\305ͭ\377\177\000\000\220\305ͭ\377\177\000\000\000\000\000\000\000\000\000\000\064\306ͭ\377\177\000\000\260\310ͭ\377\177\000\000\200\035\211\353\262\177\000\000`\311ͭ\377\177\000\000\260\311ͭ\377\177\000\000P\316ͭ\377\177\000\000穂\353\262\177\000\000\000\315ͭ\377\177\000\000P\316ͭ\377\177\000\000\260\310ͭ\377\177\000\000\002", '\000' <repeats 15 times>, "\202W\177\353\262\177\000\000\000\000\000\000\000\000\000\000"...}} #11 0x000056233ffef455 in gssd_get_servername (name=<optimized out>, sa=0x5623410090f8, addr=0x562341010e50 "130.149.115.135") at gssd.c:225 addrlen = <optimized out> err = <optimized out> hbuf = "client.domain.tu-berlin.de\000\000\000lin.de\000\000\000\060\000\000\000\354\306\317\353\262\177\000\000\000\000\000\000\000\000\000\000\353\001\000\000\000\000\000\000\000U\217\353\262\177\000\000\001\000\000\000\000\000\000\000\200Rq\353\262\177\000\000\364\312\317\353\262\177\000\000\353\001\000\000\000\000\000\000\346\370|\353\262\177\000\000\000U\217\353\262\177\000\000\001\000\000\000\000\000\000\000\260\314ͭ\377\177\000\000@\273\214\353\262\177\000\000g\314ͭ\377\177\000\000\060\316\377@#V\000\000\356\037\000\000\000\000\000\000\241\254\376?#V\000\000\233\326\067\314\v\000\000\000\005\313ͭ\377\177\000\000\001\000\000\000\000\000\000\000\220\313ͭ"... buf = "\202\225s\207#V\000\000\000\000\000\000\000\000\000" #12 0x000056233ffef82c in gssd_read_service_info (clp=0x562341008fa0, dirfd=11) at gssd.c:326 server = 0x562340ffc5f0 "130.149.115.135" fd = <optimized out> info = 0x562340ffa790 numfields = <optimized out> servername = 0x0 service = 0x562341010d80 "nfs4_cb" program = 1073741824 version = 1 address = 0x562341010e50 "130.149.115.135" --Type <RET> for more, q to quit, c to continue without paging-- protoname = 0x562341011900 "tcp" port = 0x562341011860 "41205" fd = <optimized out> info = <optimized out> numfields = <optimized out> server = <optimized out> service = <optimized out> program = <optimized out> version = <optimized out> address = <optimized out> protoname = <optimized out> port = <optimized out> servername = <optimized out> #13 gssd_scan_clnt (clp=0x562341008fa0) at gssd.c:551 clntfd = 11 gssd_was_closed = <optimized out> krb5_was_closed = <optimized out> #14 0x000056233fff01d0 in gssd_create_clnt (name=0x7fffadcdced0 "clnt50f", tdi=0x562340ffe490) at gssd.c:568 clp = <optimized out> clp = <optimized out> #15 gssd_inotify_topdir (ev=0x7fffadcdcec0, tdi=0x562340ffe490) at gssd.c:709 No locals. #16 gssd_inotify_cb (ifd=8, UNUSED_which=<optimized out>, UNUSED_data=<optimized out>) at gssd.c:793 buf = "\005\000\000\000\000\001\000@\000\000\000\000\020\000\000\000clnt50f\000\000\000\000\000\000\000\000\000K\000\000\000\000\200\000\000\000\000\000\000\000\000\000\000gssd", '\000' <repeats 12 times>, "I\000\000\000\000\002\000\000\000\000\000\000\020\000\000\000info", '\000' <repeats 12 times>, "\005\000\000\000\000\001\000@\000\000\000\000\020\000\000\000clnt50d\000\000\000\000\000\000\000\000\000\070\070.167.149.130.in-addr.arpa\000\377\177\000\000\324\317ͭ\377\177\000\000\236\353\202\353\262\177\000\000\000\000\000\000\000\000\000\000\350\306k\353\262\177\000\000\370\302k\353\262\177\000\000 "... ev = 0x7fffadcdcec0 len = <optimized out> ptr = 0x7fffadcdcec0 "\005" rescan = <optimized out> tdi = 0x562340ffe490 clp = <optimized out> #17 0x00007fb2ebab49ba in event_persist_closure (ev=<optimized out>, base=0x562340ffd410) at event.c:1580 evcb_fd = <optimized out> evcb_callback = <optimized out> evcb_res = 2 evcb_arg = 0x0 evcb_callback = <optimized out> evcb_fd = <optimized out> evcb_res = <optimized out> evcb_arg = <optimized out> run_at = <optimized out> relative_to = <optimized out> delay = <optimized out> now = <optimized out> usec_mask = <optimized out> #18 event_process_active_single_queue (base=base@entry=0x562340ffd410, activeq=0x562340ffd050, max_to_process=max_to_process@entry=2147483647, endtime=endtime@entry=0x0) at event.c:1639 ev = <optimized out> evcb = <optimized out> count = 1 __func__ = "event_process_active_single_queue" #19 0x00007fb2ebab5537 in event_process_active (base=0x562340ffd410) at event.c:1738 activeq = <optimized out> i = 0 c = 0 tv = {tv_sec = 94709414344352, tv_usec = 94709397571904} maxcb = <optimized out> endtime = 0x0 limit_after_prio = 2147483647 activeq = <optimized out> i = <optimized out> c = <optimized out> endtime = <optimized out> tv = <optimized out> maxcb = <optimized out> limit_after_prio = <optimized out> #20 event_base_loop (base=0x562340ffd410, flags=<optimized out>) at event.c:1961 n = <optimized out> evsel = 0x7fb2ebce5dc0 <epollops> tv = {tv_sec = 94709397548095, tv_usec = 94709414362848} tv_p = <optimized out> res = <optimized out> done = 0 retval = 0 __func__ = "event_base_loop" #21 0x000056233ffedeaa in main (argc=<optimized out>, argv=0x7fffadcde1e8) at gssd.c:1006 fg = 0 verbosity = 7 rpc_verbosity = 7 opt = <optimized out> i = <optimized out> --Type <RET> for more, q to quit, c to continue without paging-- progname = <optimized out> ccachedir = <optimized out> sighup_ev = {ev_evcallback = {evcb_active_next = {tqe_next = 0x7fffadd76268, tqe_prev = 0x7fb2ebd1c730}, evcb_flags = 130, evcb_pri = 0 '\000', evcb_closure = 1 '\001', evcb_cb_union = {evcb_callback = 0x56233fff0260 <gssd_scan_cb>, evcb_selfcb = 0x56233fff0260 <gssd_scan_cb>, evcb_evfinalize = 0x56233fff0260 <gssd_scan_cb>, evcb_cbfinalize = 0x56233fff0260 <gssd_scan_cb>}, evcb_arg = 0x0}, ev_timeout_pos = {ev_next_with_common_timeout = {tqe_next = 0xffffffff, tqe_prev = 0x0}, min_heap_idx = -1}, ev_fd = 1, ev_base = 0x562340ffd410, ev_ = {ev_io = {ev_io_next = { le_next = 0x0, le_prev = 0x562340ffcf50}, ev_timeout = {tv_sec = 140736109281280, tv_usec = 0}}, ev_signal = {ev_signal_next = {le_next = 0x0, le_prev = 0x562340ffcf50}, ev_ncalls = 0, ev_pncalls = 0x0}}, ev_events = 24, ev_res = 0, ev_timeout = {tv_sec = 0, tv_usec = 94709397543957}} Thread 1 (Thread 0x7fb2eaeba700 (LWP 14174)): #0 0x000056233fff038e in create_auth_rpc_client (clp=clp@entry=0x562341008fa0, tgtname=tgtname@entry=0x562341011c8f "h...@client.domain.tu-berlin.de", clnt_return=clnt_return@entry=0x7fb2eaeb9de8, auth_return=auth_return@entry=0x7fb2eaeb9d50, uid=uid@entry=0, cred=cred@entry=0x0, authtype=0) at gssd_proc.c:352 rpc_clnt = 0x0 sec = {mech = 0x56233fffc590 <krb5oid>, qop = 0, svc = RPCSEC_GSS_SVC_NONE, cred = 0x7fb2dc000d60, req_flags = 2} auth = 0x0 retval = -1 min_stat = 256 rpc_errmsg = '\000' <repeats 13 times>, "\001\000\000c\343\000\333@'\243\001\214\002\236ROO\257\303\351\267]<\033\236\250̺ \215\210o=o\316_]\v\266\301\207\307\313\377S\236\353\262\177\000\000c\301\207-\204\\\223\226\024\320E\347\025\307\021`0\314\340B\020'B\370\252\300\327Ɣ\000b3\341\064\r\320\315ۻ\026z\261\035@|⌁\023s\322&e\365\377S\236\353\262\177\000\000\347\351\233\022F\210[\274}!\225\252\213\260\377\343\016\267>$(\261:\313<ޖ\322F\314ᯨi\367\032\255\306\306Ja\203\355&\253\000\064\271M\023\026\243\266c\334\313\000\242\066p\361K\361$\320\357-\366\026)\240\063\320\f\000ܲ\177\000\000\260k"... protocol = 6 timeout = {tv_sec = -418240567167153622, tv_usec = 6354052107285749079} addr = 0x5623410090f8 salen = <optimized out> #1 0x000056233fff09f8 in krb5_use_machine_creds (clp=clp@entry=0x562341008fa0, uid=uid@entry=0, tgtname=tgtname@entry=0x562341011c8f "h...@client.domain.tu-berlin.de", service=service@entry=0x562341011cb5 "nfs", rpc_clnt=rpc_clnt@entry=0x7fb2eaeb9de8) at gssd_proc.c:558 min_stat = 0 auth = 0x0 credlist = 0x7fb2dc02a910 ccname = 0x7fb2dc02a910 nocache = <optimized out> success = 0 #2 0x000056233fff0b92 in process_krb5_upcall (clp=clp@entry=0x562341008fa0, uid=uid@entry=0, fd=11, tgtname=tgtname@entry=0x562341011c8f "h...@client.domain.tu-berlin.de", service=service@entry=0x562341011cb5 "nfs") at gssd_proc.c:646 rpc_clnt = 0x0 auth = <optimized out> pd = {pd_ctx = 0x0, pd_ctx_hndl = {length = 0, value = 0x0}, pd_seq_win = 0} token = {length = 0, value = 0x0} err = 1090591864 downcall_err = -13 maj_stat = <optimized out> min_stat = 1090560944 lifetime_rec = 22051 gacceptor = 0x0 mech = 0x7fb2dc000020 acceptor = {length = 0, value = 0x0} #3 0x000056233fff13b3 in handle_gssd_upcall (info=0x562341011c70) at gssd_proc.c:805 clp = 0x562341008fa0 uid = 0 p = 0x562341011c87 "" mech = <optimized out> uidstr = <optimized out> target = 0x562341011c8f "h...@client.domain.tu-berlin.de" service = 0x562341011cb5 "nfs" enctypes = 0x562341011cc2 "18,17,16,23,3,1,2" upcall_str = 0x7fb2dc001000 "mech=krb5 uid=0 target=h...@client.domain.tu-berlin.de service=nfs enctypes=18,17,16,23,3,1,2 " pbuf = 0x0 __func__ = "handle_gssd_upcall" #4 0x00007fb2eb8dbfa3 in start_thread (arg=<optimized out>) at pthread_create.c:486 ret = <optimized out> pd = <optimized out> now = <optimized out> unwind_buf = {cancel_jmp_buf = {{jmp_buf = {140406422218496, 1116452640567996153, 140736109338078, 140736109338079, 140406422218496, 94709414435248, -1145117077261373703, -1145118435554349319}, mask_was_saved = 0}}, priv = {pad = {0x0, 0x0, 0x0, 0x0}, data = {prev = 0x0, cleanup = 0x0, canceltype = 0}}} not_first_call = <optimized out> #5 0x00007fb2eb80c4cf in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:95 No locals. Remark: The additonal command line parameters only being used in order to debug the segfaults via syslog. The segfaults are completely idependent of these parameters. Putting the timeout switches "-T" and "-t" seems to "cure" the problem a little bit and lessens the number of segfault hits per day. For backtracing purposes, these packages were installed explicitly: systemd-coredump debian-goodies libipc-system-simple-perl libfile-which-perl elfutils dctrl-tools libcom-err2-dbgsym libevent-2.1-6-dbgsym nfs-common-dbgsym Furthermore, system journal shows the following (most interesting part framed by hashtags) debug messages from RPC stack: Jun 25 11:46:08 server rpcbind[14166]: connect from 130.149.115.135 to null() Jun 25 11:46:08 server rpcbind[14167]: connect from 130.149.115.135 to getport/addr(nfs) Jun 25 11:46:08 server rpcbind[14168]: connect from 130.149.115.135 to null() Jun 25 11:46:08 server rpcbind[14169]: connect from 130.149.115.135 to getport/addr(nfs) Jun 25 11:46:08 server rpcbind[14170]: connect from 130.149.115.135 to null() Jun 25 11:46:08 server rpcbind[14171]: connect from 130.149.115.135 to getport/addr(nfs) Jun 25 11:46:08 server rpcbind[14172]: connect from 130.149.115.135 to null() Jun 25 11:46:08 server rpcbind[14173]: connect from 130.149.115.135 to getport/addr(nfs) Jun 25 11:46:08 server rpc.svcgssd[19640]: leaving poll Jun 25 11:46:08 server rpc.svcgssd[19640]: handling null request Jun 25 11:46:08 server rpc.svcgssd[19640]: svcgssd_limit_krb5_enctypes: Calling gss_set_allowable_enctypes with 7 enctypes from the kernel Jun 25 11:46:08 server rpc.svcgssd[19640]: sname = host/client.domain.tu-berlin...@tu-berlin.de Jun 25 11:46:08 server rpc.svcgssd[19640]: nfs4_gss_princ_to_ids: calling nsswitch->princ_to_ids Jun 25 11:46:08 server rpc.svcgssd[19640]: nss_getpwnam: name 'host/client.domain.tu-berlin...@tu-berlin.de' domain '(null)': resulting localname 'host/client.domain.tu-berlin.de' Jun 25 11:46:08 server rpc.svcgssd[19640]: nss_ldap: reconnecting to LDAP server... Jun 25 11:46:08 server rpc.svcgssd[19640]: nss_ldap: reconnected to LDAP server ldaps://ldap-slaves.tu-berlin.de after 1 attempt Jun 25 11:46:08 server rpc.svcgssd[19640]: nfs4_gss_princ_to_ids: nsswitch->princ_to_ids returned -2 Jun 25 11:46:08 server rpc.svcgssd[19640]: nfs4_gss_princ_to_ids: final return value is -2 Jun 25 11:46:08 server rpc.svcgssd[19640]: DEBUG: serialize_krb5_ctx: lucid version! Jun 25 11:46:08 server rpc.svcgssd[19640]: prepare_krb5_rfc4121_buffer: protocol 1 Jun 25 11:46:08 server rpc.svcgssd[19640]: prepare_krb5_rfc4121_buffer: serializing key with enctype 18 and size 32 Jun 25 11:46:08 server rpc.svcgssd[19640]: doing downcall Jun 25 11:46:08 server rpc.svcgssd[19640]: mech: krb5, hndl len: 4, ctx len 52, timeout: 1593108198 (29830 from now), clnt: h...@client.domain.tu-berlin.de, uid: -1, gid: -1, num aux grps: 0: Jun 25 11:46:08 server rpc.svcgssd[19640]: sending null reply Jun 25 11:46:08 server rpc.svcgssd[19640]: writing message: \x \x608202cf06092a864886f71201020201006e8202be308202baa003020105a10302010ea20703050020000000a3820194618201903082018ca003020105a10e1b0c54552d4245524c494e2e4445a2273025a003020103a11e Jun 25 11:46:08 server rpc.svcgssd[19640]: finished handling null request Jun 25 11:46:08 server rpc.svcgssd[19640]: entering poll Jun 25 11:46:08 server rpc.gssd[6356]: inotify event for topdir (nfsd4_cb) - ev->wd (5) ev->name (clnt50e) ev->mask (0x40000100) Jun 25 11:46:08 server rpc.gssd[6356]: handle_gssd_upcall: 'mech=krb5 uid=0 target=h...@client.domain.tu-berlin.de service=nfs enctypes=18,17,16,23,3,1,2 ' (nfsd4_cb/clnt50e) Jun 25 11:46:08 server rpc.gssd[6356]: krb5_use_machine_creds: uid 0 tgtname h...@client.domain.tu-berlin.de Jun 25 11:46:08 server rpc.gssd[6356]: inotify event for clntdir (nfsd4_cb/clnt50e) - ev->wd (75) ev->name (krb5) ev->mask (0x00000200) Jun 25 11:46:08 server rpc.gssd[6356]: inotify event for clntdir (nfsd4_cb/clnt50e) - ev->wd (75) ev->name (gssd) ev->mask (0x00000200) Jun 25 11:46:08 server rpc.gssd[6356]: inotify event for clntdir (nfsd4_cb/clnt50e) - ev->wd (75) ev->name (info) ev->mask (0x00000200) Jun 25 11:46:08 server rpc.gssd[6356]: inotify event for clntdir (nfsd4_cb/clnt50e) - ev->wd (75) ev->name (<?>) ev->mask (0x00008000) Jun 25 11:46:08 server rpc.gssd[6356]: inotify event for topdir (nfsd4_cb) - ev->wd (5) ev->name (clnt50f) ev->mask (0x40000100) Jun 25 11:46:08 server rpc.gssd[6356]: Full hostname for '' is 'client.domain.tu-berlin.de' Jun 25 11:46:08 server rpc.gssd[6356]: Full hostname for 'server.domain.tu-berlin.de' is 'server.domain.tu-berlin.de' Jun 25 11:46:08 server rpc.gssd[6356]: Success getting keytab entry for 'nfs/server.domain.tu-berlin...@tu-berlin.de' Jun 25 11:46:08 server rpc.gssd[6356]: INFO: Credentials in CC 'FILE:/tmp/krb5ccmachine_TU-BERLIN.DE' are good until 1593101766 Jun 25 11:46:08 server rpc.gssd[6356]: INFO: Credentials in CC 'FILE:/tmp/krb5ccmachine_TU-BERLIN.DE' are good until 1593101766 ############################################### Jun 25 11:46:08 server rpc.gssd[6356]: creating (null) client for server (null) Jun 25 11:46:08 all kernel: rpc.gssd[14174]: segfault at 0 ip 000056233fff038e sp 00007fb2eaeb9880 error 4 in rpc.gssd[56233ffed000+9000] Jun 25 11:46:08 all kernel: Code: 00 00 41 89 c4 48 8d 35 b8 6b 00 00 31 c0 e8 b9 eb ff ff 48 8b 73 40 b9 04 00 00 00 48 8d 3d a1 6a 00 00 0f b7 93 58 01 00 00 <f3> a6 0f 97 c0 1c 00 0f be c0 83 f8 01 45 19 ff 41 83 e7 0b 41 83 ############################################### Jun 25 11:46:08 server rpc.svcgssd[19640]: leaving poll Jun 25 11:46:08 server rpc.svcgssd[19640]: handling null request Jun 25 11:46:08 server rpc.svcgssd[19640]: svcgssd_limit_krb5_enctypes: Calling gss_set_allowable_enctypes with 7 enctypes from the kernel Jun 25 11:46:08 server rpc.svcgssd[19640]: sname = host/client.domain.tu-berlin...@tu-berlin.de Jun 25 11:46:08 server rpc.svcgssd[19640]: nfs4_gss_princ_to_ids: calling nsswitch->princ_to_ids Jun 25 11:46:08 server rpc.svcgssd[19640]: nss_getpwnam: name 'host/client.domain.tu-berlin...@tu-berlin.de' domain '(null)': resulting localname 'host/client.domain.tu-berlin.de' Jun 25 11:46:08 server rpc.svcgssd[19640]: nfs4_gss_princ_to_ids: nsswitch->princ_to_ids returned -2 Jun 25 11:46:08 server rpc.svcgssd[19640]: nfs4_gss_princ_to_ids: final return value is -2 Jun 25 11:46:08 server rpc.svcgssd[19640]: DEBUG: serialize_krb5_ctx: lucid version! Jun 25 11:46:08 server rpc.svcgssd[19640]: prepare_krb5_rfc4121_buffer: protocol 1 Jun 25 11:46:08 server rpc.svcgssd[19640]: prepare_krb5_rfc4121_buffer: serializing key with enctype 18 and size 32 Jun 25 11:46:08 server rpc.svcgssd[19640]: doing downcall Jun 25 11:46:08 server rpc.svcgssd[19640]: mech: krb5, hndl len: 4, ctx len 52, timeout: 1593108198 (29830 from now), clnt: h...@client.domain.tu-berlin.de, uid: -1, gid: -1, num aux grps: 0: Jun 25 11:46:08 server rpc.svcgssd[19640]: sending null reply Jun 25 11:46:08 server rpc.svcgssd[19640]: writing message: \x \x608202cf06092a864886f71201020201006e8202be308202baa003020105a10302010ea20703050020000000a3820194618201903082018ca003020105a10e1b0c54552d4245524c494e2e4445a2273025a003020103a11e Jun 25 11:46:08 server rpc.svcgssd[19640]: finished handling null request Jun 25 11:46:08 server rpc.svcgssd[19640]: entering poll Jun 25 11:46:08 server rpc.svcgssd[19640]: leaving poll Jun 25 11:46:08 server rpc.svcgssd[19640]: handling null request Jun 25 11:46:08 server rpc.svcgssd[19640]: svcgssd_limit_krb5_enctypes: Calling gss_set_allowable_enctypes with 7 enctypes from the kernel Jun 25 11:46:08 server rpc.svcgssd[19640]: sname = claur...@tu-berlin.de Jun 25 11:46:08 server rpc.svcgssd[19640]: nfs4_gss_princ_to_ids: calling nsswitch->princ_to_ids Jun 25 11:46:08 server rpc.svcgssd[19640]: nss_getpwnam: name 'claur...@tu-berlin.de' domain '(null)': resulting localname 'claurjec' Jun 25 11:46:08 server rpc.svcgssd[19640]: nfs4_gss_princ_to_ids: nsswitch->princ_to_ids returned 0 Jun 25 11:46:08 server rpc.svcgssd[19640]: nfs4_gss_princ_to_ids: final return value is 0 Jun 25 11:46:08 server rpc.svcgssd[19640]: nfs4_gss_princ_to_grouplist: calling nsswitch->gss_princ_to_grouplist Jun 25 11:46:08 server rpc.svcgssd[19640]: nss_getpwnam: name 'claur...@tu-berlin.de' domain '(null)': resulting localname 'claurjec' Jun 25 11:46:08 all systemd[1]: Started Process Core Dump (PID 14175/UID 0). Jun 25 11:46:08 server rpc.svcgssd[19640]: nfs4_gss_princ_to_grouplist: nsswitch->gss_princ_to_grouplist returned 0 Jun 25 11:46:08 server rpc.svcgssd[19640]: nfs4_gss_princ_to_grouplist: final return value is 0 Jun 25 11:46:08 server rpc.svcgssd[19640]: DEBUG: serialize_krb5_ctx: lucid version! Jun 25 11:46:08 server rpc.svcgssd[19640]: prepare_krb5_rfc4121_buffer: protocol 1 Jun 25 11:46:08 server rpc.svcgssd[19640]: prepare_krb5_rfc4121_buffer: serializing key with enctype 18 and size 32 Jun 25 11:46:08 server rpc.svcgssd[19640]: doing downcall Jun 25 11:46:08 server rpc.svcgssd[19640]: mech: krb5, hndl len: 4, ctx len 52, timeout: 1593108198 (29830 from now), clnt: <null>, uid: 17942, gid: 100, num aux grps: 5: Jun 25 11:46:08 server rpc.svcgssd[19640]: ( 1) 100 Jun 25 11:46:08 server rpc.svcgssd[19640]: ( 2) 8959 Jun 25 11:46:08 server rpc.svcgssd[19640]: ( 3) 9477 Jun 25 11:46:08 server rpc.svcgssd[19640]: ( 4) 10395 Jun 25 11:46:08 server rpc.svcgssd[19640]: ( 5) 10568 Jun 25 11:46:08 server rpc.svcgssd[19640]: sending null reply Jun 25 11:46:08 server rpc.svcgssd[19640]: writing message: \x \x6082029d06092a864886f71201020201006e82028c30820288a003020105a10302010ea20703050020000000a382017c6182017830820174a003020105a10e1b0c54552d4245524c494e2e4445a2273025a003020103a11e Jun 25 11:46:08 server rpc.svcgssd[19640]: finished handling null request Jun 25 11:46:08 server rpc.svcgssd[19640]: entering poll Jun 25 11:46:08 server rpc.mountd[19637]: nss_ldap: reconnecting to LDAP server... Jun 25 11:46:08 server rpc.mountd[19637]: nss_ldap: reconnected to LDAP server ldaps://ldap-slaves.tu-berlin.de after 1 attempt Jun 25 11:46:08 all systemd[1]: rpc-gssd.service: Main process exited, code=killed, status=11/SEGV Jun 25 11:46:08 all systemd[1]: rpc-gssd.service: Failed with result 'signal'. Jun 25 11:46:09 all systemd-coredump[14176]: Process 6356 (rpc.gssd) of user 0 dumped core. Stack trace of thread 14174: #0 0x000056233fff038e n/a (rpc.gssd) #1 0x000056233fff09f8 n/a (rpc.gssd) #2 0x000056233fff0b92 n/a (rpc.gssd) #3 0x000056233fff13b3 n/a (rpc.gssd) #4 0x00007fb2eb8dbfa3 start_thread (libpthread.so.0) #5 0x00007fb2eb80c4cf __clone (libc.so.6) Stack trace of thread 6356: #0 0x00007fb2eb801819 __GI___poll (libc.so.6) #1 0x00007fb2eb6e7207 send_dg (libresolv.so.2) #2 0x00007fb2eb6e4c43 __GI___res_context_query (libresolv.so.2) #3 0x00007fb2eb6bf536 __GI__nss_dns_gethostbyaddr2_r (libnss_dns.so.2) #4 0x00007fb2eb6bf823 _nss_dns_gethostbyaddr_r (libnss_dns.so.2) #5 0x00007fb2eb81dee2 __gethostbyaddr_r (libc.so.6) #6 0x00007fb2eb8267d5 gni_host_inet_name (libc.so.6) #7 0x000056233ffef455 n/a (rpc.gssd) #8 0x000056233ffef82c n/a (rpc.gssd) #9 0x000056233fff01d0 n/a (rpc.gssd) #10 0x00007fb2ebab49ba n/a (libevent-2.1.so.6) #11 0x00007fb2ebab5537 event_base_loop (libevent-2.1.so.6) #12 0x000056233ffedeaa n/a (rpc.gssd) #13 0x00007fb2eb73709b __libc_start_main (libc.so.6) #14 0x000056233ffee03a n/a (rpc.gssd) Jun 25 11:46:09 all systemd[1]: rpc-gssd.service: Service RestartSec=100ms expired, scheduling restart. Jun 25 11:46:09 all systemd[1]: rpc-gssd.service: Scheduled restart job, restart counter is at 4. Jun 25 11:46:09 all systemd[1]: systemd-coredump@2-14175-0.service: Succeeded. Jun 25 11:46:09 all systemd[1]: Stopped RPC security service for NFS client and server. Jun 25 11:46:09 all systemd[1]: Starting Preprocess NFS configuration... Jun 25 11:46:09 all systemd[1]: nfs-config.service: Succeeded. Jun 25 11:46:09 all systemd[1]: Started Preprocess NFS configuration. Jun 25 11:46:09 all systemd[1]: Starting RPC security service for NFS client and server... Jun 25 11:46:09 server rpc.gssd[14184]: libtirpc: debug level 7 Jun 25 11:46:09 server rpc.gssd[14185]: doing a full rescan Jun 25 11:46:09 server rpc.gssd[14185]: handle_gssd_upcall: 'mech=krb5 uid=0 target=h...@client.domain.tu-berlin.de service=nfs enctypes=18,17,16,23,3,1,2 ' (nfsd4_cb/clnt50f) Jun 25 11:46:09 server rpc.gssd[14185]: krb5_use_machine_creds: uid 0 tgtname h...@client.domain.tu-berlin.de Jun 25 11:46:09 all systemd[1]: Started RPC security service for NFS client and server. Jun 25 11:46:09 server rpc.gssd[14185]: Full hostname for 'client.domain.tu-berlin.de' is 'client.domain.tu-berlin.de' Jun 25 11:46:09 server rpc.gssd[14185]: Full hostname for 'server.domain.tu-berlin.de' is 'server.domain.tu-berlin.de' Jun 25 11:46:09 server rpc.gssd[14185]: Success getting keytab entry for 'nfs/server.domain.tu-berlin...@tu-berlin.de' Jun 25 11:46:09 server rpc.gssd[14185]: gssd_get_single_krb5_cred: principal 'nfs/server.domain.tu-berlin...@tu-berlin.de' ccache:'FILE:/tmp/krb5ccmachine_TU-BERLIN.DE' Jun 25 11:46:09 server rpc.gssd[14185]: INFO: Credentials in CC 'FILE:/tmp/krb5ccmachine_TU-BERLIN.DE' are good until 1593114369 Jun 25 11:46:09 server rpc.gssd[14185]: creating tcp client for server client.domain.tu-berlin.de Jun 25 11:46:09 server rpc.gssd[14185]: DEBUG: port already set to 41205 Jun 25 11:46:09 server rpc.gssd[14185]: creating context with server h...@client.domain.tu-berlin.de Jun 25 11:46:09 server rpc.gssd[14185]: in authgss_create_default() Jun 25 11:46:09 server rpc.gssd[14185]: in authgss_create() Jun 25 11:46:09 server rpc.gssd[14185]: authgss_create: name is 0x7f864800bc60 Jun 25 11:46:09 server rpc.gssd[14185]: authgss_create: gd->name is 0x7f8648009d80 Jun 25 11:46:09 server rpc.gssd[14185]: in authgss_refresh() Jun 25 11:46:09 server rpc.gssd[14185]: gss_init_sec_context: Unspecified GSS failure. Minor code may provide more information - KDC returned error string: NO PREAUTH Jun 25 11:46:09 server rpc.gssd[14185]: authgss_create_default: freeing name 0x7f864800bc60 Jun 25 11:46:09 server rpc.gssd[14185]: WARNING: Failed to create krb5 context for user with uid 0 for server h...@client.domain.tu-berlin.de Jun 25 11:46:09 server rpc.gssd[14185]: WARNING: Failed to create machine krb5 context with cred cache FILE:/tmp/krb5ccmachine_TU-BERLIN.DE for server client.domain.tu-berlin.de Jun 25 11:46:09 server rpc.gssd[14185]: WARNING: Machine cache prematurely expired or corrupted trying to recreate cache for server client.domain.tu-berlin.de Jun 25 11:46:09 server rpc.gssd[14185]: Full hostname for 'client.domain.tu-berlin.de' is 'client.domain.tu-berlin.de' Jun 25 11:46:09 server rpc.gssd[14185]: Full hostname for 'server.domain.tu-berlin.de' is 'server.domain.tu-berlin.de' Jun 25 11:46:09 server rpc.gssd[14185]: Success getting keytab entry for 'nfs/server.domain.tu-berlin...@tu-berlin.de' Jun 25 11:46:09 server rpc.gssd[14185]: INFO: Credentials in CC 'FILE:/tmp/krb5ccmachine_TU-BERLIN.DE' are good until 1593114369 Jun 25 11:46:09 server rpc.gssd[14185]: INFO: Credentials in CC 'FILE:/tmp/krb5ccmachine_TU-BERLIN.DE' are good until 1593114369 Jun 25 11:46:09 server rpc.gssd[14185]: creating tcp client for server client.domain.tu-berlin.de Jun 25 11:46:09 server rpc.gssd[14185]: DEBUG: port already set to 41205 Jun 25 11:46:09 server rpc.gssd[14185]: creating context with server h...@client.domain.tu-berlin.de Jun 25 11:46:09 server rpc.gssd[14185]: in authgss_create_default() Jun 25 11:46:09 server rpc.gssd[14185]: in authgss_create() Jun 25 11:46:09 server rpc.gssd[14185]: authgss_create: name is 0x7f8648003170 Jun 25 11:46:09 server rpc.gssd[14185]: authgss_create: gd->name is 0x7f86480060d0 Jun 25 11:46:09 server rpc.gssd[14185]: in authgss_refresh() Jun 25 11:46:09 server rpc.gssd[14185]: gss_init_sec_context: Unspecified GSS failure. Minor code may provide more information - KDC returned error string: NO PREAUTH Jun 25 11:46:09 server rpc.gssd[14185]: authgss_create_default: freeing name 0x7f8648003170 Jun 25 11:46:09 server rpc.gssd[14185]: WARNING: Failed to create krb5 context for user with uid 0 for server h...@client.domain.tu-berlin.de Jun 25 11:46:09 server rpc.gssd[14185]: WARNING: Failed to create machine krb5 context with cred cache FILE:/tmp/krb5ccmachine_TU-BERLIN.DE for server client.domain.tu-berlin.de Jun 25 11:46:09 server rpc.gssd[14185]: ERROR: Failed to create machine krb5 context with any credentials cache for server client.domain.tu-berlin.de Jun 25 11:46:09 server rpc.gssd[14185]: doing error downcall As far as I see, there are two uncatched NULL pointers forcing rpc.gssd to segfault. Frankly spoken, I have no idea, where the NULL pointers are coming from. Saidly, the debugger shows a lot of <optimized out> within the trace. What the heck about these mysterous '(null)' strings showing up in the syslog: Jun 25 11:46:08 server rpc.gssd[6356]: creating (null) client for server (null) For time to time, I observe binary crap instead of '(null)' strings in respective log messages of rpc.gssd: Mar 09 10:11:07 server rpc.gssd[980]: creating <C2><F8><87>^FV client for server x<AB><BD><A5><F1>^? In cases of connection attemps without trouble, rpc.gssd logs the following: Mar 09 10:11:07 server rpc.gssd[980]: creating tcp client for server client.domain.tu-berlin.de In other cases, rpc.gssd experiences segfaults in libc: Mar 08 14:08:06 server kernel: rpc.gssd[12104]: segfault at ffffffffffffffc0 ip 00007f03f16e16e4 sp 00007f03f0c1c0f8 error 5 in libc-2.24.so[7f03f1661000+195000] Additonal information about system setup: Debian_version: 10.4 Kernel Version: Linux all 4.19.0-9-amd64 #1 SMP Debian 4.19.118-2+deb10u1 (2020-06-07) x86_64 GNU/Linux nfs-kernel-server: Version 1:1.3.4-2.5 libc6: Version 2.28-10 Hope, anyone of the maintainers will care about fixing this segfault issue with Debian Buster. Best and be firm Sebastian Sebastian Kraus Team IT am Institut für Chemie Gebäude C, Straße des 17. Juni 115, Raum C7 Technische Universität Berlin Fakultät II Institut für Chemie Sekretariat C3 Straße des 17. Juni 135 10623 Berlin Email: sebastian.kr...@tu-berlin.de