Hi,
I'm running 6.4 stable, with latest syspatches.
I saw ospf6d reporting this in the logs
Dec 18 08:18:10 obsd64-ic1 ospf6d[68658]: send_packet: error sending
packet on interface vmx1: No buffer space available
Searching the web, I gathered that netstat -m might shed some light, so
I proceeded :
obsd64-ic1# netstat -m
610 mbufs in use:
543 mbufs allocated to data
8 mbufs allocated to packet headers
59 mbufs allocated to socket names and addresses
13/200 mbuf 2048 byte clusters in use (current/peak)
0/30 mbuf 2112 byte clusters in use (current/peak)
1/56 mbuf 4096 byte clusters in use (current/peak)
0/48 mbuf 8192 byte clusters in use (current/peak)
475/2170 mbuf 9216 byte clusters in use (current/peak)
0/0 mbuf 12288 byte clusters in use (current/peak)
0/0 mbuf 16384 byte clusters in use (current/peak)
0/0 mbuf 65536 byte clusters in use (current/peak)
10196/23304/524288 Kbytes allocated to network (current/peak/max)
0 requests for memory denied
0 requests for memory delayed
0 calls to protocol drain routines
So if there were no requests denied or delayed and the peak was only
24MB out of 512MB max, what could cause ospf6d to complain ?
Should I be worried about this message ?
Looking at the sendto man page I get that it can return ENOBUFS in two
cases :
Case 1 - The system was unable to allocate an internal buffer
-> this seems to not be the case as shown above
This leaves only case 2 : The output queue for a network interface was
full.
Looking at netstat -id is see drops on vmx1 and vmx3.
Both of these cards are VMXNET3 cards connected to the different
VLANs/Portgroups on the same vswitch which has two 10G uplinks to the
switches.
sysctl | grep drops shows
net.inet.ip.ifq.drops=0
net.inet6.ip6.ifq.drops=0
net.pipex.inq.drops=0
net.pipex.outq.drops=0
I'm out of ideas for places where to look next.
Please, could network guru provide some insight/help ?
Or just tell me that it's not worth bothering and I should stop here ?
Thanks for your help and have a nice day !
Arnaud