Thanks for your explanations. It helped greatly. Using ktrdump and schedgraph.py and after modifying our test program to set and unset automatically debug.ktr.mask, I've been able to get useful information.

First, It made me realize that task switching, with default settings and 2 active processes, only occurs each 100ms. Knowing that, expecting a latency time around 100ms was kind of silly :)

Next, it seems most of the latency pikes are due to a process starting or waking up. For instance, it usually happens when the openssl speed test is started ( http://jflesch.kwain.net/~jflesch/sys_latence/sched/sched_graph_openssl_start.png ) or when pagedaemon wakes up (I forgot to disable the swap and my test program used too much memory to store the result values ...). I'm not sure why, but when we start openssl, it is often allowed to run for >= 300ms, even with our test program set to real time priority. My intuition is that, at first, it's considered as an interactive process, until the scheduler realizes it's not. But then, does anyone know why it would take more than 300ms for the scheduler to realize that ?

Anyway, by setting kern.sched.interact=5 (so openssl isn't considered as an interactive process), kern.sched.slice=3 (to get an high enough scheduling resolution), and our program to real-time priority, we got rid of both problems. I'm just a little bit worried about kern.sched.slice=3. Is there any known side effect when reducing slices size ?

Also, another issue remain: We were hoping to keep our program with a normal priority. However when we set our test program to a normal priority (but still an higher priority than openssl), both get 50% of the CPU (I guess this is to be expected), and from time to time we have a "hiccup" in the scheduling: http://jflesch.kwain.net/~jflesch/sys_latence/sched/sched_graph_hicups.png . Is there any way to avoid them ? In other words, is it possible to make sure that the low priority process never gets more CPU time than the high priority one ?



On 23.02.2011 02:08, Ryan Stone wrote:
To debug weird scheduling issues I find it helpful to start by looking
at a schedgraph.  schedgraph is a tool that can display a graphical
representation of what the scheduler was doing over a small slice of
time.  The one downside is that you have to recompile your kernel to
get the hooks that collect the necessary data, but it sounds like your
problem is pretty easy to reproduce and so that shouldn't be a big
problem.

To enable the hooks, put the following in your kernel conf:

options KTR
options KTR_COMPILE=KTR_SCHED
options KTR_MASK=KTR_SCHED
options KTR_ENTIRES=(128*1024)

Then rebuild and install the new kernel.  Next, run your test.  The
instant that your test has detected the long delay, set the sysctl
debug.ktr.mask to 0.  The scheduler hooks record data into a ring
buffer, so the interesting data can be flushed out within seconds.
Clearing that sysctl will stop any further events from being logged,
which means that the interesting data will stay there until you go and
collect it.

You can get the raw data by redirecting the output of "ktrdump -ct"
into a file.  Then from any machine with a graphical interface(FreeBSD
with X installed, Windows, Mac, whatever) and python installed, run:
$ python schedgraph.py /path/to/ktrdump/output

You can get schedgraph.py from /usr/src/tools/sched.

If you want to collect the data again, set the sysctl debug.ktr.mask
back to 0x20000000 to restart gathering scheduler data.

Ryan



Reply via email to