[OpenIndiana-discuss] Slowly increasing kernel load with oi_151a
Mirko Kaffka
mk at mkaffka.de
Fri Dec 23 00:08:03 UTC 2011
I recently got a new machine and installed oi_151a from scratch
(core i7-2600, intel motherboard DH67BL, 16GB RAM).
After a few days uptime I noticed a constant system load of about 10%
although the desktop was idle and I had not started anything that caused
a permanent load. There was almost no I/O activity, just a few reads
and writes every few seconds. vmstat showed 0-1% user time but
10-13% system time. prstat -v output was far below 1% or 0% user and
system time for all processes.
Over the following days the load increased further. When I took 7 cpu
cores off-line I got about 80% sys load on the remaining core. Where
does it come from?
When I switch from multi user to single user mode the load persists.
When I reboot, everything is fine for a while (0-1% sys load) but the load
slowly starts increasing again. So, I have to reboot the machine about
every 2 days what is very unpleasant.
I tried to analyze the issue using intrstat, lockstat, etc. but have not
got very far.
All following commands were run in single user mode and with only one cpu
core on-line. (I hope it's ok to put the output here?)
~ # vmstat 5
kthr memory page disk faults cpu
r b w swap free re mf pi po fr de sr s1 s2 s3 s4 in sy cs us sy id
0 0 0 9070220 2859260 2 4 0 0 0 0 0 5 -1 1 14 397 265 258 0 4 95
0 0 0 10392120 4142932 24 64 0 0 0 0 0 0 0 0 1 508 99 227 0 21 79
0 0 0 10392120 4142960 0 0 0 0 0 0 0 0 0 0 13 511 60 229 0 21 79
0 0 0 10392124 4142964 0 0 0 0 0 0 0 0 0 0 13 509 59 226 0 21 79
~ # ps -ef
UID PID PPID C STIME TTY TIME CMD
root 0 0 0 Dec 20 ? 0:01 sched
root 4 0 0 Dec 20 ? 0:00 kcfpoold
root 6 0 0 Dec 20 ? 2:34 zpool-rpool
root 1 0 0 Dec 20 ? 0:00 /sbin/init
root 2 0 0 Dec 20 ? 0:00 pageout
root 3 0 0 Dec 20 ? 8:54 fsflush
root 10 1 0 Dec 20 ? 0:03 /lib/svc/bin/svc.startd
root 12 1 0 Dec 20 ? 0:08 /lib/svc/bin/svc.configd
netadm 50 1 0 Dec 20 ? 0:00 /lib/inet/ipmgmtd
dladm 46 1 0 Dec 20 ? 0:00 /sbin/dlmgmtd
root 167 0 0 Dec 20 ? 1:50 zpool-tank
root 232 1 0 Dec 20 ? 0:00 /usr/lib/sysevent/syseventd
root 9518 10 0 21:01:56 console 0:00 -bash
root 262 1 0 Dec 20 ? 0:02 devfsadmd
root 276 1 0 Dec 20 ? 0:00 /usr/lib/power/powerd
root 10708 9518 0 21:04:53 console 0:00 ps -ef
root 3222 1 0 Dec 20 ? 0:00 -bash
~ # intrstat
device | cpu0 %tim cpu1 %tim cpu2 %tim cpu3 %tim
-------------+------------------------------------------------------------
e1000g#1 | 1 0.0 0 0.0 0 0.0 0 0.0
ehci#0 | 1 0.0 0 0.0 0 0.0 0 0.0
ehci#1 | 1 0.0 0 0.0 0 0.0 0 0.0
rtls#0 | 1 0.0 0 0.0 0 0.0 0 0.0
(cpu4..7 are all 0.0%)
~ # prstat -v
PID USERNAME USR SYS TRP TFL DFL LCK SLP LAT VCX ICX SCL SIG PROCESS/NLWP
10711 root 0.0 0.0 0.0 0.0 0.0 0.0 100 0.0 1 0 254 0 prstat/1
3222 root 0.0 0.0 0.0 0.0 0.0 0.0 100 0.0 0 0 0 0 bash/1
[cut]
Total: 14 processes, 366 lwps, load averages: 0.02, 0.32, 0.67
~ # lockstat -kIW -D20 sleep 30
Profiling interrupt: 2913 events in 30.028 seconds (97 events/sec)
Count indv cuml rcnt nsec Hottest CPU+PIL Caller
-------------------------------------------------------------------------------
2878 99% 99% 0.00 293 cpu[0] acpi_cpu_cstate
12 0% 99% 0.00 224 cpu[0] fsflush
10 0% 100% 0.00 266 cpu[0] i86_mwait
[cut]
-------------------------------------------------------------------------------
Is the high count on acpi_cpu_cstate normal?
The hotkernel script from the dtrace toolkit finally froze my system.
After the reboot hotkernel run flawlessly.
How can I further analyze this?
Thanks,
Mirko
More information about the OpenIndiana-discuss
mailing list