[OpenIndiana-discuss] Slowly increasing kernel load with oi_151a

Svavar Örn Eysteinsson svavar at fiton.is
Fri Dec 23 09:49:37 UTC 2011


I had the same problem, and pretty much the same hardware.
But as Kasper pointed out, https://www.illumos.org/issues/1333#note-21
solved my problem. no more kernel cpu rising...

put this into your /etc/system file

set apix:apic_timer_preferred_mode=0x0

and reboot.


Regards,

Svavar O





> Kasper Brink <mailto:K.Brink at cs.ru.nl>
> 23. desember 2011 08:47
>
> I'm not an Illumos expert, but the symptoms you describe sound a lot like
> those of https://www.illumos.org/issues/1333 (high sys load, dtrace hang).
> Does the workaround in comment #21
> (https://www.illumos.org/issues/1333#note-21) solve the problem for you?
>
> Regards,
>
> Kasper
>
>
> _______________________________________________
> OpenIndiana-discuss mailing list
> OpenIndiana-discuss at openindiana.org
> http://openindiana.org/mailman/listinfo/openindiana-discuss
> Chris Bünger <mailto:chris.buenger at googlemail.com>
> 23. desember 2011 07:55
> I experience the same issue. For me, a simple reboot did not help, but
> a reboot -p.
> I have similar setups with 134 and 151a and only 151a shows this behavior,
>
> Maybe this helps.
>
> Chris
>
>
> _______________________________________________
> OpenIndiana-discuss mailing list
> OpenIndiana-discuss at openindiana.org
> http://openindiana.org/mailman/listinfo/openindiana-discuss
> Mirko Kaffka <mailto:mk at mkaffka.de>
> 23. desember 2011 00:08
> I recently got a new machine and installed oi_151a from scratch
> (core i7-2600, intel motherboard DH67BL, 16GB RAM).
> After a few days uptime I noticed a constant system load of about 10%
> although the desktop was idle and I had not started anything that caused
> a permanent load. There was almost no I/O activity, just a few reads
> and writes every few seconds. vmstat showed 0-1% user time but
> 10-13% system time. prstat -v output was far below 1% or 0% user and
> system time for all processes.
> Over the following days the load increased further. When I took 7 cpu
> cores off-line I got about 80% sys load on the remaining core. Where
> does it come from?
>
> When I switch from multi user to single user mode the load persists.
> When I reboot, everything is fine for a while (0-1% sys load) but the 
> load
> slowly starts increasing again. So, I have to reboot the machine about
> every 2 days what is very unpleasant.
>
> I tried to analyze the issue using intrstat, lockstat, etc. but have not
> got very far.
>
> All following commands were run in single user mode and with only one cpu
> core on-line. (I hope it's ok to put the output here?)
>
> ~ # vmstat 5
>  kthr      memory            page            disk          faults      
> cpu
>  r b w   swap  free  re  mf pi po fr de sr s1 s2 s3 s4   in   sy   cs 
> us sy id
>  0 0 0 9070220 2859260 2  4  0  0  0  0  0  5 -1  1 14  397  265  258  
> 0  4 95
>  0 0 0 10392120 4142932 24 64 0 0  0  0  0  0  0  0  1  508   99  227  
> 0 21 79
>  0 0 0 10392120 4142960 0 0  0  0  0  0  0  0  0  0 13  511   60  229  
> 0 21 79
>  0 0 0 10392124 4142964 0 0  0  0  0  0  0  0  0  0 13  509   59  226  
> 0 21 79
>
> ~ # ps -ef
>      UID   PID  PPID   C    STIME TTY         TIME CMD
>     root     0     0   0   Dec 20 ?           0:01 sched
>     root     4     0   0   Dec 20 ?           0:00 kcfpoold
>     root     6     0   0   Dec 20 ?           2:34 zpool-rpool
>     root     1     0   0   Dec 20 ?           0:00 /sbin/init
>     root     2     0   0   Dec 20 ?           0:00 pageout
>     root     3     0   0   Dec 20 ?           8:54 fsflush
>     root    10     1   0   Dec 20 ?           0:03 
> /lib/svc/bin/svc.startd
>     root    12     1   0   Dec 20 ?           0:08 
> /lib/svc/bin/svc.configd
>   netadm    50     1   0   Dec 20 ?           0:00 /lib/inet/ipmgmtd
>    dladm    46     1   0   Dec 20 ?           0:00 /sbin/dlmgmtd
>     root   167     0   0   Dec 20 ?           1:50 zpool-tank
>     root   232     1   0   Dec 20 ?           0:00 
> /usr/lib/sysevent/syseventd
>     root  9518    10   0 21:01:56 console     0:00 -bash
>     root   262     1   0   Dec 20 ?           0:02 devfsadmd
>     root   276     1   0   Dec 20 ?           0:00 /usr/lib/power/powerd
>     root 10708  9518   0 21:04:53 console     0:00 ps -ef
>     root  3222     1   0   Dec 20 ?           0:00 -bash
>
>
> ~ # intrstat
>       device |      cpu0 %tim      cpu1 %tim      cpu2 %tim      cpu3 
> %tim
> -------------+------------------------------------------------------------ 
>
>     e1000g#1 |         1  0.0         0  0.0         0  0.0         0  
> 0.0
>       ehci#0 |         1  0.0         0  0.0         0  0.0         0  
> 0.0
>       ehci#1 |         1  0.0         0  0.0         0  0.0         0  
> 0.0
>       rtls#0 |         1  0.0         0  0.0         0  0.0         0  
> 0.0
> (cpu4..7 are all 0.0%)
>
>
> ~ # prstat -v
>    PID USERNAME USR SYS TRP TFL DFL LCK SLP LAT VCX ICX SCL SIG 
> PROCESS/NLWP
>  10711 root     0.0 0.0 0.0 0.0 0.0 0.0 100 0.0   1   0 254   0 prstat/1
>   3222 root     0.0 0.0 0.0 0.0 0.0 0.0 100 0.0   0   0   0   0 bash/1
> [cut]
> Total: 14 processes, 366 lwps, load averages: 0.02, 0.32, 0.67
>
>
> ~ # lockstat -kIW -D20 sleep 30
>
> Profiling interrupt: 2913 events in 30.028 seconds (97 events/sec)
>
> Count indv cuml rcnt     nsec Hottest CPU+PIL        Caller
> ------------------------------------------------------------------------------- 
>
>  2878  99%  99% 0.00      293 cpu[0]                 acpi_cpu_cstate
>    12   0%  99% 0.00      224 cpu[0]                 fsflush
>    10   0% 100% 0.00      266 cpu[0]                 i86_mwait
> [cut]
> ------------------------------------------------------------------------------- 
>
>
> Is the high count on acpi_cpu_cstate normal?
>
> The hotkernel script from the dtrace toolkit finally froze my system.
> After the reboot hotkernel run flawlessly.
>
> How can I further analyze this?
>
> Thanks,
> Mirko
>
>
> _______________________________________________
> OpenIndiana-discuss mailing list
> OpenIndiana-discuss at openindiana.org
> http://openindiana.org/mailman/listinfo/openindiana-discuss


More information about the OpenIndiana-discuss mailing list