[OpenIndiana-discuss] Zpool upgrade didn't seem to upgrade

Michelle Knight michelle at msknight.com
Thu Jan 20 07:05:51 UTC 2011


hi Folks, 

OI 148.

Three 1.5tb drives were replaced with three 2tb drives. They are here on 
internal SATA channels c2t2d0, c2t3d0 and c2t4d0. One is a Seagate Barracuda 
and the other two are Western Digital Greens.

mich at jaguar:~# cfgadm -lv
Ap_Id                          Receptacle   Occupant     Condition  
Information
When         Type         Busy     Phys_Id
Slot8                          connected    configured   ok         Location: 
Slot8
Jan  1  1970 unknown      n        /devices/pci at 0,0/pci8086,3b4a at 1c,4:Slot8
sata0/0::dsk/c2t0d0            connected    configured   ok         Mod: INTEL 
SSDSA2M040G2GC FRev: 2CV102HB SN: CVGB949301PH040GGN
unavailable  disk         n        /devices/pci at 0,0/pci1458,b005 at 1f,2:0
sata0/1::dsk/c2t1d0            connected    configured   ok         Mod: INTEL 
SSDSA2M040G2GC FRev: 2CV102HB SN: CVGB949301PC040GGN
unavailable  disk         n        /devices/pci at 0,0/pci1458,b005 at 1f,2:1
sata0/2::dsk/c2t2d0            connected    configured   ok         Mod: 
ST32000542AS FRev: CC34 SN: 5XW17ARW
unavailable  disk         n        /devices/pci at 0,0/pci1458,b005 at 1f,2:2
sata0/3::dsk/c2t3d0            connected    configured   ok         Mod: WDC 
WD20EARS-00MVWB0 FRev: 51.0AB51 SN: WD-WMAZA0555575
unavailable  disk         n        /devices/pci at 0,0/pci1458,b005 at 1f,2:3
sata0/4::dsk/c2t4d0            connected    configured   ok         Mod: WDC 
WD20EARS-00MVWB0 FRev: 51.0AB51 SN: WD-WMAZA0484508

A zpool export and subsequent import, which should have taken the set to 4tb 
overall storage in the raidz, appears to have not worked despite the import 
taking what must have been about ten to fifteen minutes to do the import. 
(during which time the drives were silent and the zpool process was mostly 0% 
very occasionally peaking to 25%, and the system being very slow to respond 
during that period)

Any ideas please? Or is there still some process running in the background 
that I can't see?

mich at jaguar:~# zfs list
NAME                     USED  AVAIL  REFER  MOUNTPOINT
data                    2.27T   401G  2.27T  /mirror
rpool                   7.69G  28.7G    45K  /rpool
rpool/ROOT              3.70G  28.7G    31K  legacy
rpool/ROOT/openindiana  3.70G  28.7G  3.59G  /
rpool/dump              1.93G  28.7G  1.93G  -
rpool/export            5.22M  28.7G    32K  /export
rpool/export/home       5.19M  28.7G    32K  /export/home
rpool/export/home/mich  5.16M  28.7G  5.16M  /export/home/mich
rpool/swap              2.05G  30.7G   126M  -


mich at jaguar:~# zpool status
  pool: data
 state: ONLINE
 scan: resilvered 1.13T in 12h26m with 0 errors on Wed Jan 19 23:42:23 2011
config:

        NAME        STATE     READ WRITE CKSUM
        data        ONLINE       0     0     0
          raidz1-0  ONLINE       0     0     0
            c2t2d0  ONLINE       0     0     0
            c2t3d0  ONLINE       0     0     0
            c2t4d0  ONLINE       0     0     0

errors: No known data errors


last pid:  1802;  load avg:  0.61,  0.56,  0.61;  up 0+19:55:20        
07:06:01
74 processes: 73 sleeping, 1 on cpu
CPU states: 99.8% idle,  0.0% user,  0.3% kernel,  0.0% iowait,  0.0% swap
Kernel: 375 ctxsw, 653 intr, 120 syscall
Memory: 3959M phys mem, 401M free mem, 1979M total swap, 1979M free swap

   PID USERNAME NLWP PRI NICE  SIZE   RES STATE    TIME    CPU COMMAND
  1197 gdm         1  59    0   95M   28M sleep    1:23  0.04% gdm-simple-gree
   922 root        3  59    0  102M   51M sleep    0:37  0.02% Xorg
  1801 root        1  59    0 4036K 2460K cpu/3    0:00  0.01% top
  1196 gdm         1  59    0   80M   13M sleep    0:00  0.00% metacity
  1190 gdm         1  59    0 7892K 6028K sleep    0:00  0.00% at-spi-registry
   640 root       16  59    0   14M 9072K sleep    0:09  0.00% smbd
  1737 mich        1  59    0   13M 5392K sleep    0:00  0.00% sshd
   148 root        1  59    0 8312K 1608K sleep    0:00  0.00% dhcpagent
   672 root       26  59    0   27M   15M sleep    0:01  0.00% fmd
  1198 gdm         1  59    0   87M   18M sleep    0:02  0.00% gnome-power-man
  1247 root        1  59    0 6080K 2500K sleep    0:00  0.00% sendmail
    11 root       21  59    0   15M   13M sleep    0:06  0.00% svc.configd
   272 root        6  59    0   11M 4784K sleep    0:01  0.00% devfsadm
  1220 root       24  59    0   13M 4404K sleep    0:01  0.00% nscd
    45 netcfg      5  59    0 4716K 3268K sleep    0:00  0.00% netcfgd
  1309 admin       1  59    0   16M 8404K sleep   14:30  0.00% sshd
  1312 admin       1  59    0   16M 5468K sleep    3:47  0.00% sshd
   787 root        1  59    0   12M 5928K sleep    0:05  0.00% intrd
  1237 root        4  59    0 8400K 1936K sleep    0:03  0.00% automountd
     9 root       15  59    0   20M   12M sleep    0:02  0.00% svc.startd
   387 root        5  59    0 7500K 6024K sleep    0:01  0.00% hald
  1308 root        1  59    0   13M 4788K sleep    0:01  0.00% sshd
   252 root        5  60  -20 2544K 1460K sleep    0:00  0.00% zonestatd
  1191 gdm         1  59    0  111M   42M sleep    0:00  0.00% gnome-settings-
  1175 gdm         2  59    0   20M   10M sleep    0:00  0.00% gnome-session
   790 root        4  59    0   15M 6432K sleep    0:00  0.00% rad
  1188 gdm         1  59    0   12M 5944K sleep    0:00  0.00% gconfd-2
   291 root        1  59    0   16M 5604K sleep    0:00  0.00% cupsd
  1156 mich        1  59    0   13M 5432K sleep    0:00  0.00% sshd
   132 daemon      3  59    0   13M 5104K sleep    0:00  0.00% kcfd


More information about the OpenIndiana-discuss mailing list