[OpenIndiana-discuss] Zpool upgrade didn't seem to upgrade
Michelle Knight
michelle at msknight.com
Thu Jan 20 07:05:51 UTC 2011
hi Folks,
OI 148.
Three 1.5tb drives were replaced with three 2tb drives. They are here on
internal SATA channels c2t2d0, c2t3d0 and c2t4d0. One is a Seagate Barracuda
and the other two are Western Digital Greens.
mich at jaguar:~# cfgadm -lv
Ap_Id Receptacle Occupant Condition
Information
When Type Busy Phys_Id
Slot8 connected configured ok Location:
Slot8
Jan 1 1970 unknown n /devices/pci at 0,0/pci8086,3b4a at 1c,4:Slot8
sata0/0::dsk/c2t0d0 connected configured ok Mod: INTEL
SSDSA2M040G2GC FRev: 2CV102HB SN: CVGB949301PH040GGN
unavailable disk n /devices/pci at 0,0/pci1458,b005 at 1f,2:0
sata0/1::dsk/c2t1d0 connected configured ok Mod: INTEL
SSDSA2M040G2GC FRev: 2CV102HB SN: CVGB949301PC040GGN
unavailable disk n /devices/pci at 0,0/pci1458,b005 at 1f,2:1
sata0/2::dsk/c2t2d0 connected configured ok Mod:
ST32000542AS FRev: CC34 SN: 5XW17ARW
unavailable disk n /devices/pci at 0,0/pci1458,b005 at 1f,2:2
sata0/3::dsk/c2t3d0 connected configured ok Mod: WDC
WD20EARS-00MVWB0 FRev: 51.0AB51 SN: WD-WMAZA0555575
unavailable disk n /devices/pci at 0,0/pci1458,b005 at 1f,2:3
sata0/4::dsk/c2t4d0 connected configured ok Mod: WDC
WD20EARS-00MVWB0 FRev: 51.0AB51 SN: WD-WMAZA0484508
A zpool export and subsequent import, which should have taken the set to 4tb
overall storage in the raidz, appears to have not worked despite the import
taking what must have been about ten to fifteen minutes to do the import.
(during which time the drives were silent and the zpool process was mostly 0%
very occasionally peaking to 25%, and the system being very slow to respond
during that period)
Any ideas please? Or is there still some process running in the background
that I can't see?
mich at jaguar:~# zfs list
NAME USED AVAIL REFER MOUNTPOINT
data 2.27T 401G 2.27T /mirror
rpool 7.69G 28.7G 45K /rpool
rpool/ROOT 3.70G 28.7G 31K legacy
rpool/ROOT/openindiana 3.70G 28.7G 3.59G /
rpool/dump 1.93G 28.7G 1.93G -
rpool/export 5.22M 28.7G 32K /export
rpool/export/home 5.19M 28.7G 32K /export/home
rpool/export/home/mich 5.16M 28.7G 5.16M /export/home/mich
rpool/swap 2.05G 30.7G 126M -
mich at jaguar:~# zpool status
pool: data
state: ONLINE
scan: resilvered 1.13T in 12h26m with 0 errors on Wed Jan 19 23:42:23 2011
config:
NAME STATE READ WRITE CKSUM
data ONLINE 0 0 0
raidz1-0 ONLINE 0 0 0
c2t2d0 ONLINE 0 0 0
c2t3d0 ONLINE 0 0 0
c2t4d0 ONLINE 0 0 0
errors: No known data errors
last pid: 1802; load avg: 0.61, 0.56, 0.61; up 0+19:55:20
07:06:01
74 processes: 73 sleeping, 1 on cpu
CPU states: 99.8% idle, 0.0% user, 0.3% kernel, 0.0% iowait, 0.0% swap
Kernel: 375 ctxsw, 653 intr, 120 syscall
Memory: 3959M phys mem, 401M free mem, 1979M total swap, 1979M free swap
PID USERNAME NLWP PRI NICE SIZE RES STATE TIME CPU COMMAND
1197 gdm 1 59 0 95M 28M sleep 1:23 0.04% gdm-simple-gree
922 root 3 59 0 102M 51M sleep 0:37 0.02% Xorg
1801 root 1 59 0 4036K 2460K cpu/3 0:00 0.01% top
1196 gdm 1 59 0 80M 13M sleep 0:00 0.00% metacity
1190 gdm 1 59 0 7892K 6028K sleep 0:00 0.00% at-spi-registry
640 root 16 59 0 14M 9072K sleep 0:09 0.00% smbd
1737 mich 1 59 0 13M 5392K sleep 0:00 0.00% sshd
148 root 1 59 0 8312K 1608K sleep 0:00 0.00% dhcpagent
672 root 26 59 0 27M 15M sleep 0:01 0.00% fmd
1198 gdm 1 59 0 87M 18M sleep 0:02 0.00% gnome-power-man
1247 root 1 59 0 6080K 2500K sleep 0:00 0.00% sendmail
11 root 21 59 0 15M 13M sleep 0:06 0.00% svc.configd
272 root 6 59 0 11M 4784K sleep 0:01 0.00% devfsadm
1220 root 24 59 0 13M 4404K sleep 0:01 0.00% nscd
45 netcfg 5 59 0 4716K 3268K sleep 0:00 0.00% netcfgd
1309 admin 1 59 0 16M 8404K sleep 14:30 0.00% sshd
1312 admin 1 59 0 16M 5468K sleep 3:47 0.00% sshd
787 root 1 59 0 12M 5928K sleep 0:05 0.00% intrd
1237 root 4 59 0 8400K 1936K sleep 0:03 0.00% automountd
9 root 15 59 0 20M 12M sleep 0:02 0.00% svc.startd
387 root 5 59 0 7500K 6024K sleep 0:01 0.00% hald
1308 root 1 59 0 13M 4788K sleep 0:01 0.00% sshd
252 root 5 60 -20 2544K 1460K sleep 0:00 0.00% zonestatd
1191 gdm 1 59 0 111M 42M sleep 0:00 0.00% gnome-settings-
1175 gdm 2 59 0 20M 10M sleep 0:00 0.00% gnome-session
790 root 4 59 0 15M 6432K sleep 0:00 0.00% rad
1188 gdm 1 59 0 12M 5944K sleep 0:00 0.00% gconfd-2
291 root 1 59 0 16M 5604K sleep 0:00 0.00% cupsd
1156 mich 1 59 0 13M 5432K sleep 0:00 0.00% sshd
132 daemon 3 59 0 13M 5104K sleep 0:00 0.00% kcfd
More information about the OpenIndiana-discuss
mailing list