[OpenIndiana-discuss] Zpool upgrade didn't seem to upgrade
Edward Martinez
mindbender_1 at live.com
Thu Jan 20 08:16:21 UTC 2011
On 01/19/11 23:05, Michelle Knight wrote:
> hi Folks,
>
> OI 148.
>
> Three 1.5tb drives were replaced with three 2tb drives. They are here on
> internal SATA channels c2t2d0, c2t3d0 and c2t4d0. One is a Seagate Barracuda
> and the other two are Western Digital Greens.
>
> mich at jaguar:~# cfgadm -lv
> Ap_Id Receptacle Occupant Condition
> Information
> When Type Busy Phys_Id
> Slot8 connected configured ok Location:
> Slot8
> Jan 1 1970 unknown n /devices/pci at 0,0/pci8086,3b4a at 1c,4:Slot8
> sata0/0::dsk/c2t0d0 connected configured ok Mod: INTEL
> SSDSA2M040G2GC FRev: 2CV102HB SN: CVGB949301PH040GGN
> unavailable disk n /devices/pci at 0,0/pci1458,b005 at 1f,2:0
> sata0/1::dsk/c2t1d0 connected configured ok Mod: INTEL
> SSDSA2M040G2GC FRev: 2CV102HB SN: CVGB949301PC040GGN
> unavailable disk n /devices/pci at 0,0/pci1458,b005 at 1f,2:1
> sata0/2::dsk/c2t2d0 connected configured ok Mod:
> ST32000542AS FRev: CC34 SN: 5XW17ARW
> unavailable disk n /devices/pci at 0,0/pci1458,b005 at 1f,2:2
> sata0/3::dsk/c2t3d0 connected configured ok Mod: WDC
> WD20EARS-00MVWB0 FRev: 51.0AB51 SN: WD-WMAZA0555575
> unavailable disk n /devices/pci at 0,0/pci1458,b005 at 1f,2:3
> sata0/4::dsk/c2t4d0 connected configured ok Mod: WDC
> WD20EARS-00MVWB0 FRev: 51.0AB51 SN: WD-WMAZA0484508
>
> A zpool export and subsequent import, which should have taken the set to 4tb
> overall storage in the raidz, appears to have not worked despite the import
> taking what must have been about ten to fifteen minutes to do the import.
> (during which time the drives were silent and the zpool process was mostly 0%
> very occasionally peaking to 25%, and the system being very slow to respond
> during that period)
>
> Any ideas please? Or is there still some process running in the background
> that I can't see?
>
> mich at jaguar:~# zfs list
> NAME USED AVAIL REFER MOUNTPOINT
> data 2.27T 401G 2.27T /mirror
> rpool 7.69G 28.7G 45K /rpool
> rpool/ROOT 3.70G 28.7G 31K legacy
> rpool/ROOT/openindiana 3.70G 28.7G 3.59G /
> rpool/dump 1.93G 28.7G 1.93G -
> rpool/export 5.22M 28.7G 32K /export
> rpool/export/home 5.19M 28.7G 32K /export/home
> rpool/export/home/mich 5.16M 28.7G 5.16M /export/home/mich
> rpool/swap 2.05G 30.7G 126M -
>
>
> mich at jaguar:~# zpool status
> pool: data
> state: ONLINE
> scan: resilvered 1.13T in 12h26m with 0 errors on Wed Jan 19 23:42:23 2011
> config:
>
> NAME STATE READ WRITE CKSUM
> data ONLINE 0 0 0
> raidz1-0 ONLINE 0 0 0
> c2t2d0 ONLINE 0 0 0
> c2t3d0 ONLINE 0 0 0
> c2t4d0 ONLINE 0 0 0
>
> errors: No known data errors
>
>
> last pid: 1802; load avg: 0.61, 0.56, 0.61; up 0+19:55:20
> 07:06:01
> 74 processes: 73 sleeping, 1 on cpu
> CPU states: 99.8% idle, 0.0% user, 0.3% kernel, 0.0% iowait, 0.0% swap
> Kernel: 375 ctxsw, 653 intr, 120 syscall
> Memory: 3959M phys mem, 401M free mem, 1979M total swap, 1979M free swap
>
> PID USERNAME NLWP PRI NICE SIZE RES STATE TIME CPU COMMAND
> 1197 gdm 1 59 0 95M 28M sleep 1:23 0.04% gdm-simple-gree
> 922 root 3 59 0 102M 51M sleep 0:37 0.02% Xorg
> 1801 root 1 59 0 4036K 2460K cpu/3 0:00 0.01% top
> 1196 gdm 1 59 0 80M 13M sleep 0:00 0.00% metacity
> 1190 gdm 1 59 0 7892K 6028K sleep 0:00 0.00% at-spi-registry
> 640 root 16 59 0 14M 9072K sleep 0:09 0.00% smbd
> 1737 mich 1 59 0 13M 5392K sleep 0:00 0.00% sshd
> 148 root 1 59 0 8312K 1608K sleep 0:00 0.00% dhcpagent
> 672 root 26 59 0 27M 15M sleep 0:01 0.00% fmd
> 1198 gdm 1 59 0 87M 18M sleep 0:02 0.00% gnome-power-man
> 1247 root 1 59 0 6080K 2500K sleep 0:00 0.00% sendmail
> 11 root 21 59 0 15M 13M sleep 0:06 0.00% svc.configd
> 272 root 6 59 0 11M 4784K sleep 0:01 0.00% devfsadm
> 1220 root 24 59 0 13M 4404K sleep 0:01 0.00% nscd
> 45 netcfg 5 59 0 4716K 3268K sleep 0:00 0.00% netcfgd
> 1309 admin 1 59 0 16M 8404K sleep 14:30 0.00% sshd
> 1312 admin 1 59 0 16M 5468K sleep 3:47 0.00% sshd
> 787 root 1 59 0 12M 5928K sleep 0:05 0.00% intrd
> 1237 root 4 59 0 8400K 1936K sleep 0:03 0.00% automountd
> 9 root 15 59 0 20M 12M sleep 0:02 0.00% svc.startd
> 387 root 5 59 0 7500K 6024K sleep 0:01 0.00% hald
> 1308 root 1 59 0 13M 4788K sleep 0:01 0.00% sshd
> 252 root 5 60 -20 2544K 1460K sleep 0:00 0.00% zonestatd
> 1191 gdm 1 59 0 111M 42M sleep 0:00 0.00% gnome-settings-
> 1175 gdm 2 59 0 20M 10M sleep 0:00 0.00% gnome-session
> 790 root 4 59 0 15M 6432K sleep 0:00 0.00% rad
> 1188 gdm 1 59 0 12M 5944K sleep 0:00 0.00% gconfd-2
> 291 root 1 59 0 16M 5604K sleep 0:00 0.00% cupsd
> 1156 mich 1 59 0 13M 5432K sleep 0:00 0.00% sshd
> 132 daemon 3 59 0 13M 5104K sleep 0:00 0.00% kcfd
> _______________________________________________
> OpenIndiana-discuss mailing list
> OpenIndiana-discuss at openindiana.org
> http://openindiana.org/mailman/listinfo/openindiana-discuss
>
>
SSD and HD can not be mixed.
Regards
Edward
More information about the OpenIndiana-discuss
mailing list