[OpenIndiana-discuss] Puzzling behaviour in zpools

Jan Owoc jsowoc at gmail.com
Mon Mar 4 19:22:12 UTC 2013


Hi Len,

On Mon, Mar 4, 2013 at 11:52 AM, Len Zaifman <leonardz at sickkids.ca> wrote:
> I have a system which I am configuring for maximum space to use as a low cost backup service. It has an Areca raid card , 24 2 TB drives (Format reports 1.82 TB)  and 12 4 TB drives (format reports 3.64 TB).

Yes, that is to be expected. Manufacturers say "4TB" meaning "4 000
000 000 000" bytes. Many operating systems say "3.64 TiB" meaning
"3.64 * 1024 * 1024 * 1024 * 1024". They are two different ways of
expressing the same capacity.

> using 2 2 TB drives for a mirrored rpool, I create a 2 TB pool as 2 x  11 disk raidz2 vdevs and a 4 TB pool as 1 x 12 disk raidz2 vdev

Ok, so for each set of raidz2 disks, 2 are for parity, so you have 9
for data. 2 vdevs * 9 disks * 2TB = 36.0 TB = 32.7 TiB

On the other pool, for each set of raidz2 disks, 2 are for parity, so
you have 4 for data. 2 vdevs * 4 disks * 4TB = 32.0 TB = 29.1 TiB

> df shows
>
> rpool/ROOT/openindiana
>                       1.8T  1.6G  1.8T   1% /
> rpool/smallbkup       1.6T   31K  1.6T   1% /rpool/smallbkup
> ccmbkup12TB            32T   68K   32T   1% /ccmbkup12TB
> fourTBpool             29T   56K   29T   1% /fourTBpool    <<< I think I am 7 TB short here

See calculation above :-). It matches up exactly.

As your pools fill up, people suggest using "zfs list" to get actual
usable capacity, as df gets confused by compression, deduplication,
and other things. Your pools are empty, so the numbers happen to
match.

Cheers,
Jan



More information about the OpenIndiana-discuss mailing list