[OpenIndiana-discuss] ZFS with Dedupication for NFS server

Edward Ned Harvey openindiana at nedharvey.com
Sat Apr 23 13:10:56 UTC 2011


> From: Toomas Soome [mailto:Toomas.Soome at mls.ee]
> 
> well, do a bit math. if ima correct, with 320B DTT the 1.75GB of ram can
fit
> 5.8M entries, 1TB of data, assuming 128k recordsize would produce 8M
> entries.... thats with default metadata limit.  unless i did my
calculations
> wrong, that will explain  the slowdown.

Not sure where you're getting those numbers, but rule of thumb is to add
1-3G of ram for every 1T of unique dedup data.

http://hub.opensolaris.org/bin/view/Community+Group+zfs/dedup 




More information about the OpenIndiana-discuss mailing list