My rollback "finished" yesterday after about 7.5 days. It still
wasn''t ready to receive the last snapshot, so I rm''ed all the
files (took 14 hours) and then issued the rollback command again, 2 minutes this
time.
Ok, I now have many questions, some due to a couple of responses (which
don''t appear on the http://opensolaris.org/jive website)
One response was.
"I think it has been shown by others that dedup requires LOTS of RAM and
to be safe, an SSD L2ARC, especially with large (multi-TB) datasets. Dedup
is still very new, too. People seem to forget that."
The other was
"My only suggestion is if the machine is still showing any disk activity to
try adding more RAM. I don''t know this for a fact but it seems that
destroying deduped data when the dedup table doesn''t fit in RAM is
pathologically slow because the entire table is traversed for every deletion, or
at least enough of it to hit the disk on every delete.
I''ve seen a couple of people report that the process was able to
complete in a sane amount of time after adding more RAM.
This information is based on what I remember of past conversations and is all
available in the archives as well."
I currently have 4 GB of RAM, and can''t get anymore in this box (4 x 2
TB hard drives), so it sounds like I need bigger hardware. So the question is
how much more. According to one post I have read, the poster claimed that the
dedup table would fill 13.4GB for his 1.7 TB file space, assuming this is true
(8GB per 1TB), then do modern servers have enough RAM space to use dedup
effectively. Is a SSD fast enough, or does the whole DDT need to be held in RAM?
I am currently getting a planning a new file server for the company which need
to have space for approx 16 TB of files (twice what we are currently using) and
this will need to be much more focused to performance. So would the 2 solutions
have similar performance, and what results does turning on compress give?
Both will have 20 Hard disks (2 rpool, 2 SDD cache, and 14 data as mirrored
pairs, and 2 hot spares)
non- dedup.
16 x 2 TB giving 14 TB file system space ( 2 spares)
2 x 80 GB SSD cache
16 GB RAM (2 GB for system, 14GB for ZFS, is this fine for non dedup?)
dedup ( I am getting a 2.7 ratio at the moment on the secondary backup)
14 x 1 TB giving 6 TB of file system space ( dedup of 2.3 and 2 spare slots for
upgrade)
2 x 160 GB SSD cache
64 GB RAM (2GB system, 6GB ZFS, 48 DDT, yes, I know I can''t seperate
ZFS and DDT.)
The second system will be more upgradeable/future proof, but do people think the
performance would be similar?
Thanks
John
--
This message posted from opensolaris.org