On 03/31/2016 09:53 AM, m.roth at 5-cent.us wrote:> Oddity: rsync *should* be recursing, and dealing with very large number of
> files. It works, going from box a to box b. But when I try to back b up to
> c, it fails, 100%, complaining of "out of hashtable space
[sender]". I've
> tried adding -r, and changing --delete to --delete-delay, and no joy.
>
> All boxes are current, or fairly current, CentOS 7.
The only thing I know of that's likely to cause rsync to run out of memory
is when there are a huge number of hard links and you are using the
"-H"
option to preserve them. (If you think you don't have many, look under
/var/lib/yum/yumdb and /usr/share/zoneinfo.)
And FYI, rsync doesn't do a very good job of preserving hard links when
going to a destination that already has some of the files. It doesn't
break any existing hard links at the destination, so you can end up with
a hard link topology that is somewhat different from the source. I have
to run some very messy audits to make sure all copies of my backups have
the same arrangement of hard-linked files.
--
Bob Nichols "NOSPAM" is really part of my email address.
Do NOT delete it.