Edward Ned Harvey (opensolarisisdeadlongliveopensolaris)
2012-Oct-16 13:54 UTC
[zfs-discuss] openindiana-1 filesystem, time-slider, and snapshots
Can anyone explain to me what the openindiana-1 filesystem is all about? I thought it was the "backup" copy of the openindiana filesystem, when you apply OS updates, but that doesn''t seem to be the case... I have time-slider enabled for rpool/ROOT/openindiana. It has a daily snapshot (amongst others). But every day when the new daily snap is taken, the old daily snap rotates into the rpool/ROOT/openindiana-1 filesystem. This is messing up my cron-scheduled "zfs send" script - which detects that the rpool/ROOT/openindiana filesystem no longer has the old daily snapshot, and therefore has no snapshot in common with the receiving system, and therefore sends a new full backup every night. To make matters more confusing, when I run "mount" and when I zfs get all | grep -i mount, I see / on rpool/ROOT/openindiana-1 It would seem, I shouldn''t be backing up openindiana, but instead, backup openindiana-1? I would have sworn, out-of-the-box, there was no openindiana-1. Am I simply wrong? My expectation is that rpool/ROOT/openindiana should have lots of snaps available... 3 frequent: one every 15 mins, 23 hourly: one every hour, 6 daily: one every day, 4 weekly: one every 7 days, etc. I checked to ensure auto-snapshot service is enabled. I checked svccfg to ensure I understood the correct interval, keep, and period (as described above.) I have the expected behavior (as I described, the expected behavior according to my expectations) on rpool/export/home/eharvey... But the behavior is different on rpool/ROOT/openindiana, even though, as far as I can tell, I have the same settings for both. That is, simply, com.sun:auto-snapshot=true One more comment - I recall, when I first configured time-slider, they have a threshold, default 80% pool used before they automatically bump off old snapshots (or stop taking new snaps, I''m not sure what the behavior is). I don''t see that setting anywhere I look, using svccfg or zfs get. My pools are pretty much empty right now. Nowhere near the 80% limit. -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://mail.opensolaris.org/pipermail/zfs-discuss/attachments/20121016/6f1e72fa/attachment-0001.html>
Darren J Moffat
2012-Oct-16 16:01 UTC
[zfs-discuss] openindiana-1 filesystem, time-slider, and snapshots
On 10/16/12 14:54, Edward Ned Harvey (opensolarisisdeadlongliveopensolaris) wrote:> Can anyone explain to me what the openindiana-1 filesystem is all > about?I thought it was the "backup" copy of the openindiana filesystem, > when you apply OS updates, but that doesn''t seem to be the case... > > I have time-slider enabled for rpool/ROOT/openindiana.It has a daily > snapshot (amongst others).But every day when the new daily snap is > taken, the old daily snap rotates into the rpool/ROOT/openindiana-1 > filesystem.This is messing up my cron-scheduled "zfs send" script - > which detects that the rpool/ROOT/openindiana filesystem no longer has > the old daily snapshot, and therefore has no snapshot in common with the > receiving system, and therefore sends a new full backup every night. > > To make matters more confusing, when I run "mount" and when I zfs get > all | grep -i mount, I see / on rpool/ROOT/openindiana-1It is a new boot environment see beadm(1M) - you must have done some ''pkg update'' or ''pkg install'' option that created a new BE.> It would seem, I shouldn''t be backing up openindiana, but instead, > backup openindiana-1?I would have sworn, out-of-the-box, there was no > openindiana-1.Am I simply wrong?Initially there wouldn''t have been. Are you doing the zfs send on your own or letting time-slider do it for you ? -- Darren J Moffat
Jason
2012-Oct-16 19:54 UTC
[zfs-discuss] openindiana-1 filesystem, time-slider, and snapshots
Ahh.. the illumos lists change the from address (or set reply-to), so that a plain reply simply works. Sorry about that :) But yeah, when you update you probably want to snap/backup your current BE. It won''t hurt anything to snap/backup the old ones, just that they''ll contain the non-updated versions of everything. On Tue, Oct 16, 2012 at 2:48 PM, Edward Ned Harvey (opensolarisisdeadlongliveopensolaris) < opensolarisisdeadlongliveopensolaris at nedharvey.com> wrote:> > From: jason.brian.king at gmail.com [mailto:jason.brian.king at gmail.com] On > > Behalf Of Jason > > > > When you update our install (i.e. pkg image-update) unless you explicitly > > specify a BE name, it names the BEs as ''openindiana-NN''. The way that > > pkg(5) works is that it snapshot+clones your current BE (i.e. > > rpool/ROOT/openindiana if that is your current BE), then applies the > updates > > to the clone. You can then use beadm to active the new one, or switch > back, > > etc. It is unrelated to time slider (other than both will create zfs > snapshots). > > > > That you have openindiana-1 suggests you''ve run pkg image-update at least > > once (either via cmdline or GUI). beadm should tell you which one is > active > > (i.e. you''re running on), also you can also do df -h / or mount and see > which > > fs is actually mounted as / > > (I just noticed this reply was sent to me off-list. Do you mind resending > it on-list, and I''ll send this reply again on-list?) > > That agrees with what I originally thought. So here is the confusion: > > zfs list -t snapshot | grep ''^rpool/ROOT/openindiana@'' | grep daily > rpool/ROOT/openindiana at zfs-auto-snap_daily-2012-10-16-12h04 > 0 - 3.25G - > > Every day, the above changes. And every day, a new entry appears below: > > zfs list -t snapshot | grep ''^rpool/ROOT/openindiana-1@'' | grep daily > rpool/ROOT/openindiana-1 at zfs-auto-snap_daily-2012-10-12-12h04 > 307K - 3.36G - > rpool/ROOT/openindiana-1 at zfs-auto-snap_daily-2012-10-13-12h04 > 563K - 3.36G - > rpool/ROOT/openindiana-1 at zfs-auto-snap_daily-2012-10-14-12h04 > 443K - 3.37G - > rpool/ROOT/openindiana-1 at zfs-auto-snap_daily-2012-10-15-12h04 > 1.59M - 3.37G - > rpool/ROOT/openindiana-1 at zfs-auto-snap_daily-2012-10-16-12h04 > 324K - 3.38G - > > I ran "beadm list" as suggested, and it confirmed what I''ve been > suspecting since starting this thread - The current BE is openindiana-1 > > It would seem, I simply need to be aware, I should backup whatever is > specified as "NR" by beadm. Because a new filesystem was created and > became the place where files actually are being updated... It is not safe > for me to simply backup "openindiana" and keep it that way moving forward. > Every time I "pkg update" I need to be aware, to update my backup scripts > accordingly, to backup the new BE instead of the old one. > > Sound about right? > >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://mail.opensolaris.org/pipermail/zfs-discuss/attachments/20121016/99d877b5/attachment.html>