One of the joys of working remotely is that you don't generally have access to your own local repository of distro spins and packages for doing installs. This makes virtualization development a real pain because it seems like you're always trying to whip up some VM or other and then having to download 800MB of packages from some slow mirror. The alternative to this is to maintain your own local mirror. Yum has an utility called "reposync" that supposedly does this for you. However it is not very smart -- if the same package is in f7-i386 and f7-x86_64, for example, which happens frequently, reposync will happily download both copies. Dan Berrange patched reposync to fix this at some point, although the patches didn't make it upstream as far as I know. What the patched version does is check any package directory you specify to see if the package you are about to download exists. If it does, rather than downloading it again, reposync hardlinks to it, so that as far as yum knows you have the same package in both directories. This saves many gigabytes of storage and bandwidth. To make all this work you need a current version of yum-utils (I believe that's the package name), plus the patched reposync (attached). I run the thing from a script that I have also attached to keep current copies of f7, f7-updates, f8, f8-updates, and rawhide (all for both i386 and x86_64). Dan, because he is the man, also grabs SRPMS, but I don't have the mojo for that so I don't bother with it. I run this from a cron job at 7 p.m. daily and it is usually finished by the time I come in the next morning (unless there has been a major release the night before). You need to set up yum.conf files for each repository you sync to tell reposync where to look for the packages. You can pick your favorite mirror here or just use download.fedora.redhat.com. Also note that if you need isos in a repo you have to download them separately, but you really only have to do that once so no biggie. Enjoy, --Hugh -------------- next part -------------- #!/usr/bin/python -tt # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # copyright 2006 Duke University # author seth vidal # sync all or the newest packages from a repo to the local path # TODO: # have it print out list of changes # make it work with mirrorlists (silly, really) # man page/more useful docs # deal nicely with a package changing but not changing names (ie: replacement) # criteria # if a package is not the same and smaller then reget it # if a package is not the same and larger, delete it and get it again # always replace metadata files if they're not the same. import os import sys import shutil import stat from optparse import OptionParser from urlparse import urljoin import yum import yum.Errors from yum.misc import getCacheDir from yum.constants import * from yum.packages import parsePackages from yum.packageSack import ListPackageSack import rpmUtils.arch import logging from urlgrabber.progress import TextMeter # for yum 2.4.X compat def sortPkgObj(pkg1 ,pkg2): """sorts a list of yum package objects by name""" if pkg1.name > pkg2.name: return 1 elif pkg1.name == pkg2.name: return 0 else: return -1 class RepoSync(yum.YumBase): def __init__(self, opts): yum.YumBase.__init__(self) self.logger = logging.getLogger('yum.verbose.reposync') self.opts = opts def localpkgs(dir): names = os.listdir(dir) cache = {} for name in names: fn = os.path.join(dir, name) try: st = os.lstat(fn) except os.error: continue if stat.S_ISDIR(st.st_mode): subcache= localpkgs(fn) for pkg in subcache.keys(): cache[pkg] = subcache[pkg] elif stat.S_ISREG(st.st_mode) and name.endswith(".rpm"): cache[name] = { 'path': fn, 'size': st.st_size, 'device': st.st_dev } return cache def parseArgs(): usage = """ Reposync is used to synchronize a remote yum repository to a local directory using yum to retrieve the packages. %s [options] """ % sys.argv[0] parser = OptionParser(usage=usage) parser.add_option("-c", "--config", default='/etc/yum.conf', help='config file to use (defaults to /etc/yum.conf)') parser.add_option("-a", "--arch", default=None, help='act as if running the specified arch (default: current arch, note: does not override $releasever)') parser.add_option("-r", "--repoid", default=[], action='append', help="specify repo ids to query, can be specified multiple times (default is all enabled)") parser.add_option("-e", "--cachedir", help="directory in which to store metadata") parser.add_option("-t", "--tempcache", default=False, action="store_true", help="Use a temp dir for storing/accessing yum-cache") parser.add_option("-d", "--delete", default=False, action="store_true", help="delete local packages no longer present in repository") parser.add_option("-k", "--package-cache", default=[], dest='pkgcache', action='append', help="additional directory to search for pre-existing packages") parser.add_option("-p", "--download_path", dest='destdir', default=os.getcwd(), help="Path to download packages to: defaults to current dir") parser.add_option("-g", "--gpgcheck", default=False, action="store_true", help="Remove packages that fail GPG signature checking after downloading") parser.add_option("-u", "--urls", default=False, action="store_true", help="Just list urls of what would be downloaded, don't download") parser.add_option("-n", "--newest-only", dest='newest', default=False, action="store_true", help="Download only newest packages per-repo") parser.add_option("-q", "--quiet", default=False, action="store_true", help="Output as little as possible") (opts, args) = parser.parse_args() return (opts, args) def main(): (opts, junk) = parseArgs() if not os.path.exists(opts.destdir) and not opts.urls: try: os.makedirs(opts.destdir) except OSError, e: print >> sys.stderr, "Error: Cannot create destination dir %s" % opts.destdir sys.exit(1) if not os.access(opts.destdir, os.W_OK) and not opts.urls: print >> sys.stderr, "Error: Cannot write to destination dir %s" % opts.destdir sys.exit(1) my = RepoSync(opts=opts) my.doConfigSetup(fn=opts.config, init_plugins=False) # Populate cache of existing download RPMs from other # repositories we can link to pkgcache = {} for dir in opts.pkgcache: cache = localpkgs(dir) for k in cache.keys(): pkgcache[k] = cache[k] # Force unprivileged users to have a private temporary cachedir # if they've not given an explicit cachedir if os.getuid() != 0 and not opts.cachedir: opts.tempcache = True if opts.tempcache: cachedir = getCacheDir() if cachedir is None: print >> sys.stderr, "Error: Could not make cachedir, exiting" sys.exit(50) my.repos.setCacheDir(cachedir) elif opts.cachedir: my.repos.setCacheDir(opts.cachedir) if len(opts.repoid) > 0: myrepos = [] # find the ones we want for glob in opts.repoid: myrepos.extend(my.repos.findRepos(glob)) # disable them all for repo in my.repos.repos.values(): repo.disable() # enable the ones we like for repo in myrepos: repo.enable() # Use progress bar display when downloading repo metadata # and package files if not opts.quiet: my.repos.setProgressBar(TextMeter(fo=sys.stdout)) my.doRpmDBSetup() my.doRepoSetup() my.doSackSetup(rpmUtils.arch.getArchList(opts.arch)) for repo in my.repos.listEnabled(): reposack = ListPackageSack(my.pkgSack.returnPackages(repoid=repo.id)) if opts.newest: download_list = reposack.returnNewestByNameArch() else: download_list = list(reposack) local_repo_path = opts.destdir + '/' + repo.id # make sure the repo subdir is here before we go on. if not os.path.exists(local_repo_path): try: os.makedirs(local_repo_path) except IOError, e: my.logger.error("Could not make repo subdir: %s" % e) my.closeRpmDB() sys.exit(1) # Check if there's any local files no longer on the remote # repo which need purging if opts.delete: current_pkgs = localpkgs(local_repo_path) download_set = {} for pkg in download_list: remote = pkg.returnSimple('relativepath') rpmname = os.path.basename(remote) download_set[rpmname] = 1 for pkg in current_pkgs: if download_set.has_key(pkg): continue if not opts.quiet: my.logger.info("Removing obsolete %s", pkg) os.unlink(current_pkgs[pkg]['path']) download_list.sort(sortPkgObj) n = 0 for pkg in download_list: n = n + 1 repo = my.repos.getRepo(pkg.repoid) remote = pkg.returnSimple('relativepath') rpmname = os.path.basename(remote) local = local_repo_path + '/' + remote localdir = os.path.dirname(local) pkg.localpath = local # Hack: to set the localpath we want. if not os.path.exists(localdir): os.makedirs(localdir) # If we have a local RPM with same name, and it is # on the same storage device, and it has same size # then we can hardlink it into local dir. if (not os.path.exists(local) and pkgcache.has_key(rpmname) and os.stat(local_repo_path).st_dev == pkgcache[rpmname]['device'] and pkgcache[rpmname]['size'] == int(pkg.returnSimple('packagesize'))): if not opts.quiet: my.logger.info("[%s: %-5d of %-5d ] Linking existing %s" % (repo.id, n, len(download_list), remote)) os.link(pkgcache[rpmname]['path'], local) # Optionally check gpg signature of local package if os.path.exists(local) and opts.gpgcheck: result, error = my.sigCheckPkg(pkg) if result != 0: if not opts.quiet: my.logger.error("[%s: %-5d of %-5d ] Removing non-matching %s" % (repo.id, n, len(download_list), remote)) os.unlink(local) # If we have a local pkg with same name, check its size if (os.path.exists(local)): if os.path.getsize(local) == int(pkg.returnSimple('packagesize')): if not opts.quiet: my.logger.error("[%s: %-5d of %-5d ] Skipping existing %s" % (repo.id, n, len(download_list), remote)) continue # If we're just printing URLs, skip to next repo if opts.urls: url = urljoin(repo.urls[0],remote) print '%s' % url continue # Disable cache otherwise things won't download repo.cache = 0 if not opts.quiet: my.logger.info( '[%s: %-5d of %-5d ] Downloading %s' % (repo.id, n, len(download_list), remote)) path = repo.getPackage(pkg) if opts.gpgcheck: result, error = my.sigCheckPkg(pkg) if result != 0: if result == 1: my.logger.warning('Removing %s, due to missing GPG key.' % rpmname) elif result == 2: my.logger.warning('Removing %s due to failed signature check.' % rpmname) else: my.logger.warning('Removing %s due to failed signature check: %s' % (rpmname, error)) os.unlink(path) continue if not os.path.exists(local) or not os.path.samefile(path, local): shutil.copy2(path, local) my.closeRpmDB() if __name__ == "__main__": main() -------------- next part -------------- A non-text attachment was scrubbed... Name: reposync-fedora.sh Type: application/x-sh Size: 1590 bytes Desc: not available URL: <http://listman.redhat.com/archives/ovirt-devel/attachments/20080401/8f377e87/attachment.sh>