diff options
author | pjmattal <pjmattal> | 2007-06-16 12:29:43 +0000 |
---|---|---|
committer | pjmattal <pjmattal> | 2007-06-16 12:29:43 +0000 |
commit | df8c5ef53a4d4e09fbe41ba58e0a181fc0898a2e (patch) | |
tree | 4e6818bded935c503df5d6f9be58e8d1faa63ee0 /tupkg | |
parent | d397bac2d1e41a6c23ca599b4b16e0f8280f4c03 (diff) | |
download | aurweb-df8c5ef53a4d4e09fbe41ba58e0a181fc0898a2e.tar.xz |
added modified tupkgupdate and tupkgupdate64
Diffstat (limited to 'tupkg')
-rwxr-xr-x | tupkg/update/tupkgupdate | 2 | ||||
-rwxr-xr-x | tupkg/update/tupkgupdate64 | 583 |
2 files changed, 584 insertions, 1 deletions
diff --git a/tupkg/update/tupkgupdate b/tupkg/update/tupkgupdate index 5ca7477..a047d68 100755 --- a/tupkg/update/tupkgupdate +++ b/tupkg/update/tupkgupdate @@ -575,7 +575,7 @@ for package in dbremove: error("repo-remove returned an error!") sys.exit(-1) for package in dbmodify: - retval = runRepoAdd(repo_dir, package.new.file) + retval = runRepoAdd(repo_dir, package) if (retval != 0): error("repo-add returned an error!") sys.exit(-1) diff --git a/tupkg/update/tupkgupdate64 b/tupkg/update/tupkgupdate64 new file mode 100755 index 0000000..7d90074 --- /dev/null +++ b/tupkg/update/tupkgupdate64 @@ -0,0 +1,583 @@ +#!/usr/bin/python -O + +import re, os, sys, pacman, getopt +import MySQLdb, MySQLdb.connections +import ConfigParser + +########################################################### +# Deal with configuration +########################################################### + +conffile = '/home/aur/tupkgs64.conf' + +if not os.path.isfile(conffile): + print "Error: cannot access config file ("+conffile+")" + usage(argv[0]) + sys.exit(1) + +config = ConfigParser.ConfigParser() +config.read(conffile) + +############################################################ + +# Define some classes we need +class Version: + def __init__(self): + self.version = None + self.file = None + +class Package: + def __init__(self): + self.name = None + self.category = None + self.old = None + self.new = None + self.desc = None + self.url = None + self.depends = None + self.sources = None + +class PackageDatabase: + def __init__(self, host, user, password, dbname): + self.host = host + self.user = user + self.password = password + self.dbname = dbname + self.connection = MySQLdb.connect(host=host, user=user, passwd=password, db=dbname) + def cursor(self): + try: + self.connection.ping() + except MySQLdb.OperationalError: + self.connection = MySQLdb.connect(host=self.host, user=self.user, passwd=self.password, db=self.dbname) + return self.connection.cursor() + def lookup(self, packagename): + warning("DB: Looking up package: " + packagename) + q = self.cursor() + q.execute("SELECT ID FROM Packages WHERE Name = '" + + MySQLdb.escape_string(packagename) + "'") + if (q.rowcount != 0): + row = q.fetchone() + return row[0] + return None + def getCategoryID(self, package): + category_id = self.lookupCategory(package.category) + if (category_id == None): + category_id = 1 + warning("DB: Got category ID '" + str(category_id) + "' for package '" + package.name + "'") + return category_id + def insert(self, package, locationId): + warning("DB: Inserting package: " + package.name) + global repo_dir + q = self.cursor() + q.execute("INSERT INTO Packages " + + "(Name, CategoryID, Version, FSPath, LocationID, SubmittedTS, Description, URL) VALUES ('" + + MySQLdb.escape_string(package.name) + "', " + + str(self.getCategoryID(package)) + ", '" + + MySQLdb.escape_string(package.new.version) + "', '" + + MySQLdb.escape_string( + os.path.join(repo_dir, os.path.basename(package.new.file))) + "', " + + str(locationId) + ", " + + "UNIX_TIMESTAMP(), '" + + MySQLdb.escape_string(str(package.desc)) + "', '" + + MySQLdb.escape_string(str(package.url)) + "')") + id = self.lookup(package.name) + self.insertNewInfo(package, id, locationId) + def update(self, id, package, locationId): + warning("DB: Updating package: " + package.name + " with id " + str(id)) + global repo_dir + q = self.cursor() + if (self.isdummy(package.name)): + q.execute("UPDATE Packages SET " + + "Version = '" + MySQLdb.escape_string(package.new.version) + "', " + + "CategoryID = " + str(self.getCategoryID(package)) + ", " + + "FSPath = '" + MySQLdb.escape_string( + os.path.join(repo_dir, os.path.basename(package.new.file))) + "', " + + "Description = '" + MySQLdb.escape_string(str(package.desc)) + "', " + + "DummyPkg = 0, " + + "SubmittedTS = UNIX_TIMESTAMP(), " + + "URL = '" + MySQLdb.escape_string(str(package.url)) + "' " + + "WHERE ID = " + str(id)) + else: + q.execute("UPDATE Packages SET " + + "Version = '" + MySQLdb.escape_string(package.new.version) + "', " + + "CategoryID = " + str(self.getCategoryID(package)) + ", " + + "FSPath = '" + MySQLdb.escape_string( + os.path.join(repo_dir, os.path.basename(package.new.file))) + "', " + + "Description = '" + MySQLdb.escape_string(str(package.desc)) + "', " + + "ModifiedTS = UNIX_TIMESTAMP(), " + + "URL = '" + MySQLdb.escape_string(str(package.url)) + "' " + + "WHERE ID = " + str(id)) + self.insertNewInfo(package, id, locationId) + # we must lastly check to see if this is a move of a package from + # unsupported to community, because we'd have to reset maintainer and location + q = self.cursor() + q.execute("SELECT LocationID FROM Packages WHERE ID = " + str(id)) + if (q.rowcount != 0): + row = q.fetchone() + if (row[0] != 3): + q = self.cursor() + q.execute("UPDATE Packages SET LocationID = 3, MaintainerUID = null WHERE ID = " + str(id)) + def remove(self, id, locationId): + warning("DB: Removing package with id: " + str(id)) + q = self.cursor() + q.execute("DELETE FROM Packages WHERE " + + "LocationID = " + str(locationId) + " AND ID = " + str(id)) + def clearOldInfo(self, id): + warning("DB: Clearing old info for package with id : " + str(id)) + q = self.cursor() + q.execute("DELETE FROM PackageContents WHERE PackageID = " + str(id)) + q.execute("DELETE FROM PackageDepends WHERE PackageID = " + str(id)) + q.execute("DELETE FROM PackageSources WHERE PackageID = " + str(id)) + def lookupOrDummy(self, packagename): + retval = self.lookup(packagename) + if (retval != None): + return retval + return self.createDummy(packagename) + def lookupCategory(self, categoryname): + warning("DB: Looking up category: " + categoryname) + q = self.cursor() + q.execute("SELECT ID from PackageCategories WHERE Category = '" + MySQLdb.escape_string(categoryname) + "'") + if (q.rowcount != 0): + row = q.fetchone() + return row[0] + return None + def createDummy(self, packagename): + warning("DB: Creating dummy package for: " + packagename) + q = self.cursor() + q.execute("INSERT INTO Packages " + + "(Name, Description, LocationID, DummyPkg) " + + "VALUES ('" + + MySQLdb.escape_string(packagename) + "', '" + + MySQLdb.escape_string("A dummy package") + "', 1, 1)") + return self.lookup(packagename) + def insertNewInfo(self, package, id, locationId): + q = self.cursor() + + # first delete the old; this is never bad + self.clearOldInfo(id) + + warning("DB: Inserting new package info for " + package.name + + " with id " + str(id)) + + # PackageSources + for source in package.sources: + q.execute("INSERT INTO PackageSources (PackageID, Source) " + + "VALUES (" + str(id) + ", '" + source + "')") + # PackageDepends + for dep in package.depends: + depid = self.lookupOrDummy(dep) + q.execute("INSERT INTO PackageDepends (PackageID, DepPkgID) " + + "VALUES (" + str(id) + ", " + str(depid) + ")") + def isdummy(self, packagename): + warning("DB: Looking up package: " + packagename) + q = self.cursor() + q.execute("SELECT * FROM Packages WHERE Name = '" + + MySQLdb.escape_string(packagename) + "' AND DummyPkg = 1") + if (q.rowcount != 0): + return True + return False + +############################################################ +# Functions for walking the file trees +############################################################ + +def filesForRegexp(topdir, regexp): + retval = [] + def matchfile(regexp, dirpath, namelist): + for name in namelist: + if (regexp.match(name)): + retval.append(os.path.join(dirpath, name)) + os.path.walk(topdir, matchfile, regexp) + return retval + +def packagesInTree(topdir): + return filesForRegexp(topdir, re.compile("^.*\.pkg\.tar\.gz$")) + +def pkgbuildsInTree(topdir): + return filesForRegexp(topdir, re.compile("^PKGBUILD$")) + +############################################################ +# Function for testing if two files are identical +############################################################ + +def areFilesIdentical(file_a, file_b): + command = "cmp '" + file_a + "' '" + file_b + "' >/dev/null" + retval = os.system(command) + if (retval == 0): + return True + return False + +############################################################ +# Function for fetching info from PKGBUILDs and packages +############################################################ + +def infoFromPackageFile(filename): + pkg = os.path.basename(filename) + m = re.compile("(?P<pkgname>.*)-(?P<pkgver>.*)-(?P<pkgrel>.*).pkg.tar.gz").search(pkg) + if not m: + raise Exception("Non-standard filename") + else: + return m.group('pkgname'), m.group('pkgver') + "-" + m.group('pkgrel') + +def infoFromPkgbuildFile(filename): + # first grab the category based on the file path + pkgdirectory = os.path.dirname(filename) + catdirectory = os.path.dirname(pkgdirectory) + m = re.match(r".*/([^/]+)$", catdirectory) + if (m): + category = m.group(1) + else: + category = "none" + + # open and source the file + pf_stdin, pf_stdout = os.popen2("/bin/bash", 't', 0) + print >>pf_stdin, ". " + filename + #print "PKGBUILD: " + filename + + # get pkgname + print >>pf_stdin, 'echo $pkgname' + pkgname = pf_stdout.readline().strip() + #print "PKGBUILD: pkgname: " + pkgname + + # get pkgver + print >>pf_stdin, 'echo $pkgver' + pkgver = pf_stdout.readline().strip() + #print "PKGBUILD: pkgver: " + pkgver + + # get pkgrel + print >>pf_stdin, 'echo $pkgrel' + pkgrel = pf_stdout.readline().strip() + #print "PKGBUILD: pkgrel: " + pkgrel + + # get url + print >>pf_stdin, 'echo $url' + url = pf_stdout.readline().strip() + #print "PKGBUILD: url: " + url + + # get desc + print >>pf_stdin, 'echo $pkgdesc' + pkgdesc = pf_stdout.readline().strip() + #print "PKGBUILD: pkgdesc: " + pkgdesc + + # get source array + print >>pf_stdin, 'echo ${source[*]}' + source = (pf_stdout.readline().strip()).split(" ") + + # get depends array + print >>pf_stdin, 'echo ${depends[*]}' + depends = (pf_stdout.readline().strip()).split(" ") + + # clean up + pf_stdin.close() + pf_stdout.close() + + return pkgname, pkgver + "-" + pkgrel, pkgdesc, url, depends, source, category + +def infoFromPkgbuildFileWorse(filename): + # load the file with pacman library + pkg = pacman.load(filename) + return (pkg.name, pkg.version + "-" + pkg.release, pkg.desc, + pkg.url, pkg.depends, pkg.source) + +############################################################ +# Functions for doing the final steps of execution +############################################################ + +def execute(command): + global switches + print(command) + if not (switches.get("-n") == True): + return os.system(command) + return 0 + +def copyFileToRepo(filename, repodir): + destfile = os.path.join(repodir, os.path.basename(filename)) + command = "cp --preserve=timestamps '" + filename + "' '" + destfile + "'" + return execute(command) + +def deleteFile(filename): + command = "rm '" + filename + "'" + return execute(command) + +def runGensync(repo, pkgbuild): +#target = os.path.join(repo, os.path.basename(repo) + ".db.tar.gz") + target = os.path.join(repo, "community.db.tar.gz") + command = "gensync '" + pkgbuild + "' '" + target + "'" + return execute(command) + +def runUpdatesyncUpd(repo, pkgbuild): + global havefakeroot + targetDB = os.path.join(repo, "community.db.tar.gz") + if havefakeroot: + command = "fakeroot updatesync upd '" + targetDB + "' '" + pkgbuild + "' '" + repo +"'" + else: + command = "updatesync upd '" + targetDB + "' '" + pkgbuild + "' '" + repo +"'" + return execute(command) + +def runUpdatesyncDel(repo, pkgname): + global havefakeroot + targetDB = os.path.join(repo, "community.db.tar.gz") + if havefakeroot: + command = "fakeroot updatesync del '" + targetDB + "' '" + pkgname +"'" + else: + command = "updatesync del '" + targetDB + "' '" + pkgname +"'" + return execute(command) + +############################################################ +# Functions for error handling +############################################################ + +def warning(string): + print >>sys.stderr, string + +had_error = 0 +def error(string): + global had_error + warning(string) + had_error = 1 + +############################################################ +# MAIN +############################################################ + +# ARGUMENTS +# +# tupkgupdate [-n] [--delete] [--paranoid] <repo_dir> <pkgbuild_dir> <build_dir> + +# First call getopt +switch_list,args_proper = getopt.getopt(sys.argv[1:], 'n', + [ "delete", "paranoid" ]) +switches = {} +for switch in switch_list: + switches[switch[0]] = 1 + +# Then handle the remaining arguments +if (len(args_proper) < 3): + print >>sys.stderr, "syntax: tupkgupdate64 [-n] [--delete] [--paranoid] <repo_dir> <pkgbuild_tree> <build_tree>" + sys.exit(-1) + +# Make sure we can use fakeroot, warn if not +havefakeroot = False +if os.access('/usr/bin/fakeroot', os.X_OK): + havefakeroot = True +else: + warning("Not using fakeroot for repo db generation") + +repo_dir, pkgbuild_dir, build_dir = args_proper + +# Open the database so we find out now if we can't! +db = PackageDatabase(config.get('mysql', 'host'), + config.get('mysql', 'username'), + config.get('mysql', 'password'), + config.get('mysql', 'db')) + +# Set up the lists and tables +packages = dict() +copy = list() +delete = list() + +dbremove = list() +dbmodify = list() + +# PASS 1: PARSING/LOCATING +# +# A) Go through the PKGBUILD tree +# For each PKGBUILD, create a Package with new Version containing +# parsed version and and None for file + +a_files = pkgbuildsInTree(pkgbuild_dir) +for a_file in a_files: + pkgname, ver, desc, url, depends, sources, category = infoFromPkgbuildFile(a_file) + + # Error (and skip) if we encounter any invalid PKGBUILD files + if (pkgname == None or ver == None): + error("Pkgbuild '" + a_file + "' is invalid!") + continue + + # Error (and skip) if we encounter any duplicate package names + # in the PKGBUILDs + if (packages.get(pkgname)): + error("Pkgbuild '" + a_file + "' is a duplicate!") + continue + + version = Version() + version.version = ver + version.file = None + + package = Package() + package.name = pkgname + package.category = category + package.desc = desc + package.url = url + package.depends = depends + package.sources = sources + package.new = version + +# print "Package: desc " + desc + + packages[pkgname] = package + +# B) Go through the old repo dir +# For each package file we encounter, create a Package with old +# Version containing parsed version and filepath + +b_files = packagesInTree(repo_dir) +for b_file in b_files: + pkgname, ver = infoFromPackageFile(b_file) + + version = Version() + version.version = ver + version.file = b_file + + package = packages.get(pkgname) + if (package == None): + package = Package() + package.name = pkgname + packages[pkgname] = package + package.old = version + +# C) Go through the build tree +# For each package file we encounter: +# 1 - look up the package name; if it fails, ignore the file (no error) +# 2 - if package.new == None, ignore the package (no error) +# 3 - if package.new.version doesn't match, then skip (no error) +# 4 - if package.new.file == None, point it to this file +# otherwise, log an error (and skip) + +c_files = packagesInTree(build_dir) +for c_file in c_files: + pkgname, ver = infoFromPackageFile(c_file) + + # 1 + package = packages.get(pkgname) + if (package == None): + continue + + # 2 + if (package.new == None): + continue + + # 3 + if (package.new.version != ver): + continue + + # 4 + if (package.new.file == None): + package.new.file = c_file + continue + else: + error("Duplicate new file '" + c_file + "'") + continue + +# PASS 2: CHECKING +# +# Go through the package collection +# 1 - if package has no new, place its old file on the "delete" list (and package on "dbremove") +# 2 - if package has a new but no new.file, and old file doesn't +# have the same version, then error (because gensync won't rebuild) +# 3 - if package has no old, add new file to "copy" list into repo dir (and package on "dbmodify") +# 4 - if new == old and paranoid is set, compare the files and error if not the same; +# otherwise just skip (no update) +# 5 - if we got here, it's a legit nontrivial new version which we allow +# add entry to "delete" list for old file and "copy" list for +# new file into repo dir (and package to "dbmodify") + +for package in packages.values(): + # 1 + if (package.new == None): + delete.append(package.old.file) + dbremove.append(package) + continue + + # 2 + if (package.new.file == None): + if (package.old == None or package.old.file == None or + package.old.version != package.new.version): + errstr = "No new package supplied for " + package.name + " " + package.new.version + "!" + error(errstr) + continue + + # 3 + if (package.old == None): + copy.append(package.new.file) + dbmodify.append(package) + continue + + # 4 + if (package.old.version == package.new.version): + if (switches.get("--paranoid") == True and package.new.file != None): + if not (areFilesIdentical(package.old.file, package.new.file)): + warning("New package file with identical version '" + + package.new.file + "' is different than the old one:") + if (switches.get("--delete") == True): + warning(" Deleting the new file.") + delete.append(package.new.file) + else: + warning(" Ignoring the new file.") + continue + + # 5 + delete.append(package.old.file) + copy.append(package.new.file) + dbmodify.append(package) + continue + +## IF WE HAVE HAD ANY ERRORS AT THIS POINT, ABORT! ## +if (had_error == 1): + error("Aborting due to errors.") + sys.exit(-1) + +# PASS 3: EXECUTION +# + +# First, do all the database updates +# We won't do these for x86_64 - jason Oct 1/2006 +#for package in dbremove: +# id = db.lookup(package.name) +# # Note: this could remove a package from unsupported; probably want to restrict to locationId and/or non-dummy +# if (id != None): +# db.clearOldInfo(id) +# db.remove(id, 3) +# +#for package in dbmodify: +# warning("DB: Package in dbmodify: " + package.name) +# id = db.lookup(package.name) +# if (id == None): +# db.insert(package, 3) +# else: +# db.update(id, package, 3) + +# Copy +for file in copy: + retval = copyFileToRepo(file, repo_dir) + if (retval != 0): + error("Could not copy file to repo: '" + file + "'") + sys.exit(-1) +# Delete (second, for safety's sake) +for file in delete: + deleteFile(file) +# Now that we've copied new files and deleted, we should delete the source +# files, if we're supposed to +if (switches.get("--delete") == True): + for file in copy: + deleteFile(file) +# Run gensync to build the repo index +#if (len(copy) + len(delete) > 0): +# retval = runGensync(repo_dir, pkgbuild_dir) +# if (retval != 0): +# error("Gensync returned an error!") +# sys.exit(-1) + +# Run updatesync where it is needed +for package in dbremove: + retval = runUpdatesyncDel(repo_dir, package.name) + if (retval != 0): + error("Updatesync del returned an error!") + sys.exit(-1) +for package in dbmodify: + retval = runUpdatesyncUpd(repo_dir, os.path.join(pkgbuild_dir,package.category,package.name,"PKGBUILD")) + if (retval != 0): + error("Updatesync upd returned an error!") + sys.exit(-1) + +# vim: ft=python ts=2 sw=2 noet |