X-Git-Url: http://git.sthu.org/?a=blobdiff_plain;f=paralleljobs.py;h=5c309cd7e7118f677161ebbbd0e7519b81cb3727;hb=fbca1279e2e56e71b7aa9a15a285e26375ed400a;hp=6eb92dda40b803a9543ea161360170e4f50ead0a;hpb=76576fca186993bcd55c521bd6018019fd0cc5d7;p=paralleljobs.git diff --git a/paralleljobs.py b/paralleljobs.py index 6eb92dd..5c309cd 100755 --- a/paralleljobs.py +++ b/paralleljobs.py @@ -1,4 +1,4 @@ -#!/usr/bin/env python +#!/usr/bin/env python3 """ A simple tool to run jobs from a database in parallel.""" __author__ = "Stefan Huber" @@ -23,6 +23,9 @@ def printStatusInfo(conn): c.execute("SELECT count(id) FROM jobs WHERE done=1;") nodone, = c.fetchone() + c.execute("SELECT count(id) FROM jobs WHERE started=1;") + nostarted, = c.fetchone() + c.execute("SELECT sum(workloadestm) FROM jobs WHERE done=1;") wldone, = c.fetchone() if wldone == None: @@ -31,13 +34,23 @@ def printStatusInfo(conn): c.execute("SELECT sum(workloadestm) FROM jobs;") wltotal, = c.fetchone() - c.close() - perdone = 100.0*float(nodone)/float(nototal) - perwl = 100.0*float(wldone)/float(wltotal) + perdone = 0 + perwl = 0 + if nototal > 0: + perdone = 100.0*float(nodone)/float(nototal) + if wltotal > 0: + perwl = 100.0*float(wldone)/float(wltotal) + + print("%d (%.1f%%) of %d jobs and %.1f%% of the workload done. %d jobs are running." % \ + (nodone, perdone, nototal, perwl, nostarted-nodone)) - print("%d (%.1f%%) of %d jobs done. %.1f%% of the workload finished." % \ - (nodone, perdone, nototal, perwl)) + print("Exit code stats:") + c.execute("SELECT exitcode, count(exitcode) AS cnt FROM jobs GROUP BY exitcode ORDER BY exitcode ASC;") + for code, cnt in c.fetchall(): + print(" %3s: %6s (%5.1f%%)" % (code, cnt, 100.0*float(cnt)/nodone)) + + c.close() def createPropertiesTable(conn, propdef): conn.execute("BEGIN EXCLUSIVE") @@ -152,29 +165,29 @@ def usage(): """Print usage text of this program""" print(""" -Take the jobs defined in jobs table of the given database and process one job +Take the jobs defined in the jobs table of the given database and process one after the other. Multiple instances may be launched against the same database. Usage: - {0} [OPTIONS] [COMMANDS] -d database + {0} [OPTIONS] [COMMANDS] -d DB {0} -h COMMANDS: - -c cmdfn add jobs from the file with list of commands + -c FILE add each line as a job resp. job's command to DB -h print this text -s print status information - -w work on the database + -w do work and process jobs OPTIONS: - -d database the database to process - -n num in -w mode, only perform num-many jobs - -p cols-def create properties table with SQL column spec - -v print output of the job's command + -d DB the database to process + -n NUM in -w mode, only process num-many jobs + -p COL-DEF create properties table with SQL column spec + -v verbose output Commands may be combined in one call of {0}. A list of jobs may be importet line-by-line from a file using the -c option. -Every job may output to stdout or stderr a string of the form +Every job may output to stdout or stderr one or more strings of the form DB-PROPERTIES: {{ "key": "value", "key2": 1.23, "key3": True }} It is assumed that a table 'properties' exists with the columns jobid, key, key2, and key3. The corresponding values are inserted into this table. Using @@ -183,7 +196,7 @@ column definitions in SQL style. The jobs table also contains a 'workloadestm' column that is used when estimating the finished workload so far. The entries default to 1 and may be -set externally. +adjusted. Examples: # create cmds.sh with jobs @@ -243,7 +256,7 @@ if __name__ == "__main__": print("No database given.") sys.exit(os.EX_USAGE) - conn = sqlite3.connect(dbfn) + conn = sqlite3.connect(dbfn, timeout=60) createSchema(conn) if status: