Mercurial > p > roundup > code
view test/benchmark.py @ 7879:39c482e6a246
fix: fix code to make tests of session and otks databases pass on windows
The existing code had a few issues:
sessions_dbm.py:
detect dumbdbm when used on windows python and properly delete session/otks
databases so clear() works.
make sure the Session/Otks.cache_db_type() is called even when the database
is newly created.
test/session_common.py:
close session and otks database in teardown before deleting database
directory to prevent errors from deleting open files on windows.
test/test_sqlite.py:
close the session and otks databases opened by SessionTest.setUp(self)
so the salite session/otks database are closed before the new anydbm
sessions databases are opened for testing the sqlite main db and anydbm
as session/otks db. Again this causes deletion of database test directory
to fail on windows as you can't delete open files.
| author | John Rouillard <rouilj@ieee.org> |
|---|---|
| date | Thu, 18 Apr 2024 17:39:56 -0400 |
| parents | 7a3392f1f7ac |
| children | 9c3ec0a5c7fc |
line wrap: on
line source
""" Usage: python benchmark.py ["database backend list" | backend1] [backend2] Import the backend (anydbm, sqlite by default) and run some performance tests. Example: test default anypy and sqlite backends python benchmark.py test mysql and sqlite backends python benchmark.py mysql sqlite or python benchmark.py "mysql sqlite" test all backends python benchmark.py anydbm mysql postgresql sqlite """ from __future__ import print_function import sys, os, time import importlib, signal, shutil # --- patch sys.path to make sure 'import roundup' finds correct version import os.path as osp thisdir = osp.dirname(osp.abspath(__file__)) rootdir = osp.dirname(thisdir) if (osp.exists(thisdir + '/benchmark.py') and osp.exists(rootdir + '/roundup/__init__.py')): # the script is located inside roundup source code sys.path.insert(0, rootdir) from roundup.hyperdb import String, Password, Link, Multilink, Date, \ Interval, DatabaseError, Boolean, Number from roundup import date, password from test.db_test_base import config # global for the default signal hander so # my signal handler can reset before it raises signal. int_sig_default_handler = None def setupSchema(db, module): status = module.Class(db, "status", name=String()) status.setkey("name") user = module.Class(db, "user", username=String(), password=Password(), assignable=Boolean(), age=Number(), roles=String()) user.setkey("username") file = module.FileClass(db, "file", name=String(), type=String(), comment=String(indexme="yes")) issue = module.IssueClass(db, "issue", title=String(indexme="yes"), status=Link("status"), nosy=Multilink("user"), deadline=Date(), foo=Interval(), files=Multilink("file"), assignedto=Link('user')) session = module.Class(db, 'session', title=String()) session.disableJournalling() db.post_init() db.commit() def rm_db_on_signal(sig, frame): print("removing incomplete database %s due to interruption." % config.DATABASE) shutil.rmtree(config.DATABASE) signal.signal(signal.SIGINT, int_sig_default_handler) # re-raise the signal so the normal signal handling runs. signal.raise_signal(signal.SIGTERM) def main(backendname, time=time.time, numissues=10): global int_sig_default_handler try: backend = importlib.import_module("roundup.backends.back_%s" % backendname) except ImportError: print("Unable to import %s backend." % backendname) return times = [] config.DATABASE = os.path.join('_benchmark', '%s-%s'%(backendname, numissues)) config.RDBMS_NAME = "rounduptest_%s" % numissues if not os.path.exists(config.DATABASE): int_sig_default_handler = signal.signal(signal.SIGINT, rm_db_on_signal) db = backend.Database(config, 'admin') setupSchema(db, backend) # if we are re-initializing, delete any existing db db.clear() db.commit() # create a whole bunch of stuff db.user.create(**{'username': 'admin', 'roles': 'Admin'}) db.status.create(name="unread") db.status.create(name="in-progress") db.status.create(name="testing") db.status.create(name="resolved") pc = -1 for i in range(numissues): db.user.create(**{'username': 'user %s'%i, 'roles': 'User'}) for j in range(10): db.user.set(str(i+1), assignable=1) db.user.set(str(i+1), assignable=0) db.issue.create(**{'title': 'issue %s'%i}) for j in range(10): db.issue.set(str(i+1), status='2', assignedto='2', nosy=[]) db.issue.set(str(i+1), status='1', assignedto='1', nosy=['1','2']) if (i*100//numissues) != pc and 'INCI' not in os.environ: pc = (i*100//numissues) sys.stdout.write("%d%%\r"%pc) sys.stdout.flush() db.commit() signal.signal(signal.SIGINT, int_sig_default_handler) else: db = backend.Database(config, 'admin') setupSchema(db, backend) sys.stdout.write('%10s: %-6d'%(backendname[:10], numissues)) sys.stdout.flush() times.append(('start', time())) # fetch db.clearCache() for i in db.issue.list(): db.issue.get(i, 'title') times.append(('fetch', time())) # journals db.clearCache() for i in db.issue.list(): db.issue.history(i) times.append(('journal', time())) # "calculated" props db.clearCache() for i in db.issue.list(): db.issue.get(i, 'activity') db.issue.get(i, 'creator') db.issue.get(i, 'creation') times.append(('jprops', time())) # lookup db.clearCache() for i in range(numissues): db.user.lookup('user %s'%i) times.append(('lookup', time())) # filter db.clearCache() for i in range(100): db.issue.filter(None, {'assignedto': '1', 'title':'issue'}, ('+', 'activity'), ('+', 'status')) times.append(('filter', time())) # filter with multilink db.clearCache() for i in range(100): db.issue.filter(None, {'nosy': ['1'], 'assignedto': '1', 'title':'issue'}, ('+', 'activity'), ('+', 'status')) times.append(('filtml', time())) # results last = None for event, stamp in times: if last is None: first = stamp else: sys.stdout.write(' %-6.2f'%(stamp-last)) last = stamp print(' %-6.2f'%(last-first)) sys.stdout.flush() if __name__ == '__main__': if len(sys.argv) == 2: test_databases = sys.argv[1].split() elif len(sys.argv) > 2: test_databases = sys.argv[1:] else: test_databases = ['anydbm', 'sqlite'] # 0 1 2 3 4 5 6 # 01234567890123456789012345678901234567890123456789012345678901234 print('Test name fetch journl jprops lookup filter filtml TOTAL ') for name in test_databases: main(name) for name in test_databases: main(name, numissues=20) for name in test_databases: main(name, numissues=100) # don't even bother benchmarking the dbm backends > 100! try: test_databases.remove('anydbm') except ValueError: # anydbm not present; this is fine pass for name in test_databases: main(name, numissues=1000) for name in test_databases: main(name, numissues=10000) # vim: set et sts=4 sw=4 :
