| | """Benchmark cache.incr method. |
| | """ |
| |
|
| | import json |
| | import multiprocessing as mp |
| | import shutil |
| | import time |
| |
|
| | import diskcache as dc |
| |
|
| | from .utils import secs |
| |
|
| | COUNT = int(1e3) |
| | PROCS = 8 |
| |
|
| |
|
| | def worker(num): |
| | """Rapidly increment key and time operation.""" |
| | time.sleep(0.1) |
| |
|
| | cache = dc.Cache('tmp') |
| | values = [] |
| |
|
| | for _ in range(COUNT): |
| | start = time.time() |
| | cache.incr(b'key') |
| | end = time.time() |
| | values.append(end - start) |
| |
|
| | with open('output-%s.json' % num, 'w') as writer: |
| | json.dump(values, writer) |
| |
|
| |
|
| | def main(): |
| | """Run workers and print percentile results.""" |
| | shutil.rmtree('tmp', ignore_errors=True) |
| |
|
| | processes = [ |
| | mp.Process(target=worker, args=(num,)) for num in range(PROCS) |
| | ] |
| |
|
| | for process in processes: |
| | process.start() |
| |
|
| | for process in processes: |
| | process.join() |
| |
|
| | with dc.Cache('tmp') as cache: |
| | assert cache.get(b'key') == COUNT * PROCS |
| |
|
| | for num in range(PROCS): |
| | values = [] |
| | with open('output-%s.json' % num) as reader: |
| | values += json.load(reader) |
| |
|
| | values.sort() |
| | p50 = int(len(values) * 0.50) - 1 |
| | p90 = int(len(values) * 0.90) - 1 |
| | p99 = int(len(values) * 0.99) - 1 |
| | p00 = len(values) - 1 |
| | print(['{0:9s}'.format(val) for val in 'p50 p90 p99 max'.split()]) |
| | print([secs(values[pos]) for pos in [p50, p90, p99, p00]]) |
| |
|
| |
|
| | if __name__ == '__main__': |
| | main() |
| |
|