diff options
author | 2017-02-17 16:32:26 -0800 | |
---|---|---|
committer | 2017-02-17 16:32:26 -0800 | |
commit | c74dfc328cc2d4cdb5f65755aa7e4b337dc5565d (patch) | |
tree | 0d4f976416d58bf6625f8eea126ff862ce6a2908 /tools/run_tests/run_microbenchmark.py | |
parent | dc55e1fed0bce8be0c70a7b39978fbf1b3f10abe (diff) | |
parent | ece502fa12ae9db7fc969f0bae1f7c26ece1dce3 (diff) |
Merge pull request #9782 from ctiller/faster-latency
Speed up latency profiling (and tune down the memory usage)
Diffstat (limited to 'tools/run_tests/run_microbenchmark.py')
-rwxr-xr-x | tools/run_tests/run_microbenchmark.py | 6 |
1 files changed, 4 insertions, 2 deletions
diff --git a/tools/run_tests/run_microbenchmark.py b/tools/run_tests/run_microbenchmark.py index 262c05b421..d709cfc92a 100755 --- a/tools/run_tests/run_microbenchmark.py +++ b/tools/run_tests/run_microbenchmark.py @@ -93,7 +93,9 @@ def collect_latency(bm_name, args): '--benchmark_list_tests']).splitlines(): link(line, '%s.txt' % fnize(line)) benchmarks.append( - jobset.JobSpec(['bins/basicprof/%s' % bm_name, '--benchmark_filter=^%s$' % line], + jobset.JobSpec(['bins/basicprof/%s' % bm_name, + '--benchmark_filter=^%s$' % line, + '--benchmark_min_time=0.05'], environ={'LATENCY_TRACE': '%s.trace' % fnize(line)})) profile_analysis.append( jobset.JobSpec([sys.executable, @@ -105,7 +107,7 @@ def collect_latency(bm_name, args): # consume upwards of five gigabytes of ram in some cases, and so analysing # hundreds of them at once is impractical -- but we want at least some # concurrency or the work takes too long - if len(benchmarks) >= min(4, multiprocessing.cpu_count()): + if len(benchmarks) >= min(16, multiprocessing.cpu_count()): # run up to half the cpu count: each benchmark can use up to two cores # (one for the microbenchmark, one for the data flush) jobset.run(benchmarks, maxjobs=max(1, multiprocessing.cpu_count()/2), |