diff options
author | Caroline Tice <cmtice@google.com> | 2017-04-06 17:16:05 -0700 |
---|---|---|
committer | chrome-bot <chrome-bot@chromium.org> | 2017-04-08 19:40:20 -0700 |
commit | f6ef4395fe1896ba68c80e52cb24763b0fcfe7f8 (patch) | |
tree | 7612123711db98f0746e56b66368d16b388192bd /crosperf/benchmark_run.py | |
parent | 8c18be1425c8a4ecfc059a7c637fc93f33edab1f (diff) | |
download | toolchain-utils-f6ef4395fe1896ba68c80e52cb24763b0fcfe7f8.tar.gz |
[toolchain-utils] Fix remaining lint errors in toolchain-utils.
In addition to fixing the lint errors, this also fixes the Python
formatting issues (ran tc_pyformat on nearly all the files).
BUG=chromium:570450
TEST=Ran all crosperf & bisect tool unit tests.
Ran afe_lock_machine.py (check machine status)
Ran full crosperf test (octane, speedometer, BootPerf) on alex.
Change-Id: Ic86f9192801ac67769f3de30f1c5f0d203ce0831
Reviewed-on: https://chromium-review.googlesource.com/471886
Commit-Ready: Caroline Tice <cmtice@chromium.org>
Tested-by: Caroline Tice <cmtice@chromium.org>
Reviewed-by: Manoj Gupta <manojgupta@chromium.org>
Diffstat (limited to 'crosperf/benchmark_run.py')
-rw-r--r-- | crosperf/benchmark_run.py | 26 |
1 files changed, 13 insertions, 13 deletions
diff --git a/crosperf/benchmark_run.py b/crosperf/benchmark_run.py index e53187e2..bba71a36 100644 --- a/crosperf/benchmark_run.py +++ b/crosperf/benchmark_run.py @@ -1,8 +1,6 @@ - # Copyright (c) 2013 The Chromium OS Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. - """Module of benchmark runs.""" from __future__ import print_function @@ -30,6 +28,7 @@ STATUS_PENDING = 'PENDING' class BenchmarkRun(threading.Thread): """The benchmarkrun class.""" + def __init__(self, name, benchmark, label, iteration, cache_conditions, machine_manager, logger_to_use, log_level, share_cache): threading.Thread.__init__(self) @@ -53,8 +52,8 @@ class BenchmarkRun(threading.Thread): self.test_args = benchmark.test_args self.cache = None self.profiler_args = self.GetExtraAutotestArgs() - self._ce = command_executer.GetCommandExecuter(self._logger, - log_level=self.log_level) + self._ce = command_executer.GetCommandExecuter( + self._logger, log_level=self.log_level) self.timeline = timeline.Timeline() self.timeline.Record(STATUS_PENDING) self.share_cache = share_cache @@ -96,8 +95,7 @@ class BenchmarkRun(threading.Thread): err = 'No cache hit.' self.result = Result.CreateFromRun( self._logger, self.log_level, self.label, self.machine, output, err, - retval, self.benchmark.test_name, - self.benchmark.suite) + retval, self.benchmark.test_name, self.benchmark.suite) else: self._logger.LogOutput('%s: No cache hit.' % self.name) @@ -141,8 +139,8 @@ class BenchmarkRun(threading.Thread): pass elif self.machine: if not self.machine.IsReachable(): - self._logger.LogOutput('Machine %s is not reachable, removing it.' % - self.machine.name) + self._logger.LogOutput( + 'Machine %s is not reachable, removing it.' % self.machine.name) self.machine_manager.RemoveMachine(self.machine.name) self._logger.LogOutput('Releasing machine: %s' % self.machine.name) self.machine_manager.ReleaseMachine(self.machine) @@ -190,8 +188,10 @@ class BenchmarkRun(threading.Thread): perf_args = ' '.join(perf_args_list) if not perf_args_list[0] in ['record', 'stat']: raise SyntaxError('perf_args must start with either record or stat') - extra_test_args = ['--profiler=custom_perf', - ("--profiler_args='perf_options=\"%s\"'" % perf_args)] + extra_test_args = [ + '--profiler=custom_perf', + ("--profiler_args='perf_options=\"%s\"'" % perf_args) + ] return ' '.join(extra_test_args) else: return '' @@ -254,9 +254,9 @@ class MockBenchmarkRun(BenchmarkRun): self.timeline.Record(STATUS_IMAGING) self.machine_manager.ImageMachine(machine, self.label) self.timeline.Record(STATUS_RUNNING) - [retval, out, err] = self.suite_runner.Run(machine.name, self.label, - self.benchmark, self.test_args, - self.profiler_args) + [retval, out, + err] = self.suite_runner.Run(machine.name, self.label, self.benchmark, + self.test_args, self.profiler_args) self.run_completed = True rr = MockResult('logger', self.label, self.log_level, machine) rr.out = out |