diff options
-rwxr-xr-x | afdo_redaction/redact_profile_test.py | 2 | ||||
-rw-r--r-- | android_bench_suite/autotest.diff | 8 | ||||
-rwxr-xr-x | android_bench_suite/fix_skia_results.py | 2 | ||||
-rwxr-xr-x | android_bench_suite/gen_json.py | 2 | ||||
-rwxr-xr-x | android_bench_suite/run.py | 4 | ||||
-rwxr-xr-x | binary_search_tool/test/cmd_script.py | 2 | ||||
-rwxr-xr-x | buildbot_test_toolchains.py | 2 | ||||
-rwxr-xr-x | cros_utils/buildbot_json.py | 6 | ||||
-rw-r--r-- | cros_utils/tabulator.py | 10 | ||||
-rw-r--r-- | crosperf/results_report.py | 6 | ||||
-rwxr-xr-x | crosperf/results_report_unittest.py | 2 |
11 files changed, 23 insertions, 23 deletions
diff --git a/afdo_redaction/redact_profile_test.py b/afdo_redaction/redact_profile_test.py index 5ccde3e4..27fb534e 100755 --- a/afdo_redaction/redact_profile_test.py +++ b/afdo_redaction/redact_profile_test.py @@ -55,7 +55,7 @@ def _generate_repeated_function_body(repeats, fn_name='_some_name'): num_width = len(str(repeats)) lines = [] - for i in xrange(repeats): + for i in range(repeats): num = str(i).zfill(num_width) lines.append(num + function_header) lines.extend(function_body) diff --git a/android_bench_suite/autotest.diff b/android_bench_suite/autotest.diff index ef0029ae..c2fed83b 100644 --- a/android_bench_suite/autotest.diff +++ b/android_bench_suite/autotest.diff @@ -49,7 +49,7 @@ index 000000000..b233b586a + 'binderThroughputTest > /data/local/tmp/bench_result' + % os.getenv('TEST_MODE')) + # Next 4 runs add to bench_result -+ for i in xrange(4): ++ for i in range(4): + self.client.run('taskset %s /data/local/tmp/' + 'binderThroughputTest >> ' + '/data/local/tmp/bench_result' @@ -165,11 +165,11 @@ index 000000000..dd6af0b53 + 'bench_result'), 'w') as f: + + # There are two benchmarks, chrome and camera. -+ for i in xrange(2): ++ for i in range(2): + f.write('Test %d:\n' % i) + total_time = 0 + # Run benchmark for several times for accurancy -+ for j in xrange(3): ++ for j in range(3): + f.write('Iteration %d: ' % j) + result = self.client.run('time taskset %s dex2oat' + ' --dex-file=data/local/tmp/dex2oat_input/test%d.apk' @@ -739,7 +739,7 @@ index 000000000..b317bd0f3 + ' > /data/local/tmp/bench_result' + % os.getenv('TEST_MODE')) + # Next 4 runs add to bench_result -+ for i in xrange(4): ++ for i in range(4): + self.client.run('taskset %s /data/local/tmp/synthmark' + ' >> /data/local/tmp/bench_result' + % os.getenv('TEST_MODE')) diff --git a/android_bench_suite/fix_skia_results.py b/android_bench_suite/fix_skia_results.py index 6eec6ccf..8c919d35 100755 --- a/android_bench_suite/fix_skia_results.py +++ b/android_bench_suite/fix_skia_results.py @@ -111,7 +111,7 @@ def _TransformBenchmarks(raw_benchmarks): if len(results) < len(samples): results.extend({ 'retval': 0 - } for _ in xrange(len(samples) - len(results))) + } for _ in range(len(samples) - len(results))) time_mul = _GetTimeMultiplier(friendly_name) for sample, app in itertools.izip(samples, results): diff --git a/android_bench_suite/gen_json.py b/android_bench_suite/gen_json.py index ad617ff4..e1252933 100755 --- a/android_bench_suite/gen_json.py +++ b/android_bench_suite/gen_json.py @@ -81,7 +81,7 @@ def main(argv): iteration = arguments.iterations result = [] - for i in xrange(iteration): + for i in range(iteration): result += collect_data(infile, bench, i) with get_outfile(outfile, bench) as fout: diff --git a/android_bench_suite/run.py b/android_bench_suite/run.py index 55acb663..19d9b36f 100755 --- a/android_bench_suite/run.py +++ b/android_bench_suite/run.py @@ -301,7 +301,7 @@ def test_bench(bench, setting_no, iterations, serials, remote, mode): logging.info('Start running benchmark on device...') # Run benchmark and tests on DUT - for i in xrange(iterations): + for i in range(iterations): logging.info('Iteration No.%d:', i) test_cmd = [ os.path.join( @@ -463,7 +463,7 @@ def main(argv): for bench in bench_list: logging.info('Start building and running benchmark: [%s]', bench) # Run script for each toolchain settings - for setting_no in xrange(setting_count): + for setting_no in range(setting_count): build_bench(setting_no, bench, compiler, llvm_version, build_os, cflags, ldflags) diff --git a/binary_search_tool/test/cmd_script.py b/binary_search_tool/test/cmd_script.py index 6940eaae..eb91fe9b 100755 --- a/binary_search_tool/test/cmd_script.py +++ b/binary_search_tool/test/cmd_script.py @@ -45,7 +45,7 @@ def Main(argv): if opt_bisect_limit == -1: opt_bisect_limit = total_pass - for i in xrange(1, total_pass + 1): + for i in range(1, total_pass + 1): bisect_str = 'BISECT: %srunning pass (%d) Combine redundant ' \ 'instructions on function (f1)' \ % ('NOT ' if i > opt_bisect_limit else '', i) diff --git a/buildbot_test_toolchains.py b/buildbot_test_toolchains.py index c2d88733..169b5150 100755 --- a/buildbot_test_toolchains.py +++ b/buildbot_test_toolchains.py @@ -129,7 +129,7 @@ class ToolchainComparator(object): assert mo image_dict = mo.groupdict() image_dict['image_type'] = 'chrome-pfq' - for _ in xrange(2): + for _ in range(2): image_dict['tip'] = str(int(image_dict['tip']) - 1) nonafdo_image = PFQ_IMAGE_FS.replace('\\', '').format(**image_dict) if buildbot_utils.DoesImageExist(self._chromeos_root, nonafdo_image): diff --git a/cros_utils/buildbot_json.py b/cros_utils/buildbot_json.py index 8a9d9cb8..42a27744 100755 --- a/cros_utils/buildbot_json.py +++ b/cros_utils/buildbot_json.py @@ -316,7 +316,7 @@ class NonAddressableNodeList(VirtualNodeList): # pylint: disable=W0223 @property def cached_children(self): if self.parent.cached_data is not None: - for i in xrange(len(self.parent.cached_data[self.subkey])): + for i in range(len(self.parent.cached_data[self.subkey])): yield self[i] @property @@ -352,7 +352,7 @@ class NonAddressableNodeList(VirtualNodeList): # pylint: disable=W0223 def __iter__(self): """Enables 'for i in obj:'. It returns children.""" if self.data: - for i in xrange(len(self.data)): + for i in range(len(self.data)): yield self[i] def __getitem__(self, key): @@ -868,7 +868,7 @@ class Builds(AddressableNodeList): # Only cache keys here. self.cache_keys() if self._keys: - for i in xrange(max(self._keys), -1, -1): + for i in range(max(self._keys), -1, -1): yield self[i] def cache_keys(self): diff --git a/cros_utils/tabulator.py b/cros_utils/tabulator.py index e2f27bc4..59e4d426 100644 --- a/cros_utils/tabulator.py +++ b/cros_utils/tabulator.py @@ -340,17 +340,17 @@ class SamplesTableGenerator(TableGenerator): row = [None] * len(header) row[0] = '%s (samples)' % k row[1] = 'N/A' - for label_index in xrange(2, len(row)): + for label_index in range(2, len(row)): row[label_index] = [0] * iterations for cur_row in table[1:]: # Iterate through each benchmark if len(cur_row) > 1: - for label_index in xrange(2, len(cur_row)): + for label_index in range(2, len(cur_row)): # Iterate through each run in a single benchmark # each result should look like ((pass, fail), [values_list]) bench_runs = cur_row[label_index][1] - for index in xrange(iterations): + for index in range(iterations): # Accumulate each run result to composite benchmark run # If any run fails, then we set this run for composite benchmark # to None so that we know it fails. @@ -360,11 +360,11 @@ class SamplesTableGenerator(TableGenerator): row[label_index][index] = None else: # One benchmark totally fails, no valid data will be in final result - for label_index in xrange(2, len(row)): + for label_index in range(2, len(row)): row[label_index] = [None] * iterations break # Calculate pass and fail count for composite benchmark - for label_index in xrange(2, len(row)): + for label_index in range(2, len(row)): run_pass = 0 run_fail = 0 for run in row[label_index]: diff --git a/crosperf/results_report.py b/crosperf/results_report.py index afd767aa..5f49872b 100644 --- a/crosperf/results_report.py +++ b/crosperf/results_report.py @@ -90,7 +90,7 @@ def _AppendUntilLengthIs(gen, the_list, target_len): Uses `gen` to generate elements. """ - the_list.extend(gen() for _ in xrange(target_len - len(the_list))) + the_list.extend(gen() for _ in range(target_len - len(the_list))) return the_list @@ -132,7 +132,7 @@ class _PerfTable(object): self.perf_data = {} for label in label_names: for bench_name, bench_iterations in benchmark_names_and_iterations: - for i in xrange(bench_iterations): + for i in range(bench_iterations): report = read_perf_report(label, bench_name, i) self._ProcessPerfReport(report, label, bench_name, i) @@ -177,7 +177,7 @@ def _ParseColumn(columns, iteration): else: new_column.extend( Column(LiteralResult(i), Format(), str(i + 1)) - for i in xrange(iteration)) + for i in range(iteration)) return new_column diff --git a/crosperf/results_report_unittest.py b/crosperf/results_report_unittest.py index 750dcdf4..f2fc9f60 100755 --- a/crosperf/results_report_unittest.py +++ b/crosperf/results_report_unittest.py @@ -141,7 +141,7 @@ def _InjectSuccesses(experiment, how_many, keyvals, for_benchmark=0, return run experiment.benchmark_runs.extend( - MakeSuccessfulRun(n) for n in xrange(how_many)) + MakeSuccessfulRun(n) for n in range(how_many)) return experiment |