aboutsummaryrefslogtreecommitdiff
path: root/android_bench_suite
diff options
context:
space:
mode:
Diffstat (limited to 'android_bench_suite')
-rw-r--r--android_bench_suite/autotest.diff8
-rwxr-xr-xandroid_bench_suite/fix_skia_results.py2
-rwxr-xr-xandroid_bench_suite/gen_json.py2
-rwxr-xr-xandroid_bench_suite/run.py4
4 files changed, 8 insertions, 8 deletions
diff --git a/android_bench_suite/autotest.diff b/android_bench_suite/autotest.diff
index ef0029ae..c2fed83b 100644
--- a/android_bench_suite/autotest.diff
+++ b/android_bench_suite/autotest.diff
@@ -49,7 +49,7 @@ index 000000000..b233b586a
+ 'binderThroughputTest > /data/local/tmp/bench_result'
+ % os.getenv('TEST_MODE'))
+ # Next 4 runs add to bench_result
-+ for i in xrange(4):
++ for i in range(4):
+ self.client.run('taskset %s /data/local/tmp/'
+ 'binderThroughputTest >> '
+ '/data/local/tmp/bench_result'
@@ -165,11 +165,11 @@ index 000000000..dd6af0b53
+ 'bench_result'), 'w') as f:
+
+ # There are two benchmarks, chrome and camera.
-+ for i in xrange(2):
++ for i in range(2):
+ f.write('Test %d:\n' % i)
+ total_time = 0
+ # Run benchmark for several times for accurancy
-+ for j in xrange(3):
++ for j in range(3):
+ f.write('Iteration %d: ' % j)
+ result = self.client.run('time taskset %s dex2oat'
+ ' --dex-file=data/local/tmp/dex2oat_input/test%d.apk'
@@ -739,7 +739,7 @@ index 000000000..b317bd0f3
+ ' > /data/local/tmp/bench_result'
+ % os.getenv('TEST_MODE'))
+ # Next 4 runs add to bench_result
-+ for i in xrange(4):
++ for i in range(4):
+ self.client.run('taskset %s /data/local/tmp/synthmark'
+ ' >> /data/local/tmp/bench_result'
+ % os.getenv('TEST_MODE'))
diff --git a/android_bench_suite/fix_skia_results.py b/android_bench_suite/fix_skia_results.py
index 6eec6ccf..8c919d35 100755
--- a/android_bench_suite/fix_skia_results.py
+++ b/android_bench_suite/fix_skia_results.py
@@ -111,7 +111,7 @@ def _TransformBenchmarks(raw_benchmarks):
if len(results) < len(samples):
results.extend({
'retval': 0
- } for _ in xrange(len(samples) - len(results)))
+ } for _ in range(len(samples) - len(results)))
time_mul = _GetTimeMultiplier(friendly_name)
for sample, app in itertools.izip(samples, results):
diff --git a/android_bench_suite/gen_json.py b/android_bench_suite/gen_json.py
index ad617ff4..e1252933 100755
--- a/android_bench_suite/gen_json.py
+++ b/android_bench_suite/gen_json.py
@@ -81,7 +81,7 @@ def main(argv):
iteration = arguments.iterations
result = []
- for i in xrange(iteration):
+ for i in range(iteration):
result += collect_data(infile, bench, i)
with get_outfile(outfile, bench) as fout:
diff --git a/android_bench_suite/run.py b/android_bench_suite/run.py
index 55acb663..19d9b36f 100755
--- a/android_bench_suite/run.py
+++ b/android_bench_suite/run.py
@@ -301,7 +301,7 @@ def test_bench(bench, setting_no, iterations, serials, remote, mode):
logging.info('Start running benchmark on device...')
# Run benchmark and tests on DUT
- for i in xrange(iterations):
+ for i in range(iterations):
logging.info('Iteration No.%d:', i)
test_cmd = [
os.path.join(
@@ -463,7 +463,7 @@ def main(argv):
for bench in bench_list:
logging.info('Start building and running benchmark: [%s]', bench)
# Run script for each toolchain settings
- for setting_no in xrange(setting_count):
+ for setting_no in range(setting_count):
build_bench(setting_no, bench, compiler, llvm_version, build_os, cflags,
ldflags)