#!/usr/bin/env python3 # -*- coding: utf-8 -*- # Copyright (c) 2011 The Chromium OS Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """Module of result cache unittest.""" from __future__ import print_function import io import os import shutil import tempfile import unittest import unittest.mock as mock import image_checksummer import machine_manager import test_flag from label import MockLabel from results_cache import CacheConditions from results_cache import PerfDataReadError from results_cache import PidVerificationError from results_cache import Result from results_cache import ResultsCache from results_cache import TelemetryResult from cros_utils import command_executer from cros_utils import logger from cros_utils import misc # pylint: disable=line-too-long OUTPUT = """CMD (True): ./test_that.sh\ --remote=172.17.128.241 --board=lumpy LibCBench CMD (None): cd /usr/local/google/home/yunlian/gd/src/build/images/lumpy/latest/../../../../..; cros_sdk -- ./in_chroot_cmd6X7Cxu.sh Identity added: /tmp/test_that.PO1234567/autotest_key (/tmp/test_that.PO1234567/autotest_key) INFO : Using emerged autotests already installed at /build/lumpy/usr/local/autotest. INFO : Running the following control files 1 times: INFO : * 'client/site_tests/platform_LibCBench/control' INFO : Running client test client/site_tests/platform_LibCBench/control ./server/autoserv -m 172.17.128.241 --ssh-port 22 -c client/site_tests/platform_LibCBench/control -r /tmp/test_that.PO1234567/platform_LibCBench --test-retry=0 --args ERROR:root:import statsd failed, no stats will be reported. 14:20:22 INFO | Results placed in /tmp/test_that.PO1234567/platform_LibCBench 14:20:22 INFO | Processing control file 14:20:23 INFO | Starting master ssh connection '/usr/bin/ssh -a -x -N -o ControlMaster=yes -o ControlPath=/tmp/_autotmp_VIIP67ssh-master/socket -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o BatchMode=yes -o ConnectTimeout=30 -o ServerAliveInterval=180 -o ServerAliveCountMax=3 -o ConnectionAttempts=4 -o Protocol=2 -l root -p 22 172.17.128.241' 14:20:23 ERROR| [stderr] Warning: Permanently added '172.17.128.241' (RSA) to the list of known hosts. 14:20:23 INFO | INFO ---- ---- kernel=3.8.11 localtime=May 22 14:20:23 timestamp=1369257623 14:20:23 INFO | Installing autotest on 172.17.128.241 14:20:23 INFO | Using installation dir /usr/local/autotest 14:20:23 WARNI| No job_repo_url for 14:20:23 INFO | Could not install autotest using the packaging system: No repos to install an autotest client from. Trying other methods 14:20:23 INFO | Installation of autotest completed 14:20:24 WARNI| No job_repo_url for 14:20:24 INFO | Executing /usr/local/autotest/bin/autotest /usr/local/autotest/control phase 0 14:20:24 INFO | Entered autotestd_monitor. 14:20:24 INFO | Finished launching tail subprocesses. 14:20:24 INFO | Finished waiting on autotestd to start. 14:20:26 INFO | START ---- ---- timestamp=1369257625 localtime=May 22 14:20:25 14:20:26 INFO | START platform_LibCBench platform_LibCBench timestamp=1369257625 localtime=May 22 14:20:25 14:20:30 INFO | GOOD platform_LibCBench platform_LibCBench timestamp=1369257630 localtime=May 22 14:20:30 completed successfully 14:20:30 INFO | END GOOD platform_LibCBench platform_LibCBench timestamp=1369257630 localtime=May 22 14:20:30 14:20:31 INFO | END GOOD ---- ---- timestamp=1369257630 localtime=May 22 14:20:30 14:20:31 INFO | Got lock of exit_code_file. 14:20:31 INFO | Released lock of exit_code_file and closed it. OUTPUT: ============================== OUTPUT: Current time: 2013-05-22 14:20:32.818831 Elapsed: 0:01:30 ETA: Unknown Done: 0% [ ] OUTPUT: Thread Status: RUNNING: 1 ('ttt: LibCBench (1)' 0:01:21) Machine Status: Machine Thread Lock Status Checksum 172.17.128.241 ttt: LibCBench (1) True RUNNING 3ba9f2ecbb222f20887daea5583d86ba OUTPUT: ============================== 14:20:33 INFO | Killing child processes. 14:20:33 INFO | Client complete 14:20:33 INFO | Finished processing control file 14:20:33 INFO | Starting master ssh connection '/usr/bin/ssh -a -x -N -o ControlMaster=yes -o ControlPath=/tmp/_autotmp_aVJUgmssh-master/socket -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o BatchMode=yes -o ConnectTimeout=30 -o ServerAliveInterval=180 -o ServerAliveCountMax=3 -o ConnectionAttempts=4 -o Protocol=2 -l root -p 22 172.17.128.241' 14:20:33 ERROR| [stderr] Warning: Permanently added '172.17.128.241' (RSA) to the list of known hosts. INFO : Test results: ------------------------------------------------------------------- platform_LibCBench [ PASSED ] platform_LibCBench/platform_LibCBench [ PASSED ] platform_LibCBench/platform_LibCBench b_malloc_big1__0_ 0.00375231466667 platform_LibCBench/platform_LibCBench b_malloc_big2__0_ 0.002951359 platform_LibCBench/platform_LibCBench b_malloc_bubble__0_ 0.015066374 platform_LibCBench/platform_LibCBench b_malloc_sparse__0_ 0.015053784 platform_LibCBench/platform_LibCBench b_malloc_thread_local__0_ 0.01138439 platform_LibCBench/platform_LibCBench b_malloc_thread_stress__0_ 0.0367894733333 platform_LibCBench/platform_LibCBench b_malloc_tiny1__0_ 0.000768474333333 platform_LibCBench/platform_LibCBench b_malloc_tiny2__0_ 0.000581407333333 platform_LibCBench/platform_LibCBench b_pthread_create_serial1__0_ 0.0291785246667 platform_LibCBench/platform_LibCBench b_pthread_createjoin_serial1__0_ 0.031907936 platform_LibCBench/platform_LibCBench b_pthread_createjoin_serial2__0_ 0.043485347 platform_LibCBench/platform_LibCBench b_pthread_uselesslock__0_ 0.0294113346667 platform_LibCBench/platform_LibCBench b_regex_compile____a_b_c__d_b__ 0.00529833933333 platform_LibCBench/platform_LibCBench b_regex_search____a_b_c__d_b__ 0.00165455066667 platform_LibCBench/platform_LibCBench b_regex_search___a_25_b__ 0.0496191923333 platform_LibCBench/platform_LibCBench b_stdio_putcgetc__0_ 0.100005711667 platform_LibCBench/platform_LibCBench b_stdio_putcgetc_unlocked__0_ 0.0371443833333 platform_LibCBench/platform_LibCBench b_string_memset__0_ 0.00275405066667 platform_LibCBench/platform_LibCBench b_string_strchr__0_ 0.00456903 platform_LibCBench/platform_LibCBench b_string_strlen__0_ 0.044893587 platform_LibCBench/platform_LibCBench b_string_strstr___aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaac__ 0.118360778 platform_LibCBench/platform_LibCBench b_string_strstr___aaaaaaaaaaaaaaaaaaaaaaaaac__ 0.068957325 platform_LibCBench/platform_LibCBench b_string_strstr___aaaaaaaaaaaaaacccccccccccc__ 0.0135694476667 platform_LibCBench/platform_LibCBench b_string_strstr___abcdefghijklmnopqrstuvwxyz__ 0.0134553343333 platform_LibCBench/platform_LibCBench b_string_strstr___azbycxdwevfugthsirjqkplomn__ 0.0133123556667 platform_LibCBench/platform_LibCBench b_utf8_bigbuf__0_ 0.0473772253333 platform_LibCBench/platform_LibCBench b_utf8_onebyone__0_ 0.130938538333 ------------------------------------------------------------------- Total PASS: 2/2 (100%) INFO : Elapsed time: 0m16s """ error = """ ERROR: Identity added: /tmp/test_that.Z4Ld/autotest_key (/tmp/test_that.Z4Ld/autotest_key) INFO : Using emerged autotests already installed at /build/lumpy/usr/local/autotest. INFO : Running the following control files 1 times: INFO : * 'client/site_tests/platform_LibCBench/control' INFO : Running client test client/site_tests/platform_LibCBench/control INFO : Test results: INFO : Elapsed time: 0m18s """ keyvals = { '': 'PASS', 'b_stdio_putcgetc__0_': '0.100005711667', 'b_string_strstr___azbycxdwevfugthsirjqkplomn__': '0.0133123556667', 'b_malloc_thread_local__0_': '0.01138439', 'b_string_strlen__0_': '0.044893587', 'b_malloc_sparse__0_': '0.015053784', 'b_string_memset__0_': '0.00275405066667', 'platform_LibCBench': 'PASS', 'b_pthread_uselesslock__0_': '0.0294113346667', 'b_string_strchr__0_': '0.00456903', 'b_pthread_create_serial1__0_': '0.0291785246667', 'b_string_strstr___aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaac__': '0.118360778', 'b_string_strstr___aaaaaaaaaaaaaacccccccccccc__': '0.0135694476667', 'b_pthread_createjoin_serial1__0_': '0.031907936', 'b_malloc_thread_stress__0_': '0.0367894733333', 'b_regex_search____a_b_c__d_b__': '0.00165455066667', 'b_malloc_bubble__0_': '0.015066374', 'b_malloc_big2__0_': '0.002951359', 'b_stdio_putcgetc_unlocked__0_': '0.0371443833333', 'b_pthread_createjoin_serial2__0_': '0.043485347', 'b_regex_search___a_25_b__': '0.0496191923333', 'b_utf8_bigbuf__0_': '0.0473772253333', 'b_malloc_big1__0_': '0.00375231466667', 'b_regex_compile____a_b_c__d_b__': '0.00529833933333', 'b_string_strstr___aaaaaaaaaaaaaaaaaaaaaaaaac__': '0.068957325', 'b_malloc_tiny2__0_': '0.000581407333333', 'b_utf8_onebyone__0_': '0.130938538333', 'b_malloc_tiny1__0_': '0.000768474333333', 'b_string_strstr___abcdefghijklmnopqrstuvwxyz__': '0.0134553343333' } PERF_DATA_HEADER = """ # ======== # captured on : Thu Jan 01 00:00:00 1980 # header version : 1 # data offset : 536 # data size : 737678672 # feat offset : 737679208 # hostname : localhost # os release : 5.4.61 # perf version : # arch : aarch64 # nrcpus online : 8 # nrcpus avail : 8 # total memory : 5911496 kB # cmdline : /usr/bin/perf record -e instructions -p {pid} # event : name = instructions, , id = ( 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193 ), type = 8, size = 112 # event : name = dummy:u, , id = ( 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204 ), type = 1, size = 112, config = 0x9 # CPU_TOPOLOGY info available, use -I to display # pmu mappings: software = 1, uprobe = 6, cs_etm = 8, breakpoint = 5, tracepoint = 2, armv8_pmuv3 = 7 # contains AUX area data (e.g. instruction trace) # time of first sample : 0.000000 # time of last sample : 0.000000 # sample duration : 0.000 ms # missing features: TRACING_DATA CPUDESC CPUID NUMA_TOPOLOGY BRANCH_STACK GROUP_DESC STAT CACHE MEM_TOPOLOGY CLOCKID DIR_FORMAT # ======== # """ TURBOSTAT_LOG_OUTPUT = ( """CPU Avg_MHz Busy% Bzy_MHz TSC_MHz IRQ CoreTmp - 329 12.13 2723 2393 10975 77 0 336 12.41 2715 2393 6328 77 2 323 11.86 2731 2393 4647 69 CPU Avg_MHz Busy% Bzy_MHz TSC_MHz IRQ CoreTmp - 1940 67.46 2884 2393 39920 83 0 1827 63.70 2877 2393 21184 83 2 2053 71.22 2891 2393 18736 67 CPU Avg_MHz Busy% Bzy_MHz TSC_MHz IRQ CoreTmp - 1927 66.02 2927 2393 48946 84 0 1880 64.47 2925 2393 24457 84 2 1973 67.57 2928 2393 24489 69 CPU Avg_MHz Busy% Bzy_MHz TSC_MHz IRQ CoreTmp - 1899 64.84 2937 2393 42540 72 0 2135 72.82 2940 2393 23615 65 2 1663 56.85 2934 2393 18925 72 CPU Avg_MHz Busy% Bzy_MHz TSC_MHz IRQ CoreTmp - 1908 65.24 2932 2393 43172 75 0 1876 64.25 2928 2393 20743 75 2 1939 66.24 2936 2393 22429 69 CPU Avg_MHz Busy% Bzy_MHz TSC_MHz IRQ CoreTmp - 1553 53.12 2933 2393 35488 46 0 1484 50.80 2929 2393 18246 46 2 1623 55.44 2936 2393 17242 45 CPU Avg_MHz Busy% Bzy_MHz TSC_MHz IRQ CoreTmp - 843 29.83 2832 2393 28161 47 0 827 29.35 2826 2393 16093 47 2 858 30.31 2838 2393 12068 46 """) TURBOSTAT_DATA = { 'cpufreq': { 'all': [2723, 2884, 2927, 2937, 2932, 2933, 2832] }, 'cputemp': { 'all': [77, 83, 84, 72, 75, 46, 47] }, } TOP_LOG = (""" PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 4102 chronos 12 -8 3454472 238300 118188 R 41.8 6.1 0:08.37 chrome 4204 chronos 12 -8 2492716 205728 179016 S 11.8 5.3 0:03.89 chrome 4890 root 20 0 3396 2064 1596 R 11.8 0.1 0:00.03 top 375 root 0 -20 0 0 0 S 5.9 0.0 0:00.17 kworker/u13 617 syslog 20 0 25332 8372 7888 S 5.9 0.2 0:00.77 sys-journal PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 5745 chronos 20 0 5438580 139328 67988 R 122.8 3.6 0:04.26 chrome 912 root -51 0 0 0 0 S 2.0 0.0 0:01.04 irq/cros-ec 121 root 20 0 0 0 0 S 1.0 0.0 0:00.45 spi5 4811 root 20 0 6808 4084 3492 S 1.0 0.1 0:00.02 sshd 4890 root 20 0 3364 2148 1596 R 1.0 0.1 0:00.36 top 5205 chronos 12 -8 3673780 240928 130864 S 1.0 6.2 0:07.30 chrome PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 5745 chronos 20 0 5434484 139432 63892 R 107.9 3.6 0:05.35 chrome 5713 chronos 20 0 5178652 103120 50372 S 17.8 2.6 0:01.13 chrome 7 root 20 0 0 0 0 S 1.0 0.0 0:00.73 rcu_preempt 855 root 20 0 0 0 0 S 1.0 0.0 0:00.01 kworker/4:2 """) TOP_DATA = [ { 'cmd': 'chrome-5745', 'cpu_use_avg': 115.35, 'count': 2, 'top5_cpu_use': [122.8, 107.9], }, { 'cmd': 'chrome-5713', 'cpu_use_avg': 8.9, 'count': 1, 'top5_cpu_use': [17.8] }, { 'cmd': 'irq/cros-ec-912', 'cpu_use_avg': 1.0, 'count': 1, 'top5_cpu_use': [2.0], }, { 'cmd': 'chrome-5205', 'cpu_use_avg': 0.5, 'count': 1, 'top5_cpu_use': [1.0] }, { 'cmd': 'spi5-121', 'cpu_use_avg': 0.5, 'count': 1, 'top5_cpu_use': [1.0], }, { 'cmd': 'sshd-4811', 'cpu_use_avg': 0.5, 'count': 1, 'top5_cpu_use': [1.0], }, { 'cmd': 'rcu_preempt-7', 'cpu_use_avg': 0.5, 'count': 1, 'top5_cpu_use': [1.0], }, { 'cmd': 'kworker/4:2-855', 'cpu_use_avg': 0.5, 'count': 1, 'top5_cpu_use': [1.0], }, ] TOP_OUTPUT = (""" COMMAND AVG CPU% SEEN HIGHEST 5 chrome 128.250000 6 [122.8, 107.9, 17.8, 5.0, 2.0] irq/230-cros-ec 1.000000 1 [2.0] sshd 0.500000 1 [1.0] irq/231-cros-ec 0.500000 1 [1.0] spi5 0.500000 1 [1.0] rcu_preempt 0.500000 1 [1.0] kworker/4:2 0.500000 1 [1.0] """) CPUSTATS_UNIQ_OUTPUT = (""" /sys/devices/system/cpu/cpu0/cpufreq/cpuinfo_cur_freq 1512000 /sys/devices/system/cpu/cpu1/cpufreq/cpuinfo_cur_freq 1512000 /sys/devices/system/cpu/cpu3/cpufreq/cpuinfo_cur_freq 2016000 soc-thermal 44444 little-cpu 41234 big-cpu 51234 /sys/devices/system/cpu/cpu0/cpufreq/cpuinfo_cur_freq 1500000 /sys/devices/system/cpu/cpu1/cpufreq/cpuinfo_cur_freq 1600000 /sys/devices/system/cpu/cpu3/cpufreq/cpuinfo_cur_freq 2012000 soc-thermal 45456 little-cpu 42555 big-cpu 61724 """) CPUSTATS_UNIQ_DATA = { 'cpufreq': { 'cpu0': [1512, 1500], 'cpu1': [1512, 1600], 'cpu3': [2016, 2012] }, 'cputemp': { 'soc-thermal': [44.4, 45.5], 'little-cpu': [41.2, 42.6], 'big-cpu': [51.2, 61.7] } } CPUSTATS_DUPL_OUTPUT = (""" /sys/devices/system/cpu/cpu0/cpufreq/cpuinfo_cur_freq 1512000 /sys/devices/system/cpu/cpu1/cpufreq/cpuinfo_cur_freq 1512000 /sys/devices/system/cpu/cpu2/cpufreq/cpuinfo_cur_freq 1512000 /sys/devices/system/cpu/cpu3/cpufreq/cpuinfo_cur_freq 2016000 /sys/devices/system/cpu/cpu0/cpufreq/cpuinfo_cur_freq 1500000 /sys/devices/system/cpu/cpu1/cpufreq/cpuinfo_cur_freq 1500000 /sys/devices/system/cpu/cpu2/cpufreq/cpuinfo_cur_freq 1500000 /sys/devices/system/cpu/cpu3/cpufreq/cpuinfo_cur_freq 2016000 /sys/devices/system/cpu/cpu0/cpufreq/cpuinfo_cur_freq 1614000 /sys/devices/system/cpu/cpu1/cpufreq/cpuinfo_cur_freq 1614000 /sys/devices/system/cpu/cpu2/cpufreq/cpuinfo_cur_freq 1614000 /sys/devices/system/cpu/cpu3/cpufreq/cpuinfo_cur_freq 1982000 """) CPUSTATS_DUPL_DATA = { 'cpufreq': { 'cpu0': [1512, 1500, 1614], 'cpu3': [2016, 2016, 1982] }, } TMP_DIR1 = '/tmp/tmpAbcXyz' HISTOGRAMSET = (""" [ { "values": [ "cache_temperature_cold", "typical", "cache_temperature:cold" ], "guid": "db6d463b-7c07-4873-b839-db0652ccb97e", "type": "GenericSet" }, { "values": [ "cache_temperature_warm", "typical", "cache_temperature:warm" ], "guid": "a270eb9d-3bb0-472a-951d-74ac3398b718", "type": "GenericSet" }, { "sampleValues": [ 1111.672 ], "name": "timeToFirstContentfulPaint", "diagnostics": { "storyTags": "a270eb9d-3bb0-472a-951d-74ac3398b718" }, "unit": "ms_smallerIsBetter" }, { "sampleValues": [ 1146.459 ], "name": "timeToFirstContentfulPaint", "diagnostics": { "storyTags": "db6d463b-7c07-4873-b839-db0652ccb97e" }, "unit": "ms_smallerIsBetter" }, { "sampleValues": [ 888.328 ], "name": "timeToFirstContentfulPaint", "diagnostics": { "storyTags": "a270eb9d-3bb0-472a-951d-74ac3398b718" }, "unit": "ms_smallerIsBetter" }, { "sampleValues": [ 853.541 ], "name": "timeToFirstContentfulPaint", "diagnostics": { "storyTags": "db6d463b-7c07-4873-b839-db0652ccb97e" }, "unit": "ms_smallerIsBetter" }, { "sampleValues": [ 400.000 ], "name": "timeToFirstContentfulPaint", "diagnostics": { "storyTags": "a270eb9d-3bb0-472a-951d-74ac3398b718" }, "unit": "ms_smallerIsBetter" } ] """) # pylint: enable=line-too-long class MockResult(Result): """Mock result class.""" def __init__(self, mylogger, label, logging_level, machine): super(MockResult, self).__init__(mylogger, label, logging_level, machine) def FindFilesInResultsDir(self, find_args): return '' # pylint: disable=arguments-differ def GetKeyvals(self, temp=False): if temp: pass return keyvals class ResultTest(unittest.TestCase): """Result test class.""" def __init__(self, *args, **kwargs): super(ResultTest, self).__init__(*args, **kwargs) self.callFakeProcessResults = False self.fakeCacheReturnResult = None self.callGetResultsDir = False self.callProcessResults = False self.callGetPerfReportFiles = False self.kv_dict = None self.tmpdir = '' self.callGetNewKeyvals = False self.callGetResultsFile = False self.callGetPerfDataFiles = False self.callGetTurbostatFile = False self.callGetCpustatsFile = False self.callGetTopFile = False self.callGetCpuinfoFile = False self.callGetWaitTimeFile = False self.args = None self.callGatherPerfResults = False self.mock_logger = mock.Mock(spec=logger.Logger) self.mock_cmd_exec = mock.Mock(spec=command_executer.CommandExecuter) self.mock_label = MockLabel('mock_label', 'build', 'chromeos_image', 'autotest_dir', 'debug_dir', '/tmp', 'lumpy', 'remote', 'image_args', 'cache_dir', 'average', 'gcc', False, None) def testCreateFromRun(self): result = MockResult.CreateFromRun(logger.GetLogger(), 'average', self.mock_label, 'remote1', OUTPUT, error, 0, True) self.assertEqual(result.keyvals, keyvals) self.assertEqual(result.chroot_results_dir, '/tmp/test_that.PO1234567/platform_LibCBench') self.assertEqual(result.results_dir, '/tmp/chroot/tmp/test_that.PO1234567/platform_LibCBench') self.assertEqual(result.retval, 0) def setUp(self): self.result = Result(self.mock_logger, self.mock_label, 'average', self.mock_cmd_exec) self.result.chromeos_root = '/tmp/chromeos' @mock.patch.object(os.path, 'isdir') @mock.patch.object(command_executer.CommandExecuter, 'RunCommand') @mock.patch.object(command_executer.CommandExecuter, 'CopyFiles') def test_copy_files_to(self, mock_copyfiles, mock_runcmd, mock_isdir): files = ['src_file_1', 'src_file_2', 'src_file_3'] dest_dir = '/tmp/test' self.mock_cmd_exec.RunCommand = mock_runcmd self.mock_cmd_exec.CopyFiles = mock_copyfiles mock_copyfiles.return_value = 0 # test 1. dest_dir exists; CopyFiles returns 0. mock_isdir.return_value = True self.result.CopyFilesTo(dest_dir, files) self.assertEqual(mock_runcmd.call_count, 0) self.assertEqual(mock_copyfiles.call_count, 3) first_args = mock_copyfiles.call_args_list[0][0] second_args = mock_copyfiles.call_args_list[1][0] third_args = mock_copyfiles.call_args_list[2][0] self.assertEqual(first_args, ('src_file_1', '/tmp/test/src_file_1.0')) self.assertEqual(second_args, ('src_file_2', '/tmp/test/src_file_2.1')) self.assertEqual(third_args, ('src_file_3', '/tmp/test/src_file_3.2')) mock_runcmd.reset_mock() mock_copyfiles.reset_mock() # test 2. dest_dir does not exist; CopyFiles returns 0. mock_isdir.return_value = False self.result.CopyFilesTo(dest_dir, files) self.assertEqual(mock_runcmd.call_count, 3) self.assertEqual(mock_copyfiles.call_count, 3) self.assertEqual(mock_runcmd.call_args_list[0], mock_runcmd.call_args_list[1]) self.assertEqual(mock_runcmd.call_args_list[0], mock_runcmd.call_args_list[2]) self.assertEqual(mock_runcmd.call_args_list[0][0], ('mkdir -p /tmp/test',)) # test 3. CopyFiles returns 1 (fails). mock_copyfiles.return_value = 1 self.assertRaises(Exception, self.result.CopyFilesTo, dest_dir, files) @mock.patch.object(Result, 'CopyFilesTo') def test_copy_results_to(self, mockCopyFilesTo): results_file = [ '/tmp/result.json.0', '/tmp/result.json.1', '/tmp/result.json.2' ] perf_data_files = [ '/tmp/perf.data.0', '/tmp/perf.data.1', '/tmp/perf.data.2' ] perf_report_files = [ '/tmp/perf.report.0', '/tmp/perf.report.1', '/tmp/perf.report.2' ] self.result.results_file = results_file self.result.perf_data_files = perf_data_files self.result.perf_report_files = perf_report_files self.result.CopyFilesTo = mockCopyFilesTo self.result.CopyResultsTo('/tmp/results/') self.assertEqual(mockCopyFilesTo.call_count, 3) self.assertEqual(len(mockCopyFilesTo.call_args_list), 3) self.assertEqual(mockCopyFilesTo.call_args_list[0][0], ('/tmp/results/', results_file)) self.assertEqual(mockCopyFilesTo.call_args_list[1][0], ('/tmp/results/', perf_data_files)) self.assertEqual(mockCopyFilesTo.call_args_list[2][0], ('/tmp/results/', perf_report_files)) def test_get_new_keyvals(self): kv_dict = {} def FakeGetDataMeasurementsFiles(): filename = os.path.join(os.getcwd(), 'unittest_keyval_file.txt') return [filename] self.result.GetDataMeasurementsFiles = FakeGetDataMeasurementsFiles kv_dict2, udict = self.result.GetNewKeyvals(kv_dict) self.assertEqual( kv_dict2, { u'Box2D__Box2D': 4775, u'Mandreel__Mandreel': 6620, u'Gameboy__Gameboy': 9901, u'Crypto__Crypto': 8737, u'telemetry_page_measurement_results__num_errored': 0, u'telemetry_page_measurement_results__num_failed': 0, u'PdfJS__PdfJS': 6455, u'Total__Score': 7918, u'EarleyBoyer__EarleyBoyer': 14340, u'MandreelLatency__MandreelLatency': 5188, u'CodeLoad__CodeLoad': 6271, u'DeltaBlue__DeltaBlue': 14401, u'Typescript__Typescript': 9815, u'SplayLatency__SplayLatency': 7653, u'zlib__zlib': 16094, u'Richards__Richards': 10358, u'RegExp__RegExp': 1765, u'NavierStokes__NavierStokes': 9815, u'Splay__Splay': 4425, u'RayTrace__RayTrace': 16600 }) self.assertEqual( udict, { u'Box2D__Box2D': u'score', u'Mandreel__Mandreel': u'score', u'Gameboy__Gameboy': u'score', u'Crypto__Crypto': u'score', u'telemetry_page_measurement_results__num_errored': u'count', u'telemetry_page_measurement_results__num_failed': u'count', u'PdfJS__PdfJS': u'score', u'Total__Score': u'score', u'EarleyBoyer__EarleyBoyer': u'score', u'MandreelLatency__MandreelLatency': u'score', u'CodeLoad__CodeLoad': u'score', u'DeltaBlue__DeltaBlue': u'score', u'Typescript__Typescript': u'score', u'SplayLatency__SplayLatency': u'score', u'zlib__zlib': u'score', u'Richards__Richards': u'score', u'RegExp__RegExp': u'score', u'NavierStokes__NavierStokes': u'score', u'Splay__Splay': u'score', u'RayTrace__RayTrace': u'score' }) def test_append_telemetry_units(self): kv_dict = { u'Box2D__Box2D': 4775, u'Mandreel__Mandreel': 6620, u'Gameboy__Gameboy': 9901, u'Crypto__Crypto': 8737, u'PdfJS__PdfJS': 6455, u'Total__Score': 7918, u'EarleyBoyer__EarleyBoyer': 14340, u'MandreelLatency__MandreelLatency': 5188, u'CodeLoad__CodeLoad': 6271, u'DeltaBlue__DeltaBlue': 14401, u'Typescript__Typescript': 9815, u'SplayLatency__SplayLatency': 7653, u'zlib__zlib': 16094, u'Richards__Richards': 10358, u'RegExp__RegExp': 1765, u'NavierStokes__NavierStokes': 9815, u'Splay__Splay': 4425, u'RayTrace__RayTrace': 16600 } units_dict = { u'Box2D__Box2D': u'score', u'Mandreel__Mandreel': u'score', u'Gameboy__Gameboy': u'score', u'Crypto__Crypto': u'score', u'PdfJS__PdfJS': u'score', u'Total__Score': u'score', u'EarleyBoyer__EarleyBoyer': u'score', u'MandreelLatency__MandreelLatency': u'score', u'CodeLoad__CodeLoad': u'score', u'DeltaBlue__DeltaBlue': u'score', u'Typescript__Typescript': u'score', u'SplayLatency__SplayLatency': u'score', u'zlib__zlib': u'score', u'Richards__Richards': u'score', u'RegExp__RegExp': u'score', u'NavierStokes__NavierStokes': u'score', u'Splay__Splay': u'score', u'RayTrace__RayTrace': u'score' } results_dict = self.result.AppendTelemetryUnits(kv_dict, units_dict) self.assertEqual( results_dict, { u'Box2D__Box2D': [4775, u'score'], u'Splay__Splay': [4425, u'score'], u'Gameboy__Gameboy': [9901, u'score'], u'Crypto__Crypto': [8737, u'score'], u'PdfJS__PdfJS': [6455, u'score'], u'Total__Score': [7918, u'score'], u'EarleyBoyer__EarleyBoyer': [14340, u'score'], u'MandreelLatency__MandreelLatency': [5188, u'score'], u'DeltaBlue__DeltaBlue': [14401, u'score'], u'SplayLatency__SplayLatency': [7653, u'score'], u'Mandreel__Mandreel': [6620, u'score'], u'Richards__Richards': [10358, u'score'], u'zlib__zlib': [16094, u'score'], u'CodeLoad__CodeLoad': [6271, u'score'], u'Typescript__Typescript': [9815, u'score'], u'RegExp__RegExp': [1765, u'score'], u'RayTrace__RayTrace': [16600, u'score'], u'NavierStokes__NavierStokes': [9815, u'score'] }) @mock.patch.object(misc, 'GetInsideChrootPath') @mock.patch.object(tempfile, 'mkdtemp') @mock.patch.object(command_executer.CommandExecuter, 'RunCommand') @mock.patch.object(command_executer.CommandExecuter, 'ChrootRunCommandWOutput') def test_get_keyvals(self, mock_chrootruncmd, mock_runcmd, mock_mkdtemp, mock_getpath): self.kv_dict = {} self.callGetNewKeyvals = False def reset(): self.kv_dict = {} self.callGetNewKeyvals = False mock_chrootruncmd.reset_mock() mock_runcmd.reset_mock() mock_mkdtemp.reset_mock() mock_getpath.reset_mock() def FakeGetNewKeyvals(kv_dict): self.kv_dict = kv_dict self.callGetNewKeyvals = True return_kvdict = {'first_time': 680, 'Total': 10} return_udict = {'first_time': 'ms', 'Total': 'score'} return return_kvdict, return_udict mock_mkdtemp.return_value = TMP_DIR1 mock_chrootruncmd.return_value = [ '', ('%s,PASS\n%s/telemetry_Crosperf,PASS\n') % (TMP_DIR1, TMP_DIR1), '' ] mock_getpath.return_value = TMP_DIR1 self.result.ce.ChrootRunCommandWOutput = mock_chrootruncmd self.result.ce.RunCommand = mock_runcmd self.result.GetNewKeyvals = FakeGetNewKeyvals self.result.suite = 'telemetry_Crosperf' self.result.results_dir = '/tmp/test_that_resultsNmq' # Test 1. no self.temp_dir. res = self.result.GetKeyvals() self.assertTrue(self.callGetNewKeyvals) self.assertEqual(self.kv_dict, {'': 'PASS', 'telemetry_Crosperf': 'PASS'}) self.assertEqual(mock_runcmd.call_count, 1) self.assertEqual(mock_runcmd.call_args_list[0][0], ('cp -r /tmp/test_that_resultsNmq/* %s' % TMP_DIR1,)) self.assertEqual(mock_chrootruncmd.call_count, 1) self.assertEqual( mock_chrootruncmd.call_args_list[0][0], (self.result.chromeos_root, ('./generate_test_report --no-color --csv %s') % TMP_DIR1)) self.assertEqual(mock_getpath.call_count, 1) self.assertEqual(mock_mkdtemp.call_count, 1) self.assertEqual(res, {'Total': [10, 'score'], 'first_time': [680, 'ms']}) # Test 2. self.temp_dir reset() mock_chrootruncmd.return_value = [ '', ('/tmp/tmpJCajRG,PASS\n/tmp/tmpJCajRG/' 'telemetry_Crosperf,PASS\n'), '' ] mock_getpath.return_value = '/tmp/tmpJCajRG' self.result.temp_dir = '/tmp/tmpJCajRG' res = self.result.GetKeyvals() self.assertEqual(mock_runcmd.call_count, 0) self.assertEqual(mock_mkdtemp.call_count, 0) self.assertEqual(mock_chrootruncmd.call_count, 1) self.assertTrue(self.callGetNewKeyvals) self.assertEqual(self.kv_dict, {'': 'PASS', 'telemetry_Crosperf': 'PASS'}) self.assertEqual(res, {'Total': [10, 'score'], 'first_time': [680, 'ms']}) # Test 3. suite != telemetry_Crosperf. Normally this would be for # running non-Telemetry autotests, such as BootPerfServer. In this test # case, the keyvals we have set up were returned from a Telemetry test run; # so this pass is basically testing that we don't append the units to the # test results (which we do for Telemetry autotest runs). reset() self.result.suite = '' res = self.result.GetKeyvals() self.assertEqual(res, {'Total': 10, 'first_time': 680}) @mock.patch.object(misc, 'GetInsideChrootPath') @mock.patch.object(command_executer.CommandExecuter, 'ChrootRunCommandWOutput') @mock.patch.object(os.path, 'exists') def test_get_samples(self, mock_exists, mock_get_total_samples, mock_getpath): self.result.perf_data_files = ['/tmp/results/perf.data'] self.result.board = 'samus' mock_getpath.return_value = '/usr/chromeos/chroot/tmp/results/perf.data' mock_get_total_samples.return_value = [ '', '45.42% 237210 chrome ', '' ] mock_exists.return_value = True # mock_open does not seem to support iteration. # pylint: disable=line-too-long content = """1.63% 66 dav1d-tile chrome [.] decode_coefs 1.48% 60 swapper [kernel.kallsyms] [k] intel_idle 1.16% 47 dav1d-tile chrome [.] decode_sb""" with mock.patch('builtins.open', return_value=io.StringIO(content)): samples = self.result.GetSamples() self.assertEqual(samples, [237210 - 60, u'samples']) def test_get_results_dir(self): self.result.out = '' self.assertRaises(Exception, self.result.GetResultsDir) self.result.out = OUTPUT resdir = self.result.GetResultsDir() self.assertEqual(resdir, '/tmp/test_that.PO1234567/platform_LibCBench') @mock.patch.object(command_executer.CommandExecuter, 'RunCommandGeneric') def test_find_files_in_results_dir(self, mock_runcmd): self.result.results_dir = None res = self.result.FindFilesInResultsDir('-name perf.data') self.assertEqual(res, '') self.result.ce.RunCommand = mock_runcmd self.result.results_dir = '/tmp/test_results' mock_runcmd.return_value = [0, '/tmp/test_results/perf.data', ''] res = self.result.FindFilesInResultsDir('-name perf.data') self.assertEqual(mock_runcmd.call_count, 1) self.assertEqual(mock_runcmd.call_args_list[0][0], ('find /tmp/test_results -name perf.data',)) self.assertEqual(res, '/tmp/test_results/perf.data') mock_runcmd.reset_mock() mock_runcmd.return_value = [1, '', ''] self.assertRaises(Exception, self.result.FindFilesInResultsDir, '-name perf.data') @mock.patch.object(Result, 'FindFilesInResultsDir') def test_get_perf_data_files(self, mock_findfiles): self.args = None mock_findfiles.return_value = 'line1\nline1\n' self.result.FindFilesInResultsDir = mock_findfiles res = self.result.GetPerfDataFiles() self.assertEqual(res, ['line1', 'line1']) self.assertEqual(mock_findfiles.call_args_list[0][0], ('-name perf.data',)) def test_get_perf_report_files(self): self.args = None def FakeFindFiles(find_args): self.args = find_args return 'line1\nline1\n' self.result.FindFilesInResultsDir = FakeFindFiles res = self.result.GetPerfReportFiles() self.assertEqual(res, ['line1', 'line1']) self.assertEqual(self.args, '-name perf.data.report') def test_get_data_measurement_files(self): self.args = None def FakeFindFiles(find_args): self.args = find_args return 'line1\nline1\n' self.result.FindFilesInResultsDir = FakeFindFiles res = self.result.GetDataMeasurementsFiles() self.assertEqual(res, ['line1', 'line1']) self.assertEqual(self.args, '-name perf_measurements') @mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput') def test_get_turbostat_file_finds_single_log(self, mock_runcmd): """Expected behavior when a single log file found.""" self.result.results_dir = '/tmp/test_results' self.result.ce.RunCommandWOutput = mock_runcmd mock_runcmd.return_value = (0, 'some/long/path/turbostat.log', '') found_single_log = self.result.GetTurbostatFile() self.assertEqual(found_single_log, 'some/long/path/turbostat.log') @mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput') def test_get_turbostat_file_finds_multiple_logs(self, mock_runcmd): """Error case when multiple files found.""" self.result.results_dir = '/tmp/test_results' self.result.ce.RunCommandWOutput = mock_runcmd mock_runcmd.return_value = (0, 'some/long/path/turbostat.log\nturbostat.log', '') found_first_logs = self.result.GetTurbostatFile() self.assertEqual(found_first_logs, 'some/long/path/turbostat.log') @mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput') def test_get_turbostat_file_finds_no_logs(self, mock_runcmd): """Error case when no log file found.""" self.result.results_dir = '/tmp/test_results' self.result.ce.RunCommandWOutput = mock_runcmd mock_runcmd.return_value = (0, '', '') found_no_logs = self.result.GetTurbostatFile() self.assertEqual(found_no_logs, '') @mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput') def test_get_turbostat_file_with_failing_find(self, mock_runcmd): """Error case when file search returns an error.""" self.result.results_dir = '/tmp/test_results' mock_runcmd.return_value = (-1, '', 'error') with self.assertRaises(RuntimeError): self.result.GetTurbostatFile() @mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput') def test_get_top_file_finds_single_log(self, mock_runcmd): """Expected behavior when a single top log file found.""" self.result.results_dir = '/tmp/test_results' self.result.ce.RunCommandWOutput = mock_runcmd mock_runcmd.return_value = (0, 'some/long/path/top.log', '') found_single_log = self.result.GetTopFile() self.assertEqual(found_single_log, 'some/long/path/top.log') @mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput') def test_get_top_file_finds_multiple_logs(self, mock_runcmd): """The case when multiple top files found.""" self.result.results_dir = '/tmp/test_results' self.result.ce.RunCommandWOutput = mock_runcmd mock_runcmd.return_value = (0, 'some/long/path/top.log\ntop.log', '') found_first_logs = self.result.GetTopFile() self.assertEqual(found_first_logs, 'some/long/path/top.log') @mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput') def test_get_top_file_finds_no_logs(self, mock_runcmd): """Error case when no log file found.""" self.result.results_dir = '/tmp/test_results' self.result.ce.RunCommandWOutput = mock_runcmd mock_runcmd.return_value = (0, '', '') found_no_logs = self.result.GetTopFile() self.assertEqual(found_no_logs, '') @mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput') def test_get_cpuinfo_file_finds_single_log(self, mock_runcmd): """Expected behavior when a single cpuinfo file found.""" self.result.results_dir = '/tmp/test_results' self.result.ce.RunCommandWOutput = mock_runcmd mock_runcmd.return_value = (0, 'some/long/path/cpuinfo.log', '') found_single_log = self.result.GetCpuinfoFile() self.assertEqual(found_single_log, 'some/long/path/cpuinfo.log') @mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput') def test_get_cpustats_file_finds_single_log(self, mock_runcmd): """Expected behavior when a single log file found.""" self.result.results_dir = '/tmp/test_results' self.result.ce.RunCommandWOutput = mock_runcmd mock_runcmd.return_value = (0, 'some/long/path/cpustats.log', '') found_single_log = self.result.GetCpustatsFile() self.assertEqual(found_single_log, 'some/long/path/cpustats.log') @mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput') def test_get_cpustats_file_finds_multiple_logs(self, mock_runcmd): """The case when multiple files found.""" self.result.results_dir = '/tmp/test_results' self.result.ce.RunCommandWOutput = mock_runcmd mock_runcmd.return_value = (0, 'some/long/path/cpustats.log\ncpustats.log', '') found_first_logs = self.result.GetCpustatsFile() self.assertEqual(found_first_logs, 'some/long/path/cpustats.log') @mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput') def test_get_cpustats_file_finds_no_logs(self, mock_runcmd): """Error case when no log file found.""" self.result.results_dir = '/tmp/test_results' self.result.ce.RunCommandWOutput = mock_runcmd mock_runcmd.return_value = (0, '', '') found_no_logs = self.result.GetCpustatsFile() self.assertEqual(found_no_logs, '') def test_verify_perf_data_pid_ok(self): """Verify perf PID which is present in TOP_DATA.""" self.result.top_cmds = TOP_DATA # pid is present in TOP_DATA. with mock.patch.object( Result, 'ReadPidFromPerfData', return_value=['5713']): self.result.VerifyPerfDataPID() def test_verify_perf_data_pid_fail(self): """Test perf PID missing in top raises the error.""" self.result.top_cmds = TOP_DATA # pid is not in the list of top processes. with mock.patch.object( Result, 'ReadPidFromPerfData', return_value=['9999']): with self.assertRaises(PidVerificationError): self.result.VerifyPerfDataPID() @mock.patch.object(command_executer.CommandExecuter, 'ChrootRunCommandWOutput') def test_read_pid_from_perf_data_ok(self, mock_runcmd): """Test perf header parser, normal flow.""" self.result.ce.ChrootRunCommandWOutput = mock_runcmd self.result.perf_data_files = ['/tmp/chromeos/chroot/tmp/results/perf.data'] exp_pid = '12345' mock_runcmd.return_value = (0, PERF_DATA_HEADER.format(pid=exp_pid), '') pids = self.result.ReadPidFromPerfData() self.assertEqual(pids, [exp_pid]) @mock.patch.object(command_executer.CommandExecuter, 'ChrootRunCommandWOutput') def test_read_pid_from_perf_data_mult_profiles(self, mock_runcmd): """Test multiple perf.data files with PID.""" self.result.ce.ChrootRunCommandWOutput = mock_runcmd # self.result.chromeos_root = '/tmp/chromeos' self.result.perf_data_files = [ '/tmp/chromeos/chroot/tmp/results/perf.data.0', '/tmp/chromeos/chroot/tmp/results/perf.data.1', ] # There is '-p ' in command line but it's still system-wide: '-a'. cmd_line = '# cmdline : /usr/bin/perf record -e instructions -p {pid}' exp_perf_pids = ['1111', '2222'] mock_runcmd.side_effect = [ (0, cmd_line.format(pid=exp_perf_pids[0]), ''), (0, cmd_line.format(pid=exp_perf_pids[1]), ''), ] pids = self.result.ReadPidFromPerfData() self.assertEqual(pids, exp_perf_pids) @mock.patch.object(command_executer.CommandExecuter, 'ChrootRunCommandWOutput') def test_read_pid_from_perf_data_no_pid(self, mock_runcmd): """Test perf.data without PID.""" self.result.ce.ChrootRunCommandWOutput = mock_runcmd self.result.perf_data_files = ['/tmp/chromeos/chroot/tmp/results/perf.data'] cmd_line = '# cmdline : /usr/bin/perf record -e instructions' mock_runcmd.return_value = (0, cmd_line, '') pids = self.result.ReadPidFromPerfData() # pids is empty. self.assertEqual(pids, []) @mock.patch.object(command_executer.CommandExecuter, 'ChrootRunCommandWOutput') def test_read_pid_from_perf_data_system_wide(self, mock_runcmd): """Test reading from system-wide profile with PID.""" self.result.ce.ChrootRunCommandWOutput = mock_runcmd self.result.perf_data_files = ['/tmp/chromeos/chroot/tmp/results/perf.data'] # There is '-p ' in command line but it's still system-wide: '-a'. cmd_line = '# cmdline : /usr/bin/perf record -e instructions -a -p 1234' mock_runcmd.return_value = (0, cmd_line, '') pids = self.result.ReadPidFromPerfData() # pids should be empty since it's not a per-process profiling. self.assertEqual(pids, []) @mock.patch.object(command_executer.CommandExecuter, 'ChrootRunCommandWOutput') def test_read_pid_from_perf_data_read_fail(self, mock_runcmd): """Failure to read perf.data raises the error.""" self.result.ce.ChrootRunCommandWOutput = mock_runcmd self.result.perf_data_files = ['/tmp/chromeos/chroot/tmp/results/perf.data'] # Error status of the profile read. mock_runcmd.return_value = (1, '', '') with self.assertRaises(PerfDataReadError): self.result.ReadPidFromPerfData() @mock.patch.object(command_executer.CommandExecuter, 'ChrootRunCommandWOutput') def test_read_pid_from_perf_data_fail(self, mock_runcmd): """Failure to find cmdline in perf.data header raises the error.""" self.result.ce.ChrootRunCommandWOutput = mock_runcmd self.result.perf_data_files = ['/tmp/chromeos/chroot/tmp/results/perf.data'] # Empty output. mock_runcmd.return_value = (0, '', '') with self.assertRaises(PerfDataReadError): self.result.ReadPidFromPerfData() def test_process_turbostat_results_with_valid_data(self): """Normal case when log exists and contains valid data.""" self.result.turbostat_log_file = '/tmp/somelogfile.log' with mock.patch('builtins.open', mock.mock_open(read_data=TURBOSTAT_LOG_OUTPUT)) as mo: cpustats = self.result.ProcessTurbostatResults() # Check that the log got opened and data were read/parsed. calls = [mock.call('/tmp/somelogfile.log')] mo.assert_has_calls(calls) self.assertEqual(cpustats, TURBOSTAT_DATA) def test_process_turbostat_results_from_empty_file(self): """Error case when log exists but file is empty.""" self.result.turbostat_log_file = '/tmp/emptylogfile.log' with mock.patch('builtins.open', mock.mock_open(read_data='')) as mo: cpustats = self.result.ProcessTurbostatResults() # Check that the log got opened and parsed successfully and empty data # returned. calls = [mock.call('/tmp/emptylogfile.log')] mo.assert_has_calls(calls) self.assertEqual(cpustats, {}) def test_process_turbostat_results_when_file_doesnt_exist(self): """Error case when file does not exist.""" nonexistinglog = '/tmp/1' while os.path.exists(nonexistinglog): # Extend file path if it happens to exist. nonexistinglog = os.path.join(nonexistinglog, '1') self.result.turbostat_log_file = nonexistinglog # Allow the tested function to call a 'real' open and hopefully crash. with self.assertRaises(IOError): self.result.ProcessTurbostatResults() def test_process_cpustats_results_with_uniq_data(self): """Process cpustats log which has freq unique to each core. Testing normal case when frequency data vary between different cores. Expecting that data for all cores will be present in returned cpustats. """ self.result.cpustats_log_file = '/tmp/somelogfile.log' with mock.patch('builtins.open', mock.mock_open(read_data=CPUSTATS_UNIQ_OUTPUT)) as mo: cpustats = self.result.ProcessCpustatsResults() # Check that the log got opened and data were read/parsed. calls = [mock.call('/tmp/somelogfile.log')] mo.assert_has_calls(calls) self.assertEqual(cpustats, CPUSTATS_UNIQ_DATA) def test_process_cpustats_results_with_dupl_data(self): """Process cpustats log where cores have duplicate freq. Testing normal case when frequency data on some cores are duplicated. Expecting that duplicated data is discarded in returned cpustats. """ self.result.cpustats_log_file = '/tmp/somelogfile.log' with mock.patch('builtins.open', mock.mock_open(read_data=CPUSTATS_DUPL_OUTPUT)) as mo: cpustats = self.result.ProcessCpustatsResults() # Check that the log got opened and data were read/parsed. calls = [mock.call('/tmp/somelogfile.log')] mo.assert_has_calls(calls) self.assertEqual(cpustats, CPUSTATS_DUPL_DATA) def test_process_cpustats_results_from_empty_file(self): """Error case when log exists but file is empty.""" self.result.cpustats_log_file = '/tmp/emptylogfile.log' with mock.patch('builtins.open', mock.mock_open(read_data='')) as mo: cpustats = self.result.ProcessCpustatsResults() # Check that the log got opened and parsed successfully and empty data # returned. calls = [mock.call('/tmp/emptylogfile.log')] mo.assert_has_calls(calls) self.assertEqual(cpustats, {}) def test_process_top_results_with_valid_data(self): """Process top log with valid data.""" self.result.top_log_file = '/tmp/fakelogfile.log' with mock.patch('builtins.open', mock.mock_open(read_data=TOP_LOG)) as mo: topproc = self.result.ProcessTopResults() # Check that the log got opened and data were read/parsed. calls = [mock.call('/tmp/fakelogfile.log')] mo.assert_has_calls(calls) self.assertEqual(topproc, TOP_DATA) def test_process_top_results_from_empty_file(self): """Error case when log exists but file is empty.""" self.result.top_log_file = '/tmp/emptylogfile.log' with mock.patch('builtins.open', mock.mock_open(read_data='')) as mo: topcalls = self.result.ProcessTopResults() # Check that the log got opened and parsed successfully and empty data # returned. calls = [mock.call('/tmp/emptylogfile.log')] mo.assert_has_calls(calls) self.assertEqual(topcalls, []) def test_format_string_top_cmds(self): """Test formatted string with top commands.""" self.result.top_cmds = [ { 'cmd': 'chrome-111', 'cpu_use_avg': 119.753453465, 'count': 44444, 'top5_cpu_use': [222.8, 217.9, 217.8, 191.0, 189.9], }, { 'cmd': 'chrome-222', 'cpu_use_avg': 100, 'count': 33333, 'top5_cpu_use': [200.0, 195.0, 190.0, 185.0, 180.0], }, { 'cmd': 'irq/230-cros-ec', 'cpu_use_avg': 10.000000000000001, 'count': 1000, 'top5_cpu_use': [11.5, 11.4, 11.3, 11.2, 11.1], }, { 'cmd': 'powerd', 'cpu_use_avg': 2.0, 'count': 2, 'top5_cpu_use': [3.0, 1.0] }, { 'cmd': 'cmd3', 'cpu_use_avg': 1.0, 'count': 1, 'top5_cpu_use': [1.0], }, { 'cmd': 'cmd4', 'cpu_use_avg': 1.0, 'count': 1, 'top5_cpu_use': [1.0], }, { 'cmd': 'cmd5', 'cpu_use_avg': 1.0, 'count': 1, 'top5_cpu_use': [1.0], }, { 'cmd': 'cmd6_not_for_print', 'cpu_avg': 1.0, 'count': 1, 'top5': [1.0], }, ] form_str = self.result.FormatStringTopCommands() self.assertEqual( form_str, '\n'.join([ 'Top commands with highest CPU usage:', ' COMMAND AVG CPU% COUNT HIGHEST 5', '-' * 50, ' chrome-111 119.75 44444 ' '[222.8, 217.9, 217.8, 191.0, 189.9]', ' chrome-222 100.00 33333 ' '[200.0, 195.0, 190.0, 185.0, 180.0]', ' irq/230-cros-ec 10.00 1000 ' '[11.5, 11.4, 11.3, 11.2, 11.1]', ' powerd 2.00 2 [3.0, 1.0]', ' cmd3 1.00 1 [1.0]', ' cmd4 1.00 1 [1.0]', ' cmd5 1.00 1 [1.0]', '-' * 50, ])) def test_format_string_top_calls_no_data(self): """Test formatted string of top with no data.""" self.result.top_cmds = [] form_str = self.result.FormatStringTopCommands() self.assertEqual( form_str, '\n'.join([ 'Top commands with highest CPU usage:', ' COMMAND AVG CPU% COUNT HIGHEST 5', '-' * 50, '[NO DATA FROM THE TOP LOG]', '-' * 50, ])) @mock.patch.object(misc, 'GetInsideChrootPath') @mock.patch.object(command_executer.CommandExecuter, 'ChrootRunCommand') def test_generate_perf_report_files(self, mock_chrootruncmd, mock_getpath): fake_file = '/usr/chromeos/chroot/tmp/results/fake_file' self.result.perf_data_files = ['/tmp/results/perf.data'] self.result.board = 'lumpy' mock_getpath.return_value = fake_file self.result.ce.ChrootRunCommand = mock_chrootruncmd mock_chrootruncmd.return_value = 0 # Debug path not found self.result.label.debug_path = '' tmp = self.result.GeneratePerfReportFiles() self.assertEqual(tmp, ['/tmp/chromeos/chroot%s' % fake_file]) self.assertEqual(mock_chrootruncmd.call_args_list[0][0], (self.result.chromeos_root, ('/usr/sbin/perf report -n ' '-i %s --stdio > %s') % (fake_file, fake_file))) @mock.patch.object(misc, 'GetInsideChrootPath') @mock.patch.object(command_executer.CommandExecuter, 'ChrootRunCommand') def test_generate_perf_report_files_debug(self, mock_chrootruncmd, mock_getpath): fake_file = '/usr/chromeos/chroot/tmp/results/fake_file' self.result.perf_data_files = ['/tmp/results/perf.data'] self.result.board = 'lumpy' mock_getpath.return_value = fake_file self.result.ce.ChrootRunCommand = mock_chrootruncmd mock_chrootruncmd.return_value = 0 # Debug path found self.result.label.debug_path = '/tmp/debug' tmp = self.result.GeneratePerfReportFiles() self.assertEqual(tmp, ['/tmp/chromeos/chroot%s' % fake_file]) self.assertEqual(mock_chrootruncmd.call_args_list[0][0], (self.result.chromeos_root, ('/usr/sbin/perf report -n --symfs /tmp/debug ' '--vmlinux /tmp/debug/usr/lib/debug/boot/vmlinux ' '-i %s --stdio > %s') % (fake_file, fake_file))) @mock.patch.object(misc, 'GetOutsideChrootPath') def test_populate_from_run(self, mock_getpath): def FakeGetResultsDir(): self.callGetResultsDir = True return '/tmp/results_dir' def FakeGetResultsFile(): self.callGetResultsFile = True return [] def FakeGetPerfDataFiles(): self.callGetPerfDataFiles = True return [] def FakeGetPerfReportFiles(): self.callGetPerfReportFiles = True return [] def FakeGetTurbostatFile(): self.callGetTurbostatFile = True return [] def FakeGetCpustatsFile(): self.callGetCpustatsFile = True return [] def FakeGetTopFile(): self.callGetTopFile = True return [] def FakeGetCpuinfoFile(): self.callGetCpuinfoFile = True return [] def FakeGetWaitTimeFile(): self.callGetWaitTimeFile = True return [] def FakeProcessResults(show_results=False): if show_results: pass self.callProcessResults = True if mock_getpath: pass mock.get_path = '/tmp/chromeos/tmp/results_dir' self.callGetResultsDir = False self.callGetResultsFile = False self.callGetPerfDataFiles = False self.callGetPerfReportFiles = False self.callGetTurbostatFile = False self.callGetCpustatsFile = False self.callGetTopFile = False self.callGetCpuinfoFile = False self.callGetWaitTimeFile = False self.callProcessResults = False self.result.GetResultsDir = FakeGetResultsDir self.result.GetResultsFile = FakeGetResultsFile self.result.GetPerfDataFiles = FakeGetPerfDataFiles self.result.GeneratePerfReportFiles = FakeGetPerfReportFiles self.result.GetTurbostatFile = FakeGetTurbostatFile self.result.GetCpustatsFile = FakeGetCpustatsFile self.result.GetTopFile = FakeGetTopFile self.result.GetCpuinfoFile = FakeGetCpuinfoFile self.result.GetWaitTimeFile = FakeGetWaitTimeFile self.result.ProcessResults = FakeProcessResults self.result.PopulateFromRun(OUTPUT, '', 0, 'test', 'telemetry_Crosperf', 'chrome') self.assertTrue(self.callGetResultsDir) self.assertTrue(self.callGetResultsFile) self.assertTrue(self.callGetPerfDataFiles) self.assertTrue(self.callGetPerfReportFiles) self.assertTrue(self.callGetTurbostatFile) self.assertTrue(self.callGetCpustatsFile) self.assertTrue(self.callGetTopFile) self.assertTrue(self.callGetCpuinfoFile) self.assertTrue(self.callGetWaitTimeFile) self.assertTrue(self.callProcessResults) def FakeGetKeyvals(self, show_all=False): if show_all: return {'first_time': 680, 'Total': 10} else: return {'Total': 10} def test_process_results(self): def FakeGatherPerfResults(): self.callGatherPerfResults = True def FakeGetSamples(): return (1, 'samples') # Test 1 self.callGatherPerfResults = False self.result.GetKeyvals = self.FakeGetKeyvals self.result.GatherPerfResults = FakeGatherPerfResults self.result.retval = 0 self.result.ProcessResults() self.assertTrue(self.callGatherPerfResults) self.assertEqual(len(self.result.keyvals), 2) self.assertEqual(self.result.keyvals, {'Total': 10, 'retval': 0}) # Test 2 self.result.retval = 1 self.result.ProcessResults() self.assertEqual(len(self.result.keyvals), 2) self.assertEqual(self.result.keyvals, {'Total': 10, 'retval': 1}) # Test 3 self.result.cwp_dso = 'chrome' self.result.retval = 0 self.result.GetSamples = FakeGetSamples self.result.ProcessResults() self.assertEqual(len(self.result.keyvals), 3) self.assertEqual(self.result.keyvals, { 'Total': 10, 'samples': (1, 'samples'), 'retval': 0 }) # Test 4. Parse output of benchmarks with multiple sotries in histogram # format self.result.suite = 'telemetry_Crosperf' self.result.results_file = [tempfile.mkdtemp() + '/histograms.json'] with open(self.result.results_file[0], 'w') as f: f.write(HISTOGRAMSET) self.result.ProcessResults() shutil.rmtree(os.path.dirname(self.result.results_file[0])) # Verify the summary for the story is correct self.assertEqual(self.result.keyvals['timeToFirstContentfulPaint__typical'], [880.000, u'ms_smallerIsBetter']) # Veirfy the summary for a certain stroy tag is correct self.assertEqual( self.result .keyvals['timeToFirstContentfulPaint__cache_temperature:cold'], [1000.000, u'ms_smallerIsBetter']) self.assertEqual( self.result .keyvals['timeToFirstContentfulPaint__cache_temperature:warm'], [800.000, u'ms_smallerIsBetter']) @mock.patch.object(Result, 'ProcessCpustatsResults') @mock.patch.object(Result, 'ProcessTurbostatResults') def test_process_results_with_turbostat_log(self, mock_proc_turbo, mock_proc_cpustats): self.result.GetKeyvals = self.FakeGetKeyvals self.result.retval = 0 self.result.turbostat_log_file = '/tmp/turbostat.log' mock_proc_turbo.return_value = { 'cpufreq': { 'all': [1, 2, 3] }, 'cputemp': { 'all': [5.0, 6.0, 7.0] } } self.result.ProcessResults() mock_proc_turbo.assert_has_calls([mock.call()]) mock_proc_cpustats.assert_not_called() self.assertEqual(len(self.result.keyvals), 8) self.assertEqual( self.result.keyvals, { 'Total': 10, 'cpufreq_all_avg': 2, 'cpufreq_all_max': 3, 'cpufreq_all_min': 1, 'cputemp_all_avg': 6.0, 'cputemp_all_min': 5.0, 'cputemp_all_max': 7.0, 'retval': 0 }) @mock.patch.object(Result, 'ProcessCpustatsResults') @mock.patch.object(Result, 'ProcessTurbostatResults') def test_process_results_with_cpustats_log(self, mock_proc_turbo, mock_proc_cpustats): self.result.GetKeyvals = self.FakeGetKeyvals self.result.retval = 0 self.result.cpustats_log_file = '/tmp/cpustats.log' mock_proc_cpustats.return_value = { 'cpufreq': { 'cpu0': [100, 100, 100], 'cpu1': [4, 5, 6] }, 'cputemp': { 'little': [20.2, 20.2, 20.2], 'big': [55.2, 66.1, 77.3] } } self.result.ProcessResults() mock_proc_turbo.assert_not_called() mock_proc_cpustats.assert_has_calls([mock.call()]) self.assertEqual(len(self.result.keyvals), 10) self.assertEqual( self.result.keyvals, { 'Total': 10, 'cpufreq_cpu0_avg': 100, 'cpufreq_cpu1_avg': 5, 'cpufreq_cpu1_max': 6, 'cpufreq_cpu1_min': 4, 'cputemp_big_avg': 66.2, 'cputemp_big_max': 77.3, 'cputemp_big_min': 55.2, 'cputemp_little_avg': 20.2, 'retval': 0 }) @mock.patch.object(Result, 'ProcessCpustatsResults') @mock.patch.object(Result, 'ProcessTurbostatResults') def test_process_results_with_turbostat_and_cpustats_logs( self, mock_proc_turbo, mock_proc_cpustats): self.result.GetKeyvals = self.FakeGetKeyvals self.result.retval = 0 self.result.turbostat_log_file = '/tmp/turbostat.log' self.result.cpustats_log_file = '/tmp/cpustats.log' mock_proc_turbo.return_value = { 'cpufreq': { 'all': [1, 2, 3] }, 'cputemp': { 'all': [5.0, 6.0, 7.0] } } self.result.ProcessResults() mock_proc_turbo.assert_has_calls([mock.call()]) mock_proc_cpustats.assert_not_called() self.assertEqual(len(self.result.keyvals), 8) self.assertEqual( self.result.keyvals, { 'Total': 10, 'cpufreq_all_avg': 2, 'cpufreq_all_max': 3, 'cpufreq_all_min': 1, 'cputemp_all_avg': 6.0, 'cputemp_all_min': 5.0, 'cputemp_all_max': 7.0, 'retval': 0 }) @mock.patch.object(Result, 'ProcessCpustatsResults') @mock.patch.object(Result, 'ProcessTurbostatResults') def test_process_results_without_cpu_data(self, mock_proc_turbo, mock_proc_cpustats): self.result.GetKeyvals = self.FakeGetKeyvals self.result.retval = 0 self.result.turbostat_log_file = '' self.result.cpustats_log_file = '' self.result.ProcessResults() mock_proc_turbo.assert_not_called() mock_proc_cpustats.assert_not_called() self.assertEqual(len(self.result.keyvals), 2) self.assertEqual(self.result.keyvals, {'Total': 10, 'retval': 0}) @mock.patch.object(misc, 'GetInsideChrootPath') @mock.patch.object(command_executer.CommandExecuter, 'ChrootRunCommandWOutput') def test_populate_from_cache_dir(self, mock_runchrootcmd, mock_getpath): # pylint: disable=redefined-builtin def FakeMkdtemp(dir=None): if dir: pass return self.tmpdir def FakeGetSamples(): return [1, u'samples'] current_path = os.getcwd() cache_dir = os.path.join(current_path, 'test_cache/test_input') self.result.ce = command_executer.GetCommandExecuter(log_level='average') self.result.ce.ChrootRunCommandWOutput = mock_runchrootcmd mock_runchrootcmd.return_value = [ '', ('%s,PASS\n%s/\telemetry_Crosperf,PASS\n') % (TMP_DIR1, TMP_DIR1), '' ] mock_getpath.return_value = TMP_DIR1 self.tmpdir = tempfile.mkdtemp() save_real_mkdtemp = tempfile.mkdtemp tempfile.mkdtemp = FakeMkdtemp self.result.PopulateFromCacheDir(cache_dir, 'sunspider', 'telemetry_Crosperf', '') self.assertEqual( self.result.keyvals, { u'Total__Total': [444.0, u'ms'], u'regexp-dna__regexp-dna': [16.2, u'ms'], u'telemetry_page_measurement_results__num_failed': [0, u'count'], u'telemetry_page_measurement_results__num_errored': [0, u'count'], u'string-fasta__string-fasta': [23.2, u'ms'], u'crypto-sha1__crypto-sha1': [11.6, u'ms'], u'bitops-3bit-bits-in-byte__bitops-3bit-bits-in-byte': [3.2, u'ms'], u'access-nsieve__access-nsieve': [7.9, u'ms'], u'bitops-nsieve-bits__bitops-nsieve-bits': [9.4, u'ms'], u'string-validate-input__string-validate-input': [19.3, u'ms'], u'3d-raytrace__3d-raytrace': [24.7, u'ms'], u'3d-cube__3d-cube': [28.0, u'ms'], u'string-unpack-code__string-unpack-code': [46.7, u'ms'], u'date-format-tofte__date-format-tofte': [26.3, u'ms'], u'math-partial-sums__math-partial-sums': [22.0, u'ms'], '\telemetry_Crosperf': ['PASS', ''], u'crypto-aes__crypto-aes': [15.2, u'ms'], u'bitops-bitwise-and__bitops-bitwise-and': [8.4, u'ms'], u'crypto-md5__crypto-md5': [10.5, u'ms'], u'string-tagcloud__string-tagcloud': [52.8, u'ms'], u'access-nbody__access-nbody': [8.5, u'ms'], 'retval': 0, u'math-spectral-norm__math-spectral-norm': [6.6, u'ms'], u'math-cordic__math-cordic': [8.7, u'ms'], u'access-binary-trees__access-binary-trees': [4.5, u'ms'], u'controlflow-recursive__controlflow-recursive': [4.4, u'ms'], u'access-fannkuch__access-fannkuch': [17.8, u'ms'], u'string-base64__string-base64': [16.0, u'ms'], u'date-format-xparb__date-format-xparb': [20.9, u'ms'], u'3d-morph__3d-morph': [22.1, u'ms'], u'bitops-bits-in-byte__bitops-bits-in-byte': [9.1, u'ms'] }) self.result.GetSamples = FakeGetSamples self.result.PopulateFromCacheDir(cache_dir, 'sunspider', 'telemetry_Crosperf', 'chrome') self.assertEqual( self.result.keyvals, { u'Total__Total': [444.0, u'ms'], u'regexp-dna__regexp-dna': [16.2, u'ms'], u'telemetry_page_measurement_results__num_failed': [0, u'count'], u'telemetry_page_measurement_results__num_errored': [0, u'count'], u'string-fasta__string-fasta': [23.2, u'ms'], u'crypto-sha1__crypto-sha1': [11.6, u'ms'], u'bitops-3bit-bits-in-byte__bitops-3bit-bits-in-byte': [3.2, u'ms'], u'access-nsieve__access-nsieve': [7.9, u'ms'], u'bitops-nsieve-bits__bitops-nsieve-bits': [9.4, u'ms'], u'string-validate-input__string-validate-input': [19.3, u'ms'], u'3d-raytrace__3d-raytrace': [24.7, u'ms'], u'3d-cube__3d-cube': [28.0, u'ms'], u'string-unpack-code__string-unpack-code': [46.7, u'ms'], u'date-format-tofte__date-format-tofte': [26.3, u'ms'], u'math-partial-sums__math-partial-sums': [22.0, u'ms'], '\telemetry_Crosperf': ['PASS', ''], u'crypto-aes__crypto-aes': [15.2, u'ms'], u'bitops-bitwise-and__bitops-bitwise-and': [8.4, u'ms'], u'crypto-md5__crypto-md5': [10.5, u'ms'], u'string-tagcloud__string-tagcloud': [52.8, u'ms'], u'access-nbody__access-nbody': [8.5, u'ms'], 'retval': 0, u'math-spectral-norm__math-spectral-norm': [6.6, u'ms'], u'math-cordic__math-cordic': [8.7, u'ms'], u'access-binary-trees__access-binary-trees': [4.5, u'ms'], u'controlflow-recursive__controlflow-recursive': [4.4, u'ms'], u'access-fannkuch__access-fannkuch': [17.8, u'ms'], u'string-base64__string-base64': [16.0, u'ms'], u'date-format-xparb__date-format-xparb': [20.9, u'ms'], u'3d-morph__3d-morph': [22.1, u'ms'], u'bitops-bits-in-byte__bitops-bits-in-byte': [9.1, u'ms'], u'samples': [1, u'samples'] }) # Clean up after test. tempfile.mkdtemp = save_real_mkdtemp command = 'rm -Rf %s' % self.tmpdir self.result.ce.RunCommand(command) @mock.patch.object(misc, 'GetRoot') @mock.patch.object(command_executer.CommandExecuter, 'RunCommand') def test_cleanup(self, mock_runcmd, mock_getroot): # Test 1. 'rm_chroot_tmp' is True; self.results_dir exists; # self.temp_dir exists; results_dir name contains 'test_that_results_'. mock_getroot.return_value = ['/tmp/tmp_AbcXyz', 'test_that_results_fake'] self.result.ce.RunCommand = mock_runcmd self.result.results_dir = 'test_results_dir' self.result.temp_dir = 'testtemp_dir' self.result.CleanUp(True) self.assertEqual(mock_getroot.call_count, 1) self.assertEqual(mock_runcmd.call_count, 2) self.assertEqual(mock_runcmd.call_args_list[0][0], ('rm -rf test_results_dir',)) self.assertEqual(mock_runcmd.call_args_list[1][0], ('rm -rf testtemp_dir',)) # Test 2. Same, except ath results_dir name does not contain # 'test_that_results_' mock_getroot.reset_mock() mock_runcmd.reset_mock() mock_getroot.return_value = ['/tmp/tmp_AbcXyz', 'other_results_fake'] self.result.ce.RunCommand = mock_runcmd self.result.results_dir = 'test_results_dir' self.result.temp_dir = 'testtemp_dir' self.result.CleanUp(True) self.assertEqual(mock_getroot.call_count, 1) self.assertEqual(mock_runcmd.call_count, 2) self.assertEqual(mock_runcmd.call_args_list[0][0], ('rm -rf /tmp/tmp_AbcXyz',)) self.assertEqual(mock_runcmd.call_args_list[1][0], ('rm -rf testtemp_dir',)) # Test 3. mock_getroot returns nothing; 'rm_chroot_tmp' is False. mock_getroot.reset_mock() mock_runcmd.reset_mock() self.result.CleanUp(False) self.assertEqual(mock_getroot.call_count, 0) self.assertEqual(mock_runcmd.call_count, 1) self.assertEqual(mock_runcmd.call_args_list[0][0], ('rm -rf testtemp_dir',)) # Test 4. 'rm_chroot_tmp' is True, but result_dir & temp_dir are None. mock_getroot.reset_mock() mock_runcmd.reset_mock() self.result.results_dir = None self.result.temp_dir = None self.result.CleanUp(True) self.assertEqual(mock_getroot.call_count, 0) self.assertEqual(mock_runcmd.call_count, 0) @mock.patch.object(misc, 'GetInsideChrootPath') @mock.patch.object(command_executer.CommandExecuter, 'ChrootRunCommand') def test_store_to_cache_dir(self, mock_chrootruncmd, mock_getpath): def FakeMkdtemp(directory=''): if directory: pass return self.tmpdir if mock_chrootruncmd or mock_getpath: pass current_path = os.getcwd() cache_dir = os.path.join(current_path, 'test_cache/test_output') self.result.ce = command_executer.GetCommandExecuter(log_level='average') self.result.out = OUTPUT self.result.err = error self.result.retval = 0 self.tmpdir = tempfile.mkdtemp() if not os.path.exists(self.tmpdir): os.makedirs(self.tmpdir) self.result.results_dir = os.path.join(os.getcwd(), 'test_cache') save_real_mkdtemp = tempfile.mkdtemp tempfile.mkdtemp = FakeMkdtemp mock_mm = machine_manager.MockMachineManager('/tmp/chromeos_root', 0, 'average', '') mock_mm.machine_checksum_string['mock_label'] = 'fake_machine_checksum123' mock_keylist = ['key1', 'key2', 'key3'] test_flag.SetTestMode(True) self.result.StoreToCacheDir(cache_dir, mock_mm, mock_keylist) # Check that the correct things were written to the 'cache'. test_dir = os.path.join(os.getcwd(), 'test_cache/test_output') base_dir = os.path.join(os.getcwd(), 'test_cache/compare_output') self.assertTrue(os.path.exists(os.path.join(test_dir, 'autotest.tbz2'))) self.assertTrue(os.path.exists(os.path.join(test_dir, 'machine.txt'))) self.assertTrue(os.path.exists(os.path.join(test_dir, 'results.txt'))) f1 = os.path.join(test_dir, 'machine.txt') f2 = os.path.join(base_dir, 'machine.txt') cmd = 'diff %s %s' % (f1, f2) [_, out, _] = self.result.ce.RunCommandWOutput(cmd) self.assertEqual(len(out), 0) f1 = os.path.join(test_dir, 'results.txt') f2 = os.path.join(base_dir, 'results.txt') cmd = 'diff %s %s' % (f1, f2) [_, out, _] = self.result.ce.RunCommandWOutput(cmd) self.assertEqual(len(out), 0) # Clean up after test. tempfile.mkdtemp = save_real_mkdtemp command = 'rm %s/*' % test_dir self.result.ce.RunCommand(command) TELEMETRY_RESULT_KEYVALS = { 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'math-cordic (ms)': '11.4', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'access-nbody (ms)': '6.9', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'access-fannkuch (ms)': '26.3', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'math-spectral-norm (ms)': '6.3', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'bitops-nsieve-bits (ms)': '9.3', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'math-partial-sums (ms)': '32.8', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'regexp-dna (ms)': '16.1', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' '3d-cube (ms)': '42.7', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'crypto-md5 (ms)': '10.8', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'crypto-sha1 (ms)': '12.4', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'string-tagcloud (ms)': '47.2', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'string-fasta (ms)': '36.3', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'access-binary-trees (ms)': '7.3', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'date-format-xparb (ms)': '138.1', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'crypto-aes (ms)': '19.2', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'Total (ms)': '656.5', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'string-base64 (ms)': '17.5', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'string-validate-input (ms)': '24.8', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' '3d-raytrace (ms)': '28.7', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'controlflow-recursive (ms)': '5.3', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'bitops-bits-in-byte (ms)': '9.8', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' '3d-morph (ms)': '50.2', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'bitops-bitwise-and (ms)': '8.8', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'access-nsieve (ms)': '8.6', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'date-format-tofte (ms)': '31.2', 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'bitops-3bit-bits-in-byte (ms)': '3.5', 'retval': 0, 'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html ' 'string-unpack-code (ms)': '45.0' } PURE_TELEMETRY_OUTPUT = """ page_name,3d-cube (ms),3d-morph (ms),3d-raytrace (ms),Total (ms),access-binary-trees (ms),access-fannkuch (ms),access-nbody (ms),access-nsieve (ms),bitops-3bit-bits-in-byte (ms),bitops-bits-in-byte (ms),bitops-bitwise-and (ms),bitops-nsieve-bits (ms),controlflow-recursive (ms),crypto-aes (ms),crypto-md5 (ms),crypto-sha1 (ms),date-format-tofte (ms),date-format-xparb (ms),math-cordic (ms),math-partial-sums (ms),math-spectral-norm (ms),regexp-dna (ms),string-base64 (ms),string-fasta (ms),string-tagcloud (ms),string-unpack-code (ms),string-validate-input (ms)\r\nhttp://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html,42.7,50.2,28.7,656.5,7.3,26.3,6.9,8.6,3.5,9.8,8.8,9.3,5.3,19.2,10.8,12.4,31.2,138.1,11.4,32.8,6.3,16.1,17.5,36.3,47.2,45.0,24.8\r """ class TelemetryResultTest(unittest.TestCase): """Telemetry result test.""" def __init__(self, *args, **kwargs): super(TelemetryResultTest, self).__init__(*args, **kwargs) self.callFakeProcessResults = False self.result = None self.mock_logger = mock.Mock(spec=logger.Logger) self.mock_cmd_exec = mock.Mock(spec=command_executer.CommandExecuter) self.mock_label = MockLabel('mock_label', 'build', 'chromeos_image', 'autotest_dir', 'debug_dir', '/tmp', 'lumpy', 'remote', 'image_args', 'cache_dir', 'average', 'gcc', False, None) self.mock_machine = machine_manager.MockCrosMachine('falco.cros', '/tmp/chromeos', 'average') def test_populate_from_run(self): def FakeProcessResults(): self.callFakeProcessResults = True self.callFakeProcessResults = False self.result = TelemetryResult(self.mock_logger, self.mock_label, 'average', self.mock_cmd_exec) self.result.ProcessResults = FakeProcessResults self.result.PopulateFromRun(OUTPUT, error, 3, 'fake_test', 'telemetry_Crosperf', '') self.assertTrue(self.callFakeProcessResults) self.assertEqual(self.result.out, OUTPUT) self.assertEqual(self.result.err, error) self.assertEqual(self.result.retval, 3) def test_populate_from_cache_dir_and_process_results(self): self.result = TelemetryResult(self.mock_logger, self.mock_label, 'average', self.mock_machine) current_path = os.getcwd() cache_dir = os.path.join(current_path, 'test_cache/test_puretelemetry_input') self.result.PopulateFromCacheDir(cache_dir, '', '', '') self.assertEqual(self.result.out.strip(), PURE_TELEMETRY_OUTPUT.strip()) self.assertEqual(self.result.err, '') self.assertEqual(self.result.retval, 0) self.assertEqual(self.result.keyvals, TELEMETRY_RESULT_KEYVALS) class ResultsCacheTest(unittest.TestCase): """Resultcache test class.""" def __init__(self, *args, **kwargs): super(ResultsCacheTest, self).__init__(*args, **kwargs) self.fakeCacheReturnResult = None self.mock_logger = mock.Mock(spec=logger.Logger) self.mock_label = MockLabel('mock_label', 'build', 'chromeos_image', 'autotest_dir', 'debug_dir', '/tmp', 'lumpy', 'remote', 'image_args', 'cache_dir', 'average', 'gcc', False, None) def setUp(self): self.results_cache = ResultsCache() mock_machine = machine_manager.MockCrosMachine('falco.cros', '/tmp/chromeos', 'average') mock_mm = machine_manager.MockMachineManager('/tmp/chromeos_root', 0, 'average', '') mock_mm.machine_checksum_string['mock_label'] = 'fake_machine_checksum123' self.results_cache.Init( self.mock_label.chromeos_image, self.mock_label.chromeos_root, 'sunspider', 1, # benchmark_run.iteration, '', # benchmark_run.test_args, '', # benchmark_run.profiler_args, mock_mm, mock_machine, self.mock_label.board, [CacheConditions.CACHE_FILE_EXISTS, CacheConditions.CHECKSUMS_MATCH], self.mock_logger, 'average', self.mock_label, '', # benchmark_run.share_cache 'telemetry_Crosperf', True, # benchmark_run.show_all_results False, # benchmark_run.run_local '') # benchmark_run.cwp_dso @mock.patch.object(image_checksummer.ImageChecksummer, 'Checksum') def test_get_cache_dir_for_write(self, mock_checksum): def FakeGetMachines(label): if label: pass m1 = machine_manager.MockCrosMachine('lumpy1.cros', self.results_cache.chromeos_root, 'average') m2 = machine_manager.MockCrosMachine('lumpy2.cros', self.results_cache.chromeos_root, 'average') return [m1, m2] mock_checksum.return_value = 'FakeImageChecksumabc123' self.results_cache.machine_manager.GetMachines = FakeGetMachines self.results_cache.machine_manager.machine_checksum['mock_label'] = ( 'FakeMachineChecksumabc987') # Based on the label, benchmark and machines, get the directory in which # to store the cache information for this test run. result_path = self.results_cache.GetCacheDirForWrite() # Verify that the returned directory is correct (since the label # contained a cache_dir, named 'cache_dir', that's what is expected in # the result, rather than '~/cros_scratch'). comp_path = os.path.join( os.getcwd(), 'cache_dir/54524606abaae4fdf7b02f49f7ae7127_' 'sunspider_1_fda29412ceccb72977516c4785d08e2c_' 'FakeImageChecksumabc123_FakeMachineChecksum' 'abc987__6') self.assertEqual(result_path, comp_path) def test_form_cache_dir(self): # This is very similar to the previous test (FormCacheDir is called # from GetCacheDirForWrite). cache_key_list = ('54524606abaae4fdf7b02f49f7ae7127', 'sunspider', '1', '7215ee9c7d9dc229d2921a40e899ec5f', 'FakeImageChecksumabc123', '*', '*', '6') path = self.results_cache.FormCacheDir(cache_key_list) self.assertEqual(len(path), 1) path1 = path[0] test_dirname = ('54524606abaae4fdf7b02f49f7ae7127_sunspider_1_7215ee9' 'c7d9dc229d2921a40e899ec5f_FakeImageChecksumabc123_*_*_6') comp_path = os.path.join(os.getcwd(), 'cache_dir', test_dirname) self.assertEqual(path1, comp_path) @mock.patch.object(image_checksummer.ImageChecksummer, 'Checksum') def test_get_cache_key_list(self, mock_checksum): # This tests the mechanism that generates the various pieces of the # cache directory name, based on various conditions. def FakeGetMachines(label): if label: pass m1 = machine_manager.MockCrosMachine('lumpy1.cros', self.results_cache.chromeos_root, 'average') m2 = machine_manager.MockCrosMachine('lumpy2.cros', self.results_cache.chromeos_root, 'average') return [m1, m2] mock_checksum.return_value = 'FakeImageChecksumabc123' self.results_cache.machine_manager.GetMachines = FakeGetMachines self.results_cache.machine_manager.machine_checksum['mock_label'] = ( 'FakeMachineChecksumabc987') # Test 1. Generating cache name for reading (not writing). key_list = self.results_cache.GetCacheKeyList(True) self.assertEqual(key_list[0], '*') # Machine checksum value, for read. self.assertEqual(key_list[1], 'sunspider') self.assertEqual(key_list[2], '1') self.assertEqual(key_list[3], 'fda29412ceccb72977516c4785d08e2c') self.assertEqual(key_list[4], 'FakeImageChecksumabc123') self.assertEqual(key_list[5], '*') self.assertEqual(key_list[6], '*') self.assertEqual(key_list[7], '6') # Test 2. Generating cache name for writing, with local image type. key_list = self.results_cache.GetCacheKeyList(False) self.assertEqual(key_list[0], '54524606abaae4fdf7b02f49f7ae7127') self.assertEqual(key_list[1], 'sunspider') self.assertEqual(key_list[2], '1') self.assertEqual(key_list[3], 'fda29412ceccb72977516c4785d08e2c') self.assertEqual(key_list[4], 'FakeImageChecksumabc123') self.assertEqual(key_list[5], 'FakeMachineChecksumabc987') self.assertEqual(key_list[6], '') self.assertEqual(key_list[7], '6') # Test 3. Generating cache name for writing, with trybot image type. self.results_cache.label.image_type = 'trybot' key_list = self.results_cache.GetCacheKeyList(False) self.assertEqual(key_list[0], '54524606abaae4fdf7b02f49f7ae7127') self.assertEqual(key_list[3], 'fda29412ceccb72977516c4785d08e2c') self.assertEqual(key_list[4], '54524606abaae4fdf7b02f49f7ae7127') self.assertEqual(key_list[5], 'FakeMachineChecksumabc987') # Test 4. Generating cache name for writing, with official image type. self.results_cache.label.image_type = 'official' key_list = self.results_cache.GetCacheKeyList(False) self.assertEqual(key_list[0], '54524606abaae4fdf7b02f49f7ae7127') self.assertEqual(key_list[1], 'sunspider') self.assertEqual(key_list[2], '1') self.assertEqual(key_list[3], 'fda29412ceccb72977516c4785d08e2c') self.assertEqual(key_list[4], '*') self.assertEqual(key_list[5], 'FakeMachineChecksumabc987') self.assertEqual(key_list[6], '') self.assertEqual(key_list[7], '6') # Test 5. Generating cache name for writing, with local image type, and # specifying that the image path must match the cached image path. self.results_cache.label.image_type = 'local' self.results_cache.cache_conditions.append(CacheConditions.IMAGE_PATH_MATCH) key_list = self.results_cache.GetCacheKeyList(False) self.assertEqual(key_list[0], '54524606abaae4fdf7b02f49f7ae7127') self.assertEqual(key_list[3], 'fda29412ceccb72977516c4785d08e2c') self.assertEqual(key_list[4], 'FakeImageChecksumabc123') self.assertEqual(key_list[5], 'FakeMachineChecksumabc987') @mock.patch.object(command_executer.CommandExecuter, 'RunCommand') @mock.patch.object(os.path, 'isdir') @mock.patch.object(Result, 'CreateFromCacheHit') def test_read_result(self, mock_create, mock_isdir, mock_runcmd): self.fakeCacheReturnResult = None def FakeGetCacheDirForRead(): return self.fakeCacheReturnResult def FakeGetCacheDirForWrite(): return self.fakeCacheReturnResult mock_cmd_exec = mock.Mock(spec=command_executer.CommandExecuter) fake_result = Result(self.mock_logger, self.mock_label, 'average', mock_cmd_exec) fake_result.retval = 0 # Set up results_cache _GetCacheDirFor{Read,Write} to return # self.fakeCacheReturnResult, which is initially None (see above). # So initially, no cache dir is returned. self.results_cache.GetCacheDirForRead = FakeGetCacheDirForRead self.results_cache.GetCacheDirForWrite = FakeGetCacheDirForWrite mock_isdir.return_value = True save_cc = [ CacheConditions.CACHE_FILE_EXISTS, CacheConditions.CHECKSUMS_MATCH ] self.results_cache.cache_conditions.append(CacheConditions.FALSE) # Test 1. CacheCondition.FALSE, which means do not read from the cache. # (force re-running of test). Result should be None. res = self.results_cache.ReadResult() self.assertIsNone(res) self.assertEqual(mock_runcmd.call_count, 1) # Test 2. Remove CacheCondition.FALSE. Result should still be None, # because GetCacheDirForRead is returning None at the moment. mock_runcmd.reset_mock() self.results_cache.cache_conditions = save_cc res = self.results_cache.ReadResult() self.assertIsNone(res) self.assertEqual(mock_runcmd.call_count, 0) # Test 3. Now set up cache dir to be returned by GetCacheDirForRead. # Since cache_dir is found, will call Result.CreateFromCacheHit, which # which will actually all our mock_create and should return fake_result. self.fakeCacheReturnResult = 'fake/cache/dir' mock_create.return_value = fake_result res = self.results_cache.ReadResult() self.assertEqual(mock_runcmd.call_count, 0) self.assertEqual(res, fake_result) # Test 4. os.path.isdir(cache_dir) will now return false, so result # should be None again (no cache found). mock_isdir.return_value = False res = self.results_cache.ReadResult() self.assertEqual(mock_runcmd.call_count, 0) self.assertIsNone(res) # Test 5. os.path.isdir returns true, but mock_create now returns None # (the call to CreateFromCacheHit returns None), so overal result is None. mock_isdir.return_value = True mock_create.return_value = None res = self.results_cache.ReadResult() self.assertEqual(mock_runcmd.call_count, 0) self.assertIsNone(res) # Test 6. Everything works 'as expected', result should be fake_result. mock_create.return_value = fake_result res = self.results_cache.ReadResult() self.assertEqual(mock_runcmd.call_count, 0) self.assertEqual(res, fake_result) # Test 7. The run failed; result should be None. mock_create.return_value = fake_result fake_result.retval = 1 self.results_cache.cache_conditions.append(CacheConditions.RUN_SUCCEEDED) res = self.results_cache.ReadResult() self.assertEqual(mock_runcmd.call_count, 0) self.assertIsNone(res) if __name__ == '__main__': unittest.main()