/test/vts/testcases/host/shell/ |
D | SampleShellTest.py | 38 results = self.dut.shell.Execute("which ls") 39 logging.info(str(results[const.STDOUT])) 40 asserts.assertEqual(len(results[const.STDOUT]), 1) 41 asserts.assertEqual(results[const.STDOUT][0].strip(), "/system/bin/ls") 42 asserts.assertEqual(results[const.EXIT_CODE][0], 0) 47 results = self.dut.shell.my_shell2.Execute(["which ls"] * 49 logging.info(str(results[const.STDOUT])) 50 asserts.assertEqual(len(results[const.STDOUT]), self.REPEAT_COUNT) 52 asserts.assertEqual(results[const.STDOUT][index].strip(), 54 asserts.assertEqual(results[const.EXIT_CODE][index], 0) [all …]
|
/test/vts/utils/python/cpu/ |
D | cpu_frequency_scaling.py | 71 results = self._shell.Execute("cat /sys/devices/system/cpu/present") 72 asserts.assertEqual(len(results[const.STDOUT]), 1) 73 stdout_lines = results[const.STDOUT][0].split("\n") 105 results = self._shell.Execute( 108 asserts.assertEqual(1, len(results[const.EXIT_CODE])) 109 if not results[const.EXIT_CODE][0]: 110 freq = [int(x) for x in results[const.STDOUT][0].split()] 138 results = self._shell.Execute(target_cmd) 140 len(results[const.STDOUT])) 141 if any(results[const.EXIT_CODE]): [all …]
|
/test/vts/testcases/host/shell_binary_crash_test/ |
D | ShellBinaryCrashTest.py | 41 results = self.dut.shell.my_shell1.Execute( 43 logging.info(str(results[const.STDOUT])) 44 asserts.assertEqual(len(results[const.STDOUT]), 2) 45 asserts.assertEqual(results[const.STDOUT][1].strip(), "") 47 asserts.assertEqual(results[const.EXIT_CODE][1], self.EXIT_CODE_CRASH) 56 results = self.dut.shell.my_shell1.Execute( 58 logging.info(str(results[const.STDOUT])) 59 asserts.assertEqual(len(results[const.STDOUT]), 2) 60 asserts.assertEqual(results[const.STDOUT][1].strip(), "") 63 asserts.assertEqual(results[const.EXIT_CODE][1], [all …]
|
/test/vts/utils/python/file/ |
D | target_file_utils.py | 37 results = shell.Execute(cmd) 38 return results[const.EXIT_CODE][0] == 0 87 results = shell.Execute(cmd) 88 logging.debug("%s: Shell command '%s' results: %s", path, cmd, results) 90 if results[const.EXIT_CODE][0] != 0: 91 raise IOError(results[const.STDERR][0]) 93 stdout = str(results[const.STDOUT][0]) 112 results = shell.Execute(cmd) 113 logging.debug("%s: Shell command '%s' results: %s", filepath, cmd, results) 116 if results[const.EXIT_CODE][0] != 0: [all …]
|
/test/vts/testcases/codelab/hello_world/ |
D | VtsCodelabHelloWorldTest.py | 35 results = self.shell.Execute( 37 logging.info(str(results[const.STDOUT])) # prints the stdout 38 asserts.assertEqual(results[const.STDOUT][0].strip(), 40 asserts.assertEqual(results[const.EXIT_CODE][0], 45 results = self.shell.Execute(["echo hello", "echo world"]) 46 logging.info(str(results[const.STDOUT])) 47 asserts.assertEqual(len(results[const.STDOUT]), 49 asserts.assertEqual(results[const.STDOUT][0].strip(), "hello") 50 asserts.assertEqual(results[const.STDOUT][1].strip(), "world") 51 asserts.assertEqual(results[const.EXIT_CODE][0], 0) [all …]
|
/test/framework/harnesses/host_controller/campaigns/ |
D | campaign_test.py | 36 results = vts.EmitConsoleCommands( 39 default_testcase.GenerateOutputData(test_name), results) 44 results = vts.EmitConsoleCommands( 47 default_testcase.GenerateOutputData(test_name), results) 52 results = cts.EmitConsoleCommands( 55 default_testcase.GenerateOutputData(test_name), results) 60 results = gts.EmitConsoleCommands( 63 default_testcase.GenerateOutputData(test_name), results) 68 results = sts.EmitConsoleCommands( 71 default_testcase.GenerateOutputData(test_name), results)
|
/test/vts/utils/python/mirror/ |
D | native_entity_mirror.py | 222 results = self._client.CallApi( 224 if (isinstance(results, tuple) and len(results) == 2 225 and isinstance(results[1], dict) 226 and "coverage" in results[1]): 227 self._last_raw_code_coverage_data = results[1]["coverage"] 228 results = results[0] 230 if isinstance(results, list): # Non-HIDL HAL does not return list. 232 for i, _ in enumerate(results): 233 result = results[i] 241 results[i] = None [all …]
|
/test/vts-testcase/hal/treble/platform_version/ |
D | VtsTreblePlatformVersionTest.py | 37 results = self.dut.shell.Execute("getprop " + prop) 39 asserts.assertEqual(results[const.EXIT_CODE][0], 0, 41 asserts.assertTrue(len(results[const.STDOUT][0].strip()) > 0, 44 if (results[const.EXIT_CODE][0] != 0 or 45 len(results[const.STDOUT][0].strip()) == 0): 49 result = results[const.STDOUT][0].strip() 58 results = self.dut.shell.Execute("printenv " + env) 59 if (results[const.EXIT_CODE][0] != 0 or 60 len(results[const.STDOUT][0].strip()) == 0): 64 result = results[const.STDOUT][0].strip()
|
/test/suite_harness/common/host-side/tradefed/tests/src/com/android/compatibility/common/tradefed/result/suite/ |
D | CertificationChecksumHelperTest.java | 66 Collection<TestRunResult> results = new ArrayList<>(); in testCreateChecksum() local 68 results.add(run1); in testCreateChecksum() 70 results.add(run2); in testCreateChecksum() 72 mWorkingDir, results, FINGERPRINT); in testCreateChecksum() 92 TestRunResult results = new TestRunResult(); in createFakeResults() local 93 results.testRunStarted(runName, testCount); in createFakeResults() 96 results.testStarted(test); in createFakeResults() 97 results.testEnded(test, new HashMap<String, Metric>()); in createFakeResults() 99 results.testRunEnded(500L, new HashMap<String, Metric>()); in createFakeResults() 100 return results; in createFakeResults()
|
/test/mlts/benchmark/src/com/android/nn/benchmark/core/ |
D | LatencyResult.java | 43 public LatencyResult(float[] results) { in LatencyResult() argument 44 mIterations = results.length; in LatencyResult() 48 for (float result : results) { in LatencyResult() 57 for (float result : results) { in LatencyResult() 68 for (float result : results) { in LatencyResult() 101 public void putToBundle(Bundle results, String prefix) { in putToBundle() argument 103 results.putFloat(prefix + "_avg", getMeanTimeSec() * 1000.0f); in putToBundle() 104 results.putFloat(prefix + "_std_dev", mTimeStdDeviation * 1000.0f); in putToBundle() 105 results.putFloat(prefix + "_total_time", mTotalTimeSec * 1000.0f); in putToBundle() 106 results.putInt(prefix + "_iterations", mIterations); in putToBundle()
|
D | BenchmarkResult.java | 179 List<Pair<String, Float>> results = new ArrayList<>(); in getEvaluatorResults() local 181 results.add(new Pair<>(mEvaluatorKeys[i], mEvaluatorResults[i])); in getEvaluatorResults() 183 return results; in getEvaluatorResults() 251 Bundle results = new Bundle(); in toBundle() local 253 results.putString(testName + "_error", mBenchmarkError); in toBundle() 254 return results; in toBundle() 257 mLatencyInference.putToBundle(results, testName + "_inference"); in toBundle() 258 results.putFloat(testName + "_inference_mean_square_error", in toBundle() 260 results.putFloat(testName + "_inference_max_single_error", mMaxSingleError); in toBundle() 262 results.putFloat(testName + "_inference_" + mEvaluatorKeys[i], mEvaluatorResults[i]); in toBundle() [all …]
|
/test/vti/dashboard/src/main/java/com/android/vts/entity/ |
D | TestCaseRunEntity.java | 54 private List<Integer> results; field in TestCaseRunEntity 92 this.results = new ArrayList<>(); in TestCaseRunEntity() 104 this.results = new ArrayList<>(); in TestCaseRunEntity() 139 if (testCaseNames.size() == results.size()) { in onLoad() 142 int result = results.get(index).intValue(); in onLoad() 159 this.results.add(result); in addTestCase() 180 List<Integer> results = new ArrayList<>(); in toEntity() local 183 results.add(testCase.result); in toEntity() 186 testCaseRunEntity.setUnindexedProperty(RESULTS, results); in toEntity() 211 List<Long> results = (List<Long>) e.getProperty(RESULTS); in fromEntity() local [all …]
|
/test/vts/testcases/vts_selftest/test_framework/base_test/ |
D | VtsSelfTestBaseTest.py | 77 results = self.shell.Execute( 79 AssertShellCommandSuccess(results, 1) 80 logging.info(str(results[const.STDOUT])) # prints the stdout 81 asserts.assertEqual(results[const.STDOUT][0].strip(), 86 results = self.shell.Execute(['echo hello', 'echo world']) 87 AssertShellCommandSuccess(results, 2) 88 logging.info(str(results[const.STDOUT])) 89 asserts.assertEqual(results[const.STDOUT][0].strip(), 'hello') 90 asserts.assertEqual(results[const.STDOUT][1].strip(), 'world')
|
/test/vts-testcase/kernel/ltp/shell_environment/ |
D | shell_environment.py | 76 results = self.shell.Execute('cat %s' % 78 if (not results or results[const.EXIT_CODE][0] or 79 not results[const.STDOUT][0]): 81 "\n Command results: {}".format(results)) 84 cpu_info = results[const.STDOUT][0].strip()
|
/test/suite_harness/common/util/src/com/android/compatibility/common/util/ |
D | CaseResult.java | 74 List<ITestResult> results = new ArrayList<>(); in getResults() local 77 results.add(result); in getResults() 80 return results; in getResults() 88 ArrayList<ITestResult> results = new ArrayList<>(mResults.values()); in getResults() local 89 Collections.sort(results); in getResults() 90 return results; in getResults()
|
/test/vti/dashboard/src/test/java/com/android/vts/entity/ |
D | TestCaseRunEntityTest.java | 35 List<Integer> results = Arrays.asList(1, 1, 1, 1, 1, 1, 1); in saveTest() local 47 for (int index = 0; index < results.size(); index++) { in saveTest() 49 int result = results.get(index); in saveTest() 54 assertEquals(loadedTestCaseRunEntity.getTestCases().size(), results.size()); in saveTest() 56 (Integer) loadedTestCaseRunEntity.getTestCases().get(0).result, results.get(0)); in saveTest()
|
/test/vts-testcase/performance/fmq_benchmark/ |
D | FmqPerformanceTest.py | 84 results = self.dut.shell.Execute([ 90 asserts.assertEqual(len(results[const.STDOUT]), 2) 92 any(results[const.EXIT_CODE]), 99 results = self.dut.shell.Execute([ 110 asserts.assertEqual(len(results[const.STDOUT]), 2) 112 any(results[const.EXIT_CODE]), "FmqPerformanceTest failed.") 117 stdout_lines = results[const.STDOUT][1].split("\n")
|
/test/vts-testcase/performance/binder_benchmark/ |
D | BinderPerformanceTest.py | 111 results = self.dut.shell.Execute([ 118 asserts.assertEqual(len(results[const.STDOUT]), 2) 119 logging.info("stderr: %s", results[const.STDERR][1]) 120 logging.info("stdout: %s", results[const.STDOUT][1]) 122 any(results[const.EXIT_CODE]), 125 results[const.STDOUT][1])
|
/test/vts-testcase/performance/hwbinder_benchmark/ |
D | HwBinderPerformanceTest.py | 115 results = self.dut.shell.Execute([ 125 asserts.assertEqual(len(results[const.STDOUT]), 2) 126 logging.info("stderr: %s", results[const.STDERR][1]) 127 logging.info("stdout: %s", results[const.STDOUT][1]) 129 any(results[const.EXIT_CODE]), 132 results[const.STDOUT][1])
|
/test/vts-testcase/performance/audio_loopback_test/ |
D | AudioLoopbackTest.py | 106 results = self.dut.shell.Execute( 108 while results[const.EXIT_CODE][0]: 111 results = self.dut.shell.Execute( 114 results = self.dut.shell.Execute( 116 asserts.assertFalse(results[const.EXIT_CODE][0], 118 stdout_lines = results[const.STDOUT][0].split("\n")
|
/test/mlts/benchmark/results/ |
D | generate_result.py | 161 results = [] 163 results.append(parser.read_benchmark_result()) 165 return (benchmark_info, results) 168 def group_results(results): argument 172 for result in results: 177 for name, results in groupings.items(): 179 results)) 180 other = sorted(filter(lambda x: x is not baseline, results), 425 results = [results_with_bl.baseline] + results_with_bl.other 426 tags = [result.backend_type for result in results] [all …]
|
/test/vts/harnesses/tradefed/src/com/android/tradefed/testtype/ |
D | VtsMultiDeviceTestResultParser.java | 384 JSONArray results = object.getJSONArray(RESULTS); in processJsonFile() local 385 if (results == null) { in processJsonFile() 386 results = new JSONArray(); in processJsonFile() 390 if (results.length() > 0) { in processJsonFile() 391 long firstBeginTime = (long) results.getJSONObject(0).get(BEGIN_TIME); in processJsonFile() 393 (long) results.getJSONObject(results.length() - 1).get(END_TIME); in processJsonFile() 401 for (int index = 0; index < results.length(); index++) { in processJsonFile() 402 JSONObject resultObject = results.getJSONObject(index); in processJsonFile() 411 for (int index = 0; index < results.length(); index++) { in processJsonFile() 412 JSONObject resultObject = results.getJSONObject(index); in processJsonFile() [all …]
|
/test/vts/utils/python/profiling/ |
D | profiling_utils.py | 158 results = dut.shell.Execute("ls " + target_trace_file) 159 asserts.assertTrue(results, "failed to find trace file") 160 stdout_lines = results[const.STDOUT][0].split("\n") 176 results = cmd_utils.ExecuteShellCommand(file_cmd) 177 if results[const.EXIT_CODE][0] != 0: 178 logging.error(results[const.STDERR][0]) 258 results = cmd_utils.ExecuteShellCommand(trace_processor_cmd) 259 if any(results[cmd_utils.EXIT_CODE]): 261 logging.error("stdout: %s" % results[const.STDOUT]) 262 logging.error("stderr: %s" % results[const.STDERR]) [all …]
|
/test/vts/runners/host/ |
D | base_test.py | 139 self.results = records.TestResult() 531 if self.results.class_errors: 541 message_b = self.web.GenerateReportMessage(self.results.requested, 542 self.results.executed) 698 logging.error(RESULT_LINE_TEMPLATE, self.results.progressStr, 732 logging.info(RESULT_LINE_TEMPLATE, self.results.progressStr, 754 logging.info(RESULT_LINE_TEMPLATE, self.results.progressStr, 839 logging.info(RESULT_LINE_TEMPLATE, self.results.progressStr, 966 logging.info(TEST_CASE_TEMPLATE, self.results.progressStr, test_name) 1048 self.results.removeRecord(tr_record) [all …]
|
/test/suite_harness/common/host-side/tradefed/src/com/android/compatibility/common/tradefed/targetprep/ |
D | ApkInstrumentationPreparer.java | 138 for (Entry<TestDescription, TestResult> results : result.getTestResults().entrySet()) { in instrument() 139 if (TestStatus.FAILURE.equals(results.getValue().getStatus())) { in instrument() 143 results.getKey(), results.getValue().getStackTrace()); in instrument() 147 results.getKey(), results.getValue().getStackTrace()); in instrument()
|