/external/skia/tools/ |
H A D | bench_pictures_main.cpp | 159 sk_tools::PictureBenchmark& benchmark) { 197 benchmark.run(picture); 231 static void setup_benchmark(sk_tools::PictureBenchmark* benchmark) { argument 314 benchmark->setTimersToShow(timerWall, truncatedTimerWall, timerCpu, truncatedTimerCpu, 344 benchmark->setTimeIndividualTiles(true); 347 benchmark->setPurgeDecodedTex(FLAGS_purgeDecodedTex); 348 benchmark->setPreprocess(FLAGS_preprocess); 357 benchmark->setTimerResultType(TimerData::kPerIter_Result); 359 benchmark->setTimerResultType(TimerData::kMin_Result); 361 benchmark 158 run_single_benchmark(const SkString& inputPath, sk_tools::PictureBenchmark& benchmark) argument 368 process_input(const char* input, sk_tools::PictureBenchmark& benchmark) argument 436 sk_tools::PictureBenchmark benchmark; local [all...] |
/external/okhttp/benchmarks/src/main/java/com/squareup/okhttp/benchmarks/ |
H A D | ApacheHttpClient.java | 41 @Override public void prepare(Benchmark benchmark) { argument 42 super.prepare(benchmark); 44 if (benchmark.tls) {
|
H A D | HttpClient.java | 20 /** An HTTP client to benchmark. */ 22 void prepare(Benchmark benchmark); argument
|
H A D | NettyHttpClient.java | 64 @Override public void prepare(final Benchmark benchmark) { argument 65 this.concurrencyLevel = benchmark.concurrencyLevel; 66 this.targetBacklog = benchmark.targetBacklog; 72 if (benchmark.tls) {
|
H A D | OkHttp.java | 34 @Override public void prepare(Benchmark benchmark) { argument 35 super.prepare(benchmark); 37 client.setProtocols(benchmark.protocols); 39 if (benchmark.tls) {
|
H A D | OkHttpAsync.java | 45 @Override public void prepare(final Benchmark benchmark) { argument 46 concurrencyLevel = benchmark.concurrencyLevel; 47 targetBacklog = benchmark.targetBacklog; 50 client.setProtocols(benchmark.protocols); 51 client.setDispatcher(new Dispatcher(new ThreadPoolExecutor(benchmark.concurrencyLevel, 52 benchmark.concurrencyLevel, 60, TimeUnit.SECONDS, new LinkedBlockingQueue<Runnable>()))); 54 if (benchmark.tls) {
|
H A D | SynchronousHttpClient.java | 30 @Override public void prepare(Benchmark benchmark) { argument 31 this.targetBacklog = benchmark.targetBacklog; 32 executor = new ThreadPoolExecutor(benchmark.concurrencyLevel, benchmark.concurrencyLevel,
|
H A D | UrlConnection.java | 34 @Override public void prepare(Benchmark benchmark) { argument 35 super.prepare(benchmark); 36 if (benchmark.tls) {
|
/external/llvm/utils/yaml-bench/ |
H A D | YAMLBench.cpp | 139 static void benchmark( llvm::TimerGroup &Group function 209 llvm::TimerGroup Group("YAML parser benchmark"); 210 benchmark(Group, "Fast", createJSONText(10, 500)); 212 llvm::TimerGroup Group("YAML parser benchmark"); 213 benchmark(Group, "Small Values", createJSONText(MemoryLimitMB, 5)); 214 benchmark(Group, "Medium Values", createJSONText(MemoryLimitMB, 500)); 215 benchmark(Group, "Large Values", createJSONText(MemoryLimitMB, 50000));
|
/external/chromium_org/v8/tools/ |
H A D | lexer-shell.cc | 191 std::string benchmark; local 202 } else if (strncmp(argv[i], "--benchmark=", 12) == 0) { 203 benchmark = std::string(argv[i]).substr(12); 228 if (benchmark.empty()) benchmark = "Baseline"; 229 printf("%s(RunTime): %.f ms\n", benchmark.c_str(), baseline_total);
|
H A D | parser-shell.cc | 133 std::string benchmark; local 142 } else if (strncmp(argv[i], "--benchmark=", 12) == 0) { 143 benchmark = std::string(argv[i]).substr(12); 169 if (benchmark.empty()) benchmark = "Baseline"; 170 printf("%s(FirstParseRunTime): %.f ms\n", benchmark.c_str(), 172 printf("%s(SecondParseRunTime): %.f ms\n", benchmark.c_str(),
|
/external/chromium_org/third_party/skia/tools/ |
H A D | bench_pictures_main.cpp | 171 sk_tools::PictureBenchmark& benchmark) { 221 benchmark.run(picture); 255 static void setup_benchmark(sk_tools::PictureBenchmark* benchmark) { argument 338 benchmark->setTimersToShow(timerWall, truncatedTimerWall, timerCpu, truncatedTimerCpu, 364 benchmark->setTimeIndividualTiles(true); 367 benchmark->setPurgeDecodedTex(FLAGS_purgeDecodedTex); 368 benchmark->setPreprocess(FLAGS_preprocess); 377 benchmark->setTimerResultType(TimerData::kPerIter_Result); 379 benchmark->setTimerResultType(TimerData::kMin_Result); 381 benchmark 170 run_single_benchmark(const SkString& inputPath, sk_tools::PictureBenchmark& benchmark) argument 388 process_input(const char* input, sk_tools::PictureBenchmark& benchmark) argument 463 sk_tools::PictureBenchmark benchmark; local [all...] |
/external/chromium_org/third_party/leveldatabase/src/doc/bench/ |
H A D | db_bench_sqlite3.cc | 68 // flag and also specify a benchmark that wants a fresh database, that 69 // benchmark will fail. 279 // Pretend at least one op was done in case we are running a benchmark 408 fprintf(stderr, "unknown benchmark '%s'\n", name.ToString().c_str()); 715 leveldb::Benchmark benchmark; local 716 benchmark.Run();
|
H A D | db_bench_tree_db.cc | 63 // flag and also specify a benchmark that wants a fresh database, that 64 // benchmark will fail. 255 // Pretend at least one op was done in case we are running a benchmark 379 fprintf(stderr, "unknown benchmark '%s'\n", name.ToString().c_str()); 525 leveldb::Benchmark benchmark; local 526 benchmark.Run();
|
/external/chromium_org/third_party/leveldatabase/src/db/ |
H A D | db_bench.cc | 98 // flag and also specify a benchmark that wants a fresh database, that 99 // benchmark will fail. 240 // Pretend at least one op was done in case we are running a benchmark 268 // State shared by all concurrent executions of the same benchmark. 287 // Per-thread state for concurrent executions of the same benchmark. 508 fprintf(stderr, "unknown benchmark '%s'\n", name.ToString().c_str()); 976 leveldb::Benchmark benchmark; local 977 benchmark.Run();
|
/external/chromium_org/third_party/WebKit/Tools/Scripts/ |
H A D | bencher | 247 puts "that each benchmark/VM configuration is run in a single VM invocation, and" 249 puts "one benchmark polluting the running time of another. The fine-grained" 255 puts "VM invocation, and the number of VM invocations per benchmark. By default," 266 puts "--rerun <n> Set the number of iterations of the benchmark that" 270 puts "--outer <n> Set the number of runtime invocations for each benchmark." 306 puts "--local Also do a local benchmark run even when doing --remote." 780 result = plans.select{|v| v.suite.name == suiteName and v.benchmark.name == benchName and v.vm.name == vmName and v.iteration == iteration} 798 def benchmark method in class:ParsedResult 799 plan.benchmark 1069 def add(benchmark) 1113 def benchmark method in class:BenchRunPlan 1145 def benchmark method in class:BenchmarkOnVM 1220 def benchmark method in class:BenchPlan [all...] |
/external/chromium_org/tools/chrome_proxy/integration_tests/ |
H A D | chrome_proxy_benchmark.py | 7 from telemetry import benchmark namespace 10 @benchmark.Enabled('android') 11 class ChromeProxyLatency(benchmark.Benchmark): 21 @benchmark.Enabled('android') 22 class ChromeProxyLatencyDirect(benchmark.Benchmark): 29 @benchmark.Enabled('android') 34 @benchmark.Enabled('android') 39 @benchmark.Enabled('android') 40 class ChromeProxyDataSaving(benchmark.Benchmark): 49 @benchmark [all...] |
/external/chromium_org/tools/perf/benchmarks/ |
H A D | blink_perf.py | 7 from telemetry import benchmark namespace 100 class BlinkPerfAnimation(benchmark.Benchmark): 109 class BlinkPerfBindings(benchmark.Benchmark): 118 class BlinkPerfCSS(benchmark.Benchmark): 127 class BlinkPerfCanvas(benchmark.Benchmark): 136 class BlinkPerfDOM(benchmark.Benchmark): 145 class BlinkPerfEvents(benchmark.Benchmark): 154 class BlinkPerfInteractive(benchmark.Benchmark): 163 class BlinkPerfLayout(benchmark.Benchmark): 172 class BlinkPerfMutation(benchmark [all...] |
H A D | browsermark.py | 7 Browsermark benchmark suite have five test groups: 23 from telemetry import benchmark namespace 49 @benchmark.Disabled 50 class Browsermark(benchmark.Benchmark):
|
H A D | canvasmark.py | 5 """Runs Canvasmark HTML5, Canvas 2D rendering and javascript benchmark. 14 from telemetry import benchmark namespace 52 @benchmark.Disabled 53 class CanvasMark(benchmark.Benchmark):
|
H A D | dom_perf.py | 9 from telemetry import benchmark namespace 69 @benchmark.Disabled('android', 'linux') 70 class DomPerf(benchmark.Benchmark): 74 Scores are not comparable across benchmark suite versions and higher scores
|
H A D | dromaeo.py | 9 from telemetry import benchmark namespace 35 # Start spying on POST request that will report benchmark results, and 51 # Starts benchmark. 95 class _DromaeoBenchmark(benchmark.Benchmark): 104 raise NotImplementedError('query_param or tag not in Dromaeo benchmark.') 116 """Dromaeo DOMCore attr JavaScript benchmark.""" 122 """Dromaeo DOMCore modify JavaScript benchmark.""" 128 """Dromaeo DOMCore query JavaScript benchmark.""" 134 """Dromaeo DOMCore traverse JavaScript benchmark.""" 140 """Dromaeo JSLib attr jquery JavaScript benchmark""" [all...] |
H A D | endure.py | 7 In each Endure benchmark, one page action is performed repeatedly and memory 15 from telemetry import benchmark namespace 18 class _EndureBenchmark(benchmark.Benchmark): 31 @benchmark.Disabled 36 @benchmark.Disabled 41 @benchmark.Disabled 46 @benchmark.Disabled 51 @benchmark.Disabled 56 @benchmark.Disabled 61 @benchmark [all...] |
H A D | html5gaming.py | 5 """Impact HTML5 Gaming benchmark. 10 0.1)). The benchmark automatically runs at a reasonable screen size. Final 15 from telemetry import benchmark namespace 23 tab.ExecuteJavaScript('benchmark();') 35 @benchmark.Disabled 36 class HTML5Gaming(benchmark.Benchmark): 37 """Imapct HTML5 smooth running games benchmark suite.""" 44 ps.AddPageWithDefaultRunNavigate('http://html5-benchmark.com/')
|
H A D | image_decoding.py | 5 from telemetry import benchmark namespace 11 class ImageDecodingToughImageCases(benchmark.Benchmark):
|