mirror of
https://github.com/flutter/flutter.git
synced 2025-06-03 00:51:18 +00:00

## FlutterTimeline Add a new class `FlutterTimeline` that's a drop-in replacement for `Timeline` from `dart:developer`. In addition to forwarding invocations of `startSync`, `finishSync`, `timeSync`, and `instantSync` to `dart:developer`, provides the following extra methods that make is easy to collect timings for code blocks on a frame-by-frame basis: * `debugCollect()` - aggregates timings since the last reset, or since the app launched. * `debugReset()` - forgets all data collected since the previous reset, or since the app launched. This allows clearing data from previous frames so timings can be attributed to the current frame. * `now` - this was enhanced so that it works on the web by calling `window.performance.now` (in `Timeline` this is a noop in Dart web compilers). * `collectionEnabled` - a field that controls whether `FlutterTimeline` stores timings in memory. By default this is disabled to avoid unexpected overhead (although the class is designed for minimal and predictable overhead). Specific benchmarks can enable collection to report to Skia Perf. ## Semantics benchmarks Add `BenchMaterial3Semantics` that benchmarks the cost of semantics when constructing a screen full of Material 3 widgets from nothing. It is expected that semantics will have non-trivial cost in this case, but we should strive to keep it much lower than the rendering cost. This is the case already. This benchmark shows that the cost of semantics is <10%. Add `BenchMaterial3ScrollSemantics` that benchmarks the cost of scrolling a previously constructed screen full of Material 3 widgets. The expectation should be that semantics will have trivial cost, since we're just shifting some widgets around. As of today, the numbers are not great, with semantics taking >50% of frame time, which is what prompted this PR in the first place. As we optimize this, we want to see this number improve.
480 lines
17 KiB
Dart
480 lines
17 KiB
Dart
// Copyright 2014 The Flutter Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
// found in the LICENSE file.
|
|
|
|
import 'dart:async';
|
|
import 'dart:convert' show json;
|
|
import 'dart:js_interop';
|
|
import 'dart:math' as math;
|
|
|
|
import 'package:web/web.dart' as web;
|
|
|
|
import 'src/web/bench_build_image.dart';
|
|
import 'src/web/bench_build_material_checkbox.dart';
|
|
import 'src/web/bench_card_infinite_scroll.dart';
|
|
import 'src/web/bench_child_layers.dart';
|
|
import 'src/web/bench_clipped_out_pictures.dart';
|
|
import 'src/web/bench_default_target_platform.dart';
|
|
import 'src/web/bench_draw_rect.dart';
|
|
import 'src/web/bench_dynamic_clip_on_static_picture.dart';
|
|
import 'src/web/bench_image_decoding.dart';
|
|
import 'src/web/bench_material_3.dart';
|
|
import 'src/web/bench_material_3_semantics.dart';
|
|
import 'src/web/bench_mouse_region_grid_hover.dart';
|
|
import 'src/web/bench_mouse_region_grid_scroll.dart';
|
|
import 'src/web/bench_mouse_region_mixed_grid_hover.dart';
|
|
import 'src/web/bench_pageview_scroll_linethrough.dart';
|
|
import 'src/web/bench_paths.dart';
|
|
import 'src/web/bench_picture_recording.dart';
|
|
import 'src/web/bench_platform_view_infinite_scroll.dart';
|
|
import 'src/web/bench_simple_lazy_text_scroll.dart';
|
|
import 'src/web/bench_text_layout.dart';
|
|
import 'src/web/bench_text_out_of_picture_bounds.dart';
|
|
import 'src/web/bench_wrapbox_scroll.dart';
|
|
import 'src/web/recorder.dart';
|
|
|
|
typedef RecorderFactory = Recorder Function();
|
|
|
|
const bool isCanvasKit = bool.fromEnvironment('FLUTTER_WEB_USE_SKIA');
|
|
|
|
/// List of all benchmarks that run in the devicelab.
|
|
///
|
|
/// When adding a new benchmark, add it to this map. Make sure that the name
|
|
/// of your benchmark is unique.
|
|
final Map<String, RecorderFactory> benchmarks = <String, RecorderFactory>{
|
|
// Benchmarks that run both in CanvasKit and HTML modes
|
|
BenchDefaultTargetPlatform.benchmarkName: () => BenchDefaultTargetPlatform(),
|
|
BenchBuildImage.benchmarkName: () => BenchBuildImage(),
|
|
BenchCardInfiniteScroll.benchmarkName: () => BenchCardInfiniteScroll.forward(),
|
|
BenchCardInfiniteScroll.benchmarkNameBackward: () => BenchCardInfiniteScroll.backward(),
|
|
BenchClippedOutPictures.benchmarkName: () => BenchClippedOutPictures(),
|
|
BenchDrawRect.benchmarkName: () => BenchDrawRect.staticPaint(),
|
|
BenchDrawRect.variablePaintBenchmarkName: () => BenchDrawRect.variablePaint(),
|
|
BenchPathRecording.benchmarkName: () => BenchPathRecording(),
|
|
BenchTextOutOfPictureBounds.benchmarkName: () => BenchTextOutOfPictureBounds(),
|
|
BenchSimpleLazyTextScroll.benchmarkName: () => BenchSimpleLazyTextScroll(),
|
|
BenchBuildMaterialCheckbox.benchmarkName: () => BenchBuildMaterialCheckbox(),
|
|
BenchDynamicClipOnStaticPicture.benchmarkName: () => BenchDynamicClipOnStaticPicture(),
|
|
BenchPageViewScrollLineThrough.benchmarkName: () => BenchPageViewScrollLineThrough(),
|
|
BenchPictureRecording.benchmarkName: () => BenchPictureRecording(),
|
|
BenchUpdateManyChildLayers.benchmarkName: () => BenchUpdateManyChildLayers(),
|
|
BenchMouseRegionGridScroll.benchmarkName: () => BenchMouseRegionGridScroll(),
|
|
BenchMouseRegionGridHover.benchmarkName: () => BenchMouseRegionGridHover(),
|
|
BenchMouseRegionMixedGridHover.benchmarkName: () => BenchMouseRegionMixedGridHover(),
|
|
BenchWrapBoxScroll.benchmarkName: () => BenchWrapBoxScroll(),
|
|
BenchPlatformViewInfiniteScroll.benchmarkName: () => BenchPlatformViewInfiniteScroll.forward(),
|
|
BenchPlatformViewInfiniteScroll.benchmarkNameBackward: () => BenchPlatformViewInfiniteScroll.backward(),
|
|
BenchMaterial3Components.benchmarkName: () => BenchMaterial3Components(),
|
|
BenchMaterial3Semantics.benchmarkName: () => BenchMaterial3Semantics(),
|
|
BenchMaterial3ScrollSemantics.benchmarkName: () => BenchMaterial3ScrollSemantics(),
|
|
|
|
// CanvasKit-only benchmarks
|
|
if (isCanvasKit) ...<String, RecorderFactory>{
|
|
BenchTextLayout.canvasKitBenchmarkName: () => BenchTextLayout.canvasKit(),
|
|
BenchBuildColorsGrid.canvasKitBenchmarkName: () => BenchBuildColorsGrid.canvasKit(),
|
|
BenchTextCachedLayout.canvasKitBenchmarkName: () => BenchTextCachedLayout.canvasKit(),
|
|
|
|
// The HTML renderer does not decode frame-by-frame. It just drops an <img>
|
|
// element and lets it animate automatically with no feedback to the
|
|
// framework. So this benchmark only makes sense in CanvasKit.
|
|
BenchImageDecoding.benchmarkName: () => BenchImageDecoding(),
|
|
},
|
|
|
|
// HTML-only benchmarks
|
|
if (!isCanvasKit) ...<String, RecorderFactory>{
|
|
BenchTextLayout.canvasBenchmarkName: () => BenchTextLayout.canvas(),
|
|
BenchTextCachedLayout.canvasBenchmarkName: () => BenchTextCachedLayout.canvas(),
|
|
BenchBuildColorsGrid.canvasBenchmarkName: () => BenchBuildColorsGrid.canvas(),
|
|
},
|
|
};
|
|
|
|
final LocalBenchmarkServerClient _client = LocalBenchmarkServerClient();
|
|
|
|
Future<void> main() async {
|
|
// Check if the benchmark server wants us to run a specific benchmark.
|
|
final String nextBenchmark = await _client.requestNextBenchmark();
|
|
|
|
if (nextBenchmark == LocalBenchmarkServerClient.kManualFallback) {
|
|
_fallbackToManual('The server did not tell us which benchmark to run next.');
|
|
return;
|
|
}
|
|
|
|
await _runBenchmark(nextBenchmark);
|
|
web.window.location.reload();
|
|
}
|
|
|
|
Future<void> _runBenchmark(String benchmarkName) async {
|
|
final RecorderFactory? recorderFactory = benchmarks[benchmarkName];
|
|
|
|
if (recorderFactory == null) {
|
|
_fallbackToManual('Benchmark $benchmarkName not found.');
|
|
return;
|
|
}
|
|
|
|
await runZoned<Future<void>>(
|
|
() async {
|
|
final Recorder recorder = recorderFactory();
|
|
final Runner runner = recorder.isTracingEnabled && !_client.isInManualMode
|
|
? Runner(
|
|
recorder: recorder,
|
|
setUpAllDidRun: () => _client.startPerformanceTracing(benchmarkName),
|
|
tearDownAllWillRun: _client.stopPerformanceTracing,
|
|
)
|
|
: Runner(recorder: recorder);
|
|
|
|
final Profile profile = await runner.run();
|
|
if (!_client.isInManualMode) {
|
|
await _client.sendProfileData(profile);
|
|
} else {
|
|
_printResultsToScreen(profile);
|
|
print(profile);
|
|
}
|
|
},
|
|
zoneSpecification: ZoneSpecification(
|
|
print: (Zone self, ZoneDelegate parent, Zone zone, String line) async {
|
|
if (_client.isInManualMode) {
|
|
parent.print(zone, '[$benchmarkName] $line');
|
|
} else {
|
|
await _client.printToConsole(line);
|
|
}
|
|
},
|
|
handleUncaughtError: (
|
|
Zone self,
|
|
ZoneDelegate parent,
|
|
Zone zone, Object error,
|
|
StackTrace stackTrace,
|
|
) async {
|
|
if (_client.isInManualMode) {
|
|
parent.print(zone, '[$benchmarkName] $error, $stackTrace');
|
|
parent.handleUncaughtError(zone, error, stackTrace);
|
|
} else {
|
|
await _client.reportError(error, stackTrace);
|
|
}
|
|
},
|
|
),
|
|
);
|
|
}
|
|
|
|
extension WebHTMLElementExtension on web.HTMLElement {
|
|
void appendHtml(String html) {
|
|
final web.HTMLDivElement div = web.document.createElement('div') as
|
|
web.HTMLDivElement;
|
|
div.innerHTML = html;
|
|
final web.DocumentFragment fragment = web.document.createDocumentFragment();
|
|
fragment.append(div);
|
|
web.document.adoptNode(fragment);
|
|
append(fragment);
|
|
}
|
|
}
|
|
|
|
void _fallbackToManual(String error) {
|
|
web.document.body!.appendHtml('''
|
|
<div id="manual-panel">
|
|
<h3>$error</h3>
|
|
|
|
<p>Choose one of the following benchmarks:</p>
|
|
|
|
<!-- Absolutely position it so it receives the clicks and not the glasspane -->
|
|
<ul style="position: absolute">
|
|
${
|
|
benchmarks.keys
|
|
.map((String name) => '<li><button id="$name">$name</button></li>')
|
|
.join('\n')
|
|
}
|
|
</ul>
|
|
</div>
|
|
''');
|
|
|
|
for (final String benchmarkName in benchmarks.keys) {
|
|
final web.Element button = web.document.querySelector('#$benchmarkName')!;
|
|
button.addEventListener('click', (JSObject _) {
|
|
final web.Element? manualPanel =
|
|
web.document.querySelector('#manual-panel');
|
|
manualPanel?.remove();
|
|
_runBenchmark(benchmarkName);
|
|
}.toJS);
|
|
}
|
|
}
|
|
|
|
/// Visualizes results on the Web page for manual inspection.
|
|
void _printResultsToScreen(Profile profile) {
|
|
web.document.body!.remove();
|
|
web.document.body = web.document.createElement('body') as web.HTMLBodyElement;
|
|
web.document.body!.appendHtml('<h2>${profile.name}</h2>');
|
|
|
|
profile.scoreData.forEach((String scoreKey, Timeseries timeseries) {
|
|
web.document.body!.appendHtml('<h2>$scoreKey</h2>');
|
|
web.document.body!.appendHtml('<pre>${timeseries.computeStats()}</pre>');
|
|
web.document.body!.append(TimeseriesVisualization(timeseries).render());
|
|
});
|
|
}
|
|
|
|
/// Draws timeseries data and statistics on a canvas.
|
|
class TimeseriesVisualization {
|
|
TimeseriesVisualization(this._timeseries) {
|
|
_stats = _timeseries.computeStats();
|
|
_canvas = web.document.createElement('canvas') as web.HTMLCanvasElement;
|
|
_screenWidth = web.window.screen.width;
|
|
_canvas.width = _screenWidth;
|
|
_canvas.height = (_kCanvasHeight * web.window.devicePixelRatio).round();
|
|
_canvas.style
|
|
..setProperty('width', '100%')
|
|
..setProperty('height', '${_kCanvasHeight}px')
|
|
..setProperty('outline', '1px solid green');
|
|
_ctx = _canvas.getContext('2d')! as web.CanvasRenderingContext2D;
|
|
|
|
// The amount of vertical space available on the chart. Because some
|
|
// outliers can be huge they can dwarf all the useful values. So we
|
|
// limit it to 1.5 x the biggest non-outlier.
|
|
_maxValueChartRange = 1.5 * _stats.samples
|
|
.where((AnnotatedSample sample) => !sample.isOutlier)
|
|
.map<double>((AnnotatedSample sample) => sample.magnitude)
|
|
.fold<double>(0, math.max);
|
|
}
|
|
|
|
static const double _kCanvasHeight = 200;
|
|
|
|
final Timeseries _timeseries;
|
|
late TimeseriesStats _stats;
|
|
late web.HTMLCanvasElement _canvas;
|
|
late web.CanvasRenderingContext2D _ctx;
|
|
late int _screenWidth;
|
|
|
|
// Used to normalize benchmark values to chart height.
|
|
late double _maxValueChartRange;
|
|
|
|
/// Converts a sample value to vertical canvas coordinates.
|
|
///
|
|
/// This does not work for horizontal coordinates.
|
|
double _normalized(double value) {
|
|
return _kCanvasHeight * value / _maxValueChartRange;
|
|
}
|
|
|
|
/// A utility for drawing lines.
|
|
void drawLine(num x1, num y1, num x2, num y2) {
|
|
_ctx.beginPath();
|
|
_ctx.moveTo(x1.toDouble(), y1.toDouble());
|
|
_ctx.lineTo(x2.toDouble(), y2.toDouble());
|
|
_ctx.stroke();
|
|
}
|
|
|
|
/// Renders the timeseries into a `<canvas>` and returns the canvas element.
|
|
web.HTMLCanvasElement render() {
|
|
_ctx.translate(0, _kCanvasHeight * web.window.devicePixelRatio);
|
|
_ctx.scale(1, -web.window.devicePixelRatio);
|
|
|
|
final double barWidth = _screenWidth / _stats.samples.length;
|
|
double xOffset = 0;
|
|
for (int i = 0; i < _stats.samples.length; i++) {
|
|
final AnnotatedSample sample = _stats.samples[i];
|
|
|
|
if (sample.isWarmUpValue) {
|
|
// Put gray background behind warm-up samples.
|
|
_ctx.fillStyle = 'rgba(200,200,200,1)'.toJS;
|
|
_ctx.fillRect(xOffset, 0, barWidth, _normalized(_maxValueChartRange));
|
|
}
|
|
|
|
if (sample.magnitude > _maxValueChartRange) {
|
|
// The sample value is so big it doesn't fit on the chart. Paint it purple.
|
|
_ctx.fillStyle = 'rgba(100,50,100,0.8)'.toJS;
|
|
} else if (sample.isOutlier) {
|
|
// The sample is an outlier, color it light red.
|
|
_ctx.fillStyle = 'rgba(255,50,50,0.6)'.toJS;
|
|
} else {
|
|
// A non-outlier sample, color it light blue.
|
|
_ctx.fillStyle = 'rgba(50,50,255,0.6)'.toJS;
|
|
}
|
|
|
|
_ctx.fillRect(xOffset, 0, barWidth - 1, _normalized(sample.magnitude));
|
|
xOffset += barWidth;
|
|
}
|
|
|
|
// Draw a horizontal solid line corresponding to the average.
|
|
_ctx.lineWidth = 1;
|
|
drawLine(0, _normalized(_stats.average), _screenWidth, _normalized(_stats.average));
|
|
|
|
// Draw a horizontal dashed line corresponding to the outlier cut off.
|
|
_ctx.setLineDash(<JSAny?>[5.toJS, 5.toJS].toJS);
|
|
drawLine(0, _normalized(_stats.outlierCutOff), _screenWidth, _normalized(_stats.outlierCutOff));
|
|
|
|
// Draw a light red band that shows the noise (1 stddev in each direction).
|
|
_ctx.fillStyle = 'rgba(255,50,50,0.3)'.toJS;
|
|
_ctx.fillRect(
|
|
0,
|
|
_normalized(_stats.average * (1 - _stats.noise)),
|
|
_screenWidth.toDouble(),
|
|
_normalized(2 * _stats.average * _stats.noise),
|
|
);
|
|
|
|
return _canvas;
|
|
}
|
|
}
|
|
|
|
/// Implements the client REST API for the local benchmark server.
|
|
///
|
|
/// The local server is optional. If it is not available the benchmark UI must
|
|
/// implement a manual fallback. This allows debugging benchmarks using plain
|
|
/// `flutter run`.
|
|
class LocalBenchmarkServerClient {
|
|
/// This value is returned by [requestNextBenchmark].
|
|
static const String kManualFallback = '__manual_fallback__';
|
|
|
|
/// Whether we fell back to manual mode.
|
|
///
|
|
/// This happens when you run benchmarks using plain `flutter run` rather than
|
|
/// devicelab test harness. The test harness spins up a special server that
|
|
/// provides API for automatically picking the next benchmark to run.
|
|
bool isInManualMode = false;
|
|
|
|
/// Asks the local server for the name of the next benchmark to run.
|
|
///
|
|
/// Returns [kManualFallback] if local server is not available (uses 404 as a
|
|
/// signal).
|
|
Future<String> requestNextBenchmark() async {
|
|
final web.XMLHttpRequest request = await _requestXhr(
|
|
'/next-benchmark',
|
|
method: 'POST',
|
|
mimeType: 'application/json',
|
|
sendData: json.encode(benchmarks.keys.toList()),
|
|
);
|
|
|
|
// 404 is expected in the following cases:
|
|
// - The benchmark is ran using plain `flutter run`, which does not provide "next-benchmark" handler.
|
|
// - We ran all benchmarks and the benchmark is telling us there are no more benchmarks to run.
|
|
if (request.status != 200) {
|
|
isInManualMode = true;
|
|
return kManualFallback;
|
|
}
|
|
|
|
isInManualMode = false;
|
|
return request.responseText;
|
|
}
|
|
|
|
void _checkNotManualMode() {
|
|
if (isInManualMode) {
|
|
throw StateError('Operation not supported in manual fallback mode.');
|
|
}
|
|
}
|
|
|
|
/// Asks the local server to begin tracing performance.
|
|
///
|
|
/// This uses the chrome://tracing tracer, which is not available from within
|
|
/// the page itself, and therefore must be controlled from outside using the
|
|
/// DevTools Protocol.
|
|
Future<void> startPerformanceTracing(String benchmarkName) async {
|
|
_checkNotManualMode();
|
|
await _requestXhr(
|
|
'/start-performance-tracing?label=$benchmarkName',
|
|
method: 'POST',
|
|
mimeType: 'application/json',
|
|
);
|
|
}
|
|
|
|
/// Stops the performance tracing session started by [startPerformanceTracing].
|
|
Future<void> stopPerformanceTracing() async {
|
|
_checkNotManualMode();
|
|
await _requestXhr(
|
|
'/stop-performance-tracing',
|
|
method: 'POST',
|
|
mimeType: 'application/json',
|
|
);
|
|
}
|
|
|
|
/// Sends the profile data collected by the benchmark to the local benchmark
|
|
/// server.
|
|
Future<void> sendProfileData(Profile profile) async {
|
|
_checkNotManualMode();
|
|
final web.XMLHttpRequest request = await _requestXhr(
|
|
'/profile-data',
|
|
method: 'POST',
|
|
mimeType: 'application/json',
|
|
sendData: json.encode(profile.toJson()),
|
|
);
|
|
if (request.status != 200) {
|
|
throw Exception(
|
|
'Failed to report profile data to benchmark server. '
|
|
'The server responded with status code ${request.status}.'
|
|
);
|
|
}
|
|
}
|
|
|
|
/// Reports an error to the benchmark server.
|
|
///
|
|
/// The server will halt the devicelab task and log the error.
|
|
Future<void> reportError(dynamic error, StackTrace stackTrace) async {
|
|
_checkNotManualMode();
|
|
await _requestXhr(
|
|
'/on-error',
|
|
method: 'POST',
|
|
mimeType: 'application/json',
|
|
sendData: json.encode(<String, dynamic>{
|
|
'error': '$error',
|
|
'stackTrace': '$stackTrace',
|
|
}),
|
|
);
|
|
}
|
|
|
|
/// Reports a message about the demo to the benchmark server.
|
|
Future<void> printToConsole(String report) async {
|
|
_checkNotManualMode();
|
|
await _requestXhr(
|
|
'/print-to-console',
|
|
method: 'POST',
|
|
mimeType: 'text/plain',
|
|
sendData: report,
|
|
);
|
|
}
|
|
|
|
/// This is the same as calling [html.HttpRequest.request] but it doesn't
|
|
/// crash on 404, which we use to detect `flutter run`.
|
|
Future<web.XMLHttpRequest> _requestXhr(
|
|
String url, {
|
|
String? method,
|
|
bool? withCredentials,
|
|
String? responseType,
|
|
String? mimeType,
|
|
Map<String, String>? requestHeaders,
|
|
dynamic sendData,
|
|
}) {
|
|
final Completer<web.XMLHttpRequest> completer = Completer<web.XMLHttpRequest>();
|
|
final web.XMLHttpRequest xhr = web.XMLHttpRequest();
|
|
|
|
method ??= 'GET';
|
|
xhr.open(method, url, true);
|
|
|
|
if (withCredentials != null) {
|
|
xhr.withCredentials = withCredentials;
|
|
}
|
|
|
|
if (responseType != null) {
|
|
xhr.responseType = responseType;
|
|
}
|
|
|
|
if (mimeType != null) {
|
|
xhr.overrideMimeType(mimeType);
|
|
}
|
|
|
|
if (requestHeaders != null) {
|
|
requestHeaders.forEach((String header, String value) {
|
|
xhr.setRequestHeader(header, value);
|
|
});
|
|
}
|
|
|
|
xhr.addEventListener('load', (web.ProgressEvent e) {
|
|
completer.complete(xhr);
|
|
}.toJS);
|
|
|
|
xhr.addEventListener('error', (JSObject error) {
|
|
return completer.completeError(error);
|
|
}.toJS);
|
|
|
|
if (sendData != null) {
|
|
xhr.send((sendData as Object?).jsify());
|
|
} else {
|
|
xhr.send();
|
|
}
|
|
|
|
return completer.future;
|
|
}
|
|
}
|