зеркало из https://github.com/mozilla/gecko-dev.git
Bug 1890318 - Update perfdocs linter documentation. r=perftest-reviewers,kshampur DONTBUILD
This patch updates the perfdocs linter documentation to contain the new config schema, and information about how to add documentation for the metrics. Differential Revision: https://phabricator.services.mozilla.com/D206953
This commit is contained in:
Родитель
764028466e
Коммит
154791c4a2
|
@ -14,21 +14,42 @@ The mozlint integration of PerfDocs can be run using mach:
|
|||
|
||||
$ mach lint --linter perfdocs .
|
||||
|
||||
Documentation can be regenerated for performance tests by including the ``--fix`` flag:
|
||||
|
||||
.. parsed-literal::
|
||||
|
||||
$ mach lint --linter perfdocs . --fix
|
||||
|
||||
|
||||
Configuration
|
||||
-------------
|
||||
|
||||
There are no configuration options available for this linter. It scans the full source tree under ``testing``, looking for folders named ``perfdocs`` and then validates their content. This has only been implemented for Raptor so far, but Talos will be added in the future. We also hope to expand this to search outside the ``testing`` directory.
|
||||
There are no configuration options available for this linter. It scans the full source tree under ``testing``, looking for folders named ``perfdocs``, validates their content, and regenerates the documentation (if ``--fix`` is provided). This has been implemented for all performance testing harnesses, and the documentation generated gets displayed in :ref:`Performance Testing`.
|
||||
|
||||
The ``perfdocs`` folders, there needs to be an ``index.rst`` file and it needs to contain the string ``{documentation}`` in some location in the file which is where the test documentation will be placed. The folders must also have a ``config.yml`` file following this schema:
|
||||
In the ``perfdocs`` folders, there needs to be an ``index.rst`` file and it needs to contain the string ``{documentation}`` in some location in the file which is where the test documentation will be placed. The folders must also have a ``config.yml`` file following this schema:
|
||||
|
||||
.. code-block:: python
|
||||
|
||||
CONFIG_SCHEMA = {
|
||||
"definitions": {
|
||||
"metrics_schema": {
|
||||
"metric_name": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"aliases": {"type": "array", "items": {"type": "string"}},
|
||||
"description": {"type": "string"},
|
||||
"matcher": {"type": "string"},
|
||||
},
|
||||
"required": ["description", "aliases"],
|
||||
},
|
||||
},
|
||||
},
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"name": {"type": "string"},
|
||||
"manifest": {"type": "string"},
|
||||
"static-only": {"type": "boolean"},
|
||||
"metrics": {"$ref": "#/definitions/metrics_schema"},
|
||||
"suites": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
|
@ -39,22 +60,17 @@ The ``perfdocs`` folders, there needs to be an ``index.rst`` file and it needs t
|
|||
"type": "object",
|
||||
"properties": {
|
||||
"test_name": {"type": "string"},
|
||||
}
|
||||
},
|
||||
},
|
||||
"description": {"type": "string"},
|
||||
"owner": {"type": "string"},
|
||||
},
|
||||
"required": [
|
||||
"description"
|
||||
]
|
||||
}
|
||||
}
|
||||
"required": ["description"],
|
||||
}
|
||||
},
|
||||
"required": [
|
||||
"name",
|
||||
"manifest",
|
||||
"suites"
|
||||
]
|
||||
},
|
||||
},
|
||||
"required": ["name", "manifest", "static-only", "suites"],
|
||||
}
|
||||
|
||||
Here is an example of a configuration file for the Raptor framework:
|
||||
|
@ -75,6 +91,35 @@ Here is an example of a configuration file for the Raptor framework:
|
|||
tests:
|
||||
wasm: "All wasm tests."
|
||||
|
||||
Metrics that produce alerts can also be documented like so:
|
||||
|
||||
.. parsed-literal::
|
||||
|
||||
name: raptor
|
||||
manifest: testing/raptor/raptor/raptor.toml
|
||||
metrics:
|
||||
"First Paint":
|
||||
description: "The description of the metric."
|
||||
aliases:
|
||||
- fcp
|
||||
- anAliasForFCP
|
||||
# Optional regex to match the metrics found in the tests with this
|
||||
# documented metric
|
||||
matcher: f.*
|
||||
suites:
|
||||
desktop:
|
||||
description: "Desktop tests."
|
||||
tests:
|
||||
raptor-tp6: "Raptor TP6 tests."
|
||||
mobile:
|
||||
description: "Mobile tests"
|
||||
benchmarks:
|
||||
description: "Benchmark tests."
|
||||
tests:
|
||||
wasm: "All wasm tests."
|
||||
|
||||
The documented metrics must exist in the tests for the suite. If they are not, then validation will fail. The same is true if a metric in a test is not documented. Also, if ``metrics`` are defined, then a ``metrics.rst`` file is expected to be found in the ``perfdocs`` folder for the given suite. It must contain the string ``{metrics_documentation}`` where the documentation should be added. The ``metrics.rst`` is renamed ``{suite-name}-metrics.rst`` in the generated folder, so if it needs to be linked to in the ``index.rst`` file, it should contain a ``{metrics_rst_name}`` string for where the link should be added - it's expected to be found in a toctree section.
|
||||
|
||||
Note that there needs to be a FrameworkGatherer implemented for the framework being documented since each of them may have different ways of parsing test manifests for the tests. See `RaptorGatherer <https://searchfox.org/mozilla-central/source/tools/lint/perfdocs/framework_gatherers.py>`_ for an example gatherer that was implemented for Raptor.
|
||||
|
||||
Sources
|
||||
|
|
Загрузка…
Ссылка в новой задаче