treeherder/docs/backend_tasks.md

4.2 KiB

Backend tasks

Running the tests

You can run flake8, isort and the pytest suite inside Docker, using:

docker-compose run backend ./runtests.sh

Note: The Selenium tests will be skipped unless yarn build has been manually run prior.

Or for more control, run each tool individually, by first running:

docker-compose run backend bash

...which saves having to wait for docker-compose to spin up for every test run.

NOTE: To run Selenium tests you need to run yarn build for the tests not to skip. yarn build will generate a .build directory which will be seen within the backend container. If you don't have yarn working on your host you can run this instead docker-compose run frontend sh -c "yarn && yarn build"

Then run the individual tools within that shell, like so:

  • pytest:

    pytest tests/
    pytest tests/log_parser/test_tasks.py
    pytest tests/etl/test_job_loader.py -k test_ingest_pulse_jobs
    pytest tests/selenium/test_pin_jobs.py::test_pin_all_jobs
    

    To run all tests, including slow tests that are normally skipped, use:

    pytest --runslow tests/
    

    For more options, see pytest --help or https://docs.pytest.org/en/stable/usage.html.

    To assist with debugging Selenium test failures, an HTML reporting containing screenshots can be generated using:

    pytest tests/selenium/ --html report.html
    
  • flake8:

    flake8
    
  • isort (checks the Python import style):

    To run interactively:

    isort
    

    Or to apply all changes without confirmation:

    isort --apply
    

Hide Jobs with Tiers

To hide jobs we use the job's tier setting. Jobs with tier of 3 are hidden by default. For TaskCluster, edit the task definition to include the tier setting in the Treeherder section.

Profiling API endpoint performance

On our development (vagrant) instance we have django-debug-toolbar installed, which can give information on exactly what SQL is run to generate individual API endpoints. Navigate to an endpoint (example: http://localhost:8000/api/repository/) and you should see the toolbar to your right.

Connecting to Services Running inside Docker

Treeherder uses various services to function, eg MySQL, etc. At times it can be useful to connect to them from outside the Docker environment.

The docker-compose.yml file defines how internal ports are mapped to the host OS' ports.

In the below example we're mapping the container's port 3306 (MySQL's default port) to host port 3306.

# This is a line from the docker-compose.yml file
ports:
  - '3306:3306'

!!! note Any forwarded ports will block usage of that port on the host OS even if there isn't a service running inside the VM talking to it.

With MySQL exposed at port 3306 you can connect to it from your host OS with the following credentials:

  • host: localhost
  • port: 3306
  • user: root
  • password: leave blank

Other services running inside the Compose project, can be accessed in the same way.

Releasing a new version of the Python client

  • Determine whether the patch, minor or major version should be bumped, by inspecting the client Git log.

  • File a separate bug for the version bump.

  • Open a PR to update the version listed in client.py.

  • Use Twine to publish both the sdist and the wheel to PyPI, by running the following from the root of the Treeherder repository:

    > pip install -U twine wheel
    > cd treeherder/client/
    > rm -rf dist/*
    > python setup.py sdist bdist_wheel
    > twine upload dist/*
    
  • File a Release Engineering::Buildduty bug requesting that the sdist and wheel releases (plus any new dependent packages) be added to the internal PyPI mirror. For an example, see bug 1236965.