зеркало из https://github.com/mozilla/treeherder.git
New Black changes
This commit is contained in:
Родитель
20921c7b48
Коммит
b70edaa7ea
|
@ -22,7 +22,6 @@ def prepare_responses():
|
|||
now = datetime.now().strftime("%Y-%m-%dT%H:%M:%S")
|
||||
|
||||
def releases(request):
|
||||
|
||||
data = [
|
||||
{
|
||||
"name": "ok",
|
||||
|
|
|
@ -412,7 +412,6 @@ def eleven_job_blobs(sample_data, sample_push, test_repository, mock_log_parser)
|
|||
|
||||
blobs = []
|
||||
for blob in jobs:
|
||||
|
||||
if push_index > max_index:
|
||||
push_index = 0
|
||||
|
||||
|
@ -449,7 +448,6 @@ def eleven_job_blobs_new_date(sample_data, sample_push, test_repository, mock_lo
|
|||
|
||||
blobs = []
|
||||
for blob in jobs:
|
||||
|
||||
if push_index > max_index:
|
||||
push_index = 0
|
||||
|
||||
|
@ -510,7 +508,6 @@ def test_job_with_notes(test_job, test_user):
|
|||
|
||||
@pytest.fixture
|
||||
def activate_responses(request):
|
||||
|
||||
responses.start()
|
||||
|
||||
def fin():
|
||||
|
@ -1039,7 +1036,6 @@ def test_run_data(bug_data):
|
|||
|
||||
@pytest.fixture
|
||||
def group_data(transactional_db, eleven_job_blobs, create_jobs):
|
||||
|
||||
query_string = '?manifest=/test&date=2022-10-01'
|
||||
|
||||
jt = []
|
||||
|
@ -1094,7 +1090,7 @@ def generate_enough_perf_datum(test_repository, test_perf_signature):
|
|||
# extra data on both sides to make sure we're using the proper values
|
||||
# to generate the actual alert)
|
||||
|
||||
for (push_id, value) in zip([1] * 30 + [2] * 30, [1] * 30 + [2] * 30):
|
||||
for push_id, value in zip([1] * 30 + [2] * 30, [1] * 30 + [2] * 30):
|
||||
push = th_models.Push.objects.get(id=push_id)
|
||||
perf_models.PerformanceDatum.objects.create(
|
||||
repository=test_repository,
|
||||
|
|
|
@ -58,7 +58,6 @@ def test_store_job_with_unparsed_log(
|
|||
def test_store_job_pending_to_completed_with_unparsed_log(
|
||||
test_repository, push_stored, failure_classifications, activate_responses
|
||||
):
|
||||
|
||||
job_guid = 'd22c74d4aa6d2a1dcba96d95dccbd5fdca70cf33'
|
||||
|
||||
# the first time, submit it as running (with no logs)
|
||||
|
|
|
@ -121,7 +121,7 @@ def test_ingest_running_to_retry_to_success_sample_job(
|
|||
job_guid_root = job['job_guid']
|
||||
|
||||
job_data = []
|
||||
for (state, result, job_guid) in [
|
||||
for state, result, job_guid in [
|
||||
('running', 'unknown', job_guid_root),
|
||||
('completed', 'retry', job_guid_root + "_" + str(job['end_timestamp'])[-5:]),
|
||||
('completed', 'success', job_guid_root),
|
||||
|
@ -132,7 +132,7 @@ def test_ingest_running_to_retry_to_success_sample_job(
|
|||
new_job_datum['job']['job_guid'] = job_guid
|
||||
job_data.append(new_job_datum)
|
||||
|
||||
for (i, j) in ingestion_cycles:
|
||||
for i, j in ingestion_cycles:
|
||||
store_job_data(test_repository, job_data[i:j])
|
||||
|
||||
assert Job.objects.count() == 2
|
||||
|
@ -165,7 +165,7 @@ def test_ingest_running_to_retry_to_success_sample_job_multiple_retries(
|
|||
job_guid_root = job['job_guid']
|
||||
|
||||
job_data = []
|
||||
for (state, result, job_guid) in [
|
||||
for state, result, job_guid in [
|
||||
('running', 'unknown', job_guid_root),
|
||||
('completed', 'retry', job_guid_root + "_" + str(job['end_timestamp'])[-5:]),
|
||||
('completed', 'retry', job_guid_root + "_12345"),
|
||||
|
@ -177,7 +177,7 @@ def test_ingest_running_to_retry_to_success_sample_job_multiple_retries(
|
|||
new_job_datum['job']['job_guid'] = job_guid
|
||||
job_data.append(new_job_datum)
|
||||
|
||||
for (i, j) in ingestion_cycles:
|
||||
for i, j in ingestion_cycles:
|
||||
ins = job_data[i:j]
|
||||
store_job_data(test_repository, ins)
|
||||
|
||||
|
|
|
@ -111,7 +111,7 @@ def _generate_perf_data_range(
|
|||
if reverse_push_range:
|
||||
push_range = reversed(push_range)
|
||||
|
||||
for (i, value) in zip(push_range, [1] * 15 + [2] * 15):
|
||||
for i, value in zip(push_range, [1] * 15 + [2] * 15):
|
||||
push_time = datetime.datetime.fromtimestamp(now + i)
|
||||
push = Push.objects.create(
|
||||
repository=test_repository,
|
||||
|
|
|
@ -14,7 +14,7 @@ def test_performance_log_parsing():
|
|||
"""
|
||||
|
||||
# first two have only one artifact, second has two artifacts
|
||||
for (logfile, num_perf_artifacts) in [
|
||||
for logfile, num_perf_artifacts in [
|
||||
('mozilla-inbound-android-api-11-debug-bm91-build1-build1317.txt.gz', 1),
|
||||
('try_ubuntu64_hw_test-chromez-bm103-tests1-linux-build1429.txt.gz', 1),
|
||||
('mozilla-inbound-linux64-bm72-build1-build225.txt.gz', 2),
|
||||
|
|
|
@ -96,7 +96,7 @@ def test_cycle_all_but_one_job(
|
|||
assert Job.objects.count() == 1
|
||||
assert JobLog.objects.count() == (num_job_logs_before - num_job_logs_to_be_deleted)
|
||||
|
||||
for (object_type, objects) in extra_objects.values():
|
||||
for object_type, objects in extra_objects.values():
|
||||
actual = set(item.id for item in object_type.objects.all())
|
||||
expected = set(item.id for item in objects)
|
||||
assert actual == expected
|
||||
|
|
|
@ -197,7 +197,7 @@ def test_import(mock_bugscache_bugzilla_request):
|
|||
1736534: [],
|
||||
}
|
||||
|
||||
for (open_bug, duplicates) in EXPECTED_BUG_DUPE_OF_DATA.items():
|
||||
for open_bug, duplicates in EXPECTED_BUG_DUPE_OF_DATA.items():
|
||||
assert Bugscache.objects.get(id=open_bug).dupe_of is None
|
||||
assert set(Bugscache.objects.filter(dupe_of=open_bug).values_list('id', flat=True)) == set(
|
||||
duplicates
|
||||
|
|
|
@ -42,7 +42,6 @@ def record_backfilled(test_perf_alert, record_context_sample):
|
|||
|
||||
@pytest.fixture
|
||||
def range_dates(record_context_sample):
|
||||
|
||||
from_date = datetime.fromisoformat(record_context_sample[0]['push_timestamp'])
|
||||
to_date = datetime.fromisoformat(record_context_sample[-1]['push_timestamp'])
|
||||
|
||||
|
|
|
@ -52,7 +52,7 @@ def _generate_performance_data(
|
|||
value,
|
||||
amount,
|
||||
):
|
||||
for (t, v) in zip(
|
||||
for t, v in zip(
|
||||
[i for i in range(start_id, start_id + amount)],
|
||||
[value for i in range(start_id, start_id + amount)],
|
||||
):
|
||||
|
@ -81,7 +81,6 @@ def test_detect_alerts_in_series(
|
|||
test_perf_signature,
|
||||
mock_deviance,
|
||||
):
|
||||
|
||||
base_time = time.time() # generate it based off current time
|
||||
INTERVAL = 30
|
||||
_generate_performance_data(
|
||||
|
@ -172,7 +171,6 @@ def test_detect_alerts_in_series_with_retriggers(
|
|||
generic_reference_data,
|
||||
test_perf_signature,
|
||||
):
|
||||
|
||||
# sometimes we detect an alert in the middle of a series
|
||||
# where there are retriggers, make sure we handle this case
|
||||
# gracefully by generating a sequence where the regression
|
||||
|
@ -264,7 +262,6 @@ def test_custom_alert_threshold(
|
|||
generic_reference_data,
|
||||
test_perf_signature,
|
||||
):
|
||||
|
||||
test_perf_signature.alert_threshold = 200.0
|
||||
test_perf_signature.save()
|
||||
|
||||
|
|
|
@ -59,7 +59,7 @@ def test_detect_changes():
|
|||
|
||||
times = [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15]
|
||||
values = [0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1]
|
||||
for (t, v) in zip(times, values):
|
||||
for t, v in zip(times, values):
|
||||
data.append(RevisionDatum(t, t, [float(v)]))
|
||||
|
||||
result = [
|
||||
|
|
|
@ -18,7 +18,6 @@ def push_usage(test_base_dir):
|
|||
|
||||
|
||||
def test_peak(push_usage):
|
||||
|
||||
peak = get_peak(push_usage['facets'][0])
|
||||
assert peak['needInvestigation'] == 149.0
|
||||
assert peak['time'] == 1584035553
|
||||
|
|
|
@ -36,7 +36,6 @@ def do_job_ingestion(test_repository, job_data, sample_push, verify_data=True):
|
|||
|
||||
blobs = []
|
||||
for blob in job_data:
|
||||
|
||||
if push_index > max_index:
|
||||
push_index = 0
|
||||
|
||||
|
@ -111,7 +110,6 @@ def do_job_ingestion(test_repository, job_data, sample_push, verify_data=True):
|
|||
|
||||
|
||||
def verify_build_platforms(build_platforms_ref):
|
||||
|
||||
build_platforms_set = set()
|
||||
for build_platform in models.BuildPlatform.objects.all():
|
||||
build_platforms_set.add(
|
||||
|
@ -122,7 +120,6 @@ def verify_build_platforms(build_platforms_ref):
|
|||
|
||||
|
||||
def verify_machine_platforms(machine_platforms_ref):
|
||||
|
||||
machine_platforms_set = set()
|
||||
for machine_platform in models.MachinePlatform.objects.all():
|
||||
machine_platforms_set.add(
|
||||
|
@ -135,38 +132,32 @@ def verify_machine_platforms(machine_platforms_ref):
|
|||
|
||||
|
||||
def verify_machines(machines_ref):
|
||||
|
||||
machines = models.Machine.objects.all().values_list('name', flat=True)
|
||||
assert machines_ref.issubset(machines)
|
||||
|
||||
|
||||
def verify_options(options_ref):
|
||||
|
||||
options = models.Option.objects.all().values_list('name', flat=True)
|
||||
|
||||
assert options_ref.issubset(options)
|
||||
|
||||
|
||||
def verify_job_types(job_types_ref):
|
||||
|
||||
job_types = models.JobType.objects.all().values_list('name', flat=True)
|
||||
assert job_types_ref.issubset(job_types)
|
||||
|
||||
|
||||
def verify_products(products_ref):
|
||||
|
||||
products = models.Product.objects.all().values_list('name', flat=True)
|
||||
|
||||
assert products_ref.issubset(products)
|
||||
|
||||
|
||||
def verify_pushes(pushes_ref):
|
||||
|
||||
return pushes_ref.issubset(models.Push.objects.values_list('revision', flat=True))
|
||||
|
||||
|
||||
def verify_log_urls(log_urls_ref):
|
||||
|
||||
log_urls = set(models.JobLog.objects.values_list('url', flat=True))
|
||||
|
||||
assert log_urls_ref.issubset(log_urls)
|
||||
|
|
|
@ -10,7 +10,6 @@ from treeherder.model.models import BugzillaComponent, FilesBugzillaMap
|
|||
|
||||
|
||||
def test_bugzilla_components_for_path(client, test_job):
|
||||
|
||||
BugzillaComponent.objects.create(product='Mock Product 1', component='Mock Component 1')
|
||||
|
||||
FilesBugzillaMap.objects.create(
|
||||
|
|
|
@ -53,7 +53,7 @@ def test_bug_job_map_list(client, test_repository, eleven_jobs_stored, test_user
|
|||
|
||||
expected = list()
|
||||
|
||||
for (i, job) in enumerate(jobs):
|
||||
for i, job in enumerate(jobs):
|
||||
bjm = BugJobMap.create(
|
||||
job_id=job.id,
|
||||
bug_id=bugs[i].id,
|
||||
|
|
|
@ -210,7 +210,6 @@ def test_job_detail_not_found(client, test_repository):
|
|||
|
||||
|
||||
def test_text_log_errors(client, test_job):
|
||||
|
||||
TextLogError.objects.create(job=test_job, line='failure 1', line_number=101)
|
||||
TextLogError.objects.create(job=test_job, line='failure 2', line_number=102)
|
||||
resp = client.get(
|
||||
|
|
|
@ -159,7 +159,6 @@ def alert_create_post_blob(test_perf_alert_summary, test_perf_signature):
|
|||
def test_alerts_post(
|
||||
client, alert_create_post_blob, test_user, test_sheriff, generate_enough_perf_datum
|
||||
):
|
||||
|
||||
# verify that we fail if not authenticated
|
||||
resp = client.post(reverse('performance-alerts-list'), alert_create_post_blob)
|
||||
assert resp.status_code == 403
|
||||
|
|
|
@ -366,7 +366,6 @@ def test_bug_number_and_timestamp_dont_update_from_other_modifications(
|
|||
def test_add_multiple_tags_to_alert_summary(
|
||||
authorized_sheriff_client, test_perf_alert_summary, test_perf_tag, test_perf_tag_2
|
||||
):
|
||||
|
||||
assert test_perf_alert_summary.performance_tags.count() == 1
|
||||
|
||||
resp = authorized_sheriff_client.put(
|
||||
|
|
|
@ -7,7 +7,7 @@ def test_perf_bug_template_api(client, test_perf_framework):
|
|||
framework2 = PerformanceFramework.objects.create(name='test_talos2', enabled=True)
|
||||
|
||||
template_dicts = []
|
||||
for (framework, i) in zip((test_perf_framework, framework2), range(2)):
|
||||
for framework, i in zip((test_perf_framework, framework2), range(2)):
|
||||
dict = {
|
||||
'keywords': "keyword{}".format(i),
|
||||
'status_whiteboard': "sw{}".format(i),
|
||||
|
|
|
@ -58,7 +58,6 @@ def test_perf_signature_same_hash_different_framework(test_perf_signature):
|
|||
|
||||
|
||||
def test_no_summary_performance_data(client, test_perf_signature, test_repository):
|
||||
|
||||
resp = client.get(
|
||||
reverse('performance-signatures-list', kwargs={"project": test_repository.name})
|
||||
)
|
||||
|
@ -366,7 +365,7 @@ def test_filter_data_by_interval(
|
|||
client, test_repository, test_perf_signature, interval, exp_push_ids
|
||||
):
|
||||
# create some test data
|
||||
for (i, timestamp) in enumerate(
|
||||
for i, timestamp in enumerate(
|
||||
[NOW, NOW - datetime.timedelta(days=2), NOW - datetime.timedelta(days=7)]
|
||||
):
|
||||
push = Push.objects.create(
|
||||
|
@ -404,7 +403,7 @@ def test_filter_data_by_range(
|
|||
client, test_repository, test_perf_signature, start_date, end_date, exp_push_ids
|
||||
):
|
||||
# create some test data
|
||||
for (i, timestamp) in enumerate(
|
||||
for i, timestamp in enumerate(
|
||||
[NOW, NOW - datetime.timedelta(days=2), NOW - datetime.timedelta(days=5)]
|
||||
):
|
||||
push = Push.objects.create(
|
||||
|
@ -453,7 +452,7 @@ def test_filter_data_by_signature(
|
|||
push = Push.objects.create(
|
||||
repository=test_repository, revision='abcdefghi', author='foo@bar.com', time=NOW
|
||||
)
|
||||
for (i, signature) in enumerate([test_perf_signature, summary_perf_signature]):
|
||||
for i, signature in enumerate([test_perf_signature, summary_perf_signature]):
|
||||
PerformanceDatum.objects.create(
|
||||
repository=signature.repository,
|
||||
push=push,
|
||||
|
@ -464,8 +463,8 @@ def test_filter_data_by_signature(
|
|||
|
||||
# test that we get the expected value for all different permutations of
|
||||
# passing in signature_id and signature hash
|
||||
for (i, signature) in enumerate([test_perf_signature, summary_perf_signature]):
|
||||
for (param, value) in [
|
||||
for i, signature in enumerate([test_perf_signature, summary_perf_signature]):
|
||||
for param, value in [
|
||||
('signatures', signature.signature_hash),
|
||||
('signature_id', signature.id),
|
||||
]:
|
||||
|
@ -481,7 +480,6 @@ def test_filter_data_by_signature(
|
|||
|
||||
|
||||
def test_perf_summary(client, test_perf_signature, test_perf_data):
|
||||
|
||||
query_params1 = (
|
||||
'?repository={}&framework={}&interval=172800&no_subtests=true&revision={}'.format(
|
||||
test_perf_signature.repository.name,
|
||||
|
|
|
@ -144,7 +144,7 @@ def test_push_list_filter_by_date(client, test_repository, sample_push):
|
|||
"""
|
||||
test retrieving a push list, filtered by a date range
|
||||
"""
|
||||
for (i, datestr) in zip(
|
||||
for i, datestr in zip(
|
||||
[3, 4, 5, 6, 7], ["2013-08-09", "2013-08-10", "2013-08-11", "2013-08-12", "2013-08-13"]
|
||||
):
|
||||
sample_push[i]['push_timestamp'] = utils.to_timestamp(utils.to_datetime(datestr))
|
||||
|
@ -192,7 +192,7 @@ def test_push_list_filter_by_id(client, test_repository, filter_param, exp_ids):
|
|||
"""
|
||||
test filtering by id in various ways
|
||||
"""
|
||||
for (revision, author) in [
|
||||
for revision, author in [
|
||||
('1234abcd', 'foo@bar.com'),
|
||||
('2234abcd', 'foo2@bar.com'),
|
||||
('3234abcd', 'foo3@bar.com'),
|
||||
|
@ -215,7 +215,7 @@ def test_push_list_id_in(client, test_repository):
|
|||
"""
|
||||
test the id__in parameter
|
||||
"""
|
||||
for (revision, author) in [
|
||||
for revision, author in [
|
||||
('1234abcd', 'foo@bar.com'),
|
||||
('2234abcd', 'foo2@bar.com'),
|
||||
('3234abcd', 'foo3@bar.com'),
|
||||
|
@ -260,7 +260,7 @@ def test_push_author(client, test_repository):
|
|||
"""
|
||||
test the author parameter
|
||||
"""
|
||||
for (revision, author) in [
|
||||
for revision, author in [
|
||||
('1234abcd', 'foo@bar.com'),
|
||||
('2234abcd', 'foo@bar.com'),
|
||||
('3234abcd', 'foo2@bar.com'),
|
||||
|
@ -304,7 +304,7 @@ def test_push_reviewbot(client, test_repository):
|
|||
"""
|
||||
test the reviewbot parameter
|
||||
"""
|
||||
for (revision, author) in [
|
||||
for revision, author in [
|
||||
('1234abcd', 'foo@bar.com'),
|
||||
('2234abcd', 'foo2@bar.com'),
|
||||
('3234abcd', 'reviewbot'),
|
||||
|
|
|
@ -5,7 +5,6 @@ import django.db.models.deletion
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
initial = True
|
||||
|
||||
dependencies = []
|
||||
|
|
|
@ -49,9 +49,9 @@ class PerformanceSignatureCollection(dict):
|
|||
signatures = signatures.filter(('suite', 'tp5o'), ('machine_platform', 'windowsxp'))
|
||||
"""
|
||||
filtered_signatures = {}
|
||||
for (signature, signature_value) in self.items():
|
||||
for signature, signature_value in self.items():
|
||||
skip = False
|
||||
for (key, val) in args:
|
||||
for key, val in args:
|
||||
if signature_value.get(key) != val:
|
||||
skip = True
|
||||
break
|
||||
|
@ -107,7 +107,6 @@ class PerformanceSeries(list):
|
|||
|
||||
|
||||
class PerfherderClient(TreeherderClient):
|
||||
|
||||
PERFORMANCE_SIGNATURES_ENDPOINT = 'performance/signatures'
|
||||
PERFORMANCE_DATA_ENDPOINT = 'performance/data'
|
||||
|
||||
|
|
|
@ -263,7 +263,7 @@ class BzApiBugProcess:
|
|||
# the bug against which they have been set as duplicate to prevent them
|
||||
# from getting dropped - they are still needed to match the failure line
|
||||
# against the bug summary.
|
||||
for (bug_duplicate, bug_openish) in duplicates_to_bugs.items():
|
||||
for bug_duplicate, bug_openish in duplicates_to_bugs.items():
|
||||
bug_openish_object = Bugscache.objects.filter(id=bug_openish)
|
||||
if len(bug_openish_object) == 0:
|
||||
# Script does not have access to open bug but to duplicate
|
||||
|
|
|
@ -469,7 +469,7 @@ def store_job_data(repository, originalData):
|
|||
|
||||
# Update the result/state of any jobs that were superseded by those ingested above.
|
||||
if superseded_job_guid_placeholders:
|
||||
for (job_guid, superseded_by_guid) in superseded_job_guid_placeholders:
|
||||
for job_guid, superseded_by_guid in superseded_job_guid_placeholders:
|
||||
Job.objects.filter(guid=superseded_by_guid).update(
|
||||
result='superseded', state='completed'
|
||||
)
|
||||
|
|
|
@ -128,7 +128,6 @@ class Commenter:
|
|||
if (
|
||||
[bug_info['product'], bug_info['component']] in COMPONENTS
|
||||
) and not self.check_whiteboard_status(bug_info['whiteboard']):
|
||||
|
||||
if bug_info['priority'] not in ['--', 'P1', 'P2', 'P3']:
|
||||
change_priority = '--'
|
||||
|
||||
|
|
|
@ -32,7 +32,7 @@ class Command(BaseCommand):
|
|||
c = TreeherderClient(server_url=options['server'])
|
||||
|
||||
# options / option collection hashes
|
||||
for (uuid, props) in c.get_option_collection_hash().items():
|
||||
for uuid, props in c.get_option_collection_hash().items():
|
||||
for prop in props:
|
||||
option, _ = Option.objects.get_or_create(name=prop['name'])
|
||||
OptionCollection.objects.get_or_create(option_collection_hash=uuid, option=option)
|
||||
|
|
|
@ -8,7 +8,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
initial = True
|
||||
|
||||
dependencies = [
|
||||
|
|
|
@ -5,7 +5,6 @@ import django.db.models.manager
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0001_squashed_0022_modify_bugscache_and_bugjobmap'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0002_add_bugjobmap_model_manager'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0003_add_matcher_name_fields'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0004_populate_matcher_name_fields'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0005_use_matcher_name_for_unique_constraint'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0006_drop_matcher_fks'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0007_remove_m2m_between_classified_failures_and_failure_match'),
|
||||
]
|
||||
|
|
|
@ -5,7 +5,6 @@ import django.db.models.manager
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0008_remove_failure_match'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0009_add_manager_to_push_and_job'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0010_remove_runnable_job'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0011_remove_matcher_table'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0012_branch_maxlen'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0013_add_index_to_push_revision'),
|
||||
]
|
||||
|
|
|
@ -5,7 +5,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0014_add_job_log_status_skipped_size'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0015_add_repository_tc_root_url'),
|
||||
]
|
||||
|
|
|
@ -5,7 +5,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0016_add_index_commit_revision'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0017_add_index_task_id'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0018_delete_jobdetail'),
|
||||
]
|
||||
|
|
|
@ -5,7 +5,6 @@ import django.db.models.deletion
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0019_increase_job_type_name_length'),
|
||||
]
|
||||
|
|
|
@ -5,7 +5,6 @@ import django.db.models.deletion
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0020_textlogerror_job'),
|
||||
]
|
||||
|
|
|
@ -5,7 +5,6 @@ import django.db.models.deletion
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0021_investigatedtests'),
|
||||
]
|
||||
|
|
|
@ -5,7 +5,6 @@ import django.db.models.deletion
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0022_support_group_status'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0023_add_filebugzillacomponent'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0024_add_bugzillasecuritygroup'),
|
||||
]
|
||||
|
|
|
@ -18,7 +18,6 @@ def set_update_next_run(apps, schema_editor):
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0025_remove_bugscache_os'),
|
||||
]
|
||||
|
|
|
@ -7,7 +7,6 @@ import django.utils.timezone
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0026_bugscache_add_dupe_of_and_processed_update'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0027_mozciclassification'),
|
||||
]
|
||||
|
|
|
@ -162,7 +162,7 @@ class Push(models.Model):
|
|||
)
|
||||
|
||||
status_dict = {'completed': 0, 'pending': 0, 'running': 0}
|
||||
for (state, result, total) in jobs.values_list('state', 'result').annotate(
|
||||
for state, result, total in jobs.values_list('state', 'result').annotate(
|
||||
total=Count('result')
|
||||
):
|
||||
if state == 'completed':
|
||||
|
@ -363,7 +363,7 @@ class OptionCollectionManager(models.Manager):
|
|||
return option_collection_map
|
||||
|
||||
option_collection_map = {}
|
||||
for (hash, option_name) in OptionCollection.objects.values_list(
|
||||
for hash, option_name in OptionCollection.objects.values_list(
|
||||
'option_collection_hash', 'option__name'
|
||||
):
|
||||
if not option_collection_map.get(hash):
|
||||
|
|
|
@ -88,7 +88,7 @@ def generate_new_alerts_in_series(signature):
|
|||
)
|
||||
|
||||
with transaction.atomic():
|
||||
for (prev, cur) in zip(analyzed_series, analyzed_series[1:]):
|
||||
for prev, cur in zip(analyzed_series, analyzed_series[1:]):
|
||||
if cur.change_detected:
|
||||
prev_value = cur.historical_stats['avg']
|
||||
new_value = cur.forward_stats['avg']
|
||||
|
|
|
@ -12,7 +12,6 @@ class Command(BaseCommand):
|
|||
help = "Populate a test set of data suitable for testing the perfherder UI"
|
||||
|
||||
def handle(self, *args, **options):
|
||||
|
||||
confirm = input(
|
||||
"""
|
||||
You have a requested a load of test performance data, this is a destructive
|
||||
|
@ -40,7 +39,7 @@ Type 'yes' to continue, or 'no' to cancel: """
|
|||
time=datetime.datetime.now(),
|
||||
)
|
||||
|
||||
for (t, v) in zip(
|
||||
for t, v in zip(
|
||||
[i for i in range(INTERVAL)],
|
||||
([0.5 for i in range(int(INTERVAL / 2))] + [1.0 for i in range(int(INTERVAL / 2))]),
|
||||
):
|
||||
|
|
|
@ -177,7 +177,6 @@ class MassiveData(Data):
|
|||
progress_notifier=None,
|
||||
**kwargs,
|
||||
):
|
||||
|
||||
super().__init__(source, target, progress_notifier, **kwargs)
|
||||
self.time_window = time_window
|
||||
self.num_workers = num_workers
|
||||
|
@ -312,7 +311,6 @@ class MassiveData(Data):
|
|||
self.progress_notifier(self.bring_in_alert, alerts, 'alert', 2)
|
||||
|
||||
def bring_in_alert(self, alert):
|
||||
|
||||
if alert.id in self.models_instances['performance_alert']:
|
||||
return
|
||||
|
||||
|
@ -399,7 +397,6 @@ class Command(BaseCommand):
|
|||
parser.add_argument('--repositories', nargs='+', default=None)
|
||||
|
||||
def handle(self, *args, **options):
|
||||
|
||||
time_window = datetime.timedelta(days=options['time_window'])
|
||||
num_workers = options['num_workers']
|
||||
frameworks = options['frameworks']
|
||||
|
|
|
@ -85,7 +85,7 @@ class Command(BaseCommand):
|
|||
signatures = []
|
||||
signatures_to_ignore = set()
|
||||
# if doing everything, only handle summary series
|
||||
for (signature, properties) in signature_data.items():
|
||||
for signature, properties in signature_data.items():
|
||||
signatures.append(signature)
|
||||
if 'subtest_signatures' in properties:
|
||||
# Don't alert on subtests which have a summary
|
||||
|
@ -103,7 +103,7 @@ class Command(BaseCommand):
|
|||
|
||||
data = []
|
||||
|
||||
for (timestamp, value) in zip(series['push_timestamp'], series['value']):
|
||||
for timestamp, value in zip(series['push_timestamp'], series['value']):
|
||||
data.append(RevisionDatum(timestamp, value))
|
||||
|
||||
for r in detect_changes(data):
|
||||
|
|
|
@ -7,7 +7,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
initial = True
|
||||
|
||||
dependencies = [
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0001_squashed_0005_permit_github_links'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0006_add_alert_summary_notes'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0007_star_performancealert'),
|
||||
]
|
||||
|
|
|
@ -5,7 +5,6 @@ import django.db.models.deletion
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0008_add_confirming_state'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0010_remove_runnable_job'),
|
||||
('perf', '0009_non_nullable_issue_tracker'),
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0010_fix_signature_uniqueness'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0011_inc_extra_options_length'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0012_rename_summary_last_updated'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0013_add_alert_timestamps'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0014_add_performance_datum_idx'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0015_add_bug_updated_field'),
|
||||
]
|
||||
|
|
|
@ -6,7 +6,6 @@ import django.db.models.deletion
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
|
||||
('perf', '0016_modify_alertsummary_status_choices'),
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0017_assignment_support_for_alert_summaries'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0018_add_measurement_units'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0015_add_repository_tc_root_url'),
|
||||
('perf', '0019_remove_confirming_state'),
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0015_add_repository_tc_root_url'),
|
||||
('perf', '0020_add_application_field'),
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0015_add_repository_tc_root_url'),
|
||||
('perf', '0021_remove_application_from_constraint'),
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0022_add_test_display_names'),
|
||||
]
|
||||
|
|
|
@ -5,7 +5,6 @@ import django.db.models.deletion
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0023_increase_extra_options_size'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0024_support_backfill_reports'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0025_provide_tag_support'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0026_add_backfill_record_status'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0027_support_perfherder_settings'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0028_default_application_to_empty_str'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0016_add_index_commit_revision'),
|
||||
('perf', '0029_add_frozen_to_report'),
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0017_add_index_task_id'),
|
||||
('perf', '0030_add_application_to_contraints'),
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0031_fix_application_in_constraints'),
|
||||
]
|
||||
|
|
|
@ -5,7 +5,6 @@ import django.db.models.deletion
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0020_textlogerror_job'),
|
||||
('perf', '0032_add_performance_tag'),
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0033_permit_multi_data_per_job'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0034_remove_performancedatum_fields'),
|
||||
]
|
||||
|
|
|
@ -6,7 +6,6 @@ MULTICOMMIT_CONSTRAINT_SYMBOL = 'perf_multicommitdatu_perf_datum_id_c2d7eb14_fk_
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0035_fix_deprecated_nullboolean_field'),
|
||||
]
|
||||
|
|
|
@ -5,7 +5,6 @@ import django.db.models.deletion
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0022_support_group_status'),
|
||||
('perf', '0036_cascade_perf_datum_deletion'),
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0037_extend_backfill_record'),
|
||||
]
|
||||
|
|
|
@ -5,7 +5,6 @@ import django.db.models.deletion
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('model', '0022_support_group_status'),
|
||||
('perf', '0038_update_record_status'),
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0039_store_more_job_details_on_record'),
|
||||
]
|
||||
|
|
|
@ -5,7 +5,6 @@ import django.db.models.deletion
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0040_performancealert_noise_profile'),
|
||||
]
|
||||
|
|
|
@ -4,7 +4,6 @@ from django.db import migrations, models
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0041_backfillnotificationrecord'),
|
||||
]
|
||||
|
|
|
@ -6,7 +6,6 @@ from django.db import migrations, connection
|
|||
|
||||
|
||||
def alter_perfdatum_pk(apps, schema_editor):
|
||||
|
||||
PerformanceDatum = apps.get_model('perf', 'PerformanceDatum')
|
||||
pursue = "yes"
|
||||
# Automatically pursue migration if performance_datum table is empty
|
||||
|
@ -26,7 +25,6 @@ def alter_perfdatum_pk(apps, schema_editor):
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0043_drop_multicommitdatum'),
|
||||
]
|
||||
|
|
|
@ -7,7 +7,6 @@ MULTICOMMIT_CONSTRAINT_SYMBOL = 'perf_multicommitdatu_perf_datum_id_c2d7eb14_fk_
|
|||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('perf', '0045_restore_perf_multicommitdatum_and_schema'),
|
||||
]
|
||||
|
|
Некоторые файлы не были показаны из-за слишком большого количества измененных файлов Показать больше
Загрузка…
Ссылка в новой задаче