From 6815875c465e9685e2dfd3b55382ce55cfed559e Mon Sep 17 00:00:00 2001 From: Marco Castelluccio Date: Sun, 4 Oct 2020 12:42:13 +0200 Subject: [PATCH] Allow classifying batches of more than 1000 bugs Since 336d96bdd075bf27822d43b89e79b80fbb1635d5, we no longer need to hope we don't hit the Bugzilla URL limit. --- http_service/bugbug_http/app.py | 4 +--- 1 file changed, 1 insertion(+), 3 deletions(-) diff --git a/http_service/bugbug_http/app.py b/http_service/bugbug_http/app.py index 0108b4bf..ae1dafd9 100644 --- a/http_service/bugbug_http/app.py +++ b/http_service/bugbug_http/app.py @@ -533,7 +533,6 @@ def batch_prediction(model_name): "bugs": { "type": "list", "minlength": 1, - "maxlength": 1000, "schema": {"type": "integer"}, } } @@ -566,8 +565,7 @@ def batch_prediction(model_name): if missing_bugs: # TODO: We should probably schedule chunks of bugs to avoid jobs that - # are running for too long and reduce pressure on bugzilla, it mights - # not like getting 1 million bug at a time + # are running for too long schedule_bug_classification(model_name, missing_bugs) return compress_response({"bugs": data}, status_code)