addons-server/apps/devhub/tasks.py

458 строки
16 KiB
Python
Исходник Обычный вид История

# -*- coding: utf8 -*-
import base64
from datetime import date
import json
2010-09-15 00:22:33 +04:00
import logging
import os
import path
import socket
2010-09-15 00:22:33 +04:00
import sys
import traceback
import urllib2
import uuid
2010-09-15 00:22:33 +04:00
from django.conf import settings
from django.core.management import call_command
from django.utils.http import urlencode
2010-09-15 00:22:33 +04:00
from celeryutils import task
2012-01-05 03:12:53 +04:00
from django_statsd.clients import statsd
from tower import ugettext as _
2010-09-15 00:22:33 +04:00
2010-12-31 04:02:31 +03:00
import amo
from amo.decorators import write, set_modified_on
from amo.utils import guard, resize_image, remove_icons
from addons.models import Addon
from applications.management.commands import dump_apps
from applications.models import Application, AppVersion
from devhub import perf
from files.models import FileUpload, File, FileValidation
2010-09-15 00:22:33 +04:00
from PIL import Image
2010-09-15 00:22:33 +04:00
log = logging.getLogger('z.devhub.task')
2011-07-29 23:31:54 +04:00
@task
2010-09-15 00:22:33 +04:00
@write
def validator(upload_id, **kw):
if not settings.VALIDATE_ADDONS:
return None
2010-09-15 00:22:33 +04:00
log.info('VALIDATING: %s' % upload_id)
upload = FileUpload.objects.get(pk=upload_id)
try:
result = run_validator(upload.path)
upload.validation = result
upload.save() # We want to hit the custom save().
2010-09-15 00:22:33 +04:00
except:
# Store the error with the FileUpload job, then raise
# it for normal logging.
tb = traceback.format_exception(*sys.exc_info())
upload.update(task_error=''.join(tb))
raise
2011-07-29 23:31:54 +04:00
@task
@write
def compatibility_check(upload_id, app_guid, appversion_str, **kw):
if not settings.VALIDATE_ADDONS:
return None
log.info('COMPAT CHECK for upload %s / app %s version %s'
% (upload_id, app_guid, appversion_str))
upload = FileUpload.objects.get(pk=upload_id)
app = Application.objects.get(guid=app_guid)
appver = AppVersion.objects.get(application=app, version=appversion_str)
try:
result = run_validator(upload.path,
for_appversions={app_guid: [appversion_str]},
test_all_tiers=True,
# Ensure we only check compatibility
# against this one specific version:
overrides={'targetapp_minVersion':
{app_guid: appversion_str},
'targetapp_maxVersion':
{app_guid: appversion_str}})
upload.validation = result
upload.compat_with_app = app
upload.compat_with_appver = appver
upload.save() # We want to hit the custom save().
except:
# Store the error with the FileUpload job, then raise
# it for normal logging.
tb = traceback.format_exception(*sys.exc_info())
upload.update(task_error=''.join(tb))
raise
2011-07-29 23:31:54 +04:00
@task
@write
def file_validator(file_id, **kw):
if not settings.VALIDATE_ADDONS:
return None
log.info('VALIDATING file: %s' % file_id)
file = File.objects.get(pk=file_id)
# Unlike upload validation, let the validator
# raise an exception if there is one.
result = run_validator(file.file_path)
return FileValidation.from_json(file, result)
def run_validator(file_path, for_appversions=None, test_all_tiers=False,
overrides=None):
"""A pre-configured wrapper around the addon validator.
*file_path*
Path to addon / extension file to validate.
*for_appversions=None*
An optional dict of application versions to validate this addon
for. The key is an application GUID and its value is a list of
versions.
*test_all_tiers=False*
When False (default) the validator will not continue if it
encounters fatal errors. When True, all tests in all tiers are run.
See bug 615426 for discussion on this default.
*overrides=None*
Normally the validator gets info from install.rdf but there are a
few things we need to override. See validator for supported overrides.
Example: {'targetapp_maxVersion': {'<app guid>': '<version>'}}
To validate the addon for compatibility with Firefox 5 and 6,
you'd pass in::
for_appversions={amo.FIREFOX.guid: ['5.0.*', '6.0.*']}
Not all application versions will have a set of registered
compatibility tests.
"""
from validator.validate import validate
# TODO(Kumar) remove this when validator is fixed, see bug 620503
from validator.testcases import scripting
scripting.SPIDERMONKEY_INSTALLATION = settings.SPIDERMONKEY
import validator.constants
validator.constants.SPIDERMONKEY_INSTALLATION = settings.SPIDERMONKEY
apps = dump_apps.Command.JSON_PATH
if not os.path.exists(apps):
call_command('dump_apps')
2011-12-13 00:42:04 +04:00
with statsd.timer('devhub.validator'):
return validate(file_path,
for_appversions=for_appversions,
format='json',
# When False, this flag says to stop testing after one
# tier fails.
determined=test_all_tiers,
approved_applications=apps,
spidermonkey=settings.SPIDERMONKEY,
overrides=overrides,
timeout=settings.VALIDATOR_TIMEOUT)
2010-11-12 03:39:18 +03:00
@task(rate_limit='4/m')
@write
def flag_binary(ids, **kw):
log.info('[%s@%s] Flagging binary addons starting with id: %s...'
% (len(ids), flag_binary.rate_limit, ids[0]))
addons = Addon.objects.filter(pk__in=ids).no_transforms()
latest = kw.pop('latest', True)
for addon in addons:
try:
log.info('Validating addon with id: %s' % addon.pk)
files = (File.objects.filter(version__addon=addon)
.exclude(status=amo.STATUS_DISABLED)
.order_by('-created'))
if latest:
files = [files[0]]
for file in files:
result = json.loads(run_validator(file.file_path))
metadata = result['metadata']
binary = (metadata.get('contains_binary_extension', False) or
metadata.get('contains_binary_content', False))
binary_components = metadata.get('binary_components', False)
log.info('Updating binary flags for addon with id=%s: '
'binary -> %s, binary_components -> %s' % (
addon.pk, binary, binary_components))
file.update(binary=binary, binary_components=binary_components)
except Exception, err:
log.error('Failed to run validation on addon id: %s, %s'
% (addon.pk, err))
@task
@set_modified_on
2010-11-12 03:39:18 +03:00
def resize_icon(src, dst, size, **kw):
"""Resizes addon icons."""
log.info('[1@None] Resizing icon: %s' % dst)
try:
if isinstance(size, list):
for s in size:
resize_image(src, '%s-%s.png' % (dst, s), (s, s),
remove_src=False)
os.remove(src)
else:
resize_image(src, dst, (size, size), remove_src=True)
return True
2010-11-12 03:39:18 +03:00
except Exception, e:
log.error("Error saving addon icon: %s" % e)
2010-12-31 04:02:31 +03:00
@task
@set_modified_on
2011-08-31 02:01:18 +04:00
def resize_preview(src, instance, **kw):
"""Resizes preview images and stores the sizes on the preview."""
thumb_dst, full_dst = instance.thumbnail_path, instance.image_path
sizes = {}
log.info('[1@None] Resizing preview and storing size: %s' % thumb_dst)
2010-12-31 04:02:31 +03:00
try:
sizes['thumbnail'] = resize_image(src, thumb_dst,
amo.ADDON_PREVIEW_SIZES[0],
remove_src=False)
sizes['image'] = resize_image(src, full_dst,
amo.ADDON_PREVIEW_SIZES[1],
remove_src=False)
instance.sizes = sizes
instance.save()
return True
2010-12-31 04:02:31 +03:00
except Exception, e:
log.error("Error saving preview: %s" % e)
2011-05-26 03:44:34 +04:00
@task
@write
def get_preview_sizes(ids, **kw):
log.info('[%s@%s] Getting preview sizes for addons starting at id: %s...'
% (len(ids), get_preview_sizes.rate_limit, ids[0]))
addons = Addon.objects.filter(pk__in=ids).no_transforms()
for addon in addons:
previews = addon.previews.all()
log.info('Found %s previews for: %s' % (previews.count(), addon.pk))
for preview in previews:
try:
log.info('Getting size for preview: %s' % preview.pk)
sizes = {
'thumbnail': Image.open(preview.thumbnail_path).size,
'image': Image.open(preview.image_path).size,
}
preview.update(sizes=sizes)
except Exception, err:
log.error('Failed to find size of preview: %s, error: %s'
% (addon.pk, err))
@task
@write
def convert_purified(ids, **kw):
log.info('[%s@%s] Converting fields to purified starting at id: %s...'
% (len(ids), convert_purified.rate_limit, ids[0]))
fields = ['the_reason', 'the_future']
for addon in Addon.objects.filter(pk__in=ids):
flag = False
for field in fields:
value = getattr(addon, field)
if value:
value.clean()
if (value.localized_string_clean != value.localized_string):
flag = True
if flag:
log.info('Saving addon: %s to purify fields' % addon.pk)
addon.save()
2011-05-26 03:44:34 +04:00
@task
def packager(data, feature_set, **kw):
"""Build an add-on based on input data."""
log.info('[1@None] Packaging add-on')
from devhub.views import packager_path
dest = packager_path(data['slug'])
2011-05-26 03:44:34 +04:00
with guard(u'devhub.packager.%s' % dest) as locked:
if locked:
log.error(u'Packaging in progress: %s' % dest)
return
2011-05-26 03:44:34 +04:00
with statsd.timer('devhub.packager'):
2011-10-04 00:04:43 +04:00
from packager.main import packager
log.info('Starting packaging: %s' % dest)
features = set([k for k, v in feature_set.items() if v])
try:
packager(data, dest, features)
except Exception, err:
log.error(u'Failed to package add-on: %s' % err)
raise
if os.path.exists(dest):
log.info(u'Package saved: %s' % dest)
def failed_validation(*messages):
"""Return a validation object that looks like the add-on validator."""
m = []
for msg in messages:
m.append({'type': 'error', 'message': msg, 'tier': 1})
return json.dumps({'errors': 1, 'success': False, 'messages': m})
def _fetch_content(url):
try:
return urllib2.urlopen(url, timeout=5)
2011-08-23 02:09:40 +04:00
except urllib2.HTTPError, e:
raise Exception(_('%s responded with %s (%s).') % (url, e.code, e.msg))
except urllib2.URLError, e:
# Unpack the URLError to try and find a useful message.
if isinstance(e.reason, socket.timeout):
raise Exception(_('Connection to "%s" timed out.') % url)
elif isinstance(e.reason, socket.gaierror):
raise Exception(_('Could not contact host at "%s".') % url)
else:
raise Exception(str(e.reason))
def check_content_type(response, content_type,
no_ct_message, wrong_ct_message):
if not response.headers.get('Content-Type', '').startswith(content_type):
if 'Content-Type' in response.headers:
raise Exception(wrong_ct_message %
(content_type, response.headers['Content-Type']))
else:
raise Exception(no_ct_message % content_type)
def get_content_and_check_size(response, max_size, error_message):
# Read one extra byte. Reject if it's too big so we don't have issues
# downloading huge files.
content = response.read(max_size + 1)
if len(content) > max_size:
raise Exception(error_message % max_size)
return content
def save_icon(webapp, content):
tmp_path = path.path(settings.TMP_PATH) / 'icon'
if not os.path.exists(tmp_path):
os.makedirs(tmp_path)
tmp_dst = tmp_path / uuid.uuid4().hex
with open(tmp_dst, 'wb') as fd:
fd.write(content)
dirname = webapp.get_icon_dir()
destination = os.path.join(dirname, '%s' % webapp.id)
remove_icons(destination)
resize_icon.delay(tmp_dst, destination, amo.ADDON_ICON_SIZES,
set_modified_on=[webapp])
# Need to set the icon type so .get_icon_url() works
# normally submit step 4 does it through AddonFormMedia,
# but we want to beat them to the punch.
# resize_icon outputs pngs, so we know it's 'image/png'
webapp.icon_type = 'image/png'
webapp.save()
@task
def fetch_icon(webapp, **kw):
"""Downloads a webapp icon from the location specified in the manifest.
Returns False if icon was not able to be retrieved
"""
log.info(u'[1@None] Fetching icon for webapp %s.' % webapp.name)
manifest = webapp.get_manifest_json()
if not 'icons' in manifest:
return
biggest = max([int(size) for size in manifest['icons']])
icon_url = manifest['icons'][str(biggest)]
if icon_url.startswith('data:image'):
image_string = icon_url.split('base64,')[1]
content = base64.decodestring(image_string)
else:
try:
response = _fetch_content(webapp.origin + icon_url)
except Exception, e:
log.error('Failed to fetch icon for webapp %s: %s'
% (webapp.pk, e.message))
return
size_error_message = _('Your icon must be less than %s bytes.')
content = get_content_and_check_size(response,
settings.MAX_ICON_UPLOAD_SIZE,
size_error_message)
save_icon(webapp, content)
@task
def fetch_manifest(url, upload_pk=None, **kw):
log.info(u'[1@None] Fetching manifest: %s.' % url)
upload = FileUpload.objects.get(pk=upload_pk)
try:
response = _fetch_content(url)
no_ct_message = _('Your manifest must be served with the HTTP '
'header "Content-Type: %s".')
wrong_ct_message = _('Your manifest must be served with the HTTP '
'header "Content-Type: %s". We saw "%s".')
check_content_type(response, 'application/x-web-app-manifest+json',
no_ct_message, wrong_ct_message)
size_error_message = _('Your manifest must be less than %s bytes.')
content = get_content_and_check_size(response,
settings.MAX_WEBAPP_UPLOAD_SIZE,
size_error_message)
except Exception, e:
# Drop a message in the validation slot and bail.
2011-08-17 22:51:20 +04:00
upload.update(validation=failed_validation(e.message))
return
upload.add_file([content], url, len(content))
# Send the upload to the validator.
validator(upload.pk)
@task
def start_perf_test_for_file(file_id, os_name, app_name, **kw):
log.info('[@%s] Starting perf tests for file %s on %s / %s'
% (start_perf_test_for_file.rate_limit, file_id,
os_name, app_name))
file_ = File.objects.get(pk=file_id)
# TODO(Kumar) store token to retrieve results later?
perf.start_perf_test(file_, os_name, app_name)
@task
def subscribe_to_responsys(campaign, address, format='html', source_url='',
lang='', country='', **kw):
"""
Subscribe a user to a list in responsys. There should be two
fields within the Responsys system named by the "campaign"
parameter: <campaign>_FLG and <campaign>_DATE.
"""
data = {
'LANG_LOCALE': lang,
'COUNTRY_': country,
'SOURCE_URL': source_url,
'EMAIL_ADDRESS_': address,
'EMAIL_FORMAT_': 'H' if format == 'html' else 'T',
}
data['%s_FLG' % campaign] = 'Y'
data['%s_DATE' % campaign] = date.today().strftime('%Y-%m-%d')
data['_ri_'] = settings.RESPONSYS_ID
try:
res = urllib2.urlopen('http://awesomeness.mozilla.org/pub/rf',
data=urlencode(data))
return res.code == 200
except urllib2.URLError:
return False