2016-07-18 18:19:36 +03:00
|
|
|
#!/usr/bin/env python3
|
|
|
|
|
|
|
|
# stdlib imports
|
|
|
|
import argparse
|
2016-08-17 18:30:53 +03:00
|
|
|
import base64
|
2016-08-12 19:21:33 +03:00
|
|
|
import copy
|
2016-07-18 18:19:36 +03:00
|
|
|
import datetime
|
|
|
|
import json
|
2016-07-20 19:18:31 +03:00
|
|
|
import hashlib
|
2016-07-18 18:19:36 +03:00
|
|
|
import os
|
|
|
|
import pathlib
|
2016-08-20 01:13:33 +03:00
|
|
|
import subprocess
|
2016-07-18 18:19:36 +03:00
|
|
|
import time
|
|
|
|
from typing import List
|
2016-08-20 01:13:33 +03:00
|
|
|
import urllib
|
2016-07-18 18:19:36 +03:00
|
|
|
# non-stdlib imports
|
|
|
|
import azure.batch.batch_auth as batchauth
|
|
|
|
import azure.batch.batch_service_client as batch
|
|
|
|
import azure.batch.models as batchmodels
|
|
|
|
import azure.common
|
|
|
|
import azure.storage.blob as azureblob
|
|
|
|
import azure.storage.queue as azurequeue
|
|
|
|
import azure.storage.table as azuretable
|
|
|
|
|
|
|
|
# global defines
|
|
|
|
_STORAGEACCOUNT = os.getenv('STORAGEACCOUNT')
|
|
|
|
_STORAGEACCOUNTKEY = os.getenv('STORAGEACCOUNTKEY')
|
|
|
|
_BATCHACCOUNTKEY = os.getenv('BATCHACCOUNTKEY')
|
|
|
|
_STORAGE_CONTAINERS = {
|
|
|
|
'blob_resourcefiles': None,
|
2016-08-19 01:17:04 +03:00
|
|
|
'blob_torrents': None,
|
2016-08-09 22:47:58 +03:00
|
|
|
'table_dht': None,
|
2016-07-18 18:19:36 +03:00
|
|
|
'table_registry': None,
|
|
|
|
'table_torrentinfo': None,
|
2016-08-09 22:47:58 +03:00
|
|
|
'table_services': None,
|
2016-07-18 18:19:36 +03:00
|
|
|
'table_globalresources': None,
|
2016-08-10 01:23:15 +03:00
|
|
|
'table_perf': None,
|
2016-07-18 18:19:36 +03:00
|
|
|
'queue_globalresources': None,
|
|
|
|
}
|
2016-08-20 01:13:33 +03:00
|
|
|
_AZUREFILE_DVD_VERSION = '0.4.1'
|
|
|
|
_AZUREFILE_DVD_URL = 'https://github.com/Azure/azurefile-' + \
|
|
|
|
'dockervolumedriver/releases/download/' + _AZUREFILE_DVD_VERSION + \
|
|
|
|
'/azurefile-dockervolumedriver'
|
|
|
|
_AZUREFILE_DVD_MD5 = 'f3c1750583c4842dfbf95bbd56f65ede'
|
|
|
|
_AZUREFILE_SYSTEMD_SERVICE_URL = (
|
|
|
|
'https://raw.githubusercontent.com/Azure/'
|
|
|
|
'azurefile-dockervolumedriver/master/contrib/init/systemd/'
|
|
|
|
'azurefile-dockervolumedriver.service'
|
|
|
|
)
|
|
|
|
_AZUREFILE_SYSTEMD_SERVICE_MD5 = 'd58f2f5e9f9f78216651ac28419878f1'
|
2016-07-18 18:19:36 +03:00
|
|
|
_NODEPREP_FILE = ('nodeprep.sh', 'scripts/nodeprep.sh')
|
2016-08-17 18:30:53 +03:00
|
|
|
_JOBPREP_FILE = ('jpdockerblock.sh', 'scripts/jpdockerblock.sh')
|
2016-07-18 18:19:36 +03:00
|
|
|
_CASCADE_FILE = ('cascade.py', 'cascade.py')
|
2016-07-20 19:18:31 +03:00
|
|
|
_SETUP_PR_FILE = ('setup_private_registry.py', 'setup_private_registry.py')
|
2016-08-10 01:23:15 +03:00
|
|
|
_PERF_FILE = ('perf.py', 'perf.py')
|
2016-08-12 19:21:33 +03:00
|
|
|
_REGISTRY_FILE = None
|
2016-08-19 21:28:29 +03:00
|
|
|
_GENERIC_DOCKER_TASK_PREFIX = 'dockertask-'
|
|
|
|
_TEMP_DIR = pathlib.Path('/tmp')
|
2016-07-18 18:19:36 +03:00
|
|
|
|
|
|
|
|
2016-08-20 01:13:33 +03:00
|
|
|
def _populate_global_settings(config: dict, action: str):
|
2016-07-18 18:19:36 +03:00
|
|
|
"""Populate global settings from config
|
|
|
|
:param dict config: configuration dict
|
2016-08-20 01:13:33 +03:00
|
|
|
:param str action: action
|
2016-07-18 18:19:36 +03:00
|
|
|
"""
|
2016-08-12 19:21:33 +03:00
|
|
|
global _STORAGEACCOUNT, _STORAGEACCOUNTKEY, _BATCHACCOUNTKEY, \
|
|
|
|
_REGISTRY_FILE
|
|
|
|
_STORAGEACCOUNT = config['credentials']['storage_account']
|
|
|
|
_STORAGEACCOUNTKEY = config['credentials']['storage_account_key']
|
|
|
|
_BATCHACCOUNTKEY = config['credentials']['batch_account_key']
|
2016-07-18 18:19:36 +03:00
|
|
|
try:
|
2016-08-12 19:21:33 +03:00
|
|
|
sep = config['storage_entity_prefix']
|
2016-07-18 18:19:36 +03:00
|
|
|
except KeyError:
|
|
|
|
sep = None
|
|
|
|
if sep is None:
|
|
|
|
sep = ''
|
2016-08-20 01:13:33 +03:00
|
|
|
_STORAGE_CONTAINERS['blob_resourcefiles'] = '-'.join(
|
|
|
|
(sep + 'resourcefiles',
|
|
|
|
config['credentials']['batch_account'].lower(),
|
|
|
|
config['pool_specification']['id'].lower()))
|
2016-08-19 01:17:04 +03:00
|
|
|
_STORAGE_CONTAINERS['blob_torrents'] = '-'.join(
|
|
|
|
(sep + 'torrents',
|
|
|
|
config['credentials']['batch_account'].lower(),
|
2016-08-19 21:28:29 +03:00
|
|
|
config['pool_specification']['id'].lower()))
|
2016-08-09 22:47:58 +03:00
|
|
|
_STORAGE_CONTAINERS['table_dht'] = sep + 'dht'
|
2016-07-18 18:19:36 +03:00
|
|
|
_STORAGE_CONTAINERS['table_registry'] = sep + 'registry'
|
|
|
|
_STORAGE_CONTAINERS['table_torrentinfo'] = sep + 'torrentinfo'
|
|
|
|
_STORAGE_CONTAINERS['table_services'] = sep + 'services'
|
|
|
|
_STORAGE_CONTAINERS['table_globalresources'] = sep + 'globalresources'
|
2016-08-10 01:23:15 +03:00
|
|
|
_STORAGE_CONTAINERS['table_perf'] = sep + 'perf'
|
2016-08-09 22:47:58 +03:00
|
|
|
_STORAGE_CONTAINERS['queue_globalresources'] = '-'.join(
|
|
|
|
(sep + 'globalresources',
|
2016-08-12 19:21:33 +03:00
|
|
|
config['credentials']['batch_account'].lower(),
|
2016-08-19 21:28:29 +03:00
|
|
|
config['pool_specification']['id'].lower()))
|
2016-08-20 01:13:33 +03:00
|
|
|
if action != 'addpool':
|
|
|
|
return
|
2016-08-12 19:21:33 +03:00
|
|
|
try:
|
2016-08-17 18:30:53 +03:00
|
|
|
if config['docker_registry']['private']['enabled']:
|
|
|
|
rf = config['docker_registry']['private'][
|
|
|
|
'docker_save_registry_file']
|
2016-08-20 01:13:33 +03:00
|
|
|
prf = pathlib.Path(rf)
|
|
|
|
# attempt to package if registry file doesn't exist
|
|
|
|
if not prf.exists():
|
|
|
|
print('attempting to generate docker private registry tarball')
|
|
|
|
try:
|
|
|
|
output = subprocess.check_output(
|
|
|
|
'sudo docker images -q registry:2', shell=True)
|
|
|
|
output = output.decode('utf-8').strip()
|
|
|
|
except subprocess.CalledProcessError as ex:
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
if len(output) == 12:
|
|
|
|
rf = 'resources/docker_registry_v2.tar.gz'
|
|
|
|
prf = pathlib.Path(rf)
|
|
|
|
config['docker_registry']['private'][
|
|
|
|
'docker_save_registry_image_id'] = output
|
|
|
|
subprocess.check_call(
|
|
|
|
'sudo docker save registry:2 '
|
|
|
|
'| gzip -c > {}'.format(rf), shell=True)
|
2016-08-17 18:30:53 +03:00
|
|
|
_REGISTRY_FILE = (
|
2016-08-20 01:13:33 +03:00
|
|
|
prf.name,
|
2016-08-17 18:30:53 +03:00
|
|
|
rf,
|
|
|
|
config['docker_registry']['private'][
|
|
|
|
'docker_save_registry_image_id']
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
_REGISTRY_FILE = (None, None, None)
|
2016-08-12 19:21:33 +03:00
|
|
|
except Exception:
|
|
|
|
_REGISTRY_FILE = (None, None, None)
|
2016-08-20 01:13:33 +03:00
|
|
|
print('private registry settings: {}'.format(_REGISTRY_FILE))
|
2016-07-18 18:19:36 +03:00
|
|
|
|
|
|
|
|
|
|
|
def _wrap_commands_in_shell(commands: List[str], wait: bool=True) -> str:
|
|
|
|
"""Wrap commands in a shell
|
|
|
|
:param list commands: list of commands to wrap
|
|
|
|
:param bool wait: add wait for background processes
|
|
|
|
:rtype: str
|
|
|
|
:return: wrapped commands
|
|
|
|
"""
|
2016-08-19 21:28:29 +03:00
|
|
|
return '/bin/bash -c \'set -e; set -o pipefail; {}{}\''.format(
|
2016-07-18 18:19:36 +03:00
|
|
|
';'.join(commands), '; wait' if wait else '')
|
|
|
|
|
|
|
|
|
|
|
|
def _create_credentials(config: dict) -> tuple:
|
|
|
|
"""Create authenticated clients
|
|
|
|
:param dict config: configuration dict
|
|
|
|
:rtype: tuple
|
|
|
|
:return: (batch client, blob client, queue client, table client)
|
|
|
|
"""
|
|
|
|
credentials = batchauth.SharedKeyCredentials(
|
2016-08-12 19:21:33 +03:00
|
|
|
config['credentials']['batch_account'],
|
2016-07-18 18:19:36 +03:00
|
|
|
_BATCHACCOUNTKEY)
|
|
|
|
batch_client = batch.BatchServiceClient(
|
|
|
|
credentials,
|
2016-08-13 01:31:05 +03:00
|
|
|
base_url=config['credentials']['batch_account_service_url'])
|
2016-07-18 18:19:36 +03:00
|
|
|
blob_client = azureblob.BlockBlobService(
|
|
|
|
account_name=_STORAGEACCOUNT,
|
|
|
|
account_key=_STORAGEACCOUNTKEY,
|
2016-08-12 19:21:33 +03:00
|
|
|
endpoint_suffix=config['credentials']['storage_endpoint'])
|
2016-07-18 18:19:36 +03:00
|
|
|
queue_client = azurequeue.QueueService(
|
|
|
|
account_name=_STORAGEACCOUNT,
|
|
|
|
account_key=_STORAGEACCOUNTKEY,
|
2016-08-12 19:21:33 +03:00
|
|
|
endpoint_suffix=config['credentials']['storage_endpoint'])
|
2016-07-18 18:19:36 +03:00
|
|
|
table_client = azuretable.TableService(
|
|
|
|
account_name=_STORAGEACCOUNT,
|
|
|
|
account_key=_STORAGEACCOUNTKEY,
|
2016-08-12 19:21:33 +03:00
|
|
|
endpoint_suffix=config['credentials']['storage_endpoint'])
|
2016-07-18 18:19:36 +03:00
|
|
|
return batch_client, blob_client, queue_client, table_client
|
|
|
|
|
|
|
|
|
2016-08-20 01:13:33 +03:00
|
|
|
def compute_md5_for_file(
|
|
|
|
file: pathlib.Path, as_base64: bool, blocksize: int=65536) -> str:
|
|
|
|
"""Compute MD5 hash for file
|
2016-08-17 18:30:53 +03:00
|
|
|
:param pathlib.Path file: file to compute md5 for
|
2016-08-20 01:13:33 +03:00
|
|
|
:param bool as_base64: return as base64 encoded string
|
2016-08-17 18:30:53 +03:00
|
|
|
:param int blocksize: block size in bytes
|
|
|
|
:rtype: str
|
2016-08-20 01:13:33 +03:00
|
|
|
:return: md5 for file
|
2016-08-17 18:30:53 +03:00
|
|
|
"""
|
|
|
|
hasher = hashlib.md5()
|
|
|
|
with file.open('rb') as filedesc:
|
|
|
|
while True:
|
|
|
|
buf = filedesc.read(blocksize)
|
|
|
|
if not buf:
|
|
|
|
break
|
|
|
|
hasher.update(buf)
|
2016-08-20 01:13:33 +03:00
|
|
|
if as_base64:
|
|
|
|
return str(base64.b64encode(hasher.digest()), 'ascii')
|
|
|
|
else:
|
|
|
|
return hasher.digest()
|
2016-08-17 18:30:53 +03:00
|
|
|
|
|
|
|
|
2016-07-18 18:19:36 +03:00
|
|
|
def upload_resource_files(
|
|
|
|
blob_client: azure.storage.blob.BlockBlobService, config: dict,
|
|
|
|
files: List[tuple]) -> dict:
|
|
|
|
"""Upload resource files to blob storage
|
|
|
|
:param azure.storage.blob.BlockBlobService blob_client: blob client
|
|
|
|
:param dict config: configuration dict
|
|
|
|
:rtype: dict
|
|
|
|
:return: sas url dict
|
|
|
|
"""
|
|
|
|
sas_urls = {}
|
|
|
|
for file in files:
|
2016-08-12 19:21:33 +03:00
|
|
|
# skip if no file is specified
|
|
|
|
if file[0] is None:
|
|
|
|
continue
|
2016-07-20 19:18:31 +03:00
|
|
|
upload = True
|
2016-08-20 01:13:33 +03:00
|
|
|
fp = pathlib.Path(file[1])
|
|
|
|
if fp.name == _REGISTRY_FILE[0] and not fp.exists():
|
|
|
|
print('skipping optional docker registry image: {}'.format(
|
|
|
|
_REGISTRY_FILE[0]))
|
|
|
|
continue
|
|
|
|
else:
|
|
|
|
# check if blob exists
|
|
|
|
try:
|
|
|
|
prop = blob_client.get_blob_properties(
|
|
|
|
_STORAGE_CONTAINERS['blob_resourcefiles'], file[0])
|
|
|
|
if (prop.properties.content_settings.content_md5 ==
|
|
|
|
compute_md5_for_file(fp, True)):
|
|
|
|
print(('remote file is the same '
|
|
|
|
'for {}, skipping').format(file[0]))
|
|
|
|
upload = False
|
|
|
|
except azure.common.AzureMissingResourceHttpError:
|
|
|
|
pass
|
2016-07-18 18:19:36 +03:00
|
|
|
if upload:
|
|
|
|
print('uploading file: {}'.format(file[1]))
|
|
|
|
blob_client.create_blob_from_path(
|
|
|
|
_STORAGE_CONTAINERS['blob_resourcefiles'], file[0], file[1])
|
|
|
|
sas_urls[file[0]] = 'https://{}.blob.{}/{}/{}?{}'.format(
|
|
|
|
_STORAGEACCOUNT,
|
2016-08-12 19:21:33 +03:00
|
|
|
config['credentials']['storage_endpoint'],
|
2016-07-18 18:19:36 +03:00
|
|
|
_STORAGE_CONTAINERS['blob_resourcefiles'], file[0],
|
|
|
|
blob_client.generate_blob_shared_access_signature(
|
|
|
|
_STORAGE_CONTAINERS['blob_resourcefiles'], file[0],
|
|
|
|
permission=azureblob.BlobPermissions.READ,
|
|
|
|
expiry=datetime.datetime.utcnow() +
|
|
|
|
datetime.timedelta(hours=2)
|
|
|
|
)
|
|
|
|
)
|
|
|
|
return sas_urls
|
|
|
|
|
|
|
|
|
2016-08-20 01:13:33 +03:00
|
|
|
def setup_azurefile_volume_driver(
|
|
|
|
blob_client: azure.storage.blob.BlockBlobService,
|
|
|
|
config: dict) -> tuple:
|
|
|
|
# check to see if binary is downloaded
|
|
|
|
bin = pathlib.Path('resources/azurefile-dockervolumedriver')
|
|
|
|
if (not bin.exists() or
|
|
|
|
compute_md5_for_file(bin, False) != _AZUREFILE_DVD_MD5):
|
|
|
|
response = urllib.request.urlopen(_AZUREFILE_DVD_URL)
|
|
|
|
with bin.open('wb') as f:
|
|
|
|
f.write(response.read())
|
|
|
|
srv = pathlib.Path('resources/azurefile-dockervolumedriver.service')
|
|
|
|
if (not srv.exists() or
|
|
|
|
compute_md5_for_file(srv, False) !=
|
|
|
|
_AZUREFILE_SYSTEMD_SERVICE_MD5):
|
|
|
|
response = urllib.request.urlopen(_AZUREFILE_SYSTEMD_SERVICE_URL)
|
|
|
|
with srv.open('wb') as f:
|
|
|
|
f.write(response.read())
|
|
|
|
# construct systemd env file
|
|
|
|
sa = None
|
|
|
|
sakey = None
|
|
|
|
for svkey in config[
|
|
|
|
'global_resources']['docker_volumes']['shared_data_volumes']:
|
|
|
|
conf = config[
|
|
|
|
'global_resources']['docker_volumes']['shared_data_volumes'][svkey]
|
|
|
|
if conf['volume_driver'] == 'azurefile':
|
|
|
|
# check every entry to ensure the same storage account
|
|
|
|
_sa = conf['storage_account']
|
|
|
|
_sakey = conf['storage_account_key']
|
|
|
|
if sa is not None and sa != _sa:
|
|
|
|
raise ValueError(
|
|
|
|
'multiple storage accounts are not supported for '
|
|
|
|
'azurefile docker volume driver')
|
|
|
|
sa = _sa
|
|
|
|
sakey = _sakey
|
|
|
|
if sa is None or sakey is None:
|
|
|
|
raise RuntimeError(
|
|
|
|
'storage account or storage account key not specified for '
|
|
|
|
'azurefile docker volume driver')
|
|
|
|
srvenv = pathlib.Path('resources/azurefile-dockervolumedriver.env')
|
|
|
|
with srvenv.open('w') as f:
|
|
|
|
f.write('AZURE_STORAGE_ACCOUNT={}\n'.format(sa))
|
|
|
|
f.write('AZURE_STORAGE_ACCOUNT_KEY={}\n'.format(sakey))
|
|
|
|
# create docker volume mount command script
|
|
|
|
volcreate = pathlib.Path('resources/azurefile-dockervolume-create.sh')
|
|
|
|
with volcreate.open('w') as f:
|
|
|
|
f.write('#!/usr/bin/env bash\n\n')
|
|
|
|
for svkey in config[
|
|
|
|
'global_resources']['docker_volumes']['shared_data_volumes']:
|
|
|
|
conf = config[
|
|
|
|
'global_resources']['docker_volumes'][
|
|
|
|
'shared_data_volumes'][svkey]
|
|
|
|
opts = [
|
|
|
|
'-o share={}'.format(conf['azure_file_share_name'])
|
|
|
|
]
|
|
|
|
for opt in conf['mount_options']:
|
|
|
|
opts.append('-o {}'.format(opt))
|
|
|
|
f.write('docker volume create -d azurefile --name {} {}\n'.format(
|
|
|
|
svkey, ' '.join(opts)))
|
|
|
|
return bin, srv, srvenv, volcreate
|
|
|
|
|
|
|
|
|
2016-07-18 18:19:36 +03:00
|
|
|
def add_pool(
|
|
|
|
batch_client: azure.batch.batch_service_client.BatchServiceClient,
|
|
|
|
blob_client: azure.storage.blob.BlockBlobService, config: dict):
|
|
|
|
"""Add a Batch pool to account
|
|
|
|
:param azure.batch.batch_service_client.BatchServiceClient: batch client
|
|
|
|
:param azure.storage.blob.BlockBlobService blob_client: blob client
|
|
|
|
:param dict config: configuration dict
|
|
|
|
"""
|
2016-08-19 21:28:29 +03:00
|
|
|
publisher = config['pool_specification']['publisher']
|
|
|
|
offer = config['pool_specification']['offer']
|
|
|
|
sku = config['pool_specification']['sku']
|
2016-08-19 01:17:04 +03:00
|
|
|
# peer-to-peer settings
|
2016-07-18 18:19:36 +03:00
|
|
|
try:
|
2016-08-13 01:31:05 +03:00
|
|
|
p2p = config['data_replication']['peer_to_peer']['enabled']
|
2016-07-18 18:19:36 +03:00
|
|
|
except KeyError:
|
|
|
|
p2p = True
|
2016-08-17 18:30:53 +03:00
|
|
|
if p2p:
|
|
|
|
nonp2pcd = False
|
|
|
|
try:
|
|
|
|
p2psbias = config['data_replication'][
|
|
|
|
'peer_to_peer']['direct_download_seed_bias']
|
|
|
|
except KeyError:
|
|
|
|
p2psbias = 3
|
|
|
|
try:
|
|
|
|
p2pcomp = config[
|
|
|
|
'data_replication']['peer_to_peer']['compression']
|
|
|
|
except KeyError:
|
|
|
|
p2pcomp = True
|
|
|
|
else:
|
2016-08-13 01:31:05 +03:00
|
|
|
try:
|
|
|
|
nonp2pcd = config[
|
|
|
|
'data_replication']['non_peer_to_peer_concurrent_downloading']
|
|
|
|
except KeyError:
|
|
|
|
nonp2pcd = True
|
2016-08-19 01:17:04 +03:00
|
|
|
# private registry settings
|
2016-07-18 18:19:36 +03:00
|
|
|
try:
|
2016-08-17 18:30:53 +03:00
|
|
|
preg = config['docker_registry']['private']['enabled']
|
2016-08-12 19:21:33 +03:00
|
|
|
pcont = config['docker_registry']['private']['container']
|
2016-07-18 18:19:36 +03:00
|
|
|
except KeyError:
|
|
|
|
preg = False
|
2016-08-19 01:17:04 +03:00
|
|
|
if preg:
|
|
|
|
preg = ' -r {}:{}:{}'.format(
|
|
|
|
pcont, _REGISTRY_FILE[0], _REGISTRY_FILE[2])
|
|
|
|
else:
|
|
|
|
preg = ''
|
|
|
|
# docker settings
|
2016-07-18 18:19:36 +03:00
|
|
|
try:
|
2016-08-12 19:21:33 +03:00
|
|
|
dockeruser = config['docker_registry']['login']['username']
|
|
|
|
dockerpw = config['docker_registry']['login']['password']
|
2016-07-18 18:19:36 +03:00
|
|
|
except KeyError:
|
|
|
|
dockeruser = None
|
|
|
|
dockerpw = None
|
2016-08-20 01:13:33 +03:00
|
|
|
# check volume mounts for azurefile
|
|
|
|
azurefile_vd = False
|
|
|
|
try:
|
|
|
|
shared_data_volumes = config[
|
|
|
|
'global_resources']['docker_volumes']['shared_data_volumes']
|
|
|
|
for key in shared_data_volumes:
|
|
|
|
if shared_data_volumes[key]['volume_driver'] == 'azurefile':
|
|
|
|
azurefile_vd = True
|
|
|
|
except KeyError:
|
|
|
|
pass
|
2016-08-19 01:17:04 +03:00
|
|
|
# prefix settings
|
2016-07-18 18:19:36 +03:00
|
|
|
try:
|
2016-08-12 19:21:33 +03:00
|
|
|
prefix = config['storage_entity_prefix']
|
2016-07-18 18:19:36 +03:00
|
|
|
if len(prefix) == 0:
|
|
|
|
prefix = None
|
|
|
|
except KeyError:
|
|
|
|
prefix = None
|
|
|
|
# TODO for now, only support Ubuntu 16.04
|
|
|
|
if (publisher != 'Canonical' or offer != 'UbuntuServer' or
|
|
|
|
sku < '16.04.0-LTS'):
|
|
|
|
raise ValueError('Unsupported Docker Host VM Config')
|
|
|
|
# pick latest sku
|
|
|
|
node_agent_skus = batch_client.account.list_node_agent_skus()
|
|
|
|
skus_to_use = [
|
|
|
|
(nas, image_ref) for nas in node_agent_skus for image_ref in sorted(
|
|
|
|
nas.verified_image_references, key=lambda item: item.sku)
|
|
|
|
if image_ref.publisher.lower() == publisher.lower() and
|
|
|
|
image_ref.offer.lower() == offer.lower() and
|
|
|
|
image_ref.sku.lower() == sku.lower()
|
|
|
|
]
|
|
|
|
sku_to_use, image_ref_to_use = skus_to_use[-1]
|
2016-08-20 01:13:33 +03:00
|
|
|
# create resource files list
|
|
|
|
_rflist = [
|
|
|
|
_NODEPREP_FILE, _JOBPREP_FILE, _CASCADE_FILE, _SETUP_PR_FILE,
|
|
|
|
_PERF_FILE, _REGISTRY_FILE
|
|
|
|
]
|
|
|
|
# handle azurefile docker volume driver
|
|
|
|
if azurefile_vd:
|
|
|
|
if (publisher != 'Canonical' or offer != 'UbuntuServer' or
|
|
|
|
sku < '16.04.0-LTS'):
|
|
|
|
raise ValueError(
|
|
|
|
'Unsupported Docker Host VM Config with Azurefile '
|
|
|
|
'Docker Volume Driver')
|
|
|
|
afbin, afsrv, afenv, afvc = setup_azurefile_volume_driver(
|
|
|
|
blob_client, config)
|
|
|
|
_rflist.append((str(afbin.name), str(afbin)))
|
|
|
|
_rflist.append((str(afsrv.name), str(afsrv)))
|
|
|
|
_rflist.append((str(afenv.name), str(afenv)))
|
|
|
|
_rflist.append((str(afvc.name), str(afvc)))
|
2016-07-18 18:19:36 +03:00
|
|
|
# upload resource files
|
2016-08-20 01:13:33 +03:00
|
|
|
sas_urls = upload_resource_files(blob_client, config, _rflist)
|
|
|
|
del _rflist
|
2016-07-18 18:19:36 +03:00
|
|
|
# create pool param
|
|
|
|
pool = batchmodels.PoolAddParameter(
|
2016-08-19 21:28:29 +03:00
|
|
|
id=config['pool_specification']['id'],
|
2016-07-18 18:19:36 +03:00
|
|
|
virtual_machine_configuration=batchmodels.VirtualMachineConfiguration(
|
|
|
|
image_reference=image_ref_to_use,
|
|
|
|
node_agent_sku_id=sku_to_use.id),
|
2016-08-19 21:28:29 +03:00
|
|
|
vm_size=config['pool_specification']['vm_size'],
|
|
|
|
target_dedicated=config['pool_specification']['vm_count'],
|
2016-08-08 19:38:12 +03:00
|
|
|
enable_inter_node_communication=True,
|
2016-07-18 18:19:36 +03:00
|
|
|
start_task=batchmodels.StartTask(
|
2016-08-20 01:13:33 +03:00
|
|
|
command_line='{} -o {} -s {}{}{}{}{}{}'.format(
|
2016-08-19 01:17:04 +03:00
|
|
|
_NODEPREP_FILE[0],
|
|
|
|
offer,
|
|
|
|
sku,
|
|
|
|
preg,
|
2016-07-18 18:19:36 +03:00
|
|
|
' -p {}'.format(prefix) if prefix else '',
|
2016-08-17 18:30:53 +03:00
|
|
|
' -t {}:{}'.format(p2pcomp, p2psbias) if p2p else '',
|
2016-08-13 01:31:05 +03:00
|
|
|
' -c' if nonp2pcd else '',
|
2016-08-20 01:13:33 +03:00
|
|
|
' -a' if azurefile_vd else '',
|
2016-07-18 18:19:36 +03:00
|
|
|
),
|
|
|
|
run_elevated=True,
|
|
|
|
wait_for_success=True,
|
|
|
|
environment_settings=[
|
|
|
|
batchmodels.EnvironmentSetting('LC_ALL', 'en_US.UTF-8'),
|
|
|
|
batchmodels.EnvironmentSetting('CASCADE_SA', _STORAGEACCOUNT),
|
|
|
|
batchmodels.EnvironmentSetting(
|
|
|
|
'CASCADE_SAKEY', _STORAGEACCOUNTKEY),
|
|
|
|
],
|
|
|
|
resource_files=[],
|
|
|
|
),
|
|
|
|
)
|
|
|
|
for rf in sas_urls:
|
|
|
|
pool.start_task.resource_files.append(
|
|
|
|
batchmodels.ResourceFile(
|
|
|
|
file_path=rf,
|
|
|
|
blob_source=sas_urls[rf])
|
|
|
|
)
|
|
|
|
if preg:
|
|
|
|
pool.start_task.environment_settings.append(
|
|
|
|
batchmodels.EnvironmentSetting(
|
|
|
|
'PRIVATE_REGISTRY_SA',
|
2016-08-12 19:21:33 +03:00
|
|
|
config['docker_registry']['private']['storage_account'])
|
2016-07-18 18:19:36 +03:00
|
|
|
)
|
|
|
|
pool.start_task.environment_settings.append(
|
|
|
|
batchmodels.EnvironmentSetting(
|
|
|
|
'PRIVATE_REGISTRY_SAKEY',
|
2016-08-12 19:21:33 +03:00
|
|
|
config['docker_registry']['private']['storage_account_key'])
|
2016-07-18 18:19:36 +03:00
|
|
|
)
|
|
|
|
if (dockeruser is not None and len(dockeruser) > 0 and
|
|
|
|
dockerpw is not None and len(dockerpw) > 0):
|
|
|
|
pool.start_task.environment_settings.append(
|
|
|
|
batchmodels.EnvironmentSetting('DOCKER_LOGIN_USERNAME', dockeruser)
|
|
|
|
)
|
|
|
|
pool.start_task.environment_settings.append(
|
|
|
|
batchmodels.EnvironmentSetting('DOCKER_LOGIN_PASSWORD', dockerpw)
|
|
|
|
)
|
|
|
|
# create pool if not exists
|
|
|
|
try:
|
|
|
|
print('Attempting to create pool:', pool.id)
|
|
|
|
batch_client.pool.add(pool)
|
|
|
|
print('Created pool:', pool.id)
|
|
|
|
except batchmodels.BatchErrorException as e:
|
|
|
|
if e.error.code != 'PoolExists':
|
|
|
|
raise
|
|
|
|
else:
|
|
|
|
print('Pool {!r} already exists'.format(pool.id))
|
|
|
|
# wait for pool idle
|
|
|
|
node_state = frozenset(
|
|
|
|
(batchmodels.ComputeNodeState.starttaskfailed,
|
|
|
|
batchmodels.ComputeNodeState.unusable,
|
|
|
|
batchmodels.ComputeNodeState.idle)
|
|
|
|
)
|
|
|
|
print('waiting for all nodes in pool {} to reach one of: {!r}'.format(
|
|
|
|
pool.id, node_state))
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
# refresh pool to ensure that there is no resize error
|
|
|
|
pool = batch_client.pool.get(pool.id)
|
|
|
|
if pool.resize_error is not None:
|
|
|
|
raise RuntimeError(
|
2016-08-09 22:47:58 +03:00
|
|
|
'resize error encountered for pool {}: code={} msg={}'.format(
|
|
|
|
pool.id, pool.resize_error.code,
|
|
|
|
pool.resize_error.message))
|
2016-07-18 18:19:36 +03:00
|
|
|
nodes = list(batch_client.compute_node.list(pool.id))
|
|
|
|
if (len(nodes) >= pool.target_dedicated and
|
|
|
|
all(node.state in node_state for node in nodes)):
|
|
|
|
break
|
|
|
|
i += 1
|
|
|
|
if i % 3 == 0:
|
2016-08-09 22:47:58 +03:00
|
|
|
print('waiting for {} nodes to reach desired state:'.format(
|
2016-07-18 18:19:36 +03:00
|
|
|
pool.target_dedicated))
|
2016-08-09 22:47:58 +03:00
|
|
|
for node in nodes:
|
|
|
|
print(' > {}: {}'.format(node.id, node.state))
|
2016-07-18 18:19:36 +03:00
|
|
|
time.sleep(10)
|
|
|
|
get_remote_login_settings(batch_client, pool.id, nodes)
|
|
|
|
if any(node.state != batchmodels.ComputeNodeState.idle for node in nodes):
|
|
|
|
raise RuntimeError('node(s) of pool {} not in idle state'.format(
|
|
|
|
pool.id))
|
|
|
|
|
|
|
|
|
|
|
|
def resize_pool(batch_client, pool_id, vm_count):
|
|
|
|
print('Resizing pool {} to {}'.format(pool_id, vm_count))
|
|
|
|
batch_client.pool.resize(
|
|
|
|
pool_id=pool_id,
|
2016-08-09 22:47:58 +03:00
|
|
|
pool_resize_parameter=batchmodels.PoolResizeParameter(
|
|
|
|
target_dedicated=vm_count,
|
|
|
|
resize_timeout=datetime.timedelta(minutes=20),
|
|
|
|
)
|
2016-07-18 18:19:36 +03:00
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
def del_pool(batch_client, pool_id):
|
|
|
|
print('Deleting pool: {}'.format(pool_id))
|
|
|
|
batch_client.pool.delete(pool_id)
|
|
|
|
|
|
|
|
|
2016-08-17 18:30:53 +03:00
|
|
|
def del_node(batch_client, pool_id, node_id):
|
|
|
|
print('Deleting node {} from pool {}'.format(node_id, pool_id))
|
|
|
|
batch_client.pool.remove_nodes(
|
|
|
|
pool_id=pool_id,
|
|
|
|
node_remove_parameter=batchmodels.NodeRemoveParameter(
|
|
|
|
node_list=[node_id],
|
|
|
|
)
|
|
|
|
)
|
|
|
|
|
|
|
|
|
2016-08-19 21:28:29 +03:00
|
|
|
def add_job(batch_client, blob_client, config):
|
|
|
|
pool_id = config['pool_specification']['id']
|
2016-08-17 18:30:53 +03:00
|
|
|
global_resources = []
|
2016-08-19 21:28:29 +03:00
|
|
|
for gr in config['global_resources']['docker_images']:
|
|
|
|
global_resources.append(gr)
|
|
|
|
# TODO add global resources for non-docker resources
|
|
|
|
jpcmdline = _wrap_commands_in_shell([
|
|
|
|
'$AZ_BATCH_NODE_SHARED_DIR/{} {}'.format(
|
|
|
|
_JOBPREP_FILE[0], ' '.join(global_resources))])
|
|
|
|
for jobspec in config['job_specifications']:
|
|
|
|
job = batchmodels.JobAddParameter(
|
|
|
|
id=jobspec['id'],
|
|
|
|
pool_info=batchmodels.PoolInformation(pool_id=pool_id),
|
|
|
|
job_preparation_task=batchmodels.JobPreparationTask(
|
|
|
|
command_line=jpcmdline,
|
|
|
|
wait_for_success=True,
|
|
|
|
run_elevated=True,
|
|
|
|
rerun_on_node_reboot_after_success=False,
|
|
|
|
)
|
2016-08-17 18:30:53 +03:00
|
|
|
)
|
2016-08-19 21:28:29 +03:00
|
|
|
print('adding job: {}'.format(job.id))
|
|
|
|
try:
|
|
|
|
batch_client.job.add(job)
|
|
|
|
except batchmodels.batch_error.BatchErrorException as ex:
|
|
|
|
if 'The specified job already exists' not in ex.message.value:
|
|
|
|
raise
|
|
|
|
# add all tasks under job
|
|
|
|
for task in jobspec['tasks']:
|
|
|
|
image = task['image']
|
|
|
|
# get or generate task id
|
|
|
|
try:
|
|
|
|
task_id = task['id']
|
|
|
|
if task_id is None:
|
|
|
|
raise KeyError()
|
|
|
|
except KeyError:
|
|
|
|
# get filtered, sorted list of generic docker task ids
|
|
|
|
try:
|
|
|
|
tasklist = sorted(
|
|
|
|
filter(lambda x: x.id.startswith(
|
|
|
|
_GENERIC_DOCKER_TASK_PREFIX), list(
|
|
|
|
batch_client.task.list(job.id))),
|
|
|
|
key=lambda y: y.id)
|
|
|
|
tasknum = int(tasklist[-1].id.split('-')[-1]) + 1
|
|
|
|
except (batchmodels.batch_error.BatchErrorException,
|
|
|
|
IndexError):
|
|
|
|
tasknum = 0
|
|
|
|
task_id = '{0}{1:03d}'.format(
|
|
|
|
_GENERIC_DOCKER_TASK_PREFIX, tasknum)
|
|
|
|
# get generic run opts
|
|
|
|
try:
|
|
|
|
run_opts = task['additional_docker_run_options']
|
|
|
|
except KeyError:
|
|
|
|
run_opts = []
|
|
|
|
# parse remove container option
|
|
|
|
try:
|
|
|
|
rm_container = task['remove_container_after_exit']
|
|
|
|
except KeyError:
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
if rm_container and '--rm' not in run_opts:
|
|
|
|
run_opts.append('--rm')
|
|
|
|
del rm_container
|
|
|
|
# parse name option
|
|
|
|
try:
|
|
|
|
name = task['name']
|
|
|
|
if name is not None:
|
|
|
|
run_opts.append('-n {}'.format(name))
|
|
|
|
del name
|
|
|
|
except KeyError:
|
|
|
|
pass
|
|
|
|
# parse labels option
|
|
|
|
try:
|
|
|
|
labels = task['labels']
|
|
|
|
if labels is not None and len(labels) > 0:
|
|
|
|
for label in labels:
|
|
|
|
run_opts.append('-l {}'.format(label))
|
|
|
|
del labels
|
|
|
|
except KeyError:
|
|
|
|
pass
|
|
|
|
# parse ports option
|
|
|
|
try:
|
|
|
|
ports = task['ports']
|
|
|
|
if ports is not None and len(ports) > 0:
|
|
|
|
for port in ports:
|
|
|
|
run_opts.append('-p {}'.format(port))
|
|
|
|
del ports
|
|
|
|
except KeyError:
|
|
|
|
pass
|
|
|
|
# parse entrypoint
|
|
|
|
try:
|
|
|
|
entrypoint = task['entrypoint']
|
|
|
|
if entrypoint is not None:
|
|
|
|
run_opts.append('--entrypoint {}'.format(entrypoint))
|
|
|
|
del entrypoint
|
|
|
|
except KeyError:
|
|
|
|
pass
|
|
|
|
# parse data volumes
|
|
|
|
try:
|
|
|
|
data_volumes = task['data_volumes']
|
|
|
|
except KeyError:
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
if data_volumes is not None and len(data_volumes) > 0:
|
|
|
|
for key in data_volumes:
|
|
|
|
dvspec = config[
|
|
|
|
'global_resources']['docker_volumes'][
|
|
|
|
'data_volumes'][key]
|
|
|
|
try:
|
|
|
|
hostpath = dvspec['host_path']
|
|
|
|
except KeyError:
|
|
|
|
hostpath = None
|
|
|
|
if hostpath is not None:
|
|
|
|
run_opts.append('-v {}:{}'.format(
|
|
|
|
hostpath, dvspec['container_path']))
|
|
|
|
else:
|
|
|
|
run_opts.append('-v {}'.format(
|
|
|
|
dvspec['container_path']))
|
|
|
|
# parse shared data volumes
|
|
|
|
try:
|
|
|
|
shared_data_volumes = task['shared_data_volumes']
|
|
|
|
except KeyError:
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
if (shared_data_volumes is not None and
|
|
|
|
len(shared_data_volumes) > 0):
|
|
|
|
for key in shared_data_volumes:
|
|
|
|
dvspec = config[
|
|
|
|
'global_resources']['docker_volumes'][
|
|
|
|
'shared_data_volumes'][key]
|
|
|
|
run_opts.append('-v {}:{}'.format(
|
|
|
|
key, dvspec['container_path']))
|
|
|
|
# get command
|
|
|
|
try:
|
|
|
|
command = task['command']
|
|
|
|
except KeyError:
|
|
|
|
command = None
|
|
|
|
# get and create env var file
|
2016-08-20 01:13:33 +03:00
|
|
|
envfile = '.env.list'.format(task_id)
|
2016-08-19 21:28:29 +03:00
|
|
|
sas_urls = None
|
|
|
|
try:
|
2016-08-20 01:13:33 +03:00
|
|
|
env_vars = jobspec['environment_variables']
|
|
|
|
except KeyError:
|
|
|
|
env_vars = None
|
|
|
|
try:
|
|
|
|
task_ev = task['environment_variables']
|
|
|
|
if env_vars is None:
|
|
|
|
env_vars = task_ev
|
|
|
|
else:
|
|
|
|
env_vars = merge_dict(env_vars, task_ev)
|
2016-08-19 21:28:29 +03:00
|
|
|
except KeyError:
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
if env_vars is not None and len(env_vars) > 0:
|
2016-08-20 01:13:33 +03:00
|
|
|
envfiletmp = _TEMP_DIR / '{}{}'.format(task_id, envfile)
|
|
|
|
envfileloc = '{}taskrf-{}/{}{}'.format(
|
|
|
|
config['storage_entity_prefix'], job.id, task_id,
|
|
|
|
envfile)
|
2016-08-19 21:28:29 +03:00
|
|
|
with envfiletmp.open('w') as f:
|
|
|
|
for key in env_vars:
|
|
|
|
f.write('{}={}{}'.format(
|
|
|
|
key, env_vars[key], os.linesep))
|
|
|
|
# upload env var file if exists
|
|
|
|
sas_urls = upload_resource_files(
|
|
|
|
blob_client, config, [(envfileloc, str(envfiletmp))])
|
|
|
|
envfiletmp.unlink()
|
2016-08-20 01:13:33 +03:00
|
|
|
if len(sas_urls) != 1:
|
|
|
|
raise RuntimeError('unexpected number of sas urls')
|
2016-08-19 21:28:29 +03:00
|
|
|
# always add option for envfile
|
|
|
|
run_opts.append('--env-file {}'.format(envfile))
|
|
|
|
# mount batch root dir
|
|
|
|
run_opts.append(
|
|
|
|
'-v $AZ_BATCH_NODE_ROOT_DIR:$AZ_BATCH_NODE_ROOT_DIR')
|
|
|
|
# set working directory
|
|
|
|
run_opts.append('-w $AZ_BATCH_TASK_WORKING_DIR')
|
|
|
|
# add task
|
|
|
|
task_commands = [
|
|
|
|
'env | grep AZ_BATCH_ >> {}'.format(envfile),
|
|
|
|
'docker run {} {}{}'.format(
|
|
|
|
' '.join(run_opts),
|
|
|
|
image,
|
|
|
|
' {}'.format(command) if command else '')
|
|
|
|
]
|
|
|
|
task = batchmodels.TaskAddParameter(
|
|
|
|
id=task_id,
|
|
|
|
command_line=_wrap_commands_in_shell(task_commands),
|
|
|
|
run_elevated=True,
|
|
|
|
resource_files=[],
|
|
|
|
)
|
2016-08-20 01:13:33 +03:00
|
|
|
task.resource_files.append(
|
|
|
|
batchmodels.ResourceFile(
|
|
|
|
file_path=str(envfile),
|
|
|
|
blob_source=next(iter(sas_urls.values())),
|
|
|
|
file_mode='0640',
|
2016-08-19 21:28:29 +03:00
|
|
|
)
|
2016-08-20 01:13:33 +03:00
|
|
|
)
|
2016-08-19 21:28:29 +03:00
|
|
|
print('adding task {}: {}'.format(
|
|
|
|
task_id, task.command_line))
|
|
|
|
batch_client.task.add(job_id=job.id, task=task)
|
2016-07-18 18:19:36 +03:00
|
|
|
|
|
|
|
|
|
|
|
def del_job(batch_client, job_id):
|
|
|
|
print('Deleting job: {}'.format(job_id))
|
|
|
|
batch_client.job.delete(job_id)
|
|
|
|
|
|
|
|
|
2016-08-19 21:28:29 +03:00
|
|
|
def terminate_job(batch_client, job_id):
|
|
|
|
print('Terminating job: {}'.format(job_id))
|
|
|
|
batch_client.job.terminate(job_id)
|
|
|
|
|
|
|
|
|
2016-07-18 18:19:36 +03:00
|
|
|
def del_all_jobs(batch_client):
|
|
|
|
print('Listing jobs...')
|
|
|
|
jobs = batch_client.job.list()
|
|
|
|
for job in jobs:
|
|
|
|
del_job(batch_client, job.id)
|
|
|
|
|
|
|
|
|
|
|
|
def get_remote_login_settings(batch_client, pool_id, nodes=None):
|
|
|
|
if nodes is None:
|
|
|
|
nodes = batch_client.compute_node.list(pool_id)
|
|
|
|
for node in nodes:
|
|
|
|
rls = batch_client.compute_node.get_remote_login_settings(
|
|
|
|
pool_id, node.id)
|
|
|
|
print('node {}: {}'.format(node.id, rls))
|
|
|
|
|
|
|
|
|
|
|
|
def delete_storage_containers(blob_client, queue_client, table_client, config):
|
2016-08-09 22:47:58 +03:00
|
|
|
for key in _STORAGE_CONTAINERS:
|
|
|
|
if key.startswith('blob_'):
|
|
|
|
blob_client.delete_container(_STORAGE_CONTAINERS[key])
|
|
|
|
elif key.startswith('table_'):
|
|
|
|
table_client.delete_table(_STORAGE_CONTAINERS[key])
|
|
|
|
elif key.startswith('queue_'):
|
|
|
|
queue_client.delete_queue(_STORAGE_CONTAINERS[key])
|
2016-07-18 18:19:36 +03:00
|
|
|
|
|
|
|
|
|
|
|
def _clear_blobs(blob_client, container):
|
|
|
|
print('deleting blobs: {}'.format(container))
|
|
|
|
blobs = blob_client.list_blobs(container)
|
|
|
|
for blob in blobs:
|
|
|
|
blob_client.delete_blob(container, blob.name)
|
|
|
|
|
|
|
|
|
2016-08-20 01:13:33 +03:00
|
|
|
def _clear_blob_task_resourcefiles(blob_client, container, config):
|
|
|
|
envfileloc = '{}taskrf-'.format(config['storage_entity_prefix'])
|
|
|
|
print('deleting blobs with prefix: {}'.format(envfileloc))
|
|
|
|
blobs = blob_client.list_blobs(container, prefix=envfileloc)
|
|
|
|
for blob in blobs:
|
|
|
|
blob_client.delete_blob(container, blob.name)
|
|
|
|
|
|
|
|
|
2016-08-09 22:47:58 +03:00
|
|
|
def _clear_table(table_client, table_name, config):
|
2016-07-18 18:19:36 +03:00
|
|
|
print('clearing table: {}'.format(table_name))
|
2016-08-09 22:47:58 +03:00
|
|
|
ents = table_client.query_entities(
|
|
|
|
table_name, filter='PartitionKey eq \'{}${}\''.format(
|
2016-08-12 19:21:33 +03:00
|
|
|
config['credentials']['batch_account'],
|
2016-08-19 21:28:29 +03:00
|
|
|
config['pool_specification']['id'])
|
2016-08-09 22:47:58 +03:00
|
|
|
)
|
|
|
|
# batch delete entities
|
2016-08-10 01:23:15 +03:00
|
|
|
i = 0
|
|
|
|
bet = azuretable.TableBatch()
|
|
|
|
for ent in ents:
|
|
|
|
bet.delete_entity(ent['PartitionKey'], ent['RowKey'])
|
|
|
|
i += 1
|
|
|
|
if i == 100:
|
|
|
|
table_client.commit_batch(table_name, bet)
|
|
|
|
bet = azuretable.TableBatch()
|
|
|
|
i = 0
|
|
|
|
if i > 0:
|
|
|
|
table_client.commit_batch(table_name, bet)
|
2016-07-18 18:19:36 +03:00
|
|
|
|
|
|
|
|
|
|
|
def clear_storage_containers(blob_client, queue_client, table_client, config):
|
2016-08-09 22:47:58 +03:00
|
|
|
for key in _STORAGE_CONTAINERS:
|
|
|
|
if key.startswith('blob_'):
|
2016-08-19 01:17:04 +03:00
|
|
|
# TODO this is temp to preserve registry upload
|
|
|
|
if key != 'blob_resourcefiles':
|
|
|
|
_clear_blobs(blob_client, _STORAGE_CONTAINERS[key])
|
2016-08-20 01:13:33 +03:00
|
|
|
else:
|
|
|
|
_clear_blob_task_resourcefiles(
|
|
|
|
blob_client, _STORAGE_CONTAINERS[key], config)
|
2016-08-09 22:47:58 +03:00
|
|
|
elif key.startswith('table_'):
|
|
|
|
_clear_table(table_client, _STORAGE_CONTAINERS[key], config)
|
|
|
|
elif key.startswith('queue_'):
|
|
|
|
print('clearing queue: {}'.format(_STORAGE_CONTAINERS[key]))
|
|
|
|
queue_client.clear_messages(_STORAGE_CONTAINERS[key])
|
2016-07-18 18:19:36 +03:00
|
|
|
|
|
|
|
|
|
|
|
def create_storage_containers(blob_client, queue_client, table_client, config):
|
2016-08-09 22:47:58 +03:00
|
|
|
for key in _STORAGE_CONTAINERS:
|
|
|
|
if key.startswith('blob_'):
|
|
|
|
print('creating container: {}'.format(_STORAGE_CONTAINERS[key]))
|
|
|
|
blob_client.create_container(_STORAGE_CONTAINERS[key])
|
|
|
|
elif key.startswith('table_'):
|
|
|
|
print('creating table: {}'.format(_STORAGE_CONTAINERS[key]))
|
|
|
|
table_client.create_table(_STORAGE_CONTAINERS[key])
|
|
|
|
elif key.startswith('queue_'):
|
|
|
|
print('creating queue: {}'.format(_STORAGE_CONTAINERS[key]))
|
|
|
|
queue_client.create_queue(_STORAGE_CONTAINERS[key])
|
2016-07-18 18:19:36 +03:00
|
|
|
|
|
|
|
|
2016-08-19 21:28:29 +03:00
|
|
|
def _add_global_resource(
|
|
|
|
queue_client, table_client, config, pk, p2pcsd, grtype):
|
|
|
|
try:
|
|
|
|
for gr in config['global_resources'][grtype]:
|
|
|
|
if grtype == 'docker_images':
|
|
|
|
prefix = 'docker'
|
|
|
|
else:
|
|
|
|
raise NotImplementedError()
|
|
|
|
resource = '{}:{}'.format(prefix, gr)
|
|
|
|
table_client.insert_or_replace_entity(
|
|
|
|
_STORAGE_CONTAINERS['table_globalresources'],
|
|
|
|
{
|
|
|
|
'PartitionKey': pk,
|
|
|
|
'RowKey': hashlib.sha1(
|
|
|
|
resource.encode('utf8')).hexdigest(),
|
|
|
|
'Resource': resource,
|
|
|
|
}
|
|
|
|
)
|
|
|
|
for _ in range(0, p2pcsd):
|
|
|
|
queue_client.put_message(
|
|
|
|
_STORAGE_CONTAINERS['queue_globalresources'], resource)
|
|
|
|
except KeyError:
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
2016-07-18 18:19:36 +03:00
|
|
|
def populate_queues(queue_client, table_client, config):
|
|
|
|
try:
|
2016-08-17 18:30:53 +03:00
|
|
|
preg = config['docker_registry']['private']['enabled']
|
2016-07-18 18:19:36 +03:00
|
|
|
except KeyError:
|
2016-08-17 18:30:53 +03:00
|
|
|
preg = False
|
2016-07-18 18:19:36 +03:00
|
|
|
pk = '{}${}'.format(
|
2016-08-12 19:21:33 +03:00
|
|
|
config['credentials']['batch_account'],
|
2016-08-19 21:28:29 +03:00
|
|
|
config['pool_specification']['id'])
|
2016-07-18 18:19:36 +03:00
|
|
|
# if using docker public hub, then populate registry table with hub
|
2016-08-19 01:17:04 +03:00
|
|
|
if not preg:
|
2016-08-17 18:30:53 +03:00
|
|
|
table_client.insert_or_replace_entity(
|
|
|
|
_STORAGE_CONTAINERS['table_registry'],
|
|
|
|
{
|
|
|
|
'PartitionKey': pk,
|
|
|
|
'RowKey': 'registry.hub.docker.com',
|
|
|
|
'Port': 80,
|
|
|
|
}
|
|
|
|
)
|
2016-08-19 21:28:29 +03:00
|
|
|
# get p2pcsd setting
|
2016-08-13 01:31:05 +03:00
|
|
|
try:
|
|
|
|
p2p = config['data_replication']['peer_to_peer']['enabled']
|
|
|
|
except KeyError:
|
|
|
|
p2p = True
|
|
|
|
if p2p:
|
|
|
|
try:
|
|
|
|
p2pcsd = config['data_replication']['peer_to_peer'][
|
|
|
|
'concurrent_source_downloads']
|
|
|
|
except KeyError:
|
|
|
|
p2pcsd = 1
|
|
|
|
else:
|
|
|
|
p2pcsd = 1
|
2016-08-19 21:28:29 +03:00
|
|
|
# add global resources
|
|
|
|
_add_global_resource(
|
|
|
|
queue_client, table_client, config, pk, p2pcsd, 'docker_images')
|
2016-07-18 18:19:36 +03:00
|
|
|
|
|
|
|
|
2016-08-12 19:21:33 +03:00
|
|
|
def merge_dict(dict1, dict2):
|
|
|
|
"""Recursively merge dictionaries: dict2 on to dict1. This differs
|
|
|
|
from dict.update() in that values that are dicts are recursively merged.
|
|
|
|
Note that only dict value types are merged, not lists, etc.
|
|
|
|
|
|
|
|
Code adapted from:
|
|
|
|
https://www.xormedia.com/recursively-merge-dictionaries-in-python/
|
|
|
|
|
|
|
|
:param dict dict1: dictionary to merge to
|
|
|
|
:param dict dict2: dictionary to merge with
|
|
|
|
:rtype: dict
|
|
|
|
:return: merged dictionary
|
|
|
|
"""
|
|
|
|
if not isinstance(dict1, dict) or not isinstance(dict2, dict):
|
|
|
|
raise ValueError('dict1 or dict2 is not a dictionary')
|
|
|
|
result = copy.deepcopy(dict1)
|
|
|
|
for k, v in dict2.items():
|
|
|
|
if k in result and isinstance(result[k], dict):
|
|
|
|
result[k] = merge_dict(result[k], v)
|
|
|
|
else:
|
|
|
|
result[k] = copy.deepcopy(v)
|
|
|
|
return result
|
|
|
|
|
|
|
|
|
2016-07-18 18:19:36 +03:00
|
|
|
def main():
|
|
|
|
"""Main function"""
|
|
|
|
# get command-line args
|
|
|
|
args = parseargs()
|
|
|
|
args.action = args.action.lower()
|
|
|
|
|
2016-08-12 19:21:33 +03:00
|
|
|
if args.settings is None:
|
|
|
|
raise ValueError('global settings not specified')
|
|
|
|
if args.config is None:
|
|
|
|
raise ValueError('config settings for action not specified')
|
|
|
|
|
|
|
|
with open(args.settings, 'r') as f:
|
|
|
|
config = json.load(f)
|
|
|
|
with open(args.config, 'r') as f:
|
|
|
|
config = merge_dict(config, json.load(f))
|
|
|
|
print('config:')
|
|
|
|
print(json.dumps(config, indent=4))
|
2016-08-20 01:13:33 +03:00
|
|
|
_populate_global_settings(config, args.action)
|
2016-07-18 18:19:36 +03:00
|
|
|
|
|
|
|
batch_client, blob_client, queue_client, table_client = \
|
|
|
|
_create_credentials(config)
|
|
|
|
|
|
|
|
if args.action == 'addpool':
|
|
|
|
create_storage_containers(
|
|
|
|
blob_client, queue_client, table_client, config)
|
2016-08-09 22:47:58 +03:00
|
|
|
clear_storage_containers(
|
|
|
|
blob_client, queue_client, table_client, config)
|
2016-07-18 18:19:36 +03:00
|
|
|
populate_queues(queue_client, table_client, config)
|
|
|
|
add_pool(batch_client, blob_client, config)
|
|
|
|
elif args.action == 'resizepool':
|
|
|
|
resize_pool(batch_client, args.poolid, args.numvms)
|
|
|
|
elif args.action == 'delpool':
|
|
|
|
del_pool(batch_client, args.poolid)
|
2016-08-17 18:30:53 +03:00
|
|
|
elif args.action == 'delnode':
|
|
|
|
del_node(batch_client, args.poolid, args.nodeid)
|
2016-07-18 18:19:36 +03:00
|
|
|
elif args.action == 'addjob':
|
2016-08-19 21:28:29 +03:00
|
|
|
add_job(batch_client, blob_client, config)
|
|
|
|
elif args.action == 'termjob':
|
|
|
|
terminate_job(batch_client, blob_client, args.jobid)
|
2016-07-18 18:19:36 +03:00
|
|
|
elif args.action == 'deljob':
|
|
|
|
del_job(batch_client, args.jobid)
|
|
|
|
elif args.action == 'delalljobs':
|
|
|
|
del_all_jobs(batch_client)
|
|
|
|
elif args.action == 'grl':
|
|
|
|
get_remote_login_settings(batch_client, args.poolid)
|
|
|
|
elif args.action == 'delstorage':
|
|
|
|
delete_storage_containers(
|
|
|
|
blob_client, queue_client, table_client, config)
|
|
|
|
elif args.action == 'clearstorage':
|
|
|
|
clear_storage_containers(
|
|
|
|
blob_client, queue_client, table_client, config)
|
|
|
|
else:
|
|
|
|
raise ValueError('Unknown action: {}'.format(args.action))
|
|
|
|
|
|
|
|
|
|
|
|
def parseargs():
|
|
|
|
"""Parse program arguments
|
|
|
|
:rtype: argparse.Namespace
|
|
|
|
:return: parsed arguments
|
|
|
|
"""
|
|
|
|
parser = argparse.ArgumentParser(
|
|
|
|
description='Shipyard: Azure Batch to Docker Bridge')
|
|
|
|
parser.add_argument(
|
2016-08-19 21:28:29 +03:00
|
|
|
'action', help='action: addpool, addjob, termjob, delpool, '
|
|
|
|
'delnode, deljob, delalljobs, grl, delstorage, clearstorage')
|
2016-07-18 18:19:36 +03:00
|
|
|
parser.add_argument(
|
2016-08-12 19:21:33 +03:00
|
|
|
'--settings',
|
|
|
|
help='global settings json file config. required for all actions')
|
|
|
|
parser.add_argument(
|
|
|
|
'--config',
|
|
|
|
help='json file config for option. required for all actions')
|
2016-07-18 18:19:36 +03:00
|
|
|
parser.add_argument('--poolid', help='pool id')
|
|
|
|
parser.add_argument('--jobid', help='job id')
|
2016-08-17 18:30:53 +03:00
|
|
|
parser.add_argument('--nodeid', help='node id')
|
2016-07-18 18:19:36 +03:00
|
|
|
return parser.parse_args()
|
|
|
|
|
|
|
|
if __name__ == '__main__':
|
|
|
|
main()
|