3048 строки
114 KiB
Python
3048 строки
114 KiB
Python
#!/usr/bin/env python
|
|
# Copyright 2019 The Emscripten Authors. All rights reserved.
|
|
# Emscripten is available under two separate licenses, the MIT license and the
|
|
# University of Illinois/NCSA Open Source License. Both these licenses can be
|
|
# found in the LICENSE file.
|
|
|
|
from __future__ import print_function
|
|
|
|
import copy
|
|
from collections import OrderedDict
|
|
import errno
|
|
import json
|
|
import multiprocessing
|
|
import os
|
|
import os.path
|
|
import platform
|
|
import re
|
|
import shutil
|
|
import stat
|
|
import subprocess
|
|
import sys
|
|
import sysconfig
|
|
import zipfile
|
|
if os.name == 'nt':
|
|
try:
|
|
import winreg
|
|
except ImportError:
|
|
# old python 2 name
|
|
import _winreg as winreg
|
|
import ctypes.wintypes
|
|
|
|
if sys.version_info >= (3,):
|
|
from urllib.parse import urljoin
|
|
from urllib.request import urlopen
|
|
import functools
|
|
else:
|
|
from urlparse import urljoin
|
|
from urllib2 import urlopen
|
|
|
|
|
|
emsdk_packages_url = 'https://storage.googleapis.com/webassembly/emscripten-releases-builds/deps/'
|
|
|
|
emscripten_releases_repo = 'https://chromium.googlesource.com/emscripten-releases'
|
|
|
|
emscripten_releases_download_url_template = "https://storage.googleapis.com/webassembly/emscripten-releases-builds/%s/%s/wasm-binaries.%s"
|
|
|
|
emsdk_zip_download_url = 'https://github.com/emscripten-core/emsdk/archive/master.zip'
|
|
|
|
zips_subdir = 'zips/'
|
|
|
|
extra_release_tag = None
|
|
|
|
# Enable this to do very verbose printing about the different steps that are
|
|
# being run. Useful for debugging.
|
|
VERBOSE = int(os.getenv('EMSDK_VERBOSE', '0'))
|
|
TTY_OUTPUT = not os.getenv('EMSDK_NOTTY', not sys.stdout.isatty())
|
|
|
|
WINDOWS = False
|
|
if os.name == 'nt' or (os.getenv('SYSTEMROOT') is not None and 'windows' in os.getenv('SYSTEMROOT').lower()) or (os.getenv('COMSPEC') is not None and 'windows' in os.getenv('COMSPEC').lower()):
|
|
WINDOWS = True
|
|
|
|
|
|
def errlog(msg):
|
|
print(msg, file=sys.stderr)
|
|
|
|
|
|
MINGW = False
|
|
MSYS = False
|
|
if os.getenv('MSYSTEM'):
|
|
MSYS = True
|
|
# Some functions like os.path.normpath() exhibit different behavior between
|
|
# different versions of Python, so we need to distinguish between the MinGW
|
|
# and MSYS versions of Python
|
|
if sysconfig.get_platform() == 'mingw':
|
|
MINGW = True
|
|
if os.getenv('MSYSTEM') != 'MSYS' and os.getenv('MSYSTEM') != 'MINGW64':
|
|
# https://stackoverflow.com/questions/37460073/msys-vs-mingw-internal-environment-variables
|
|
errlog('Warning: MSYSTEM environment variable is present, and is set to "' + os.getenv('MSYSTEM') + '". This shell has not been tested with emsdk and may not work.')
|
|
|
|
MACOS = False
|
|
if platform.mac_ver()[0] != '':
|
|
MACOS = True
|
|
|
|
LINUX = False
|
|
if not MACOS and (platform.system() == 'Linux'):
|
|
LINUX = True
|
|
|
|
UNIX = (MACOS or LINUX)
|
|
|
|
|
|
# Pick which shell of 4 shells to use
|
|
POWERSHELL = bool(os.getenv('EMSDK_POWERSHELL'))
|
|
CSH = bool(os.getenv('EMSDK_CSH'))
|
|
CMD = bool(os.getenv('EMSDK_CMD'))
|
|
BASH = bool(os.getenv('EMSDK_BASH'))
|
|
if WINDOWS and BASH:
|
|
MSYS = True
|
|
|
|
if not CSH and not POWERSHELL and not BASH and not CMD:
|
|
# Fall back to default of `cmd` on windows and `bash` otherwise
|
|
if WINDOWS and not MSYS:
|
|
CMD = True
|
|
else:
|
|
BASH = True
|
|
|
|
if WINDOWS:
|
|
ENVPATH_SEPARATOR = ';'
|
|
else:
|
|
ENVPATH_SEPARATOR = ':'
|
|
|
|
ARCH = 'unknown'
|
|
# platform.machine() may return AMD64 on windows, so standardize the case.
|
|
machine = platform.machine().lower()
|
|
if machine.startswith('x64') or machine.startswith('amd64') or machine.startswith('x86_64'):
|
|
ARCH = 'x86_64'
|
|
elif machine.endswith('86'):
|
|
ARCH = 'x86'
|
|
elif machine.startswith('aarch64') or machine.lower().startswith('arm64'):
|
|
ARCH = 'aarch64'
|
|
elif platform.machine().startswith('arm'):
|
|
ARCH = 'arm'
|
|
else:
|
|
errlog("Warning: unknown machine architecture " + machine)
|
|
errlog()
|
|
|
|
# Don't saturate all cores to not steal the whole system, but be aggressive.
|
|
CPU_CORES = int(os.environ.get('EMSDK_NUM_CORES', max(multiprocessing.cpu_count() - 1, 1)))
|
|
|
|
CMAKE_BUILD_TYPE_OVERRIDE = None
|
|
|
|
# If true, perform a --shallow clone of git.
|
|
GIT_CLONE_SHALLOW = False
|
|
|
|
# If true, LLVM backend is built with tests enabled, and Binaryen is built with
|
|
# Visual Studio static analyzer enabled.
|
|
BUILD_FOR_TESTING = False
|
|
|
|
# If 'auto', assertions are decided by the build type
|
|
# (Release&MinSizeRel=disabled, Debug&RelWithDebInfo=enabled)
|
|
# Other valid values are 'ON' and 'OFF'
|
|
ENABLE_LLVM_ASSERTIONS = 'auto'
|
|
|
|
|
|
def os_name():
|
|
if WINDOWS:
|
|
return 'win'
|
|
elif LINUX:
|
|
return 'linux'
|
|
elif MACOS:
|
|
return 'macos'
|
|
else:
|
|
raise Exception('unknown OS')
|
|
|
|
|
|
def os_name_for_emscripten_releases():
|
|
if WINDOWS:
|
|
return 'win'
|
|
elif LINUX:
|
|
return 'linux'
|
|
elif MACOS:
|
|
return 'mac'
|
|
else:
|
|
raise Exception('unknown OS')
|
|
|
|
|
|
def debug_print(msg):
|
|
if VERBOSE:
|
|
errlog(msg)
|
|
|
|
|
|
def to_unix_path(p):
|
|
return p.replace('\\', '/')
|
|
|
|
|
|
def emsdk_path():
|
|
return to_unix_path(os.path.dirname(os.path.realpath(__file__)))
|
|
|
|
|
|
EMSDK_SET_ENV = ""
|
|
if POWERSHELL:
|
|
EMSDK_SET_ENV = os.path.join(emsdk_path(), 'emsdk_set_env.ps1')
|
|
else:
|
|
EMSDK_SET_ENV = os.path.join(emsdk_path(), 'emsdk_set_env.bat')
|
|
|
|
ARCHIVE_SUFFIXES = ('zip', '.tar', '.gz', '.xz', '.tbz2', '.bz2')
|
|
|
|
|
|
# Finds the given executable 'program' in PATH. Operates like the Unix tool 'which'.
|
|
def which(program):
|
|
def is_exe(fpath):
|
|
return os.path.isfile(fpath) and (WINDOWS or os.access(fpath, os.X_OK))
|
|
|
|
fpath, fname = os.path.split(program)
|
|
if fpath:
|
|
if is_exe(program):
|
|
return program
|
|
else:
|
|
for path in os.environ["PATH"].split(os.pathsep):
|
|
path = path.strip('"')
|
|
exe_file = os.path.join(path, program)
|
|
if is_exe(exe_file):
|
|
return exe_file
|
|
|
|
if WINDOWS and '.' not in fname:
|
|
if is_exe(exe_file + '.exe'):
|
|
return exe_file + '.exe'
|
|
if is_exe(exe_file + '.cmd'):
|
|
return exe_file + '.cmd'
|
|
if is_exe(exe_file + '.bat'):
|
|
return exe_file + '.bat'
|
|
|
|
return None
|
|
|
|
|
|
def vswhere(version):
|
|
try:
|
|
program_files = os.environ['ProgramFiles(x86)'] if 'ProgramFiles(x86)' in os.environ else os.environ['ProgramFiles']
|
|
vswhere_path = os.path.join(program_files, 'Microsoft Visual Studio', 'Installer', 'vswhere.exe')
|
|
output = json.loads(subprocess.check_output([vswhere_path, '-latest', '-version', '[%s.0,%s.0)' % (version, version + 1), '-requires', 'Microsoft.VisualStudio.Component.VC.Tools.x86.x64', '-property', 'installationPath', '-format', 'json']))
|
|
# Visual Studio 2017 Express is not included in the above search, and it
|
|
# does not have the VC.Tools.x86.x64 tool, so do a catch-all attempt as a
|
|
# fallback, to detect Express version.
|
|
if not output:
|
|
output = json.loads(subprocess.check_output([vswhere_path, '-latest', '-version', '[%s.0,%s.0)' % (version, version + 1), '-products', '*', '-property', 'installationPath', '-format', 'json']))
|
|
if not output:
|
|
return ''
|
|
return str(output[0]['installationPath'])
|
|
except Exception:
|
|
return ''
|
|
|
|
|
|
def vs_filewhere(installation_path, platform, file):
|
|
try:
|
|
vcvarsall = os.path.join(installation_path, 'VC\\Auxiliary\\Build\\vcvarsall.bat')
|
|
env = subprocess.check_output('cmd /c "%s" %s & where %s' % (vcvarsall, platform, file))
|
|
paths = [path[:-len(file)] for path in env.split('\r\n') if path.endswith(file)]
|
|
return paths[0]
|
|
except Exception:
|
|
return ''
|
|
|
|
|
|
CMAKE_GENERATOR = 'Unix Makefiles'
|
|
if WINDOWS:
|
|
# Detect which CMake generator to use when building on Windows
|
|
if '--mingw' in sys.argv:
|
|
CMAKE_GENERATOR = 'MinGW Makefiles'
|
|
elif '--vs2017' in sys.argv:
|
|
CMAKE_GENERATOR = 'Visual Studio 15'
|
|
elif '--vs2019' in sys.argv:
|
|
CMAKE_GENERATOR = 'Visual Studio 16'
|
|
else:
|
|
program_files = os.environ['ProgramFiles(x86)'] if 'ProgramFiles(x86)' in os.environ else os.environ['ProgramFiles']
|
|
vs2019_exists = len(vswhere(16)) > 0
|
|
vs2017_exists = len(vswhere(15)) > 0
|
|
mingw_exists = which('mingw32-make') is not None and which('g++') is not None
|
|
if vs2019_exists:
|
|
CMAKE_GENERATOR = 'Visual Studio 16'
|
|
elif vs2017_exists:
|
|
# VS2017 has an LLVM build issue, see
|
|
# https://github.com/kripken/emscripten-fastcomp/issues/185
|
|
CMAKE_GENERATOR = 'Visual Studio 15'
|
|
elif mingw_exists:
|
|
CMAKE_GENERATOR = 'MinGW Makefiles'
|
|
else:
|
|
# No detected generator
|
|
CMAKE_GENERATOR = ''
|
|
|
|
|
|
sys.argv = [a for a in sys.argv if a not in ('--mingw', '--vs2017', '--vs2019')]
|
|
|
|
|
|
# Computes a suitable path prefix to use when building with a given generator.
|
|
def cmake_generator_prefix():
|
|
if CMAKE_GENERATOR == 'Visual Studio 16':
|
|
return '_vs2019'
|
|
if CMAKE_GENERATOR == 'Visual Studio 15':
|
|
return '_vs2017'
|
|
elif CMAKE_GENERATOR == 'MinGW Makefiles':
|
|
return '_mingw'
|
|
# Unix Makefiles do not specify a path prefix for backwards path compatibility
|
|
return ''
|
|
|
|
|
|
# Removes a directory tree even if it was readonly, and doesn't throw exception
|
|
# on failure.
|
|
def remove_tree(d):
|
|
debug_print('remove_tree(' + str(d) + ')')
|
|
if not os.path.exists(d):
|
|
return
|
|
try:
|
|
def remove_readonly_and_try_again(func, path, exc_info):
|
|
if not (os.stat(path).st_mode & stat.S_IWRITE):
|
|
os.chmod(path, stat.S_IWRITE)
|
|
func(path)
|
|
else:
|
|
raise
|
|
shutil.rmtree(d, onerror=remove_readonly_and_try_again)
|
|
except Exception as e:
|
|
debug_print('remove_tree threw an exception, ignoring: ' + str(e))
|
|
|
|
|
|
def win_set_environment_variable_direct(key, value, system=True):
|
|
folder = None
|
|
try:
|
|
if system:
|
|
# Read globally from ALL USERS section.
|
|
folder = winreg.OpenKeyEx(winreg.HKEY_LOCAL_MACHINE, 'SYSTEM\\CurrentControlSet\\Control\\Session Manager\\Environment', 0, winreg.KEY_ALL_ACCESS)
|
|
else:
|
|
# Register locally from CURRENT USER section.
|
|
folder = winreg.OpenKeyEx(winreg.HKEY_CURRENT_USER, 'Environment', 0, winreg.KEY_ALL_ACCESS)
|
|
winreg.SetValueEx(folder, key, 0, winreg.REG_EXPAND_SZ, value)
|
|
debug_print('Set key=' + key + ' with value ' + value + ' in registry.')
|
|
return True
|
|
except Exception as e:
|
|
# 'Access is denied.'
|
|
if e.args[3] == 5:
|
|
exit_with_error('Error! Failed to set the environment variable \'' + key + '\'! Setting environment variables permanently requires administrator access. Please rerun this command with administrative privileges. This can be done for example by holding down the Ctrl and Shift keys while opening a command prompt in start menu.')
|
|
errlog('Failed to write environment variable ' + key + ':')
|
|
errlog(str(e))
|
|
return False
|
|
finally:
|
|
if folder is not None:
|
|
folder.Close()
|
|
|
|
|
|
def win_get_environment_variable(key, system=True, user=True, fallback=True):
|
|
if (not system and not user and fallback):
|
|
# if no --system or --permanent flag is provided use shell's value
|
|
return os.environ[key]
|
|
try:
|
|
folder = None
|
|
try:
|
|
if system:
|
|
# Read globally from ALL USERS section.
|
|
folder = winreg.OpenKey(winreg.HKEY_LOCAL_MACHINE, 'SYSTEM\\CurrentControlSet\\Control\\Session Manager\\Environment')
|
|
else:
|
|
# Register locally from CURRENT USER section.
|
|
folder = winreg.OpenKey(winreg.HKEY_CURRENT_USER, 'Environment')
|
|
value = str(winreg.QueryValueEx(folder, key)[0])
|
|
except Exception:
|
|
# If reading registry fails for some reason - read via os.environ. This has the drawback
|
|
# that expansion items such as %PROGRAMFILES% will have been expanded, so
|
|
# need to be precise not to set these back to system registry, or
|
|
# expansion items would be lost.
|
|
if fallback:
|
|
return os.environ[key]
|
|
return None
|
|
finally:
|
|
if folder is not None:
|
|
folder.Close()
|
|
|
|
except Exception as e:
|
|
# this catch is if both the registry key threw an exception and the key is not in os.environ
|
|
if e.args[0] != 2:
|
|
# 'The system cannot find the file specified.'
|
|
errlog('Failed to read environment variable ' + key + ':')
|
|
errlog(str(e))
|
|
return None
|
|
return value
|
|
|
|
|
|
def win_set_environment_variable(key, value, system, user):
|
|
debug_print('set ' + str(key) + '=' + str(value) + ', in system=' + str(system))
|
|
previous_value = win_get_environment_variable(key, system=system, user=user)
|
|
if previous_value == value:
|
|
debug_print(' no need to set, since same value already exists.')
|
|
# No need to elevate UAC for nothing to set the same value, skip.
|
|
return False
|
|
|
|
if not value:
|
|
try:
|
|
if system:
|
|
cmd = ['REG', 'DELETE', 'SYSTEM\\CurrentControlSet\\Control\\Session Manager\\Environment', '/V', key, '/f']
|
|
else:
|
|
cmd = ['REG', 'DELETE', 'HKCU\\Environment', '/V', key, '/f']
|
|
debug_print(str(cmd))
|
|
value = subprocess.call(cmd, stdout=subprocess.PIPE)
|
|
except Exception:
|
|
return False
|
|
return True
|
|
|
|
try:
|
|
if win_set_environment_variable_direct(key, value, system):
|
|
return True
|
|
# Escape % signs so that we don't expand references to environment variables.
|
|
value = value.replace('%', '^%')
|
|
if len(value) >= 1024:
|
|
exit_with_error('ERROR! The new environment variable ' + key + ' is more than 1024 characters long! A value this long cannot be set via command line: please add the environment variable specified above to system environment manually via Control Panel.')
|
|
cmd = ['SETX', key, value]
|
|
debug_print(str(cmd))
|
|
retcode = subprocess.call(cmd, stdout=subprocess.PIPE)
|
|
if retcode != 0:
|
|
errlog('ERROR! Failed to set environment variable ' + key + '=' + value + '. You may need to set it manually.')
|
|
else:
|
|
return True
|
|
except Exception as e:
|
|
errlog('ERROR! Failed to set environment variable ' + key + '=' + value + ':')
|
|
errlog(str(e))
|
|
errlog('You may need to set it manually.')
|
|
|
|
return False
|
|
|
|
|
|
def win_set_environment_variables(env_vars_to_add, system, user):
|
|
if not env_vars_to_add:
|
|
return
|
|
|
|
changed = False
|
|
|
|
for key, value in env_vars_to_add:
|
|
if win_set_environment_variable(key, value, system, user):
|
|
if not changed:
|
|
changed = True
|
|
print('Setting global environment variables:')
|
|
|
|
print(key + ' = ' + value)
|
|
|
|
if not changed:
|
|
print('Global environment variables up to date')
|
|
return
|
|
|
|
# if changes were made then we need to notify other processes
|
|
try:
|
|
HWND_BROADCAST = ctypes.wintypes.HWND(0xFFFF) # win32con.HWND_BROADCAST == 65535
|
|
WM_SETTINGCHANGE = 0x001A # win32con.WM_SETTINGCHANGE == 26
|
|
SMTO_BLOCK = 0x0001 # win32con.SMTO_BLOCK == 1
|
|
ctypes.windll.user32.SendMessageTimeoutA(
|
|
HWND_BROADCAST, # hWnd: notify everyone
|
|
WM_SETTINGCHANGE, # Msg: registry changed
|
|
0, # wParam: Must be 0 when setting changed is sent by users
|
|
'Environment', # lParam: Specifically environment variables changed
|
|
SMTO_BLOCK, # fuFlags: Wait for message to be sent or timeout
|
|
100) # uTimeout: 100ms
|
|
except Exception as e:
|
|
errlog('SendMessageTimeout failed with error: ' + str(e))
|
|
|
|
|
|
def win_delete_environment_variable(key, system=True, user=True):
|
|
debug_print('win_delete_environment_variable(key=' + key + ', system=' + str(system) + ')')
|
|
return win_set_environment_variable(key, None, system, user)
|
|
|
|
|
|
# Returns the absolute pathname to the given path inside the Emscripten SDK.
|
|
def sdk_path(path):
|
|
if os.path.isabs(path):
|
|
return path
|
|
|
|
return to_unix_path(os.path.join(emsdk_path(), path))
|
|
|
|
|
|
# Modifies the given file in-place to contain '\r\n' line endings.
|
|
def file_to_crlf(filename):
|
|
text = open(filename, 'r').read()
|
|
text = text.replace('\r\n', '\n').replace('\n', '\r\n')
|
|
open(filename, 'wb').write(text)
|
|
|
|
|
|
# Modifies the given file in-place to contain '\n' line endings.
|
|
def file_to_lf(filename):
|
|
text = open(filename, 'r').read()
|
|
text = text.replace('\r\n', '\n')
|
|
open(filename, 'wb').write(text)
|
|
|
|
|
|
# Removes a single file, suppressing exceptions on failure.
|
|
def rmfile(filename):
|
|
debug_print('rmfile(' + filename + ')')
|
|
try:
|
|
os.remove(filename)
|
|
except:
|
|
pass
|
|
|
|
|
|
def fix_lineendings(filename):
|
|
if WINDOWS:
|
|
file_to_crlf(filename)
|
|
else:
|
|
file_to_lf(filename)
|
|
|
|
|
|
# http://stackoverflow.com/questions/600268/mkdir-p-functionality-in-python
|
|
def mkdir_p(path):
|
|
debug_print('mkdir_p(' + path + ')')
|
|
if os.path.exists(path):
|
|
return
|
|
try:
|
|
os.makedirs(path)
|
|
except OSError as exc: # Python >2.5
|
|
if exc.errno == errno.EEXIST and os.path.isdir(path):
|
|
pass
|
|
else:
|
|
raise
|
|
|
|
|
|
def num_files_in_directory(path):
|
|
if not os.path.isdir(path):
|
|
return 0
|
|
return len([name for name in os.listdir(path) if os.path.exists(os.path.join(path, name))])
|
|
|
|
|
|
def run(cmd, cwd=None):
|
|
debug_print('run(cmd=' + str(cmd) + ', cwd=' + str(cwd) + ')')
|
|
process = subprocess.Popen(cmd, cwd=cwd, env=os.environ.copy())
|
|
process.communicate()
|
|
if process.returncode != 0:
|
|
errlog(str(cmd) + ' failed with error code ' + str(process.returncode) + '!')
|
|
return process.returncode
|
|
|
|
|
|
# http://pythonicprose.blogspot.fi/2009/10/python-extract-targz-archive.html
|
|
def untargz(source_filename, dest_dir, unpack_even_if_exists=False):
|
|
debug_print('untargz(source_filename=' + source_filename + ', dest_dir=' + dest_dir + ')')
|
|
if not unpack_even_if_exists and num_files_in_directory(dest_dir) > 0:
|
|
print("File '" + source_filename + "' has already been unpacked, skipping.")
|
|
return True
|
|
print("Unpacking '" + source_filename + "' to '" + dest_dir + "'")
|
|
mkdir_p(dest_dir)
|
|
run(['tar', '-xvf' if VERBOSE else '-xf', sdk_path(source_filename), '--strip', '1'], cwd=dest_dir)
|
|
# tfile = tarfile.open(source_filename, 'r:gz')
|
|
# tfile.extractall(dest_dir)
|
|
return True
|
|
|
|
|
|
# On Windows, it is not possible to reference path names that are longer than
|
|
# ~260 characters, unless the path is referenced via a "\\?\" prefix.
|
|
# See https://msdn.microsoft.com/en-us/library/aa365247.aspx#maxpath and http://stackoverflow.com/questions/3555527/python-win32-filename-length-workaround
|
|
# In that mode, forward slashes cannot be used as delimiters.
|
|
def fix_potentially_long_windows_pathname(pathname):
|
|
if not WINDOWS:
|
|
return pathname
|
|
# Test if emsdk calls fix_potentially_long_windows_pathname() with long relative paths (which is problematic)
|
|
if not os.path.isabs(pathname) and len(pathname) > 200:
|
|
errlog('Warning: Seeing a relative path "' + pathname + '" which is dangerously long for being referenced as a short Windows path name. Refactor emsdk to be able to handle this!')
|
|
if pathname.startswith('\\\\?\\'):
|
|
return pathname
|
|
pathname = os.path.normpath(pathname.replace('/', '\\'))
|
|
if MINGW:
|
|
# MinGW versions of Python return normalized paths with backslashes
|
|
# converted to forward slashes, so we must use forward slashes in our
|
|
# prefix
|
|
return '//?/' + pathname
|
|
return '\\\\?\\' + pathname
|
|
|
|
|
|
# On windows, rename/move will fail if the destination exists, and there is no
|
|
# race-free way to do it. This method removes the destination if it exists, so
|
|
# the move always works
|
|
def move_with_overwrite(src, dest):
|
|
if os.path.exists(dest):
|
|
os.remove(dest)
|
|
os.rename(src, dest)
|
|
|
|
|
|
# http://stackoverflow.com/questions/12886768/simple-way-to-unzip-file-in-python-on-all-oses
|
|
def unzip(source_filename, dest_dir, unpack_even_if_exists=False):
|
|
debug_print('unzip(source_filename=' + source_filename + ', dest_dir=' + dest_dir + ')')
|
|
if not unpack_even_if_exists and num_files_in_directory(dest_dir) > 0:
|
|
print("File '" + source_filename + "' has already been unpacked, skipping.")
|
|
return True
|
|
print("Unpacking '" + source_filename + "' to '" + dest_dir + "'")
|
|
mkdir_p(dest_dir)
|
|
common_subdir = None
|
|
try:
|
|
with zipfile.ZipFile(source_filename) as zf:
|
|
# Implement '--strip 1' behavior to unzipping by testing if all the files
|
|
# in the zip reside in a common subdirectory, and if so, we move the
|
|
# output tree at the end of uncompression step.
|
|
for member in zf.infolist():
|
|
words = member.filename.split('/')
|
|
if len(words) > 1: # If there is a directory component?
|
|
if common_subdir is None:
|
|
common_subdir = words[0]
|
|
elif common_subdir != words[0]:
|
|
common_subdir = None
|
|
break
|
|
else:
|
|
common_subdir = None
|
|
break
|
|
|
|
unzip_to_dir = dest_dir
|
|
if common_subdir:
|
|
unzip_to_dir = os.path.join(os.path.dirname(dest_dir), 'unzip_temp')
|
|
|
|
# Now do the actual decompress.
|
|
for member in zf.infolist():
|
|
zf.extract(member, fix_potentially_long_windows_pathname(unzip_to_dir))
|
|
dst_filename = os.path.join(unzip_to_dir, member.filename)
|
|
|
|
# See: https://stackoverflow.com/questions/42326428/zipfile-in-python-file-permission
|
|
unix_attributes = member.external_attr >> 16
|
|
if unix_attributes:
|
|
os.chmod(dst_filename, unix_attributes)
|
|
|
|
# Move the extracted file to its final location without the base
|
|
# directory name, if we are stripping that away.
|
|
if common_subdir:
|
|
if not member.filename.startswith(common_subdir):
|
|
raise Exception('Unexpected filename "' + member.filename + '"!')
|
|
stripped_filename = '.' + member.filename[len(common_subdir):]
|
|
final_dst_filename = os.path.join(dest_dir, stripped_filename)
|
|
# Check if a directory
|
|
if stripped_filename.endswith('/'):
|
|
d = fix_potentially_long_windows_pathname(final_dst_filename)
|
|
if not os.path.isdir(d):
|
|
os.mkdir(d)
|
|
else:
|
|
parent_dir = os.path.dirname(fix_potentially_long_windows_pathname(final_dst_filename))
|
|
if parent_dir and not os.path.exists(parent_dir):
|
|
os.makedirs(parent_dir)
|
|
move_with_overwrite(fix_potentially_long_windows_pathname(dst_filename), fix_potentially_long_windows_pathname(final_dst_filename))
|
|
|
|
if common_subdir:
|
|
remove_tree(unzip_to_dir)
|
|
except zipfile.BadZipfile as e:
|
|
errlog("Unzipping file '" + source_filename + "' failed due to reason: " + str(e) + "! Removing the corrupted zip file.")
|
|
rmfile(source_filename)
|
|
return False
|
|
except Exception as e:
|
|
errlog("Unzipping file '" + source_filename + "' failed due to reason: " + str(e))
|
|
return False
|
|
|
|
return True
|
|
|
|
|
|
# This function interprets whether the given string looks like a path to a
|
|
# directory instead of a file, without looking at the actual filesystem.
|
|
# 'a/b/c' points to directory, so does 'a/b/c/', but 'a/b/c.x' is parsed as a
|
|
# filename
|
|
def path_points_to_directory(path):
|
|
if path == '.':
|
|
return True
|
|
last_slash = max(path.rfind('/'), path.rfind('\\'))
|
|
last_dot = path.rfind('.')
|
|
no_suffix = last_dot < last_slash or last_dot == -1
|
|
if no_suffix:
|
|
return True
|
|
suffix = path[last_dot:]
|
|
# Very simple logic for the only file suffixes used by emsdk downloader. Other
|
|
# suffixes, like 'clang-3.2' are treated as dirs.
|
|
if suffix in ('.exe', '.zip', '.txt'):
|
|
return False
|
|
else:
|
|
return True
|
|
|
|
|
|
def get_content_length(download):
|
|
try:
|
|
meta = download.info()
|
|
if hasattr(meta, "getheaders") and hasattr(meta.getheaders, "Content-Length"):
|
|
return int(meta.getheaders("Content-Length")[0])
|
|
elif hasattr(download, "getheader") and download.getheader('Content-Length'):
|
|
return int(download.getheader('Content-Length'))
|
|
elif hasattr(meta, "getheader") and meta.getheader('Content-Length'):
|
|
return int(meta.getheader('Content-Length'))
|
|
except Exception:
|
|
pass
|
|
|
|
return 0
|
|
|
|
|
|
def get_download_target(url, dstpath, filename_prefix=''):
|
|
file_name = filename_prefix + url.split('/')[-1]
|
|
if path_points_to_directory(dstpath):
|
|
file_name = os.path.join(dstpath, file_name)
|
|
else:
|
|
file_name = dstpath
|
|
|
|
# Treat all relative destination paths as relative to the SDK root directory,
|
|
# not the current working directory.
|
|
file_name = sdk_path(file_name)
|
|
|
|
return file_name
|
|
|
|
|
|
# On success, returns the filename on the disk pointing to the destination file that was produced
|
|
# On failure, returns None.
|
|
def download_file(url, dstpath, download_even_if_exists=False, filename_prefix=''):
|
|
debug_print('download_file(url=' + url + ', dstpath=' + dstpath + ')')
|
|
file_name = get_download_target(url, dstpath, filename_prefix)
|
|
|
|
if os.path.exists(file_name) and not download_even_if_exists:
|
|
print("File '" + file_name + "' already downloaded, skipping.")
|
|
return file_name
|
|
try:
|
|
u = urlopen(url)
|
|
mkdir_p(os.path.dirname(file_name))
|
|
with open(file_name, 'wb') as f:
|
|
file_size = get_content_length(u)
|
|
if file_size > 0:
|
|
print("Downloading: %s from %s, %s Bytes" % (file_name, url, file_size))
|
|
else:
|
|
print("Downloading: %s from %s" % (file_name, url))
|
|
|
|
file_size_dl = 0
|
|
# Draw a progress bar 80 chars wide (in non-TTY mode)
|
|
progress_max = 80 - 4
|
|
progress_shown = 0
|
|
block_sz = 256 * 1024
|
|
if not TTY_OUTPUT:
|
|
print(' [', end='')
|
|
while True:
|
|
buffer = u.read(block_sz)
|
|
if not buffer:
|
|
break
|
|
|
|
file_size_dl += len(buffer)
|
|
f.write(buffer)
|
|
if file_size:
|
|
percent = file_size_dl * 100.0 / file_size
|
|
if TTY_OUTPUT:
|
|
status = r" %10d [%3.02f%%]" % (file_size_dl, percent)
|
|
print(status, end='\r')
|
|
else:
|
|
while progress_shown < progress_max * percent / 100:
|
|
print('-', end='')
|
|
sys.stdout.flush()
|
|
progress_shown += 1
|
|
if not TTY_OUTPUT:
|
|
print(']')
|
|
sys.stdout.flush()
|
|
except Exception as e:
|
|
errlog("Error: Downloading URL '" + url + "': " + str(e))
|
|
if "SSL: CERTIFICATE_VERIFY_FAILED" in str(e) or "urlopen error unknown url type: https" in str(e):
|
|
errlog("Warning: Possibly SSL/TLS issue. Update or install Python SSL root certificates (2048-bit or greater) supplied in Python folder or https://pypi.org/project/certifi/ and try again.")
|
|
rmfile(file_name)
|
|
return None
|
|
except KeyboardInterrupt:
|
|
rmfile(file_name)
|
|
exit_with_error("Aborted by User, exiting")
|
|
return file_name
|
|
|
|
|
|
def run_get_output(cmd, cwd=None):
|
|
debug_print('run_get_output(cmd=' + str(cmd) + ', cwd=' + str(cwd) + ')')
|
|
process = subprocess.Popen(cmd, cwd=cwd, stdout=subprocess.PIPE, stderr=subprocess.PIPE, stdin=subprocess.PIPE, env=os.environ.copy(), universal_newlines=True)
|
|
stdout, stderr = process.communicate()
|
|
return (process.returncode, stdout, stderr)
|
|
|
|
|
|
# must_succeed: If false, the search is performed silently without printing out
|
|
# errors if not found. Empty string is returned if git is not found.
|
|
# If true, the search is required to succeed, and the execution
|
|
# will terminate with sys.exit(1) if not found.
|
|
def GIT(must_succeed=True):
|
|
# The order in the following is important, and specifies the preferred order
|
|
# of using the git tools. Primarily use git from emsdk if installed. If not,
|
|
# use system git.
|
|
gits = ['git/1.9.4/bin/git.exe', which('git')]
|
|
for git in gits:
|
|
try:
|
|
ret, stdout, stderr = run_get_output([git, '--version'])
|
|
if ret == 0:
|
|
return git
|
|
except:
|
|
pass
|
|
if must_succeed:
|
|
if WINDOWS:
|
|
msg = "ERROR: git executable was not found. Please install it by typing 'emsdk install git-1.9.4', or alternatively by installing it manually from http://git-scm.com/downloads . If you install git manually, remember to add it to PATH"
|
|
elif MACOS:
|
|
msg = "ERROR: git executable was not found. Please install git for this operation! This can be done from http://git-scm.com/ , or by installing XCode and then the XCode Command Line Tools (see http://stackoverflow.com/questions/9329243/xcode-4-4-command-line-tools )"
|
|
elif LINUX:
|
|
msg = "ERROR: git executable was not found. Please install git for this operation! This can be probably be done using your package manager, see http://git-scm.com/book/en/Getting-Started-Installing-Git"
|
|
else:
|
|
msg = "ERROR: git executable was not found. Please install git for this operation!"
|
|
exit_with_error(msg)
|
|
# Not found
|
|
return ''
|
|
|
|
|
|
def git_repo_version(repo_path):
|
|
returncode, stdout, stderr = run_get_output([GIT(), 'log', '-n', '1', '--pretty="%aD %H"'], cwd=repo_path)
|
|
if returncode == 0:
|
|
return stdout.strip()
|
|
else:
|
|
return ""
|
|
|
|
|
|
def git_recent_commits(repo_path, n=20):
|
|
returncode, stdout, stderr = run_get_output([GIT(), 'log', '-n', str(n), '--pretty="%H"'], cwd=repo_path)
|
|
if returncode == 0:
|
|
return stdout.strip().replace('\r', '').replace('"', '').split('\n')
|
|
else:
|
|
return []
|
|
|
|
|
|
def git_clone(url, dstpath):
|
|
debug_print('git_clone(url=' + url + ', dstpath=' + dstpath + ')')
|
|
if os.path.isdir(os.path.join(dstpath, '.git')):
|
|
debug_print("Repository '" + url + "' already cloned to directory '" + dstpath + "', skipping.")
|
|
return True
|
|
mkdir_p(dstpath)
|
|
git_clone_args = []
|
|
if GIT_CLONE_SHALLOW:
|
|
git_clone_args += ['--depth', '1']
|
|
return run([GIT(), 'clone'] + git_clone_args + [url, dstpath]) == 0
|
|
|
|
|
|
def git_checkout_and_pull(repo_path, branch):
|
|
debug_print('git_checkout_and_pull(repo_path=' + repo_path + ', branch=' + branch + ')')
|
|
ret = run([GIT(), 'fetch', '--quiet', 'origin'], repo_path)
|
|
if ret != 0:
|
|
return False
|
|
try:
|
|
print("Fetching latest changes to the branch '" + branch + "' for '" + repo_path + "'...")
|
|
ret = run([GIT(), 'fetch', '--quiet', 'origin'], repo_path)
|
|
if ret != 0:
|
|
return False
|
|
# run([GIT, 'checkout', '-b', branch, '--track', 'origin/'+branch], repo_path)
|
|
# this line assumes that the user has not gone and manually messed with the
|
|
# repo and added new remotes to ambiguate the checkout.
|
|
ret = run([GIT(), 'checkout', '--quiet', branch], repo_path)
|
|
if ret != 0:
|
|
return False
|
|
# this line assumes that the user has not gone and made local changes to the repo
|
|
ret = run([GIT(), 'merge', '--ff-only', 'origin/' + branch], repo_path)
|
|
if ret != 0:
|
|
return False
|
|
except:
|
|
errlog('git operation failed!')
|
|
return False
|
|
print("Successfully updated and checked out branch '" + branch + "' on repository '" + repo_path + "'")
|
|
print("Current repository version: " + git_repo_version(repo_path))
|
|
return True
|
|
|
|
|
|
def git_clone_checkout_and_pull(url, dstpath, branch):
|
|
debug_print('git_clone_checkout_and_pull(url=' + url + ', dstpath=' + dstpath + ', branch=' + branch + ')')
|
|
success = git_clone(url, dstpath)
|
|
if not success:
|
|
return False
|
|
success = git_checkout_and_pull(dstpath, branch)
|
|
return success
|
|
|
|
|
|
# Each tool can have its own build type, or it can be overridden on the command
|
|
# line.
|
|
def decide_cmake_build_type(tool):
|
|
if CMAKE_BUILD_TYPE_OVERRIDE:
|
|
return CMAKE_BUILD_TYPE_OVERRIDE
|
|
else:
|
|
return tool.cmake_build_type
|
|
|
|
|
|
# The root directory of the build.
|
|
def llvm_build_dir(tool):
|
|
generator_suffix = ''
|
|
if CMAKE_GENERATOR == 'Visual Studio 15':
|
|
generator_suffix = '_vs2017'
|
|
elif CMAKE_GENERATOR == 'Visual Studio 16':
|
|
generator_suffix = '_vs2019'
|
|
elif CMAKE_GENERATOR == 'MinGW Makefiles':
|
|
generator_suffix = '_mingw'
|
|
|
|
bitness_suffix = '_32' if tool.bitness == 32 else '_64'
|
|
|
|
if hasattr(tool, 'git_branch'):
|
|
build_dir = 'build_' + tool.git_branch.replace(os.sep, '-') + generator_suffix + bitness_suffix
|
|
else:
|
|
build_dir = 'build_' + tool.version + generator_suffix + bitness_suffix
|
|
return build_dir
|
|
|
|
|
|
def exe_suffix(filename):
|
|
if WINDOWS and not filename.endswith('.exe'):
|
|
filename += '.exe'
|
|
return filename
|
|
|
|
|
|
# The directory where the binaries are produced. (relative to the installation
|
|
# root directory of the tool)
|
|
def fastcomp_build_bin_dir(tool):
|
|
build_dir = llvm_build_dir(tool)
|
|
if WINDOWS and 'Visual Studio' in CMAKE_GENERATOR:
|
|
old_llvm_bin_dir = os.path.join(build_dir, 'bin', decide_cmake_build_type(tool))
|
|
|
|
new_llvm_bin_dir = None
|
|
default_cmake_build_type = decide_cmake_build_type(tool)
|
|
cmake_build_types = [default_cmake_build_type, 'Release', 'RelWithDebInfo', 'MinSizeRel', 'Debug']
|
|
for build_type in cmake_build_types:
|
|
d = os.path.join(build_dir, build_type, 'bin')
|
|
if os.path.isfile(os.path.join(tool.installation_path(), d, exe_suffix('clang'))):
|
|
new_llvm_bin_dir = d
|
|
break
|
|
|
|
if new_llvm_bin_dir and os.path.exists(os.path.join(tool.installation_path(), new_llvm_bin_dir)):
|
|
return new_llvm_bin_dir
|
|
elif os.path.exists(os.path.join(tool.installation_path(), old_llvm_bin_dir)):
|
|
return old_llvm_bin_dir
|
|
return os.path.join(build_dir, default_cmake_build_type, 'bin')
|
|
else:
|
|
return os.path.join(build_dir, 'bin')
|
|
|
|
|
|
def build_env(generator):
|
|
build_env = os.environ.copy()
|
|
|
|
# To work around a build issue with older Mac OS X builds, add -stdlib=libc++ to all builds.
|
|
# See https://groups.google.com/forum/#!topic/emscripten-discuss/5Or6QIzkqf0
|
|
if MACOS:
|
|
build_env['CXXFLAGS'] = ((build_env['CXXFLAGS'] + ' ') if hasattr(build_env, 'CXXFLAGS') else '') + '-stdlib=libc++'
|
|
elif 'Visual Studio 15' in generator or 'Visual Studio 16' in generator:
|
|
if 'Visual Studio 16' in generator:
|
|
path = vswhere(16)
|
|
else:
|
|
path = vswhere(15)
|
|
build_env['VCTargetsPath'] = os.path.join(path, 'Common7\\IDE\\VC\\VCTargets')
|
|
|
|
# CMake and VS2017 cl.exe needs to have mspdb140.dll et al. in its PATH.
|
|
vc_bin_paths = [vs_filewhere(path, 'amd64', 'cl.exe'),
|
|
vs_filewhere(path, 'x86', 'cl.exe')]
|
|
for path in vc_bin_paths:
|
|
if os.path.isdir(path):
|
|
build_env['PATH'] = build_env['PATH'] + ';' + path
|
|
|
|
return build_env
|
|
|
|
|
|
def get_generator_for_sln_file(sln_file):
|
|
contents = open(sln_file, 'r').read()
|
|
if '# Visual Studio 16' in contents or '# Visual Studio Version 16' in contents: # VS2019
|
|
return 'Visual Studio 16'
|
|
if '# Visual Studio 15' in contents: # VS2017
|
|
return 'Visual Studio 15'
|
|
raise Exception('Unknown generator used to build solution file ' + sln_file)
|
|
|
|
|
|
def find_msbuild(sln_file):
|
|
# The following logic attempts to find a Visual Studio version specific
|
|
# MSBuild.exe from a list of known locations.
|
|
generator = get_generator_for_sln_file(sln_file)
|
|
debug_print('find_msbuild looking for generator ' + str(generator))
|
|
if generator == 'Visual Studio 16': # VS2019
|
|
path = vswhere(16)
|
|
search_paths = [os.path.join(path, 'MSBuild/Current/Bin'),
|
|
os.path.join(path, 'MSBuild/15.0/Bin/amd64'),
|
|
os.path.join(path, 'MSBuild/15.0/Bin')]
|
|
elif generator == 'Visual Studio 15': # VS2017
|
|
path = vswhere(15)
|
|
search_paths = [os.path.join(path, 'MSBuild/15.0/Bin/amd64'),
|
|
os.path.join(path, 'MSBuild/15.0/Bin')]
|
|
else:
|
|
raise Exception('Unknown generator!')
|
|
|
|
for path in search_paths:
|
|
p = os.path.join(path, 'MSBuild.exe')
|
|
debug_print('Searching for MSBuild.exe: ' + p)
|
|
if os.path.isfile(p):
|
|
return p
|
|
debug_print('MSBuild.exe in PATH? ' + str(which('MSBuild.exe')))
|
|
# Last fallback, try any MSBuild from PATH (might not be compatible, but best effort)
|
|
return which('MSBuild.exe')
|
|
|
|
|
|
def make_build(build_root, build_type, build_target_platform='x64'):
|
|
debug_print('make_build(build_root=' + build_root + ', build_type=' + build_type + ', build_target_platform=' + build_target_platform + ')')
|
|
if CPU_CORES > 1:
|
|
print('Performing a parallel build with ' + str(CPU_CORES) + ' cores.')
|
|
else:
|
|
print('Performing a singlethreaded build.')
|
|
|
|
generator_to_use = CMAKE_GENERATOR
|
|
|
|
if WINDOWS:
|
|
if 'Visual Studio' in CMAKE_GENERATOR:
|
|
solution_name = str(subprocess.check_output(['dir', '/b', '*.sln'], shell=True, cwd=build_root).decode('utf-8').strip())
|
|
generator_to_use = get_generator_for_sln_file(os.path.join(build_root, solution_name))
|
|
# Disabled for now: Don't pass /maxcpucount argument to msbuild, since it
|
|
# looks like when building, msbuild already automatically spawns the full
|
|
# amount of logical cores the system has, and passing the number of
|
|
# logical cores here has been observed to give a quadratic N*N explosion
|
|
# on the number of spawned processes (e.g. on a Core i7 5960X with 16
|
|
# logical cores, it would spawn 16*16=256 cl.exe processes, which would
|
|
# start crashing when running out of system memory)
|
|
# make = [find_msbuild(os.path.join(build_root, solution_name)), '/maxcpucount:' + str(CPU_CORES), '/t:Build', '/p:Configuration=' + build_type, '/nologo', '/verbosity:minimal', solution_name]
|
|
make = [find_msbuild(os.path.join(build_root, solution_name)), '/t:Build', '/p:Configuration=' + build_type, '/p:Platform=' + build_target_platform, '/nologo', '/verbosity:minimal', solution_name]
|
|
else:
|
|
make = ['mingw32-make', '-j' + str(CPU_CORES)]
|
|
else:
|
|
make = ['cmake', '--build', '.', '--', '-j' + str(CPU_CORES)]
|
|
|
|
# Build
|
|
try:
|
|
print('Running build: ' + str(make))
|
|
ret = subprocess.check_call(make, cwd=build_root, env=build_env(generator_to_use))
|
|
if ret != 0:
|
|
errlog('Build failed with exit code ' + ret + '!')
|
|
errlog('Working directory: ' + build_root)
|
|
return False
|
|
except Exception as e:
|
|
errlog('Build failed due to exception!')
|
|
errlog('Working directory: ' + build_root)
|
|
errlog(str(e))
|
|
return False
|
|
|
|
return True
|
|
|
|
|
|
def cmake_configure(generator, build_root, src_root, build_type, extra_cmake_args=[]):
|
|
debug_print('cmake_configure(generator=' + str(generator) + ', build_root=' + str(build_root) + ', src_root=' + str(src_root) + ', build_type=' + str(build_type) + ', extra_cmake_args=' + str(extra_cmake_args) + ')')
|
|
# Configure
|
|
if not os.path.isdir(build_root):
|
|
# Create build output directory if it doesn't yet exist.
|
|
os.mkdir(build_root)
|
|
try:
|
|
if generator:
|
|
generator = ['-G', generator]
|
|
else:
|
|
generator = []
|
|
cmdline = ['cmake'] + generator + ['-DCMAKE_BUILD_TYPE=' + build_type, '-DPYTHON_EXECUTABLE=' + sys.executable] + extra_cmake_args + [src_root]
|
|
print('Running CMake: ' + str(cmdline))
|
|
|
|
def quote_parens(x):
|
|
if ' ' in x:
|
|
return '"' + x.replace('"', '\\"') + '"'
|
|
else:
|
|
return x
|
|
|
|
# Create a file 'recmake.bat/sh' in the build root that user can call to
|
|
# manually recmake the build tree with the previous build params
|
|
open(os.path.join(build_root, 'recmake.' + ('bat' if WINDOWS else 'sh')), 'w').write(' '.join(map(quote_parens, cmdline)))
|
|
ret = subprocess.check_call(cmdline, cwd=build_root, env=build_env(CMAKE_GENERATOR))
|
|
if ret != 0:
|
|
errlog('CMake invocation failed with exit code ' + ret + '!')
|
|
errlog('Working directory: ' + build_root)
|
|
return False
|
|
except OSError as e:
|
|
if e.errno == errno.ENOENT:
|
|
errlog(str(e))
|
|
errlog('Could not run CMake, perhaps it has not been installed?')
|
|
if WINDOWS:
|
|
errlog('Installing this package requires CMake. Get it from http://www.cmake.org/')
|
|
elif LINUX:
|
|
errlog('Installing this package requires CMake. Get it via your system package manager (e.g. sudo apt-get install cmake), or from http://www.cmake.org/')
|
|
elif MACOS:
|
|
errlog('Installing this package requires CMake. Get it via a macOS package manager (Homebrew: "brew install cmake", or MacPorts: "sudo port install cmake"), or from http://www.cmake.org/')
|
|
return False
|
|
raise
|
|
except Exception as e:
|
|
errlog('CMake invocation failed due to exception!')
|
|
errlog('Working directory: ' + build_root)
|
|
errlog(str(e))
|
|
return False
|
|
|
|
return True
|
|
|
|
|
|
def xcode_sdk_version():
|
|
try:
|
|
output = subprocess.check_output(['xcrun', '--show-sdk-version'])
|
|
if sys.version_info >= (3,):
|
|
output = output.decode('utf8')
|
|
return output.strip().split('.')
|
|
except:
|
|
return subprocess.checkplatform.mac_ver()[0].split('.')
|
|
|
|
|
|
def build_fastcomp(tool):
|
|
debug_print('build_fastcomp(' + str(tool) + ')')
|
|
fastcomp_root = tool.installation_path()
|
|
fastcomp_src_root = os.path.join(fastcomp_root, 'src')
|
|
# Does this tool want to be git cloned from github?
|
|
if hasattr(tool, 'git_branch'):
|
|
success = git_clone_checkout_and_pull(tool.download_url(), fastcomp_src_root, tool.git_branch)
|
|
if not success:
|
|
return False
|
|
if hasattr(tool, 'clang_url'):
|
|
clang_root = os.path.join(fastcomp_src_root, 'tools/clang')
|
|
success = git_clone_checkout_and_pull(tool.clang_url, clang_root, tool.git_branch)
|
|
if not success:
|
|
return False
|
|
if hasattr(tool, 'lld_url'):
|
|
lld_root = os.path.join(fastcomp_src_root, 'tools/lld')
|
|
success = git_clone_checkout_and_pull(tool.lld_url, lld_root, tool.git_branch)
|
|
if not success:
|
|
return False
|
|
else:
|
|
# Not a git cloned tool, so instead download from git tagged releases
|
|
success = download_and_unzip(tool.download_url(), fastcomp_src_root, filename_prefix='llvm-e')
|
|
if not success:
|
|
return False
|
|
success = download_and_unzip(tool.windows_clang_url if WINDOWS else tool.unix_clang_url, os.path.join(fastcomp_src_root, 'tools/clang'), filename_prefix='clang-e')
|
|
if not success:
|
|
return False
|
|
|
|
args = []
|
|
|
|
cmake_generator = CMAKE_GENERATOR
|
|
if 'Visual Studio 16' in CMAKE_GENERATOR: # VS2019
|
|
# With Visual Studio 16 2019, CMake changed the way they specify target arch.
|
|
# Instead of appending it into the CMake generator line, it is specified
|
|
# with a -A arch parameter.
|
|
args += ['-A', 'x64' if tool.bitness == 64 else 'x86']
|
|
elif 'Visual Studio' in CMAKE_GENERATOR and tool.bitness == 64:
|
|
cmake_generator += ' Win64'
|
|
|
|
build_dir = llvm_build_dir(tool)
|
|
build_root = os.path.join(fastcomp_root, build_dir)
|
|
|
|
build_type = decide_cmake_build_type(tool)
|
|
|
|
# Configure
|
|
tests_arg = 'ON' if BUILD_FOR_TESTING else 'OFF'
|
|
|
|
enable_assertions = ENABLE_LLVM_ASSERTIONS.lower() == 'on' or (ENABLE_LLVM_ASSERTIONS == 'auto' and build_type.lower() != 'release' and build_type.lower() != 'minsizerel')
|
|
|
|
only_supports_wasm = hasattr(tool, 'only_supports_wasm')
|
|
if ARCH == 'x86' or ARCH == 'x86_64':
|
|
targets_to_build = 'X86'
|
|
elif ARCH == 'arm':
|
|
targets_to_build = 'ARM'
|
|
elif ARCH == 'aarch64':
|
|
targets_to_build = 'AArch64'
|
|
else:
|
|
# May have problems with emconfigure
|
|
targets_to_build = ''
|
|
if not only_supports_wasm:
|
|
if targets_to_build != '':
|
|
targets_to_build += ';'
|
|
targets_to_build += 'JSBackend'
|
|
args += ['-DLLVM_TARGETS_TO_BUILD=' + targets_to_build, '-DLLVM_INCLUDE_EXAMPLES=OFF', '-DCLANG_INCLUDE_EXAMPLES=OFF', '-DLLVM_INCLUDE_TESTS=' + tests_arg, '-DCLANG_INCLUDE_TESTS=' + tests_arg, '-DLLVM_ENABLE_ASSERTIONS=' + ('ON' if enable_assertions else 'OFF')]
|
|
if os.environ.get('LLVM_CMAKE_ARGS'):
|
|
extra_args = os.environ['LLVM_CMAKE_ARGS'].split(',')
|
|
print('Passing the following extra arguments to LLVM CMake configuration: ' + str(extra_args))
|
|
args += extra_args
|
|
|
|
# MacOS < 10.13 workaround for LLVM build bug https://github.com/kripken/emscripten/issues/5418:
|
|
# specify HAVE_FUTIMENS=0 in the build if building with target SDK that is older than 10.13.
|
|
if MACOS and (not os.environ.get('LLVM_CMAKE_ARGS') or 'HAVE_FUTIMENS' not in os.environ.get('LLVM_CMAKE_ARGS')) and xcode_sdk_version() < ['10', '13']:
|
|
print('Passing -DHAVE_FUTIMENS=0 to LLVM CMake configure to workaround https://github.com/kripken/emscripten/issues/5418. Please update to macOS 10.13 or newer')
|
|
args += ['-DHAVE_FUTIMENS=0']
|
|
|
|
success = cmake_configure(cmake_generator, build_root, fastcomp_src_root, build_type, args)
|
|
if not success:
|
|
return False
|
|
|
|
# Make
|
|
success = make_build(build_root, build_type, 'x64' if tool.bitness == 64 else 'Win32')
|
|
return success
|
|
|
|
|
|
# LLVM git source tree migrated to a single repository instead of multiple
|
|
# ones, build_llvm() builds via that repository structure
|
|
def build_llvm(tool):
|
|
debug_print('build_llvm(' + str(tool) + ')')
|
|
llvm_root = tool.installation_path()
|
|
llvm_src_root = os.path.join(llvm_root, 'src')
|
|
success = git_clone_checkout_and_pull(tool.download_url(), llvm_src_root, tool.git_branch)
|
|
if not success:
|
|
return False
|
|
|
|
build_dir = llvm_build_dir(tool)
|
|
build_root = os.path.join(llvm_root, build_dir)
|
|
|
|
build_type = decide_cmake_build_type(tool)
|
|
|
|
# Configure
|
|
tests_arg = 'ON' if BUILD_FOR_TESTING else 'OFF'
|
|
|
|
enable_assertions = ENABLE_LLVM_ASSERTIONS.lower() == 'on' or (ENABLE_LLVM_ASSERTIONS == 'auto' and build_type.lower() != 'release' and build_type.lower() != 'minsizerel')
|
|
|
|
if ARCH == 'x86' or ARCH == 'x86_64':
|
|
targets_to_build = 'WebAssembly;X86'
|
|
elif ARCH == 'arm':
|
|
targets_to_build = 'WebAssembly;ARM'
|
|
elif ARCH == 'aarch64':
|
|
targets_to_build = 'WebAssembly;AArch64'
|
|
else:
|
|
targets_to_build = 'WebAssembly'
|
|
args = ['-DLLVM_TARGETS_TO_BUILD=' + targets_to_build,
|
|
'-DLLVM_INCLUDE_EXAMPLES=OFF',
|
|
'-DCLANG_INCLUDE_EXAMPLES=OFF',
|
|
'-DLLVM_INCLUDE_TESTS=' + tests_arg,
|
|
'-DCLANG_INCLUDE_TESTS=' + tests_arg,
|
|
'-DLLVM_ENABLE_ASSERTIONS=' + ('ON' if enable_assertions else 'OFF')]
|
|
# LLVM build system bug: looks like everything needs to be passed to LLVM_ENABLE_PROJECTS twice. (or every second field is ignored?)
|
|
|
|
# LLVM build system bug #2: compiler-rt does not build on Windows. It insists on performing a CMake install step that writes to C:\Program Files. Attempting
|
|
# to reroute that to build_root directory then fails on an error
|
|
# file INSTALL cannot find
|
|
# "C:/code/emsdk/llvm/git/build_master_vs2017_64/$(Configuration)/lib/clang/10.0.0/lib/windows/clang_rt.ubsan_standalone-x86_64.lib".
|
|
# (there instead of $(Configuration), one would need ${CMAKE_BUILD_TYPE} ?)
|
|
# It looks like compiler-rt is not compatible to build on Windows?
|
|
args += ['-DLLVM_ENABLE_PROJECTS="clang;clang;lld;lld"']
|
|
cmake_generator = CMAKE_GENERATOR
|
|
if 'Visual Studio 16' in CMAKE_GENERATOR: # VS2019
|
|
# With Visual Studio 16 2019, CMake changed the way they specify target arch.
|
|
# Instead of appending it into the CMake generator line, it is specified
|
|
# with a -A arch parameter.
|
|
args += ['-A', 'x64' if tool.bitness == 64 else 'x86']
|
|
args += ['-Thost=x64']
|
|
elif 'Visual Studio' in CMAKE_GENERATOR and tool.bitness == 64:
|
|
cmake_generator += ' Win64'
|
|
args += ['-Thost=x64']
|
|
|
|
if os.environ.get('LLVM_CMAKE_ARGS'):
|
|
extra_args = os.environ['LLVM_CMAKE_ARGS'].split(',')
|
|
print('Passing the following extra arguments to LLVM CMake configuration: ' + str(extra_args))
|
|
args += extra_args
|
|
|
|
cmakelists_dir = os.path.join(llvm_src_root, 'llvm')
|
|
success = cmake_configure(cmake_generator, build_root, cmakelists_dir, build_type, args)
|
|
if not success:
|
|
return False
|
|
|
|
# Make
|
|
success = make_build(build_root, build_type, 'x64' if tool.bitness == 64 else 'Win32')
|
|
return success
|
|
|
|
|
|
# Emscripten asm.js optimizer build scripts:
|
|
def optimizer_build_root(tool):
|
|
build_root = tool.installation_path().strip()
|
|
if build_root.endswith('/') or build_root.endswith('\\'):
|
|
build_root = build_root[:-1]
|
|
generator_prefix = cmake_generator_prefix()
|
|
build_root = build_root + generator_prefix + '_' + str(tool.bitness) + 'bit_optimizer'
|
|
return build_root
|
|
|
|
|
|
def uninstall_optimizer(tool):
|
|
debug_print('uninstall_optimizer(' + str(tool) + ')')
|
|
build_root = optimizer_build_root(tool)
|
|
print("Deleting path '" + build_root + "'")
|
|
remove_tree(build_root)
|
|
|
|
|
|
def is_optimizer_installed(tool):
|
|
build_root = optimizer_build_root(tool)
|
|
return os.path.exists(build_root)
|
|
|
|
|
|
# Finds the newest installed version of a given tool
|
|
def find_latest_installed_tool(name):
|
|
for t in reversed(tools):
|
|
if t.id == name and t.is_installed():
|
|
return t
|
|
|
|
|
|
# npm install in Emscripten root directory
|
|
def emscripten_npm_install(tool, directory):
|
|
node_tool = find_latest_installed_tool('node')
|
|
if not node_tool:
|
|
npm_fallback = which('npm')
|
|
if not npm_fallback:
|
|
errlog('Failed to find npm command!')
|
|
errlog('Running "npm ci" in installed Emscripten root directory ' + tool.installation_path() + ' is required!')
|
|
errlog('Please install node.js first!')
|
|
return False
|
|
node_path = os.path.dirname(npm_fallback)
|
|
else:
|
|
node_path = os.path.join(node_tool.installation_path(), 'bin')
|
|
|
|
npm = os.path.join(node_path, 'npm' + ('.cmd' if WINDOWS else ''))
|
|
env = os.environ.copy()
|
|
env["PATH"] = node_path + os.pathsep + env["PATH"]
|
|
print('Running post-install step: npm ci ...')
|
|
try:
|
|
subprocess.check_output(
|
|
[npm, 'ci', '--production', '--no-optional'],
|
|
cwd=directory, stderr=subprocess.STDOUT, env=env,
|
|
universal_newlines=True)
|
|
except subprocess.CalledProcessError as e:
|
|
errlog('Error running %s:\n%s' % (e.cmd, e.output))
|
|
return False
|
|
|
|
# Manually install the appropriate native Closure Compiler package
|
|
# This is currently needed because npm ci will install the packages
|
|
# for Closure for all platforms, adding 180MB to the download size
|
|
# There are two problems here:
|
|
# 1. npm ci does not consider the platform of optional dependencies
|
|
# https://github.com/npm/cli/issues/558
|
|
# 2. A bug with the native compiler has bloated the packages from
|
|
# 30MB to almost 300MB
|
|
# https://github.com/google/closure-compiler-npm/issues/186
|
|
# If either of these bugs are fixed then we can remove this exception
|
|
closure_compiler_native = ''
|
|
if LINUX and ARCH in ('x86', 'x86_64'):
|
|
closure_compiler_native = 'google-closure-compiler-linux'
|
|
if MACOS and ARCH in ('x86', 'x86_64'):
|
|
closure_compiler_native = 'google-closure-compiler-osx'
|
|
if WINDOWS and ARCH == 'x86_64':
|
|
closure_compiler_native = 'google-closure-compiler-windows'
|
|
if closure_compiler_native:
|
|
print('Running post-install step: npm install', closure_compiler_native)
|
|
try:
|
|
subprocess.check_output(
|
|
[npm, 'install', closure_compiler_native],
|
|
cwd=directory, stderr=subprocess.STDOUT, env=env,
|
|
universal_newlines=True)
|
|
except subprocess.CalledProcessError as e:
|
|
errlog('Error running %s:\n%s' % (e.cmd, e.output))
|
|
return False
|
|
|
|
print('Done running: npm ci')
|
|
return True
|
|
|
|
|
|
def emscripten_post_install(tool):
|
|
debug_print('emscripten_post_install(' + str(tool) + ')')
|
|
src_root = os.path.join(tool.installation_path(), 'tools', 'optimizer')
|
|
build_root = optimizer_build_root(tool)
|
|
build_type = decide_cmake_build_type(tool)
|
|
|
|
args = []
|
|
|
|
# Configure
|
|
cmake_generator = CMAKE_GENERATOR
|
|
if 'Visual Studio 16' in CMAKE_GENERATOR: # VS2019
|
|
# With Visual Studio 16 2019, CMake changed the way they specify target arch.
|
|
# Instead of appending it into the CMake generator line, it is specified
|
|
# with a -A arch parameter.
|
|
args += ['-A', 'x64' if tool.bitness == 64 else 'x86']
|
|
elif 'Visual Studio' in CMAKE_GENERATOR and tool.bitness == 64:
|
|
cmake_generator += ' Win64'
|
|
|
|
success = cmake_configure(cmake_generator, build_root, src_root, build_type, args)
|
|
if not success:
|
|
return False
|
|
|
|
# Make
|
|
success = make_build(build_root, build_type, 'x64' if tool.bitness == 64 else 'Win32')
|
|
if not success:
|
|
return False
|
|
|
|
success = emscripten_npm_install(tool, tool.installation_path())
|
|
|
|
return True
|
|
|
|
|
|
# Binaryen build scripts:
|
|
def binaryen_build_root(tool):
|
|
build_root = tool.installation_path().strip()
|
|
if build_root.endswith('/') or build_root.endswith('\\'):
|
|
build_root = build_root[:-1]
|
|
generator_prefix = cmake_generator_prefix()
|
|
build_root = build_root + generator_prefix + '_' + str(tool.bitness) + 'bit_binaryen'
|
|
return build_root
|
|
|
|
|
|
def uninstall_binaryen(tool):
|
|
debug_print('uninstall_binaryen(' + str(tool) + ')')
|
|
build_root = binaryen_build_root(tool)
|
|
print("Deleting path '" + build_root + "'")
|
|
remove_tree(build_root)
|
|
|
|
|
|
def is_binaryen_installed(tool):
|
|
build_root = binaryen_build_root(tool)
|
|
return os.path.exists(build_root)
|
|
|
|
|
|
def build_binaryen_tool(tool):
|
|
debug_print('build_binaryen_tool(' + str(tool) + ')')
|
|
src_root = tool.installation_path()
|
|
build_root = binaryen_build_root(tool)
|
|
build_type = decide_cmake_build_type(tool)
|
|
|
|
# Configure
|
|
args = []
|
|
|
|
cmake_generator = CMAKE_GENERATOR
|
|
if 'Visual Studio 16' in CMAKE_GENERATOR: # VS2019
|
|
# With Visual Studio 16 2019, CMake changed the way they specify target arch.
|
|
# Instead of appending it into the CMake generator line, it is specified
|
|
# with a -A arch parameter.
|
|
args += ['-A', 'x64' if tool.bitness == 64 else 'x86']
|
|
elif 'Visual Studio' in CMAKE_GENERATOR and tool.bitness == 64:
|
|
cmake_generator += ' Win64'
|
|
|
|
if 'Visual Studio' in CMAKE_GENERATOR:
|
|
if BUILD_FOR_TESTING:
|
|
args += ['-DRUN_STATIC_ANALYZER=1']
|
|
|
|
success = cmake_configure(cmake_generator, build_root, src_root, build_type, args)
|
|
if not success:
|
|
return False
|
|
|
|
# Make
|
|
success = make_build(build_root, build_type, 'x64' if tool.bitness == 64 else 'Win32')
|
|
|
|
# Deploy scripts needed from source repository to build directory
|
|
remove_tree(os.path.join(build_root, 'scripts'))
|
|
shutil.copytree(os.path.join(src_root, 'scripts'), os.path.join(build_root, 'scripts'))
|
|
remove_tree(os.path.join(build_root, 'src', 'js'))
|
|
shutil.copytree(os.path.join(src_root, 'src', 'js'), os.path.join(build_root, 'src', 'js'))
|
|
|
|
return success
|
|
|
|
|
|
def download_and_unzip(zipfile, dest_dir, download_even_if_exists=False,
|
|
filename_prefix='', clobber=True):
|
|
debug_print('download_and_unzip(zipfile=' + zipfile + ', dest_dir=' + dest_dir + ')')
|
|
|
|
url = urljoin(emsdk_packages_url, zipfile)
|
|
download_target = get_download_target(url, zips_subdir, filename_prefix)
|
|
|
|
# If the archive was already downloaded, and the directory it would be
|
|
# unpacked to has contents, assume it's the same contents and skip.
|
|
if not download_even_if_exists and num_files_in_directory(dest_dir) > 0:
|
|
print("The contents of file '" + zipfile + "' already exist in destination '" + dest_dir + "', skipping.")
|
|
return True
|
|
# Otherwise, if the archive must be downloaded, always write into the
|
|
# target directory, since it may be a new version of a tool that gets
|
|
# installed to the same place (that is, a different download name
|
|
# indicates different contents).
|
|
download_even_if_exists = True
|
|
|
|
received_download_target = download_file(url, zips_subdir, download_even_if_exists, filename_prefix)
|
|
if not received_download_target:
|
|
return False
|
|
assert received_download_target == download_target
|
|
|
|
# Remove the old directory, since we have some SDKs that install into the
|
|
# same directory. If we didn't do this contents of the previous install
|
|
# could remain.
|
|
if clobber:
|
|
remove_tree(dest_dir)
|
|
if zipfile.endswith('.zip'):
|
|
return unzip(download_target, dest_dir, unpack_even_if_exists=download_even_if_exists)
|
|
else:
|
|
return untargz(download_target, dest_dir, unpack_even_if_exists=download_even_if_exists)
|
|
|
|
|
|
def to_native_path(p):
|
|
if WINDOWS and not MSYS:
|
|
return to_unix_path(p).replace('/', '\\')
|
|
else:
|
|
return to_unix_path(p)
|
|
|
|
|
|
# Finds and returns a list of the directories that need to be added to PATH for
|
|
# the given set of tools.
|
|
def get_required_path(active_tools):
|
|
path_add = [to_native_path(emsdk_path())]
|
|
for tool in active_tools:
|
|
if hasattr(tool, 'activated_path'):
|
|
path = to_native_path(tool.expand_vars(tool.activated_path))
|
|
path_add.append(path)
|
|
return path_add
|
|
|
|
|
|
# Returns the absolute path to the file '.emscripten' for the current user on
|
|
# this system.
|
|
def dot_emscripten_path():
|
|
return os.path.join(emsdk_path(), ".emscripten")
|
|
|
|
|
|
dot_emscripten = {}
|
|
|
|
|
|
def parse_key_value(line):
|
|
if not line:
|
|
return ('', '')
|
|
eq = line.find('=')
|
|
if eq != -1:
|
|
key = line[0:eq].strip()
|
|
value = line[eq + 1:].strip()
|
|
return (key, value)
|
|
else:
|
|
return (key, '')
|
|
|
|
|
|
def load_dot_emscripten():
|
|
dot_emscripten.clear()
|
|
lines = []
|
|
try:
|
|
lines = open(dot_emscripten_path(), "r").read().split('\n')
|
|
except:
|
|
pass
|
|
for line in lines:
|
|
try:
|
|
key, value = parse_key_value(line)
|
|
if value != '':
|
|
dot_emscripten[key] = value
|
|
except:
|
|
pass
|
|
|
|
|
|
def generate_dot_emscripten(active_tools):
|
|
temp_dir = sdk_path('tmp')
|
|
mkdir_p(temp_dir)
|
|
|
|
cfg = 'import os\n'
|
|
cfg += "emsdk_path = os.path.dirname(os.environ.get('EM_CONFIG')).replace('\\\\', '/')\n"
|
|
|
|
# Different tools may provide the same activated configs; the latest to be
|
|
# activated is the relevant one.
|
|
activated_config = OrderedDict()
|
|
for tool in active_tools:
|
|
for name, value in tool.activated_config().items():
|
|
activated_config[name] = value
|
|
|
|
if 'NODE_JS' not in activated_config:
|
|
node_fallback = which('nodejs')
|
|
if not node_fallback:
|
|
node_fallback = 'node'
|
|
activated_config['NODE_JS'] = node_fallback
|
|
|
|
for name, value in activated_config.items():
|
|
cfg += name + " = '" + value + "'\n"
|
|
|
|
cfg += '''\
|
|
TEMP_DIR = '%s'
|
|
COMPILER_ENGINE = NODE_JS
|
|
JS_ENGINES = [NODE_JS]
|
|
''' % temp_dir
|
|
|
|
cfg = cfg.replace("'" + emsdk_path(), "emsdk_path + '")
|
|
|
|
if os.path.exists(dot_emscripten_path()):
|
|
backup_path = dot_emscripten_path() + ".old"
|
|
move_with_overwrite(dot_emscripten_path(), backup_path)
|
|
|
|
with open(dot_emscripten_path(), "w") as text_file:
|
|
text_file.write(cfg)
|
|
|
|
# Clear old emscripten content.
|
|
rmfile(os.path.join(emsdk_path(), ".emscripten_sanity"))
|
|
|
|
path_add = get_required_path(active_tools)
|
|
if not WINDOWS:
|
|
emsdk_env = sdk_path('emsdk_env.sh')
|
|
print('Next steps:')
|
|
print('- To conveniently access emsdk tools from the command line,')
|
|
print(' consider adding the following directories to your PATH:')
|
|
for p in path_add:
|
|
print(' ' + p)
|
|
print('- This can be done for the current shell by running:')
|
|
print(' source "%s"' % emsdk_env)
|
|
print('- Configure emsdk in your bash profile by running:')
|
|
print(' echo \'source "%s"\' >> $HOME/.bash_profile' % emsdk_env)
|
|
|
|
|
|
def find_msbuild_dir():
|
|
if 'ProgramFiles' in os.environ and os.environ['ProgramFiles']:
|
|
program_files = os.environ['ProgramFiles']
|
|
else:
|
|
program_files = 'C:/Program Files'
|
|
if 'ProgramFiles(x86)' in os.environ and os.environ['ProgramFiles(x86)']:
|
|
program_files_x86 = os.environ['ProgramFiles(x86)']
|
|
else:
|
|
program_files_x86 = 'C:/Program Files (x86)'
|
|
MSBUILDX86_DIR = os.path.join(program_files_x86, "MSBuild/Microsoft.Cpp/v4.0/Platforms")
|
|
MSBUILD_DIR = os.path.join(program_files, "MSBuild/Microsoft.Cpp/v4.0/Platforms")
|
|
if os.path.exists(MSBUILDX86_DIR):
|
|
return MSBUILDX86_DIR
|
|
if os.path.exists(MSBUILD_DIR):
|
|
return MSBUILD_DIR
|
|
# No MSbuild installed.
|
|
return ''
|
|
|
|
|
|
class Tool(object):
|
|
def __init__(self, data):
|
|
# Convert the dictionary representation of the tool in 'data' to members of
|
|
# this class for convenience.
|
|
for key, value in data.items():
|
|
# Python2 compat, convert unicode to str
|
|
if sys.version_info < (3,) and isinstance(value, unicode): # noqa
|
|
value = value.encode('Latin-1')
|
|
setattr(self, key, value)
|
|
|
|
# Cache the name ID of this Tool (these are read very often)
|
|
self.name = self.id + '-' + self.version
|
|
if hasattr(self, 'bitness'):
|
|
self.name += '-' + str(self.bitness) + 'bit'
|
|
|
|
def __str__(self):
|
|
return self.name
|
|
|
|
def __repr__(self):
|
|
return self.name
|
|
|
|
def expand_vars(self, str):
|
|
if WINDOWS and '%MSBuildPlatformsDir%' in str:
|
|
str = str.replace('%MSBuildPlatformsDir%', find_msbuild_dir())
|
|
if '%cmake_build_type_on_win%' in str:
|
|
str = str.replace('%cmake_build_type_on_win%', (decide_cmake_build_type(self) + '/') if WINDOWS else '')
|
|
if '%installation_dir%' in str:
|
|
str = str.replace('%installation_dir%', sdk_path(self.installation_dir()))
|
|
if '%generator_prefix%' in str:
|
|
str = str.replace('%generator_prefix%', cmake_generator_prefix())
|
|
str = str.replace('%.exe%', '.exe' if WINDOWS else '')
|
|
if '%fastcomp_build_dir%' in str:
|
|
str = str.replace('%fastcomp_build_dir%', llvm_build_dir(self))
|
|
if '%fastcomp_build_bin_dir%' in str:
|
|
str = str.replace('%fastcomp_build_bin_dir%', fastcomp_build_bin_dir(self))
|
|
return str
|
|
|
|
# Return true if this tool requires building from source, and false if this is a precompiled tool.
|
|
def needs_compilation(self):
|
|
if hasattr(self, 'cmake_build_type'):
|
|
return True
|
|
|
|
if hasattr(self, 'uses'):
|
|
for tool_name in self.uses:
|
|
tool = find_tool(tool_name)
|
|
if not tool:
|
|
debug_print('Tool ' + str(self) + ' depends on ' + tool_name + ' which does not exist!')
|
|
continue
|
|
if tool.needs_compilation():
|
|
return True
|
|
|
|
return False
|
|
|
|
# Specifies the target path where this tool will be installed to. This could
|
|
# either be a directory or a filename (e.g. in case of node.js)
|
|
def installation_path(self):
|
|
if WINDOWS and hasattr(self, 'windows_install_path'):
|
|
pth = self.expand_vars(self.windows_install_path)
|
|
return sdk_path(pth)
|
|
if hasattr(self, 'install_path'):
|
|
pth = self.expand_vars(self.install_path)
|
|
return sdk_path(pth)
|
|
p = self.version
|
|
if hasattr(self, 'bitness') and (not hasattr(self, 'append_bitness') or self.append_bitness):
|
|
p += '_' + str(self.bitness) + 'bit'
|
|
return sdk_path(os.path.join(self.id, p))
|
|
|
|
# Specifies the target directory this tool will be installed to.
|
|
def installation_dir(self):
|
|
dir = self.installation_path()
|
|
if path_points_to_directory(dir):
|
|
return dir
|
|
else:
|
|
return os.path.dirname(dir)
|
|
|
|
# Returns the configuration item that needs to be added to .emscripten to make
|
|
# this Tool active for the current user.
|
|
def activated_config(self):
|
|
if not hasattr(self, 'activated_cfg'):
|
|
return {}
|
|
config = OrderedDict()
|
|
expanded = to_unix_path(self.expand_vars(self.activated_cfg))
|
|
for specific_cfg in expanded.split(';'):
|
|
name, value = specific_cfg.split('=')
|
|
config[name] = value.strip("'")
|
|
return config
|
|
|
|
def activated_environment(self):
|
|
if hasattr(self, 'activated_env'):
|
|
return self.expand_vars(self.activated_env).split(';')
|
|
else:
|
|
return []
|
|
|
|
def compatible_with_this_arch(self):
|
|
if hasattr(self, 'arch'):
|
|
if self.arch != ARCH:
|
|
return False
|
|
return True
|
|
|
|
def compatible_with_this_os(self):
|
|
if hasattr(self, 'os'):
|
|
if self.os == 'all':
|
|
return True
|
|
if self.compatible_with_this_arch() and ((WINDOWS and 'win' in self.os) or (LINUX and ('linux' in self.os or 'unix' in self.os)) or (MACOS and ('macos' in self.os or 'unix' in self.os))):
|
|
return True
|
|
else:
|
|
return False
|
|
else:
|
|
if not hasattr(self, 'macos_url') and not hasattr(self, 'windows_url') and not hasattr(self, 'unix_url') and not hasattr(self, 'linux_url'):
|
|
return True
|
|
|
|
if MACOS and hasattr(self, 'macos_url') and self.compatible_with_this_arch():
|
|
return True
|
|
|
|
if LINUX and hasattr(self, 'linux_url') and self.compatible_with_this_arch():
|
|
return True
|
|
|
|
if WINDOWS and (hasattr(self, 'windows_url') or hasattr(self, 'windows_install_path')) and self.compatible_with_this_arch():
|
|
return True
|
|
|
|
if UNIX and hasattr(self, 'unix_url'):
|
|
return True
|
|
|
|
return hasattr(self, 'url')
|
|
|
|
# the "version file" is a file inside install dirs that indicates the
|
|
# version installed there. this helps disambiguate when there is more than
|
|
# one version that may be installed to the same directory (which is used
|
|
# to avoid accumulating builds over time in some cases, with new builds
|
|
# overwriting the old)
|
|
def get_version_file_path(self):
|
|
return os.path.join(self.installation_path(), '.emsdk_version')
|
|
|
|
def is_installed_version(self):
|
|
version_file_path = self.get_version_file_path()
|
|
if os.path.isfile(version_file_path):
|
|
with open(version_file_path, 'r') as version_file:
|
|
return version_file.read().strip() == self.name
|
|
return False
|
|
|
|
def update_installed_version(self):
|
|
with open(self.get_version_file_path(), 'w') as version_file:
|
|
version_file.write(self.name + '\n')
|
|
return None
|
|
|
|
def is_installed(self, skip_version_check=False):
|
|
# If this tool/sdk depends on other tools, require that all dependencies are
|
|
# installed for this tool to count as being installed.
|
|
if hasattr(self, 'uses'):
|
|
for tool_name in self.uses:
|
|
tool = find_tool(tool_name)
|
|
if tool is None:
|
|
errlog("Manifest error: No tool by name '" + tool_name + "' found! This may indicate an internal SDK error!")
|
|
return False
|
|
if not tool.is_installed():
|
|
return False
|
|
|
|
if self.download_url() is None:
|
|
# This tool does not contain downloadable elements, so it is installed by default.
|
|
return True
|
|
|
|
content_exists = os.path.exists(self.installation_path()) and (os.path.isfile(self.installation_path()) or num_files_in_directory(self.installation_path()) > 0)
|
|
|
|
# For e.g. fastcomp clang from git repo, the activated PATH is the
|
|
# directory where the compiler is built to, and installation_path is
|
|
# the directory where the source tree exists. To distinguish between
|
|
# multiple packages sharing the same source (clang-main-32bit,
|
|
# clang-main-64bit, clang-main-32bit and clang-main-64bit each
|
|
# share the same git repo), require that in addition to the installation
|
|
# directory, each item in the activated PATH must exist.
|
|
if hasattr(self, 'activated_path') and not os.path.exists(self.expand_vars(self.activated_path)):
|
|
content_exists = False
|
|
|
|
if hasattr(self, 'custom_is_installed_script'):
|
|
if self.custom_is_installed_script == 'is_optimizer_installed':
|
|
return is_optimizer_installed(self)
|
|
elif self.custom_is_installed_script == 'is_binaryen_installed':
|
|
return is_binaryen_installed(self)
|
|
else:
|
|
raise Exception('Unknown custom_is_installed_script directive "' + self.custom_is_installed_script + '"!')
|
|
|
|
return content_exists and (skip_version_check or self.is_installed_version())
|
|
|
|
def is_active(self):
|
|
if not self.is_installed():
|
|
return False
|
|
|
|
if self.id == 'vs-tool':
|
|
# vs-tool is a special tool since all versions must be installed to the
|
|
# same dir, which means that if this tool is installed, it is also active.
|
|
return True
|
|
|
|
# All dependencies of this tool must be active as well.
|
|
deps = self.dependencies()
|
|
for tool in deps:
|
|
if not tool.is_active():
|
|
return False
|
|
|
|
activated_cfg = self.activated_config()
|
|
if not activated_cfg:
|
|
return len(deps) > 0
|
|
|
|
for key, value in activated_cfg.items():
|
|
if key not in dot_emscripten:
|
|
debug_print(str(self) + ' is not active, because key="' + key + '" does not exist in .emscripten')
|
|
return False
|
|
|
|
# all paths are stored dynamically relative to the emsdk root, so
|
|
# normalize those first.
|
|
dot_emscripten_key = dot_emscripten[key].replace("emsdk_path + '", "'" + emsdk_path())
|
|
dot_emscripten_key = dot_emscripten_key.strip("'")
|
|
if dot_emscripten_key != value:
|
|
debug_print(str(self) + ' is not active, because key="' + key + '" has value "' + dot_emscripten_key + '" but should have value "' + value + '"')
|
|
return False
|
|
return True
|
|
|
|
# Returns true if the system environment variables requires by this tool are currently active.
|
|
def is_env_active(self):
|
|
envs = self.activated_environment()
|
|
for env in envs:
|
|
key, value = parse_key_value(env)
|
|
if key not in os.environ or to_unix_path(os.environ[key]) != to_unix_path(value):
|
|
debug_print(str(self) + ' is not active, because environment variable key="' + key + '" has value "' + str(os.getenv(key)) + '" but should have value "' + value + '"')
|
|
return False
|
|
|
|
if hasattr(self, 'activated_path'):
|
|
path = to_unix_path(self.expand_vars(self.activated_path))
|
|
for p in path:
|
|
path_items = os.environ['PATH'].replace('\\', '/').split(ENVPATH_SEPARATOR)
|
|
if not normalized_contains(path_items, p):
|
|
debug_print(str(self) + ' is not active, because environment variable PATH item "' + p + '" is not present (PATH=' + os.environ['PATH'] + ')')
|
|
return False
|
|
return True
|
|
|
|
# If this tool can be installed on this system, this function returns True.
|
|
# Otherwise, this function returns a string that describes the reason why this
|
|
# tool is not available.
|
|
def can_be_installed(self):
|
|
if hasattr(self, 'bitness'):
|
|
if self.bitness == 64 and not is_os_64bit():
|
|
return "this tool is only provided for 64-bit OSes"
|
|
|
|
if self.id == 'vs-tool':
|
|
msbuild_dir = find_msbuild_dir()
|
|
if msbuild_dir:
|
|
return True
|
|
else:
|
|
return "Visual Studio was not found!"
|
|
else:
|
|
return True
|
|
|
|
def download_url(self):
|
|
if WINDOWS and hasattr(self, 'windows_url'):
|
|
return self.windows_url
|
|
elif MACOS and hasattr(self, 'macos_url'):
|
|
return self.macos_url
|
|
elif LINUX and hasattr(self, 'linux_url'):
|
|
return self.linux_url
|
|
elif UNIX and hasattr(self, 'unix_url'):
|
|
return self.unix_url
|
|
elif hasattr(self, 'url'):
|
|
return self.url
|
|
else:
|
|
return None
|
|
|
|
def install(self):
|
|
"""Returns True if the Tool was installed of False if was skipped due to
|
|
already being installed.
|
|
"""
|
|
if self.can_be_installed() is not True:
|
|
exit_with_error("The tool '" + str(self) + "' is not available due to the reason: " + self.can_be_installed())
|
|
|
|
if self.id == 'sdk':
|
|
return self.install_sdk()
|
|
else:
|
|
return self.install_tool()
|
|
|
|
def install_sdk(self):
|
|
"""Returns True if any SDK component was installed of False all componented
|
|
were already installed.
|
|
"""
|
|
print("Installing SDK '" + str(self) + "'..")
|
|
installed = False
|
|
|
|
for tool_name in self.uses:
|
|
tool = find_tool(tool_name)
|
|
if tool is None:
|
|
exit_with_error("Manifest error: No tool by name '" + tool_name + "' found! This may indicate an internal SDK error!")
|
|
installed |= tool.install()
|
|
|
|
if not installed:
|
|
print("All SDK components already installed: '" + str(self) + "'.")
|
|
return False
|
|
|
|
if getattr(self, 'custom_install_script', None) == 'emscripten_npm_install':
|
|
# upstream tools have hardcoded paths that are not stored in emsdk_manifest.json registry
|
|
install_path = 'upstream' if 'releases-upstream' in self.version else 'fastcomp'
|
|
emscripten_dir = os.path.join(emsdk_path(), install_path, 'emscripten')
|
|
# Older versions of the sdk did not include the node_modules directory
|
|
# and require `npm ci` to be run post-install
|
|
if not os.path.exists(os.path.join(emscripten_dir, 'node_modules')):
|
|
if not emscripten_npm_install(self, emscripten_dir):
|
|
exit_with_error('post-install step failed: emscripten_npm_install')
|
|
|
|
print("Done installing SDK '" + str(self) + "'.")
|
|
return True
|
|
|
|
def install_tool(self):
|
|
"""Returns True if the SDK was installed of False if was skipped due to
|
|
already being installed.
|
|
"""
|
|
# Avoid doing a redundant reinstall of the tool, if it has already been installed.
|
|
# However all tools that are sourced directly from git branches do need to be
|
|
# installed every time when requested, since the install step is then used to git
|
|
# pull the tool to a newer version.
|
|
if self.is_installed() and not hasattr(self, 'git_branch'):
|
|
print("Skipped installing " + self.name + ", already installed.")
|
|
return False
|
|
|
|
print("Installing tool '" + str(self) + "'..")
|
|
url = self.download_url()
|
|
|
|
if hasattr(self, 'custom_install_script') and self.custom_install_script == 'build_fastcomp':
|
|
success = build_fastcomp(self)
|
|
elif hasattr(self, 'custom_install_script') and self.custom_install_script == 'build_llvm':
|
|
success = build_llvm(self)
|
|
elif hasattr(self, 'git_branch'):
|
|
success = git_clone_checkout_and_pull(url, self.installation_path(), self.git_branch)
|
|
elif url.endswith(ARCHIVE_SUFFIXES):
|
|
# TODO: explain the vs-tool special-casing
|
|
download_even_if_exists = (self.id == 'vs-tool')
|
|
# The 'releases' sdk is doesn't include a verion number in the directory
|
|
# name and instead only one version can be install at the time and each
|
|
# one will clobber the other. This means we always need to extract this
|
|
# archive even when the target directory exists.
|
|
download_even_if_exists = (self.id == 'releases')
|
|
filename_prefix = getattr(self, 'zipfile_prefix', '')
|
|
success = download_and_unzip(url, self.installation_path(), download_even_if_exists=download_even_if_exists, filename_prefix=filename_prefix)
|
|
else:
|
|
dst_file = download_file(urljoin(emsdk_packages_url, self.download_url()), self.installation_path())
|
|
if dst_file:
|
|
success = True
|
|
else:
|
|
success = False
|
|
|
|
if not success:
|
|
exit_with_error("Installation failed!")
|
|
|
|
if hasattr(self, 'custom_install_script'):
|
|
if self.custom_install_script == 'emscripten_post_install':
|
|
success = emscripten_post_install(self)
|
|
elif self.custom_install_script == 'emscripten_npm_install':
|
|
success = emscripten_npm_install(self, self.installation_path())
|
|
elif self.custom_install_script in ('build_fastcomp', 'build_llvm'):
|
|
# 'build_fastcomp' is a special one that does the download on its
|
|
# own, others do the download manually.
|
|
pass
|
|
elif self.custom_install_script == 'build_binaryen':
|
|
success = build_binaryen_tool(self)
|
|
else:
|
|
raise Exception('Unknown custom_install_script command "' + self.custom_install_script + '"!')
|
|
|
|
if not success:
|
|
exit_with_error("Installation failed!")
|
|
|
|
# Install an emscripten-version.txt file if told to, and if there is one.
|
|
# (If this is not an actual release, but some other build, then we do not
|
|
# write anything.)
|
|
if hasattr(self, 'emscripten_releases_hash'):
|
|
emscripten_version_file_path = os.path.join(to_native_path(self.expand_vars(self.activated_path)), 'emscripten-version.txt')
|
|
version = get_emscripten_release_version(self.emscripten_releases_hash)
|
|
if version:
|
|
open(emscripten_version_file_path, 'w').write('"%s"' % version)
|
|
|
|
print("Done installing tool '" + str(self) + "'.")
|
|
|
|
# Sanity check that the installation succeeded, and if so, remove unneeded
|
|
# leftover installation files.
|
|
if not self.is_installed(skip_version_check=True):
|
|
exit_with_error("Installation of '" + str(self) + "' failed, but no error was detected. Either something went wrong with the installation, or this may indicate an internal emsdk error.")
|
|
|
|
self.cleanup_temp_install_files()
|
|
self.update_installed_version()
|
|
return True
|
|
|
|
def cleanup_temp_install_files(self):
|
|
url = self.download_url()
|
|
if url.endswith(ARCHIVE_SUFFIXES):
|
|
download_target = get_download_target(url, zips_subdir, getattr(self, 'zipfile_prefix', ''))
|
|
debug_print("Deleting temporary zip file " + download_target)
|
|
rmfile(download_target)
|
|
|
|
def uninstall(self):
|
|
if not self.is_installed():
|
|
print("Tool '" + str(self) + "' was not installed. No need to uninstall.")
|
|
return
|
|
print("Uninstalling tool '" + str(self) + "'..")
|
|
if hasattr(self, 'custom_uninstall_script'):
|
|
if self.custom_uninstall_script == 'uninstall_optimizer':
|
|
uninstall_optimizer(self)
|
|
elif self.custom_uninstall_script == 'uninstall_binaryen':
|
|
uninstall_binaryen(self)
|
|
else:
|
|
raise Exception('Unknown custom_uninstall_script directive "' + self.custom_uninstall_script + '"!')
|
|
print("Deleting path '" + self.installation_path() + "'")
|
|
remove_tree(self.installation_path())
|
|
print("Done uninstalling '" + str(self) + "'.")
|
|
|
|
def dependencies(self):
|
|
if not hasattr(self, 'uses'):
|
|
return []
|
|
deps = []
|
|
|
|
for tool_name in self.uses:
|
|
tool = find_tool(tool_name)
|
|
if tool:
|
|
deps += [tool]
|
|
return deps
|
|
|
|
def recursive_dependencies(self):
|
|
if not hasattr(self, 'uses'):
|
|
return []
|
|
deps = []
|
|
for tool_name in self.uses:
|
|
tool = find_tool(tool_name)
|
|
if tool:
|
|
deps += [tool]
|
|
deps += tool.recursive_dependencies()
|
|
return deps
|
|
|
|
|
|
# A global registry of all known Emscripten SDK tools available in the SDK manifest.
|
|
tools = []
|
|
tools_map = {}
|
|
|
|
|
|
def add_tool(tool):
|
|
tool.is_sdk = False
|
|
tools.append(tool)
|
|
if find_tool(str(tool)):
|
|
raise Exception('Duplicate tool ' + str(tool) + '! Existing:\n{' + ', '.join("%s: %s" % item for item in vars(find_tool(str(tool))).items()) + '}, New:\n{' + ', '.join("%s: %s" % item for item in vars(tool).items()) + '}')
|
|
tools_map[str(tool)] = tool
|
|
|
|
|
|
# A global registry of all known SDK toolsets.
|
|
sdks = []
|
|
sdks_map = {}
|
|
|
|
|
|
def add_sdk(sdk):
|
|
sdk.is_sdk = True
|
|
sdks.append(sdk)
|
|
if find_sdk(str(sdk)):
|
|
raise Exception('Duplicate sdk ' + str(sdk) + '! Existing:\n{' + ', '.join("%s: %s" % item for item in vars(find_sdk(str(sdk))).items()) + '}, New:\n{' + ', '.join("%s: %s" % item for item in vars(sdk).items()) + '}')
|
|
sdks_map[str(sdk)] = sdk
|
|
|
|
|
|
# N.B. In both tools and sdks list above, we take the convention that the newest
|
|
# items are at the back of the list (ascending chronological order)
|
|
|
|
def find_tool(name):
|
|
return tools_map.get(name)
|
|
|
|
|
|
def find_sdk(name):
|
|
return sdks_map.get(name)
|
|
|
|
|
|
def is_os_64bit():
|
|
# http://stackoverflow.com/questions/2208828/detect-64bit-os-windows-in-python
|
|
return platform.machine().endswith('64')
|
|
|
|
|
|
def find_latest_releases_version():
|
|
releases_info = load_releases_info()
|
|
return releases_info['latest']
|
|
|
|
|
|
def find_latest_releases_hash():
|
|
releases_info = load_releases_info()
|
|
return releases_info['releases'][find_latest_releases_version()]
|
|
|
|
|
|
def find_latest_releases_sdk(which):
|
|
return 'sdk-releases-%s-%s-64bit' % (which, find_latest_releases_hash())
|
|
|
|
|
|
def find_tot_sdk():
|
|
debug_print('Fetching emscripten-releases repository...')
|
|
global extra_release_tag
|
|
extra_release_tag = get_emscripten_releases_tot()
|
|
return 'sdk-releases-upstream-%s-64bit' % (extra_release_tag)
|
|
|
|
|
|
# Given a git hash in emscripten-releases, find the emscripten
|
|
# version for it. There may not be one if this is not the hash of
|
|
# a release, in which case we return None.
|
|
def get_emscripten_release_version(emscripten_releases_hash):
|
|
releases_info = load_releases_info()
|
|
for key, value in dict(releases_info['releases']).items():
|
|
if value == emscripten_releases_hash:
|
|
return key
|
|
return None
|
|
|
|
|
|
# Get the tip-of-tree build identifier.
|
|
def get_emscripten_releases_tot():
|
|
git_clone_checkout_and_pull(emscripten_releases_repo, sdk_path('releases'), 'master')
|
|
recent_releases = git_recent_commits(sdk_path('releases'))
|
|
# The recent releases are the latest hashes in the git repo. There
|
|
# may not be a build for the most recent ones yet; find the last
|
|
# that does.
|
|
for release in recent_releases:
|
|
url = emscripten_releases_download_url_template % (
|
|
os_name_for_emscripten_releases(),
|
|
release,
|
|
'tbz2' if not WINDOWS else 'zip'
|
|
)
|
|
try:
|
|
urlopen(url)
|
|
except:
|
|
continue
|
|
return release
|
|
exit_with_error('failed to find build of any recent emsdk revision')
|
|
|
|
|
|
def get_release_hash(arg, releases_info):
|
|
return releases_info.get(arg, None) or releases_info.get('sdk-' + arg + '-64bit')
|
|
|
|
|
|
def version_key(ver):
|
|
return tuple(map(int, re.split('[._-]', ver)))
|
|
|
|
|
|
# A sort function that is compatible with both Python 2 and Python 3 using a
|
|
# custom comparison function.
|
|
def python_2_3_sorted(arr, cmp):
|
|
if sys.version_info >= (3,):
|
|
return sorted(arr, key=functools.cmp_to_key(cmp))
|
|
else:
|
|
return sorted(arr, cmp=cmp)
|
|
|
|
|
|
def is_emsdk_sourced_from_github():
|
|
return os.path.exists(os.path.join(emsdk_path(), '.git'))
|
|
|
|
|
|
def update_emsdk():
|
|
if is_emsdk_sourced_from_github():
|
|
errlog('You seem to have bootstrapped Emscripten SDK by cloning from GitHub. In this case, use "git pull" instead of "emsdk update" to update emsdk. (Not doing that automatically in case you have local changes)')
|
|
sys.exit(1)
|
|
if not download_and_unzip(emsdk_zip_download_url, emsdk_path(), download_even_if_exists=True, clobber=False):
|
|
sys.exit(1)
|
|
|
|
|
|
# Lists all legacy (pre-emscripten-releases) tagged versions directly in the Git
|
|
# repositories. These we can pull and compile from source.
|
|
def load_legacy_emscripten_tags():
|
|
return open(sdk_path('legacy-emscripten-tags.txt'), 'r').read().split('\n')
|
|
|
|
|
|
def load_legacy_binaryen_tags():
|
|
return open(sdk_path('legacy-binaryen-tags.txt'), 'r').read().split('\n')
|
|
|
|
|
|
def remove_prefix(s, prefix):
|
|
if s.startswith(prefix):
|
|
return s[len(prefix):]
|
|
else:
|
|
return s
|
|
|
|
|
|
def remove_suffix(s, suffix):
|
|
if s.endswith(suffix):
|
|
return s[:len(s) - len(suffix)]
|
|
else:
|
|
return s
|
|
|
|
|
|
# filename should be one of: 'llvm-precompiled-tags-32bit.txt', 'llvm-precompiled-tags-64bit.txt'
|
|
def load_file_index_list(filename):
|
|
items = open(sdk_path(filename)).read().splitlines()
|
|
items = [remove_suffix(remove_suffix(remove_prefix(x, 'emscripten-llvm-e'), '.tar.gz'), '.zip').strip() for x in items]
|
|
items = [x for x in items if 'latest' not in x and len(x) > 0]
|
|
|
|
# Sort versions from oldest to newest (the default sort would be
|
|
# lexicographic, i.e. '1.37.1 < 1.37.10 < 1.37.2')
|
|
return sorted(items, key=version_key)
|
|
|
|
|
|
def exit_with_error(msg):
|
|
errlog(str(msg))
|
|
sys.exit(1)
|
|
|
|
|
|
# Load the json info for emscripten-releases.
|
|
def load_releases_info():
|
|
if not hasattr(load_releases_info, 'cached_info'):
|
|
try:
|
|
text = open(sdk_path('emscripten-releases-tags.txt'), 'r').read()
|
|
load_releases_info.cached_info = json.loads(text)
|
|
except Exception as e:
|
|
print('Error parsing emscripten-releases-tags.txt!')
|
|
exit_with_error(str(e))
|
|
|
|
return load_releases_info.cached_info
|
|
|
|
|
|
def get_installed_sdk_version():
|
|
version_file = sdk_path(os.path.join('upstream', '.emsdk_version'))
|
|
if not os.path.exists(version_file):
|
|
return None
|
|
with open(version_file) as f:
|
|
version = f.read()
|
|
return version.split('-')[2]
|
|
|
|
|
|
# Get a list of tags for emscripten-releases.
|
|
def load_releases_tags():
|
|
tags = []
|
|
tags_fastcomp = []
|
|
info = load_releases_info()
|
|
|
|
for version, sha in sorted(info['releases'].items(), key=lambda x: version_key(x[0])):
|
|
tags.append(sha)
|
|
# Only include versions older than 1.39.0 in fastcomp releases
|
|
if version_key(version) < (2, 0, 0):
|
|
tags_fastcomp.append(sha)
|
|
|
|
if extra_release_tag:
|
|
tags.append(extra_release_tag)
|
|
|
|
# Explicitly add the currently installed SDK version. This could be a custom
|
|
# version (installed explicitly) so it might not be part of the main list loaded above.
|
|
installed = get_installed_sdk_version()
|
|
if installed and installed not in tags:
|
|
tags.append(installed)
|
|
|
|
return tags, tags_fastcomp
|
|
|
|
|
|
def load_releases_versions():
|
|
info = load_releases_info()
|
|
versions = list(info['releases'].keys())
|
|
return versions
|
|
|
|
|
|
def is_string(s):
|
|
if sys.version_info[0] >= 3:
|
|
return isinstance(s, str)
|
|
return isinstance(s, basestring) # noqa
|
|
|
|
|
|
def load_sdk_manifest():
|
|
try:
|
|
manifest = json.loads(open(sdk_path("emsdk_manifest.json"), "r").read())
|
|
except Exception as e:
|
|
print('Error parsing emsdk_manifest.json!')
|
|
print(str(e))
|
|
return
|
|
|
|
emscripten_tags = load_legacy_emscripten_tags()
|
|
llvm_precompiled_tags_32bit = []
|
|
llvm_precompiled_tags_64bit = load_file_index_list('llvm-tags-64bit.txt')
|
|
llvm_precompiled_tags = llvm_precompiled_tags_32bit + llvm_precompiled_tags_64bit
|
|
binaryen_tags = load_legacy_binaryen_tags()
|
|
releases_tags, releases_tags_fastcomp = load_releases_tags()
|
|
|
|
def dependencies_exist(sdk):
|
|
for tool_name in sdk.uses:
|
|
tool = find_tool(tool_name)
|
|
if not tool:
|
|
debug_print('missing dependency: ' + tool_name)
|
|
return False
|
|
return True
|
|
|
|
def cmp_version(ver, cmp_operand, reference):
|
|
if cmp_operand == '<=':
|
|
return version_key(ver) <= version_key(reference)
|
|
if cmp_operand == '<':
|
|
return version_key(ver) < version_key(reference)
|
|
if cmp_operand == '>=':
|
|
return version_key(ver) >= version_key(reference)
|
|
if cmp_operand == '>':
|
|
return version_key(ver) > version_key(reference)
|
|
if cmp_operand == '==':
|
|
return version_key(ver) == version_key(reference)
|
|
if cmp_operand == '!=':
|
|
return version_key(ver) != version_key(reference)
|
|
raise Exception('Invalid cmp_operand "' + cmp_operand + '"!')
|
|
|
|
def passes_filters(param, ver, filters):
|
|
for v in filters:
|
|
if v[0] == param and not cmp_version(ver, v[1], v[2]):
|
|
return False
|
|
return True
|
|
|
|
# A 'category parameter' is a %foo%-encoded identifier that specifies
|
|
# a class of tools instead of just one tool, e.g. %tag%
|
|
def expand_category_param(param, category_list, t, is_sdk):
|
|
for i, ver in enumerate(category_list):
|
|
if not ver.strip():
|
|
continue
|
|
t2 = copy.copy(t)
|
|
found_param = False
|
|
for p, v in vars(t2).items():
|
|
if is_string(v) and param in v:
|
|
t2.__dict__[p] = v.replace(param, ver)
|
|
found_param = True
|
|
if not found_param:
|
|
continue
|
|
t2.is_old = i < len(category_list) - 2
|
|
if hasattr(t2, 'uses'):
|
|
t2.uses = [x.replace(param, ver) for x in t2.uses]
|
|
|
|
# Filter out expanded tools by version requirements, such as ["tag", "<=", "1.37.22"]
|
|
if hasattr(t2, 'version_filter'):
|
|
passes = passes_filters(param, ver, t2.version_filter)
|
|
if not passes:
|
|
continue
|
|
|
|
if is_sdk:
|
|
if dependencies_exist(t2):
|
|
if not find_sdk(t2.name):
|
|
add_sdk(t2)
|
|
else:
|
|
debug_print('SDK ' + str(t2) + ' already existed in manifest, not adding twice')
|
|
else:
|
|
if not find_tool(t2.name):
|
|
add_tool(t2)
|
|
else:
|
|
debug_print('Tool ' + str(t2) + ' already existed in manifest, not adding twice')
|
|
|
|
for tool in manifest['tools']:
|
|
t = Tool(tool)
|
|
if t.compatible_with_this_os():
|
|
if not hasattr(t, 'is_old'):
|
|
t.is_old = False
|
|
|
|
# Expand the metapackages that refer to tags
|
|
if '%tag%' in t.version:
|
|
expand_category_param('%tag%', emscripten_tags, t, is_sdk=False)
|
|
elif '%precompiled_tag%' in t.version:
|
|
expand_category_param('%precompiled_tag%', llvm_precompiled_tags, t, is_sdk=False)
|
|
elif '%precompiled_tag32%' in t.version:
|
|
expand_category_param('%precompiled_tag32%', llvm_precompiled_tags_32bit, t, is_sdk=False)
|
|
elif '%precompiled_tag64%' in t.version:
|
|
expand_category_param('%precompiled_tag64%', llvm_precompiled_tags_64bit, t, is_sdk=False)
|
|
elif '%binaryen_tag%' in t.version:
|
|
expand_category_param('%binaryen_tag%', binaryen_tags, t, is_sdk=False)
|
|
elif '%releases-tag%' in t.version and 'fastcomp' in t.version:
|
|
expand_category_param('%releases-tag%', releases_tags_fastcomp, t, is_sdk=False)
|
|
elif '%releases-tag%' in t.version:
|
|
expand_category_param('%releases-tag%', releases_tags, t, is_sdk=False)
|
|
else:
|
|
add_tool(t)
|
|
|
|
for sdk_str in manifest['sdks']:
|
|
sdk_str['id'] = 'sdk'
|
|
sdk = Tool(sdk_str)
|
|
if sdk.compatible_with_this_os():
|
|
if not hasattr(sdk, 'is_old'):
|
|
sdk.is_old = False
|
|
|
|
if '%tag%' in sdk.version:
|
|
expand_category_param('%tag%', emscripten_tags, sdk, is_sdk=True)
|
|
elif '%precompiled_tag%' in sdk.version:
|
|
expand_category_param('%precompiled_tag%', llvm_precompiled_tags, sdk, is_sdk=True)
|
|
elif '%precompiled_tag32%' in sdk.version:
|
|
expand_category_param('%precompiled_tag32%', llvm_precompiled_tags_32bit, sdk, is_sdk=True)
|
|
elif '%precompiled_tag64%' in sdk.version:
|
|
expand_category_param('%precompiled_tag64%', llvm_precompiled_tags_64bit, sdk, is_sdk=True)
|
|
elif '%releases-tag%' in sdk.version and 'fastcomp' in sdk.version:
|
|
expand_category_param('%releases-tag%', releases_tags_fastcomp, sdk, is_sdk=True)
|
|
elif '%releases-tag%' in sdk.version:
|
|
expand_category_param('%releases-tag%', releases_tags, sdk, is_sdk=True)
|
|
else:
|
|
add_sdk(sdk)
|
|
|
|
|
|
# Tests if the two given tools can be active at the same time.
|
|
# Currently only a simple check for name for same tool with different versions,
|
|
# possibly adds more logic in the future.
|
|
def can_simultaneously_activate(tool1, tool2):
|
|
return tool1.id != tool2.id
|
|
|
|
|
|
def remove_nonexisting_tools(tool_list, log_errors=True):
|
|
i = 0
|
|
while i < len(tool_list):
|
|
tool = tool_list[i]
|
|
if not tool.is_installed():
|
|
if log_errors:
|
|
errlog("Warning: The SDK/tool '" + str(tool) + "' cannot be activated since it is not installed! Skipping this tool...")
|
|
tool_list.pop(i)
|
|
continue
|
|
i += 1
|
|
return tool_list
|
|
|
|
|
|
# Expands dependencies for each tool, and removes ones that don't exist.
|
|
def process_tool_list(tools_to_activate, log_errors=True):
|
|
i = 0
|
|
# Gather dependencies for each tool
|
|
while i < len(tools_to_activate):
|
|
tool = tools_to_activate[i]
|
|
deps = tool.recursive_dependencies()
|
|
tools_to_activate = tools_to_activate[:i] + deps + tools_to_activate[i:]
|
|
i += len(deps) + 1
|
|
|
|
tools_to_activate = remove_nonexisting_tools(tools_to_activate, log_errors=log_errors)
|
|
|
|
# Remove conflicting tools
|
|
i = 0
|
|
while i < len(tools_to_activate):
|
|
j = 0
|
|
while j < i:
|
|
secondary_tool = tools_to_activate[j]
|
|
primary_tool = tools_to_activate[i]
|
|
if not can_simultaneously_activate(primary_tool, secondary_tool):
|
|
tools_to_activate.pop(j)
|
|
j -= 1
|
|
i -= 1
|
|
j += 1
|
|
i += 1
|
|
return tools_to_activate
|
|
|
|
|
|
def write_set_env_script(env_string):
|
|
assert(WINDOWS)
|
|
open(EMSDK_SET_ENV, 'w').write(env_string)
|
|
|
|
|
|
# Reconfigure .emscripten to choose the currently activated toolset, set PATH
|
|
# and other environment variables.
|
|
# Returns the full list of deduced tools that are now active.
|
|
def set_active_tools(tools_to_activate, permanently_activate, system):
|
|
tools_to_activate = process_tool_list(tools_to_activate, log_errors=True)
|
|
|
|
if tools_to_activate:
|
|
tools = [x for x in tools_to_activate if not x.is_sdk]
|
|
print('Setting the following tools as active:\n ' + '\n '.join(map(lambda x: str(x), tools)))
|
|
print('')
|
|
|
|
generate_dot_emscripten(tools_to_activate)
|
|
|
|
# Construct a .bat script that will be invoked to set env. vars and PATH
|
|
# We only do this on windows since emsdk.bat is able to modify the
|
|
# calling shell environment. On other platform `source emsdk_env.sh` is
|
|
# required.
|
|
if WINDOWS:
|
|
# always set local environment variables since permanently activating will only set the registry settings and
|
|
# will not affect the current session
|
|
env_vars_to_add = get_env_vars_to_add(tools_to_activate, system, user=permanently_activate)
|
|
env_string = construct_env_with_vars(env_vars_to_add)
|
|
write_set_env_script(env_string)
|
|
|
|
if permanently_activate:
|
|
win_set_environment_variables(env_vars_to_add, system, user=permanently_activate)
|
|
|
|
return tools_to_activate
|
|
|
|
|
|
def currently_active_sdk():
|
|
for sdk in reversed(sdks):
|
|
if sdk.is_active():
|
|
return sdk
|
|
return None
|
|
|
|
|
|
def currently_active_tools():
|
|
active_tools = []
|
|
for tool in tools:
|
|
if tool.is_active():
|
|
active_tools += [tool]
|
|
return active_tools
|
|
|
|
|
|
# http://stackoverflow.com/questions/480214/how-do-you-remove-duplicates-from-a-list-in-python-whilst-preserving-order
|
|
def unique_items(seq):
|
|
seen = set()
|
|
seen_add = seen.add
|
|
return [x for x in seq if x not in seen and not seen_add(x)]
|
|
|
|
|
|
# Tests if a path is contained in the given list, but with separators normalized.
|
|
def normalized_contains(lst, elem):
|
|
elem = to_unix_path(elem)
|
|
for e in lst:
|
|
if elem == to_unix_path(e):
|
|
return True
|
|
return False
|
|
|
|
|
|
def to_msys_path(p):
|
|
p = to_unix_path(p)
|
|
new_path = re.sub(r'([a-zA-Z]):/(.*)', r'/\1/\2', p)
|
|
if len(new_path) > 3 and new_path[0] == '/' and new_path[2] == '/':
|
|
new_path = new_path[0] + new_path[1].lower() + new_path[2:]
|
|
return new_path
|
|
|
|
|
|
# Looks at the current PATH and adds and removes entries so that the PATH reflects
|
|
# the set of given active tools.
|
|
def adjusted_path(tools_to_activate, system=False, user=False):
|
|
# These directories should be added to PATH
|
|
path_add = get_required_path(tools_to_activate)
|
|
# These already exist.
|
|
if WINDOWS and not MSYS:
|
|
existing_path = win_get_environment_variable('PATH', system=system, user=user, fallback=True).split(ENVPATH_SEPARATOR)
|
|
else:
|
|
existing_path = os.environ['PATH'].split(ENVPATH_SEPARATOR)
|
|
emsdk_root_path = to_unix_path(emsdk_path())
|
|
|
|
existing_emsdk_tools = []
|
|
existing_nonemsdk_path = []
|
|
for entry in existing_path:
|
|
if to_unix_path(entry).startswith(emsdk_root_path):
|
|
existing_emsdk_tools.append(entry)
|
|
else:
|
|
existing_nonemsdk_path.append(entry)
|
|
|
|
new_emsdk_tools = []
|
|
kept_emsdk_tools = []
|
|
for entry in path_add:
|
|
if not normalized_contains(existing_emsdk_tools, entry):
|
|
new_emsdk_tools.append(entry)
|
|
else:
|
|
kept_emsdk_tools.append(entry)
|
|
|
|
whole_path = unique_items(new_emsdk_tools + kept_emsdk_tools + existing_nonemsdk_path)
|
|
|
|
if MSYS:
|
|
# XXX Hack: If running native Windows Python in MSYS prompt where PATH
|
|
# entries look like "/c/Windows/System32", os.environ['PATH']
|
|
# in Python will transform to show them as "C:\\Windows\\System32", so need
|
|
# to reconvert path delimiter back to forward slashes.
|
|
whole_path = [to_msys_path(p) for p in whole_path]
|
|
new_emsdk_tools = [to_msys_path(p) for p in new_emsdk_tools]
|
|
|
|
separator = ':' if MSYS else ENVPATH_SEPARATOR
|
|
return (separator.join(whole_path), new_emsdk_tools)
|
|
|
|
|
|
def get_env_vars_to_add(tools_to_activate, system, user):
|
|
env_vars_to_add = []
|
|
|
|
newpath, added_path = adjusted_path(tools_to_activate, system, user)
|
|
|
|
# Don't bother setting the path if there are no changes.
|
|
if os.environ['PATH'] != newpath:
|
|
env_vars_to_add += [('PATH', newpath)]
|
|
|
|
if added_path:
|
|
errlog('Adding directories to PATH:')
|
|
for item in added_path:
|
|
errlog('PATH += ' + item)
|
|
errlog('')
|
|
|
|
# A core variable EMSDK points to the root of Emscripten SDK directory.
|
|
env_vars_to_add += [('EMSDK', to_unix_path(emsdk_path()))]
|
|
env_vars_to_add += [('EM_CONFIG', os.path.normpath(dot_emscripten_path()))]
|
|
|
|
for tool in tools_to_activate:
|
|
config = tool.activated_config()
|
|
if 'EMSCRIPTEN_ROOT' in config:
|
|
# For older emscripten versions that don't use this default we export
|
|
# EM_CACHE.
|
|
em_cache_dir = os.path.join(config['EMSCRIPTEN_ROOT'], 'cache')
|
|
env_vars_to_add += [('EM_CACHE', em_cache_dir)]
|
|
envs = tool.activated_environment()
|
|
for env in envs:
|
|
key, value = parse_key_value(env)
|
|
value = to_native_path(tool.expand_vars(value))
|
|
env_vars_to_add += [(key, value)]
|
|
|
|
return env_vars_to_add
|
|
|
|
|
|
def construct_env(tools_to_activate, system, user):
|
|
return construct_env_with_vars(get_env_vars_to_add(tools_to_activate, system, user))
|
|
|
|
|
|
def construct_env_with_vars(env_vars_to_add):
|
|
env_string = ''
|
|
if env_vars_to_add:
|
|
errlog('Setting environment variables:')
|
|
|
|
for key, value in env_vars_to_add:
|
|
# Don't set env vars which are already set to the correct value.
|
|
if key not in os.environ or to_unix_path(os.environ[key]) != to_unix_path(value):
|
|
errlog(key + ' = ' + value)
|
|
if POWERSHELL:
|
|
env_string += '$env:' + key + '="' + value + '"\n'
|
|
elif CMD:
|
|
env_string += 'SET ' + key + '=' + value + '\n'
|
|
elif CSH:
|
|
env_string += 'setenv ' + key + ' "' + value + '"\n'
|
|
elif BASH:
|
|
env_string += 'export ' + key + '="' + value + '"\n'
|
|
else:
|
|
assert False
|
|
if 'EMSDK_PYTHON' in env_vars_to_add:
|
|
# When using our bundled python we never want the user's
|
|
# PYTHONHOME or PYTHONPATH
|
|
# See https://github.com/emscripten-core/emsdk/issues/598
|
|
if POWERSHELL:
|
|
env_string += 'Remove-Item env:PYTHONHOME\n'
|
|
env_string += 'Remove-Item env:PYTHONPATH\n'
|
|
elif CMD:
|
|
env_string += 'set PYTHONHOME=\n'
|
|
env_string += 'set PYTHONPATH=\n'
|
|
elif CSH:
|
|
env_string += 'unsetenv PYTHONHOME\n'
|
|
env_string += 'unsetenv PYTHONPATH\n'
|
|
elif BASH:
|
|
env_string += 'unset PYTHONHOME\n'
|
|
env_string += 'unset PYTHONPATH\n'
|
|
else:
|
|
assert False
|
|
|
|
return env_string
|
|
|
|
|
|
def error_on_missing_tool(name):
|
|
if name.endswith('-64bit') and not is_os_64bit():
|
|
errlog("Error: '%s' is only provided for 64-bit OSes." % name)
|
|
else:
|
|
errlog("Error: No tool or SDK found by name '%s'." % name)
|
|
return 1
|
|
|
|
|
|
def exit_with_fastcomp_error():
|
|
exit_with_error('The fastcomp backend is not getting new builds or releases. Please use the upstream llvm backend or use an older version than 2.0.0 (such as 1.40.1).')
|
|
|
|
|
|
def expand_sdk_name(name, activating):
|
|
if 'upstream-master' in name:
|
|
errlog('upstream-master SDK has been renamed upstream-main')
|
|
name = name.replace('upstream-master', 'upstream-main')
|
|
if name in ('latest-fastcomp', 'latest-releases-fastcomp', 'tot-fastcomp', 'sdk-nightly-latest'):
|
|
exit_with_fastcomp_error()
|
|
if name in ('latest', 'sdk-latest', 'latest-64bit', 'sdk-latest-64bit'):
|
|
# This is effectly the default SDK
|
|
return str(find_latest_releases_sdk('upstream'))
|
|
elif name in ('latest-upstream', 'latest-clang-upstream', 'latest-releases-upstream'):
|
|
return str(find_latest_releases_sdk('upstream'))
|
|
elif name in ('tot', 'sdk-tot', 'tot-upstream'):
|
|
if activating:
|
|
# When we are activating a tot release, assume that the currently
|
|
# installed SDK, if any, is the tot release we want to activate.
|
|
# Without this `install tot && activate tot` will race with the builders
|
|
# that are producing new builds.
|
|
installed = get_installed_sdk_version()
|
|
if installed:
|
|
debug_print('activating currently installed SDK; not updating tot version')
|
|
return 'sdk-releases-upstream-%s-64bit' % installed
|
|
return str(find_tot_sdk())
|
|
else:
|
|
# check if it's a release handled by an emscripten-releases version,
|
|
# and if so use that by using the right hash. we support a few notations,
|
|
# x.y.z[-(upstream|fastcomp_])
|
|
# sdk-x.y.z[-(upstream|fastcomp_])-64bit
|
|
# TODO: support short notation for old builds too?
|
|
backend = None
|
|
fullname = name
|
|
if '-upstream' in fullname:
|
|
fullname = name.replace('-upstream', '')
|
|
backend = 'upstream'
|
|
elif '-fastcomp' in fullname:
|
|
fullname = fullname.replace('-fastcomp', '')
|
|
backend = 'fastcomp'
|
|
version = fullname.replace('sdk-', '').replace('releases-', '').replace('-64bit', '').replace('tag-', '')
|
|
releases_info = load_releases_info()['releases']
|
|
release_hash = get_release_hash(version, releases_info)
|
|
if release_hash:
|
|
# Known release hash
|
|
if backend == 'fastcomp' and version_key(version) >= (2, 0, 0):
|
|
exit_with_fastcomp_error()
|
|
if backend is None:
|
|
if version_key(version) >= (1, 39, 0):
|
|
backend = 'upstream'
|
|
else:
|
|
backend = 'fastcomp'
|
|
return 'sdk-releases-%s-%s-64bit' % (backend, release_hash)
|
|
elif len(version) == 40:
|
|
global extra_release_tag
|
|
extra_release_tag = version
|
|
return 'sdk-releases-%s-%s-64bit' % (backend, version)
|
|
return name
|
|
|
|
|
|
def main(args):
|
|
if not args:
|
|
errlog("Missing command; Type 'emsdk help' to get a list of commands.")
|
|
return 1
|
|
|
|
cmd = args.pop(0)
|
|
|
|
if cmd in ('help', '--help', '-h'):
|
|
print(' emsdk: Available commands:')
|
|
|
|
print('''
|
|
emsdk list [--old] [--uses] - Lists all available SDKs and tools and their
|
|
current installation status. With the --old
|
|
parameter, also historical versions are
|
|
shown. If --uses is passed, displays the
|
|
composition of different SDK packages and
|
|
dependencies.
|
|
|
|
emsdk update - Updates emsdk to the newest version. If you have
|
|
bootstrapped emsdk via cloning directly from
|
|
GitHub, call "git pull" instead to update emsdk.
|
|
|
|
emsdk install [options] <tool 1> <tool 2> <tool 3> ...
|
|
- Downloads and installs given tools or SDKs.
|
|
Options can contain:
|
|
|
|
-j<num>: Specifies the number of cores to use when
|
|
building the tool. Default: use one less
|
|
than the # of detected cores.
|
|
|
|
--build=<type>: Controls what kind of build of LLVM to
|
|
perform. Pass either 'Debug', 'Release',
|
|
'MinSizeRel' or 'RelWithDebInfo'. Default:
|
|
'RelWithDebInfo'.
|
|
|
|
--generator=<type>: Specifies the CMake Generator to be used
|
|
during the build. Possible values are the
|
|
same as what your CMake supports and whether
|
|
the generator is valid depends on the tools
|
|
you have installed. Defaults to 'Unix Makefiles'
|
|
on *nix systems. If generator name is multiple
|
|
words, enclose with single or double quotes.
|
|
|
|
--shallow: When installing tools from one of the git
|
|
development branches, this parameter can be
|
|
passed to perform a shallow git clone instead
|
|
of a full one. This reduces the amount of
|
|
network transfer that is needed. This option
|
|
should only be used when you are interested in
|
|
downloading one of the development branches,
|
|
but are not looking to develop Emscripten
|
|
yourself. Default: disabled, i.e. do a full
|
|
clone.
|
|
|
|
--build-tests: If enabled, LLVM is built with internal tests
|
|
included. Pass this to enable running test
|
|
other.test_llvm_lit in the Emscripten test
|
|
suite. Default: disabled.
|
|
--enable-assertions: If specified, LLVM is built with assert()
|
|
checks enabled. Useful for development
|
|
purposes. Default: Enabled
|
|
--disable-assertions: Forces assertions off during the build.
|
|
|
|
--vs2017/--vs2019: If building from source, overrides to build
|
|
using the specified compiler. When installing
|
|
precompiled packages, this has no effect.
|
|
Note: The same compiler specifier must be
|
|
passed to the emsdk activate command to
|
|
activate the desired version.
|
|
|
|
Notes on building from source:
|
|
|
|
To pass custom CMake directives when configuring
|
|
LLVM build, specify the environment variable
|
|
LLVM_CMAKE_ARGS="param1=value1,param2=value2"
|
|
in the environment where the build is invoked.
|
|
See README.md for details.
|
|
|
|
emsdk uninstall <tool/sdk> - Removes the given tool or SDK from disk.''')
|
|
|
|
if WINDOWS:
|
|
print('''
|
|
emsdk activate [--permanent] [--system] [--build=type] [--vs2017/--vs2019] <tool/sdk>
|
|
|
|
- Activates the given tool or SDK in the
|
|
environment of the current shell.
|
|
|
|
- If the `--permanent` option is passed, then the environment
|
|
variables are set permanently for the current user.
|
|
|
|
- If the `--system` option is passed, the registration
|
|
is done for all users of the system.
|
|
This needs admin privileges
|
|
(uses Machine environment variables).
|
|
|
|
- If a custom compiler version was used to override
|
|
the compiler to use, pass the same --vs2017/--vs2019 parameter
|
|
here to choose which version to activate.
|
|
|
|
emcmdprompt.bat - Spawns a new command prompt window with the
|
|
Emscripten environment active.''')
|
|
else:
|
|
print(''' emsdk activate [--build=type] <tool/sdk>
|
|
|
|
- Activates the given tool or SDK in the
|
|
environment of the current shell.''')
|
|
|
|
print('''
|
|
Both commands 'install' and 'activate' accept an optional parameter
|
|
'--build=type', which can be used to override what kind of installation
|
|
or activation to perform. Possible values for type are Debug, Release,
|
|
MinSizeRel or RelWithDebInfo. Note: When overriding a custom build type,
|
|
be sure to match the same --build= option to both 'install' and
|
|
'activate' commands and the invocation of 'emsdk_env', or otherwise
|
|
these commands will default to operating on the default build type
|
|
which in and RelWithDebInfo.''')
|
|
return 0
|
|
|
|
# Extracts a boolean command line argument from args and returns True if it was present
|
|
def extract_bool_arg(name):
|
|
if name in args:
|
|
args.remove(name)
|
|
return True
|
|
return False
|
|
|
|
arg_old = extract_bool_arg('--old')
|
|
arg_uses = extract_bool_arg('--uses')
|
|
arg_permanent = extract_bool_arg('--permanent')
|
|
arg_global = extract_bool_arg('--global')
|
|
arg_system = extract_bool_arg('--system')
|
|
if arg_global:
|
|
print('--global is deprecated. Use `--system` to set the environment variables for all users')
|
|
arg_system = True
|
|
if arg_system:
|
|
arg_permanent = True
|
|
if extract_bool_arg('--embedded'):
|
|
errlog('embedded mode is now the only mode available')
|
|
if extract_bool_arg('--no-embedded'):
|
|
errlog('embedded mode is now the only mode available')
|
|
return 1
|
|
|
|
arg_notty = extract_bool_arg('--notty')
|
|
if arg_notty:
|
|
global TTY_OUTPUT
|
|
TTY_OUTPUT = False
|
|
|
|
# Replace meta-packages with the real package names.
|
|
if cmd in ('update', 'install', 'activate'):
|
|
activating = cmd == 'activate'
|
|
args = [expand_sdk_name(a, activating=activating) for a in args]
|
|
|
|
load_dot_emscripten()
|
|
load_sdk_manifest()
|
|
|
|
# Process global args
|
|
for i in range(len(args)):
|
|
if args[i].startswith('--generator='):
|
|
build_generator = re.match(r'''^--generator=['"]?([^'"]+)['"]?$''', args[i])
|
|
if build_generator:
|
|
global CMAKE_GENERATOR
|
|
CMAKE_GENERATOR = build_generator.group(1)
|
|
args[i] = ''
|
|
else:
|
|
errlog("Cannot parse CMake generator string: " + args[i] + ". Try wrapping generator string with quotes")
|
|
return 1
|
|
elif args[i].startswith('--build='):
|
|
build_type = re.match(r'^--build=(.+)$', args[i])
|
|
if build_type:
|
|
global CMAKE_BUILD_TYPE_OVERRIDE
|
|
build_type = build_type.group(1)
|
|
build_types = ['Debug', 'MinSizeRel', 'RelWithDebInfo', 'Release']
|
|
try:
|
|
build_type_index = [x.lower() for x in build_types].index(build_type.lower())
|
|
CMAKE_BUILD_TYPE_OVERRIDE = build_types[build_type_index]
|
|
args[i] = ''
|
|
except:
|
|
errlog('Unknown CMake build type "' + build_type + '" specified! Please specify one of ' + str(build_types))
|
|
return 1
|
|
else:
|
|
errlog("Invalid command line parameter " + args[i] + ' specified!')
|
|
return 1
|
|
args = [x for x in args if x]
|
|
|
|
if cmd == 'list':
|
|
print('')
|
|
|
|
def installed_sdk_text(name):
|
|
sdk = find_sdk(name)
|
|
return 'INSTALLED' if sdk and sdk.is_installed() else ''
|
|
|
|
if (LINUX or MACOS or WINDOWS) and (ARCH == 'x86' or ARCH == 'x86_64'):
|
|
print('The *recommended* precompiled SDK download is %s (%s).' % (find_latest_releases_version(), find_latest_releases_hash()))
|
|
print()
|
|
print('To install/activate it, use one of:')
|
|
print(' latest [default (llvm) backend]')
|
|
print(' latest-fastcomp [legacy (fastcomp) backend]')
|
|
print('')
|
|
print('Those are equivalent to installing/activating the following:')
|
|
print(' %s %s' % (find_latest_releases_version(), installed_sdk_text(find_latest_releases_sdk('upstream'))))
|
|
print(' %s-fastcomp %s' % (find_latest_releases_version(), installed_sdk_text(find_latest_releases_sdk('fastcomp'))))
|
|
print('')
|
|
else:
|
|
print('Warning: your platform does not have precompiled SDKs available.')
|
|
print('You may install components from source.')
|
|
print('')
|
|
|
|
print('All recent (non-legacy) installable versions are:')
|
|
releases_versions = sorted(
|
|
load_releases_versions(),
|
|
key=lambda x: [int(v) if v.isdigit() else -1 for v in x.split('.')],
|
|
reverse=True,
|
|
)
|
|
releases_info = load_releases_info()['releases']
|
|
for ver in releases_versions:
|
|
print(' %s %s' % (ver, installed_sdk_text('sdk-releases-upstream-%s-64bit' % get_release_hash(ver, releases_info))))
|
|
print()
|
|
|
|
# Use array to work around the lack of being able to mutate from enclosing
|
|
# function.
|
|
has_partially_active_tools = [False]
|
|
|
|
if sdks:
|
|
def find_sdks(needs_compilation):
|
|
s = []
|
|
for sdk in sdks:
|
|
if sdk.is_old and not arg_old:
|
|
continue
|
|
if sdk.needs_compilation() == needs_compilation:
|
|
s += [sdk]
|
|
return s
|
|
|
|
def print_sdks(s):
|
|
for sdk in s:
|
|
installed = '\tINSTALLED' if sdk.is_installed() else ''
|
|
active = '*' if sdk.is_active() else ' '
|
|
print(' ' + active + ' {0: <25}'.format(str(sdk)) + installed)
|
|
if arg_uses:
|
|
for dep in sdk.uses:
|
|
print(' - {0: <25}'.format(dep))
|
|
print('')
|
|
print('The additional following precompiled SDKs are also available for download:')
|
|
print_sdks(find_sdks(False))
|
|
|
|
print('The following SDKs can be compiled from source:')
|
|
print_sdks(find_sdks(True))
|
|
|
|
if tools:
|
|
def find_tools(needs_compilation):
|
|
t = []
|
|
for tool in tools:
|
|
if tool.is_old and not arg_old:
|
|
continue
|
|
if tool.needs_compilation() != needs_compilation:
|
|
continue
|
|
t += [tool]
|
|
return t
|
|
|
|
def print_tools(t):
|
|
for tool in t:
|
|
if tool.is_old and not arg_old:
|
|
continue
|
|
if tool.can_be_installed() is True:
|
|
installed = '\tINSTALLED' if tool.is_installed() else ''
|
|
else:
|
|
installed = '\tNot available: ' + tool.can_be_installed()
|
|
tool_is_active = tool.is_active()
|
|
tool_is_env_active = tool_is_active and tool.is_env_active()
|
|
if tool_is_env_active:
|
|
active = ' * '
|
|
elif tool_is_active:
|
|
active = '(*)'
|
|
has_partially_active_tools[0] = has_partially_active_tools[0] or True
|
|
else:
|
|
active = ' '
|
|
print(' ' + active + ' {0: <25}'.format(str(tool)) + installed)
|
|
print('')
|
|
|
|
print('The following precompiled tool packages are available for download:')
|
|
print_tools(find_tools(needs_compilation=False))
|
|
print('The following tools can be compiled from source:')
|
|
print_tools(find_tools(needs_compilation=True))
|
|
else:
|
|
if is_emsdk_sourced_from_github():
|
|
print("There are no tools available. Run 'git pull' to fetch the latest set of tools.")
|
|
else:
|
|
print("There are no tools available. Run 'emsdk update' to fetch the latest set of tools.")
|
|
print('')
|
|
|
|
print('Items marked with * are activated for the current user.')
|
|
if has_partially_active_tools[0]:
|
|
env_cmd = 'emsdk_env.bat' if WINDOWS else 'source ./emsdk_env.sh'
|
|
print('Items marked with (*) are selected for use, but your current shell environment is not configured to use them. Type "' + env_cmd + '" to set up your current shell to use them' + (', or call "emsdk activate --permanent <name_of_sdk>" to permanently activate them.' if WINDOWS else '.'))
|
|
if not arg_old:
|
|
print('')
|
|
print("To access the historical archived versions, type 'emsdk list --old'")
|
|
|
|
print('')
|
|
if is_emsdk_sourced_from_github():
|
|
print('Run "git pull" to pull in the latest list.')
|
|
else:
|
|
print('Run "./emsdk update" to pull in the latest list.')
|
|
|
|
return 0
|
|
elif cmd == 'construct_env':
|
|
# Clean up old temp file up front, in case of failure later before we get
|
|
# to write out the new one.
|
|
tools_to_activate = currently_active_tools()
|
|
tools_to_activate = process_tool_list(tools_to_activate, log_errors=True)
|
|
env_string = construct_env(tools_to_activate, arg_system, arg_permanent)
|
|
if WINDOWS and not BASH:
|
|
write_set_env_script(env_string)
|
|
else:
|
|
sys.stdout.write(env_string)
|
|
return 0
|
|
elif cmd == 'update':
|
|
update_emsdk()
|
|
if WINDOWS:
|
|
# Clean up litter after old emsdk update which may have left this temp
|
|
# file around.
|
|
rmfile(sdk_path(EMSDK_SET_ENV))
|
|
return 0
|
|
elif cmd == 'update-tags':
|
|
errlog('`update-tags` is not longer needed. To install the latest tot release just run `install tot`')
|
|
return 0
|
|
elif cmd == 'activate':
|
|
if arg_permanent:
|
|
print('Registering active Emscripten environment permanently')
|
|
print('')
|
|
|
|
tools_to_activate = currently_active_tools()
|
|
args = [x for x in args if not x.startswith('--')]
|
|
for arg in args:
|
|
tool = find_tool(arg)
|
|
if tool is None:
|
|
tool = find_sdk(arg)
|
|
if tool is None:
|
|
return error_on_missing_tool(arg)
|
|
tools_to_activate += [tool]
|
|
if not tools_to_activate:
|
|
errlog('No tools/SDKs specified to activate! Usage:\n emsdk activate tool/sdk1 [tool/sdk2] [...]')
|
|
return 1
|
|
active_tools = set_active_tools(tools_to_activate, permanently_activate=arg_permanent, system=arg_system)
|
|
if not active_tools:
|
|
errlog('No tools/SDKs found to activate! Usage:\n emsdk activate tool/sdk1 [tool/sdk2] [...]')
|
|
return 1
|
|
if WINDOWS and not arg_permanent:
|
|
errlog('The changes made to environment variables only apply to the currently running shell instance. Use the \'emsdk_env.bat\' to re-enter this environment later, or if you\'d like to permanently register this environment permanently, rerun this command with the option --permanent.')
|
|
return 0
|
|
elif cmd == 'install':
|
|
global BUILD_FOR_TESTING, ENABLE_LLVM_ASSERTIONS, CPU_CORES, GIT_CLONE_SHALLOW
|
|
|
|
# Process args
|
|
for i in range(len(args)):
|
|
if args[i].startswith('-j'):
|
|
multicore = re.match(r'^-j(\d+)$', args[i])
|
|
if multicore:
|
|
CPU_CORES = int(multicore.group(1))
|
|
args[i] = ''
|
|
else:
|
|
errlog("Invalid command line parameter " + args[i] + ' specified!')
|
|
return 1
|
|
elif args[i] == '--shallow':
|
|
GIT_CLONE_SHALLOW = True
|
|
args[i] = ''
|
|
elif args[i] == '--build-tests':
|
|
BUILD_FOR_TESTING = True
|
|
args[i] = ''
|
|
elif args[i] == '--enable-assertions':
|
|
ENABLE_LLVM_ASSERTIONS = 'ON'
|
|
args[i] = ''
|
|
elif args[i] == '--disable-assertions':
|
|
ENABLE_LLVM_ASSERTIONS = 'OFF'
|
|
args[i] = ''
|
|
args = [x for x in args if x]
|
|
if not args:
|
|
errlog("Missing parameter. Type 'emsdk install <tool name>' to install a tool or an SDK. Type 'emsdk list' to obtain a list of available tools. Type 'emsdk install latest' to automatically install the newest version of the SDK.")
|
|
return 1
|
|
|
|
for t in args:
|
|
tool = find_tool(t)
|
|
if tool is None:
|
|
tool = find_sdk(t)
|
|
if tool is None:
|
|
return error_on_missing_tool(t)
|
|
tool.install()
|
|
return 0
|
|
elif cmd == 'uninstall':
|
|
if not args:
|
|
errlog("Syntax error. Call 'emsdk uninstall <tool name>'. Call 'emsdk list' to obtain a list of available tools.")
|
|
return 1
|
|
tool = find_tool(args[0])
|
|
if tool is None:
|
|
errlog("Error: Tool by name '" + args[0] + "' was not found.")
|
|
return 1
|
|
tool.uninstall()
|
|
return 0
|
|
|
|
errlog("Unknown command '" + cmd + "' given! Type 'emsdk help' to get a list of commands.")
|
|
return 1
|
|
|
|
|
|
if __name__ == '__main__':
|
|
sys.exit(main(sys.argv[1:]))
|