2019-04-17 03:30:30 -04:00
|
|
|
#!/usr/bin/env python
|
|
|
|
# coding=utf-8
|
|
|
|
#
|
|
|
|
# This script helps installing tools required to use the ESP-IDF, and updating PATH
|
|
|
|
# to use the installed tools. It can also create a Python virtual environment,
|
|
|
|
# and install Python requirements into it.
|
|
|
|
# It does not install OS dependencies. It does install tools such as the Xtensa
|
|
|
|
# GCC toolchain and ESP32 ULP coprocessor toolchain.
|
|
|
|
#
|
|
|
|
# By default, downloaded tools will be installed under $HOME/.espressif directory
|
|
|
|
# (%USERPROFILE%/.espressif on Windows). This path can be modified by setting
|
|
|
|
# IDF_TOOLS_PATH variable prior to running this tool.
|
|
|
|
#
|
|
|
|
# Users do not need to interact with this script directly. In IDF root directory,
|
|
|
|
# install.sh (.bat) and export.sh (.bat) scripts are provided to invoke this script.
|
|
|
|
#
|
|
|
|
# Usage:
|
|
|
|
#
|
|
|
|
# * To install the tools, run `idf_tools.py install`.
|
|
|
|
#
|
|
|
|
# * To install the Python environment, run `idf_tools.py install-python-env`.
|
|
|
|
#
|
|
|
|
# * To start using the tools, run `eval "$(idf_tools.py export)"` — this will update
|
|
|
|
# the PATH to point to the installed tools and set up other environment variables
|
|
|
|
# needed by the tools.
|
|
|
|
#
|
|
|
|
###
|
|
|
|
#
|
|
|
|
# Copyright 2019 Espressif Systems (Shanghai) PTE LTD
|
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
2021-01-25 21:49:01 -05:00
|
|
|
import argparse
|
|
|
|
import contextlib
|
|
|
|
import copy
|
|
|
|
import errno
|
|
|
|
import functools
|
|
|
|
import hashlib
|
2019-04-17 03:30:30 -04:00
|
|
|
import json
|
|
|
|
import os
|
2021-01-25 21:49:01 -05:00
|
|
|
import platform
|
|
|
|
import re
|
|
|
|
import shutil
|
|
|
|
import ssl
|
2019-04-17 03:30:30 -04:00
|
|
|
import subprocess
|
|
|
|
import sys
|
|
|
|
import tarfile
|
|
|
|
import zipfile
|
|
|
|
from collections import OrderedDict, namedtuple
|
|
|
|
|
2020-01-17 11:38:50 -05:00
|
|
|
try:
|
|
|
|
import typing # noqa: F401
|
|
|
|
except ImportError:
|
|
|
|
pass
|
|
|
|
|
2019-04-17 03:30:30 -04:00
|
|
|
try:
|
2020-05-20 03:50:57 -04:00
|
|
|
from urllib.error import ContentTooShortError
|
2021-01-25 21:49:01 -05:00
|
|
|
from urllib.request import urlopen
|
2019-04-17 03:30:30 -04:00
|
|
|
except ImportError:
|
2021-01-25 21:49:01 -05:00
|
|
|
from urllib import ContentTooShortError, urlopen
|
2019-04-17 03:30:30 -04:00
|
|
|
|
2019-09-23 00:01:09 -04:00
|
|
|
try:
|
|
|
|
from exceptions import WindowsError
|
|
|
|
except ImportError:
|
|
|
|
class WindowsError(OSError):
|
|
|
|
pass
|
|
|
|
|
2019-04-17 03:30:30 -04:00
|
|
|
|
|
|
|
TOOLS_FILE = 'tools/tools.json'
|
|
|
|
TOOLS_SCHEMA_FILE = 'tools/tools_schema.json'
|
|
|
|
TOOLS_FILE_NEW = 'tools/tools.new.json'
|
|
|
|
TOOLS_FILE_VERSION = 1
|
|
|
|
IDF_TOOLS_PATH_DEFAULT = os.path.join('~', '.espressif')
|
|
|
|
UNKNOWN_VERSION = 'unknown'
|
|
|
|
SUBST_TOOL_PATH_REGEX = re.compile(r'\${TOOL_PATH}')
|
|
|
|
VERSION_REGEX_REPLACE_DEFAULT = r'\1'
|
|
|
|
IDF_MAINTAINER = os.environ.get('IDF_MAINTAINER') or False
|
|
|
|
TODO_MESSAGE = 'TODO'
|
|
|
|
DOWNLOAD_RETRY_COUNT = 3
|
|
|
|
URL_PREFIX_MAP_SEPARATOR = ','
|
|
|
|
IDF_TOOLS_INSTALL_CMD = os.environ.get('IDF_TOOLS_INSTALL_CMD')
|
|
|
|
IDF_TOOLS_EXPORT_CMD = os.environ.get('IDF_TOOLS_INSTALL_CMD')
|
|
|
|
|
|
|
|
PYTHON_PLATFORM = platform.system() + '-' + platform.machine()
|
|
|
|
|
|
|
|
# Identifiers used in tools.json for different platforms.
|
|
|
|
PLATFORM_WIN32 = 'win32'
|
|
|
|
PLATFORM_WIN64 = 'win64'
|
|
|
|
PLATFORM_MACOS = 'macos'
|
|
|
|
PLATFORM_LINUX32 = 'linux-i686'
|
|
|
|
PLATFORM_LINUX64 = 'linux-amd64'
|
|
|
|
PLATFORM_LINUX_ARM32 = 'linux-armel'
|
|
|
|
PLATFORM_LINUX_ARM64 = 'linux-arm64'
|
|
|
|
|
|
|
|
|
|
|
|
# Mappings from various other names these platforms are known as, to the identifiers above.
|
|
|
|
# This includes strings produced from "platform.system() + '-' + platform.machine()", see PYTHON_PLATFORM
|
|
|
|
# definition above.
|
|
|
|
# This list also includes various strings used in release archives of xtensa-esp32-elf-gcc, OpenOCD, etc.
|
|
|
|
PLATFORM_FROM_NAME = {
|
|
|
|
# Windows
|
|
|
|
PLATFORM_WIN32: PLATFORM_WIN32,
|
|
|
|
'Windows-i686': PLATFORM_WIN32,
|
2019-07-09 07:00:16 -04:00
|
|
|
'Windows-x86': PLATFORM_WIN32,
|
2019-04-17 03:30:30 -04:00
|
|
|
PLATFORM_WIN64: PLATFORM_WIN64,
|
|
|
|
'Windows-x86_64': PLATFORM_WIN64,
|
|
|
|
'Windows-AMD64': PLATFORM_WIN64,
|
|
|
|
# macOS
|
|
|
|
PLATFORM_MACOS: PLATFORM_MACOS,
|
|
|
|
'osx': PLATFORM_MACOS,
|
|
|
|
'darwin': PLATFORM_MACOS,
|
|
|
|
'Darwin-x86_64': PLATFORM_MACOS,
|
2021-01-07 15:21:18 -05:00
|
|
|
# pretend it is x86_64 until Darwin-arm64 tool builds are available:
|
|
|
|
'Darwin-arm64': PLATFORM_MACOS,
|
2019-04-17 03:30:30 -04:00
|
|
|
# Linux
|
|
|
|
PLATFORM_LINUX64: PLATFORM_LINUX64,
|
|
|
|
'linux64': PLATFORM_LINUX64,
|
|
|
|
'Linux-x86_64': PLATFORM_LINUX64,
|
|
|
|
PLATFORM_LINUX32: PLATFORM_LINUX32,
|
|
|
|
'linux32': PLATFORM_LINUX32,
|
|
|
|
'Linux-i686': PLATFORM_LINUX32,
|
|
|
|
PLATFORM_LINUX_ARM32: PLATFORM_LINUX_ARM32,
|
|
|
|
'Linux-arm': PLATFORM_LINUX_ARM32,
|
|
|
|
'Linux-armv7l': PLATFORM_LINUX_ARM32,
|
|
|
|
PLATFORM_LINUX_ARM64: PLATFORM_LINUX_ARM64,
|
|
|
|
'Linux-arm64': PLATFORM_LINUX_ARM64,
|
|
|
|
'Linux-aarch64': PLATFORM_LINUX_ARM64,
|
|
|
|
'Linux-armv8l': PLATFORM_LINUX_ARM64,
|
|
|
|
}
|
|
|
|
|
|
|
|
UNKNOWN_PLATFORM = 'unknown'
|
|
|
|
CURRENT_PLATFORM = PLATFORM_FROM_NAME.get(PYTHON_PLATFORM, UNKNOWN_PLATFORM)
|
|
|
|
|
|
|
|
EXPORT_SHELL = 'shell'
|
|
|
|
EXPORT_KEY_VALUE = 'key-value'
|
|
|
|
|
2020-10-14 12:24:55 -04:00
|
|
|
ISRG_X1_ROOT_CERT = u"""
|
2020-05-20 03:50:57 -04:00
|
|
|
-----BEGIN CERTIFICATE-----
|
|
|
|
MIIFazCCA1OgAwIBAgIRAIIQz7DSQONZRGPgu2OCiwAwDQYJKoZIhvcNAQELBQAw
|
|
|
|
TzELMAkGA1UEBhMCVVMxKTAnBgNVBAoTIEludGVybmV0IFNlY3VyaXR5IFJlc2Vh
|
|
|
|
cmNoIEdyb3VwMRUwEwYDVQQDEwxJU1JHIFJvb3QgWDEwHhcNMTUwNjA0MTEwNDM4
|
|
|
|
WhcNMzUwNjA0MTEwNDM4WjBPMQswCQYDVQQGEwJVUzEpMCcGA1UEChMgSW50ZXJu
|
|
|
|
ZXQgU2VjdXJpdHkgUmVzZWFyY2ggR3JvdXAxFTATBgNVBAMTDElTUkcgUm9vdCBY
|
|
|
|
MTCCAiIwDQYJKoZIhvcNAQEBBQADggIPADCCAgoCggIBAK3oJHP0FDfzm54rVygc
|
|
|
|
h77ct984kIxuPOZXoHj3dcKi/vVqbvYATyjb3miGbESTtrFj/RQSa78f0uoxmyF+
|
|
|
|
0TM8ukj13Xnfs7j/EvEhmkvBioZxaUpmZmyPfjxwv60pIgbz5MDmgK7iS4+3mX6U
|
|
|
|
A5/TR5d8mUgjU+g4rk8Kb4Mu0UlXjIB0ttov0DiNewNwIRt18jA8+o+u3dpjq+sW
|
|
|
|
T8KOEUt+zwvo/7V3LvSye0rgTBIlDHCNAymg4VMk7BPZ7hm/ELNKjD+Jo2FR3qyH
|
|
|
|
B5T0Y3HsLuJvW5iB4YlcNHlsdu87kGJ55tukmi8mxdAQ4Q7e2RCOFvu396j3x+UC
|
|
|
|
B5iPNgiV5+I3lg02dZ77DnKxHZu8A/lJBdiB3QW0KtZB6awBdpUKD9jf1b0SHzUv
|
|
|
|
KBds0pjBqAlkd25HN7rOrFleaJ1/ctaJxQZBKT5ZPt0m9STJEadao0xAH0ahmbWn
|
|
|
|
OlFuhjuefXKnEgV4We0+UXgVCwOPjdAvBbI+e0ocS3MFEvzG6uBQE3xDk3SzynTn
|
|
|
|
jh8BCNAw1FtxNrQHusEwMFxIt4I7mKZ9YIqioymCzLq9gwQbooMDQaHWBfEbwrbw
|
|
|
|
qHyGO0aoSCqI3Haadr8faqU9GY/rOPNk3sgrDQoo//fb4hVC1CLQJ13hef4Y53CI
|
|
|
|
rU7m2Ys6xt0nUW7/vGT1M0NPAgMBAAGjQjBAMA4GA1UdDwEB/wQEAwIBBjAPBgNV
|
|
|
|
HRMBAf8EBTADAQH/MB0GA1UdDgQWBBR5tFnme7bl5AFzgAiIyBpY9umbbjANBgkq
|
|
|
|
hkiG9w0BAQsFAAOCAgEAVR9YqbyyqFDQDLHYGmkgJykIrGF1XIpu+ILlaS/V9lZL
|
|
|
|
ubhzEFnTIZd+50xx+7LSYK05qAvqFyFWhfFQDlnrzuBZ6brJFe+GnY+EgPbk6ZGQ
|
|
|
|
3BebYhtF8GaV0nxvwuo77x/Py9auJ/GpsMiu/X1+mvoiBOv/2X/qkSsisRcOj/KK
|
|
|
|
NFtY2PwByVS5uCbMiogziUwthDyC3+6WVwW6LLv3xLfHTjuCvjHIInNzktHCgKQ5
|
|
|
|
ORAzI4JMPJ+GslWYHb4phowim57iaztXOoJwTdwJx4nLCgdNbOhdjsnvzqvHu7Ur
|
|
|
|
TkXWStAmzOVyyghqpZXjFaH3pO3JLF+l+/+sKAIuvtd7u+Nxe5AW0wdeRlN8NwdC
|
|
|
|
jNPElpzVmbUq4JUagEiuTDkHzsxHpFKVK7q4+63SM1N95R1NbdWhscdCb+ZAJzVc
|
|
|
|
oyi3B43njTOQ5yOf+1CceWxG1bQVs5ZufpsMljq4Ui0/1lvh+wjChP4kqKOJ2qxq
|
|
|
|
4RgqsahDYVvTH9w7jXbyLeiNdd8XM2w9U/t7y0Ff/9yi0GE44Za4rF2LN9d11TPA
|
|
|
|
mRGunUHBcnWEvgJBQl9nJEiU0Zsnvgc/ubhPgXRR4Xq37Z0j4r7g1SgEEzwxA57d
|
|
|
|
emyPxgcYxn/eR44/KJ4EBs+lVDR3veyJm+kXQ99b21/+jh5Xos1AnX5iItreGCc=
|
|
|
|
-----END CERTIFICATE-----
|
|
|
|
"""
|
|
|
|
|
2019-04-17 03:30:30 -04:00
|
|
|
|
|
|
|
global_quiet = False
|
2019-06-05 03:16:19 -04:00
|
|
|
global_non_interactive = False
|
2020-01-17 11:38:50 -05:00
|
|
|
global_idf_path = None # type: typing.Optional[str]
|
|
|
|
global_idf_tools_path = None # type: typing.Optional[str]
|
|
|
|
global_tools_json = None # type: typing.Optional[str]
|
2019-04-17 03:30:30 -04:00
|
|
|
|
|
|
|
|
|
|
|
def fatal(text, *args):
|
|
|
|
if not global_quiet:
|
|
|
|
sys.stderr.write('ERROR: ' + text + '\n', *args)
|
|
|
|
|
|
|
|
|
|
|
|
def warn(text, *args):
|
|
|
|
if not global_quiet:
|
|
|
|
sys.stderr.write('WARNING: ' + text + '\n', *args)
|
|
|
|
|
|
|
|
|
|
|
|
def info(text, f=None, *args):
|
|
|
|
if not global_quiet:
|
|
|
|
if f is None:
|
|
|
|
f = sys.stdout
|
|
|
|
f.write(text + '\n', *args)
|
|
|
|
|
|
|
|
|
|
|
|
def run_cmd_check_output(cmd, input_text=None, extra_paths=None):
|
|
|
|
# If extra_paths is given, locate the executable in one of these directories.
|
|
|
|
# Note: it would seem logical to add extra_paths to env[PATH], instead, and let OS do the job of finding the
|
|
|
|
# executable for us. However this does not work on Windows: https://bugs.python.org/issue8557.
|
|
|
|
if extra_paths:
|
|
|
|
found = False
|
|
|
|
extensions = ['']
|
|
|
|
if sys.platform == 'win32':
|
|
|
|
extensions.append('.exe')
|
|
|
|
for path in extra_paths:
|
|
|
|
for ext in extensions:
|
|
|
|
fullpath = os.path.join(path, cmd[0] + ext)
|
|
|
|
if os.path.exists(fullpath):
|
|
|
|
cmd[0] = fullpath
|
|
|
|
found = True
|
|
|
|
break
|
|
|
|
if found:
|
|
|
|
break
|
|
|
|
|
|
|
|
try:
|
|
|
|
if input_text:
|
|
|
|
input_text = input_text.encode()
|
2020-12-06 23:01:10 -05:00
|
|
|
result = subprocess.run(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE, check=True, input=input_text)
|
2019-04-17 03:30:30 -04:00
|
|
|
return result.stdout + result.stderr
|
tools: fix idf_tools.py exception with python3
use `subprocess.Popen` when catch TypeError:
```
Traceback (most recent call last):
File "tools/idf_tools.py", line 1249, in <module>
main(sys.argv[1:])
File "tools/idf_tools.py", line 1245, in main
action_func(args)
File "tools/idf_tools.py", line 1038, in action_install
tool_obj.find_installed_versions()
File "tools/idf_tools.py", line 468, in find_installed_versions
ver_str = self.check_version()
File "tools/idf_tools.py", line 426, in check_version
version_cmd_result = run_cmd_check_output(cmd, None, extra_paths)
File "tools/idf_tools.py", line 176, in run_cmd_check_output
result = subprocess.run(cmd, capture_output=True, check=True, input=input_text)
File "/opt/pyenv/pyenv-1.2.6/versions/3.5.5/lib/python3.5/subprocess.py", line 383, in run
with Popen(*popenargs, **kwargs) as process:
TypeError: __init__() got an unexpected keyword argument 'capture_output'
```
2019-07-04 03:06:00 -04:00
|
|
|
except (AttributeError, TypeError):
|
2019-04-17 03:30:30 -04:00
|
|
|
p = subprocess.Popen(cmd, stdout=subprocess.PIPE, stdin=subprocess.PIPE, stderr=subprocess.PIPE)
|
|
|
|
stdout, stderr = p.communicate(input_text)
|
|
|
|
if p.returncode != 0:
|
|
|
|
try:
|
|
|
|
raise subprocess.CalledProcessError(p.returncode, cmd, stdout, stderr)
|
|
|
|
except TypeError:
|
|
|
|
raise subprocess.CalledProcessError(p.returncode, cmd, stdout)
|
|
|
|
return stdout + stderr
|
|
|
|
|
|
|
|
|
|
|
|
def to_shell_specific_paths(paths_list):
|
|
|
|
if sys.platform == 'win32':
|
|
|
|
paths_list = [p.replace('/', os.path.sep) if os.path.sep in p else p for p in paths_list]
|
|
|
|
|
|
|
|
if 'MSYSTEM' in os.environ:
|
|
|
|
paths_msys = run_cmd_check_output(['cygpath', '-u', '-f', '-'],
|
|
|
|
input_text='\n'.join(paths_list))
|
|
|
|
paths_list = paths_msys.decode().strip().split('\n')
|
|
|
|
|
|
|
|
return paths_list
|
|
|
|
|
|
|
|
|
|
|
|
def get_env_for_extra_paths(extra_paths):
|
|
|
|
"""
|
|
|
|
Return a copy of environment variables dict, prepending paths listed in extra_paths
|
|
|
|
to the PATH environment variable.
|
|
|
|
"""
|
|
|
|
env_arg = os.environ.copy()
|
|
|
|
new_path = os.pathsep.join(extra_paths) + os.pathsep + env_arg['PATH']
|
|
|
|
if sys.version_info.major == 2:
|
|
|
|
env_arg['PATH'] = new_path.encode('utf8')
|
|
|
|
else:
|
|
|
|
env_arg['PATH'] = new_path
|
|
|
|
return env_arg
|
|
|
|
|
|
|
|
|
|
|
|
def get_file_size_sha256(filename, block_size=65536):
|
|
|
|
sha256 = hashlib.sha256()
|
|
|
|
size = 0
|
|
|
|
with open(filename, 'rb') as f:
|
|
|
|
for block in iter(lambda: f.read(block_size), b''):
|
|
|
|
sha256.update(block)
|
|
|
|
size += len(block)
|
|
|
|
return size, sha256.hexdigest()
|
|
|
|
|
|
|
|
|
|
|
|
def report_progress(count, block_size, total_size):
|
|
|
|
percent = int(count * block_size * 100 / total_size)
|
|
|
|
percent = min(100, percent)
|
2021-01-25 21:49:01 -05:00
|
|
|
sys.stdout.write('\r%d%%' % percent)
|
2019-04-17 03:30:30 -04:00
|
|
|
sys.stdout.flush()
|
|
|
|
|
|
|
|
|
|
|
|
def mkdir_p(path):
|
|
|
|
try:
|
|
|
|
os.makedirs(path)
|
|
|
|
except OSError as exc:
|
|
|
|
if exc.errno != errno.EEXIST or not os.path.isdir(path):
|
|
|
|
raise
|
|
|
|
|
|
|
|
|
|
|
|
def unpack(filename, destination):
|
|
|
|
info('Extracting {0} to {1}'.format(filename, destination))
|
2020-06-08 09:34:05 -04:00
|
|
|
if filename.endswith(('.tar.gz', '.tgz')):
|
2019-04-17 03:30:30 -04:00
|
|
|
archive_obj = tarfile.open(filename, 'r:gz')
|
|
|
|
elif filename.endswith('zip'):
|
|
|
|
archive_obj = zipfile.ZipFile(filename)
|
|
|
|
else:
|
|
|
|
raise NotImplementedError('Unsupported archive type')
|
|
|
|
if sys.version_info.major == 2:
|
|
|
|
# This is a workaround for the issue that unicode destination is not handled:
|
|
|
|
# https://bugs.python.org/issue17153
|
|
|
|
destination = str(destination)
|
|
|
|
archive_obj.extractall(destination)
|
|
|
|
|
|
|
|
|
2021-01-11 10:58:21 -05:00
|
|
|
def splittype(url):
|
|
|
|
match = re.match('([^/:]+):(.*)', url, re.DOTALL)
|
|
|
|
if match:
|
|
|
|
scheme, data = match.groups()
|
|
|
|
return scheme.lower(), data
|
|
|
|
return None, url
|
|
|
|
|
|
|
|
|
2020-05-20 03:50:57 -04:00
|
|
|
# An alternative version of urlretrieve which takes SSL context as an argument
|
|
|
|
def urlretrieve_ctx(url, filename, reporthook=None, data=None, context=None):
|
|
|
|
url_type, path = splittype(url)
|
|
|
|
|
2020-10-14 14:19:32 -04:00
|
|
|
# urlopen doesn't have context argument in Python <=2.7.9
|
|
|
|
extra_urlopen_args = {}
|
|
|
|
if context:
|
2021-01-25 21:49:01 -05:00
|
|
|
extra_urlopen_args['context'] = context
|
2020-10-14 14:19:32 -04:00
|
|
|
with contextlib.closing(urlopen(url, data, **extra_urlopen_args)) as fp:
|
2020-05-20 03:50:57 -04:00
|
|
|
headers = fp.info()
|
|
|
|
|
|
|
|
# Just return the local path and the "headers" for file://
|
|
|
|
# URLs. No sense in performing a copy unless requested.
|
2021-01-25 21:49:01 -05:00
|
|
|
if url_type == 'file' and not filename:
|
2020-05-20 03:50:57 -04:00
|
|
|
return os.path.normpath(path), headers
|
|
|
|
|
|
|
|
# Handle temporary file setup.
|
|
|
|
tfp = open(filename, 'wb')
|
|
|
|
|
|
|
|
with tfp:
|
|
|
|
result = filename, headers
|
|
|
|
bs = 1024 * 8
|
2021-01-25 21:49:01 -05:00
|
|
|
size = int(headers.get('content-length', -1))
|
2020-05-20 03:50:57 -04:00
|
|
|
read = 0
|
|
|
|
blocknum = 0
|
|
|
|
|
|
|
|
if reporthook:
|
|
|
|
reporthook(blocknum, bs, size)
|
|
|
|
|
|
|
|
while True:
|
|
|
|
block = fp.read(bs)
|
|
|
|
if not block:
|
|
|
|
break
|
|
|
|
read += len(block)
|
|
|
|
tfp.write(block)
|
|
|
|
blocknum += 1
|
|
|
|
if reporthook:
|
|
|
|
reporthook(blocknum, bs, size)
|
|
|
|
|
|
|
|
if size >= 0 and read < size:
|
|
|
|
raise ContentTooShortError(
|
2021-01-25 21:49:01 -05:00
|
|
|
'retrieval incomplete: got only %i out of %i bytes'
|
2020-05-20 03:50:57 -04:00
|
|
|
% (read, size), result)
|
|
|
|
|
|
|
|
return result
|
|
|
|
|
|
|
|
|
2019-09-23 00:01:09 -04:00
|
|
|
# Sometimes renaming a directory on Windows (randomly?) causes a PermissionError.
|
|
|
|
# This is confirmed to be a workaround:
|
|
|
|
# https://github.com/espressif/esp-idf/issues/3819#issuecomment-515167118
|
|
|
|
# https://github.com/espressif/esp-idf/issues/4063#issuecomment-531490140
|
|
|
|
# https://stackoverflow.com/a/43046729
|
|
|
|
def rename_with_retry(path_from, path_to):
|
|
|
|
if sys.platform.startswith('win'):
|
|
|
|
retry_count = 100
|
|
|
|
else:
|
|
|
|
retry_count = 1
|
|
|
|
|
|
|
|
for retry in range(retry_count):
|
|
|
|
try:
|
|
|
|
os.rename(path_from, path_to)
|
|
|
|
return
|
|
|
|
except (OSError, WindowsError): # WindowsError until Python 3.3, then OSError
|
|
|
|
if retry == retry_count - 1:
|
|
|
|
raise
|
|
|
|
warn('Rename {} to {} failed, retrying...'.format(path_from, path_to))
|
|
|
|
|
|
|
|
|
2019-04-17 03:30:30 -04:00
|
|
|
def strip_container_dirs(path, levels):
|
|
|
|
assert levels > 0
|
|
|
|
# move the original directory out of the way (add a .tmp suffix)
|
|
|
|
tmp_path = path + '.tmp'
|
|
|
|
if os.path.exists(tmp_path):
|
|
|
|
shutil.rmtree(tmp_path)
|
2019-09-23 00:01:09 -04:00
|
|
|
rename_with_retry(path, tmp_path)
|
2019-04-17 03:30:30 -04:00
|
|
|
os.mkdir(path)
|
|
|
|
base_path = tmp_path
|
|
|
|
# walk given number of levels down
|
|
|
|
for level in range(levels):
|
|
|
|
contents = os.listdir(base_path)
|
|
|
|
if len(contents) > 1:
|
|
|
|
raise RuntimeError('at level {}, expected 1 entry, got {}'.format(level, contents))
|
|
|
|
base_path = os.path.join(base_path, contents[0])
|
|
|
|
if not os.path.isdir(base_path):
|
|
|
|
raise RuntimeError('at level {}, {} is not a directory'.format(level, contents[0]))
|
|
|
|
# get the list of directories/files to move
|
|
|
|
contents = os.listdir(base_path)
|
|
|
|
for name in contents:
|
|
|
|
move_from = os.path.join(base_path, name)
|
|
|
|
move_to = os.path.join(path, name)
|
2019-09-23 00:01:09 -04:00
|
|
|
rename_with_retry(move_from, move_to)
|
2019-04-17 03:30:30 -04:00
|
|
|
shutil.rmtree(tmp_path)
|
|
|
|
|
|
|
|
|
|
|
|
class ToolNotFound(RuntimeError):
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
class ToolExecError(RuntimeError):
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
class DownloadError(RuntimeError):
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
class IDFToolDownload(object):
|
|
|
|
def __init__(self, platform_name, url, size, sha256):
|
|
|
|
self.platform_name = platform_name
|
|
|
|
self.url = url
|
|
|
|
self.size = size
|
|
|
|
self.sha256 = sha256
|
|
|
|
self.platform_name = platform_name
|
|
|
|
|
|
|
|
|
|
|
|
@functools.total_ordering
|
|
|
|
class IDFToolVersion(object):
|
|
|
|
STATUS_RECOMMENDED = 'recommended'
|
|
|
|
STATUS_SUPPORTED = 'supported'
|
|
|
|
STATUS_DEPRECATED = 'deprecated'
|
|
|
|
|
|
|
|
STATUS_VALUES = [STATUS_RECOMMENDED, STATUS_SUPPORTED, STATUS_DEPRECATED]
|
|
|
|
|
|
|
|
def __init__(self, version, status):
|
|
|
|
self.version = version
|
|
|
|
self.status = status
|
|
|
|
self.downloads = OrderedDict()
|
|
|
|
self.latest = False
|
|
|
|
|
|
|
|
def __lt__(self, other):
|
|
|
|
if self.status != other.status:
|
|
|
|
return self.status > other.status
|
|
|
|
else:
|
|
|
|
assert not (self.status == IDFToolVersion.STATUS_RECOMMENDED
|
|
|
|
and other.status == IDFToolVersion.STATUS_RECOMMENDED)
|
|
|
|
return self.version < other.version
|
|
|
|
|
|
|
|
def __eq__(self, other):
|
|
|
|
return self.status == other.status and self.version == other.version
|
|
|
|
|
|
|
|
def add_download(self, platform_name, url, size, sha256):
|
|
|
|
self.downloads[platform_name] = IDFToolDownload(platform_name, url, size, sha256)
|
|
|
|
|
2020-01-17 11:38:50 -05:00
|
|
|
def get_download_for_platform(self, platform_name): # type: (str) -> IDFToolDownload
|
2019-04-17 03:30:30 -04:00
|
|
|
if platform_name in PLATFORM_FROM_NAME.keys():
|
|
|
|
platform_name = PLATFORM_FROM_NAME[platform_name]
|
|
|
|
if platform_name in self.downloads.keys():
|
|
|
|
return self.downloads[platform_name]
|
|
|
|
if 'any' in self.downloads.keys():
|
|
|
|
return self.downloads['any']
|
|
|
|
return None
|
|
|
|
|
|
|
|
def compatible_with_platform(self, platform_name=PYTHON_PLATFORM):
|
|
|
|
return self.get_download_for_platform(platform_name) is not None
|
|
|
|
|
2020-01-17 11:38:50 -05:00
|
|
|
def get_supported_platforms(self): # type: () -> typing.Set[str]
|
|
|
|
return set(self.downloads.keys())
|
|
|
|
|
2019-04-17 03:30:30 -04:00
|
|
|
|
|
|
|
OPTIONS_LIST = ['version_cmd',
|
|
|
|
'version_regex',
|
|
|
|
'version_regex_replace',
|
|
|
|
'export_paths',
|
|
|
|
'export_vars',
|
|
|
|
'install',
|
|
|
|
'info_url',
|
|
|
|
'license',
|
|
|
|
'strip_container_dirs']
|
|
|
|
|
|
|
|
IDFToolOptions = namedtuple('IDFToolOptions', OPTIONS_LIST)
|
|
|
|
|
|
|
|
|
|
|
|
class IDFTool(object):
|
|
|
|
# possible values of 'install' field
|
|
|
|
INSTALL_ALWAYS = 'always'
|
|
|
|
INSTALL_ON_REQUEST = 'on_request'
|
|
|
|
INSTALL_NEVER = 'never'
|
|
|
|
|
|
|
|
def __init__(self, name, description, install, info_url, license, version_cmd, version_regex, version_regex_replace=None,
|
|
|
|
strip_container_dirs=0):
|
|
|
|
self.name = name
|
|
|
|
self.description = description
|
2020-01-17 11:38:50 -05:00
|
|
|
self.versions = OrderedDict() # type: typing.Dict[str, IDFToolVersion]
|
2019-04-17 03:30:30 -04:00
|
|
|
self.version_in_path = None
|
|
|
|
self.versions_installed = []
|
|
|
|
if version_regex_replace is None:
|
|
|
|
version_regex_replace = VERSION_REGEX_REPLACE_DEFAULT
|
|
|
|
self.options = IDFToolOptions(version_cmd, version_regex, version_regex_replace,
|
|
|
|
[], OrderedDict(), install, info_url, license, strip_container_dirs)
|
|
|
|
self.platform_overrides = []
|
2019-07-18 11:20:40 -04:00
|
|
|
self._platform = CURRENT_PLATFORM
|
2019-04-17 03:30:30 -04:00
|
|
|
self._update_current_options()
|
|
|
|
|
2020-01-17 11:38:50 -05:00
|
|
|
def copy_for_platform(self, platform): # type: (str) -> IDFTool
|
2019-07-18 11:20:40 -04:00
|
|
|
result = copy.deepcopy(self)
|
|
|
|
result._platform = platform
|
|
|
|
result._update_current_options()
|
|
|
|
return result
|
|
|
|
|
2019-04-17 03:30:30 -04:00
|
|
|
def _update_current_options(self):
|
|
|
|
self._current_options = IDFToolOptions(*self.options)
|
|
|
|
for override in self.platform_overrides:
|
2019-07-18 11:20:40 -04:00
|
|
|
if self._platform not in override['platforms']:
|
2019-04-17 03:30:30 -04:00
|
|
|
continue
|
|
|
|
override_dict = override.copy()
|
|
|
|
del override_dict['platforms']
|
|
|
|
self._current_options = self._current_options._replace(**override_dict)
|
|
|
|
|
|
|
|
def add_version(self, version):
|
|
|
|
assert(type(version) is IDFToolVersion)
|
|
|
|
self.versions[version.version] = version
|
|
|
|
|
2020-01-17 11:38:50 -05:00
|
|
|
def get_path(self): # type: () -> str
|
2019-04-17 03:30:30 -04:00
|
|
|
return os.path.join(global_idf_tools_path, 'tools', self.name)
|
|
|
|
|
2020-01-17 11:38:50 -05:00
|
|
|
def get_path_for_version(self, version): # type: (str) -> str
|
2019-04-17 03:30:30 -04:00
|
|
|
assert(version in self.versions)
|
|
|
|
return os.path.join(self.get_path(), version)
|
|
|
|
|
2020-01-17 11:38:50 -05:00
|
|
|
def get_export_paths(self, version): # type: (str) -> typing.List[str]
|
2019-04-17 03:30:30 -04:00
|
|
|
tool_path = self.get_path_for_version(version)
|
|
|
|
return [os.path.join(tool_path, *p) for p in self._current_options.export_paths]
|
|
|
|
|
2020-01-17 11:38:50 -05:00
|
|
|
def get_export_vars(self, version): # type: (str) -> typing.Dict[str]
|
2019-04-17 03:30:30 -04:00
|
|
|
"""
|
|
|
|
Get the dictionary of environment variables to be exported, for the given version.
|
|
|
|
Expands:
|
|
|
|
- ${TOOL_PATH} => the actual path where the version is installed
|
|
|
|
"""
|
|
|
|
result = {}
|
|
|
|
for k, v in self._current_options.export_vars.items():
|
|
|
|
replace_path = self.get_path_for_version(version).replace('\\', '\\\\')
|
|
|
|
v_repl = re.sub(SUBST_TOOL_PATH_REGEX, replace_path, v)
|
|
|
|
if v_repl != v:
|
|
|
|
v_repl = to_shell_specific_paths([v_repl])[0]
|
|
|
|
result[k] = v_repl
|
|
|
|
return result
|
|
|
|
|
2020-01-17 11:38:50 -05:00
|
|
|
def check_version(self, extra_paths=None): # type: (typing.Optional[typing.List[str]]) -> str
|
2019-04-17 03:30:30 -04:00
|
|
|
"""
|
|
|
|
Execute the tool, optionally prepending extra_paths to PATH,
|
|
|
|
extract the version string and return it as a result.
|
|
|
|
Raises ToolNotFound if the tool is not found (not present in the paths).
|
|
|
|
Raises ToolExecError if the tool returns with a non-zero exit code.
|
|
|
|
Returns 'unknown' if tool returns something from which version string
|
|
|
|
can not be extracted.
|
|
|
|
"""
|
2019-07-18 11:20:40 -04:00
|
|
|
# this function can not be called for a different platform
|
|
|
|
assert self._platform == CURRENT_PLATFORM
|
2019-04-17 03:30:30 -04:00
|
|
|
cmd = self._current_options.version_cmd
|
|
|
|
try:
|
|
|
|
version_cmd_result = run_cmd_check_output(cmd, None, extra_paths)
|
|
|
|
except OSError:
|
|
|
|
# tool is not on the path
|
|
|
|
raise ToolNotFound('Tool {} not found'.format(self.name))
|
|
|
|
except subprocess.CalledProcessError as e:
|
2021-04-12 04:06:56 -04:00
|
|
|
raise ToolExecError('returned non-zero exit code ({}) with error message:\n{}'.format(
|
|
|
|
e.returncode, e.stderr.decode('utf-8',errors='ignore'))) # type: ignore
|
2019-04-17 03:30:30 -04:00
|
|
|
|
2021-01-25 21:49:01 -05:00
|
|
|
in_str = version_cmd_result.decode('utf-8')
|
2019-04-17 03:30:30 -04:00
|
|
|
match = re.search(self._current_options.version_regex, in_str)
|
|
|
|
if not match:
|
|
|
|
return UNKNOWN_VERSION
|
|
|
|
return re.sub(self._current_options.version_regex, self._current_options.version_regex_replace, match.group(0))
|
|
|
|
|
|
|
|
def get_install_type(self):
|
|
|
|
return self._current_options.install
|
|
|
|
|
2019-07-23 00:55:30 -04:00
|
|
|
def compatible_with_platform(self):
|
|
|
|
return any([v.compatible_with_platform() for v in self.versions.values()])
|
|
|
|
|
2020-01-17 11:38:50 -05:00
|
|
|
def get_supported_platforms(self): # type: () -> typing.Set[str]
|
|
|
|
result = set()
|
|
|
|
for v in self.versions.values():
|
|
|
|
result.update(v.get_supported_platforms())
|
|
|
|
return result
|
|
|
|
|
2019-04-17 03:30:30 -04:00
|
|
|
def get_recommended_version(self):
|
|
|
|
recommended_versions = [k for k, v in self.versions.items()
|
|
|
|
if v.status == IDFToolVersion.STATUS_RECOMMENDED
|
2019-07-18 11:20:40 -04:00
|
|
|
and v.compatible_with_platform(self._platform)]
|
2019-04-17 03:30:30 -04:00
|
|
|
assert len(recommended_versions) <= 1
|
|
|
|
if recommended_versions:
|
|
|
|
return recommended_versions[0]
|
|
|
|
return None
|
|
|
|
|
|
|
|
def get_preferred_installed_version(self):
|
|
|
|
recommended_versions = [k for k in self.versions_installed
|
|
|
|
if self.versions[k].status == IDFToolVersion.STATUS_RECOMMENDED
|
2019-07-18 11:20:40 -04:00
|
|
|
and self.versions[k].compatible_with_platform(self._platform)]
|
2019-04-17 03:30:30 -04:00
|
|
|
assert len(recommended_versions) <= 1
|
|
|
|
if recommended_versions:
|
|
|
|
return recommended_versions[0]
|
|
|
|
return None
|
|
|
|
|
|
|
|
def find_installed_versions(self):
|
|
|
|
"""
|
|
|
|
Checks whether the tool can be found in PATH and in global_idf_tools_path.
|
|
|
|
Writes results to self.version_in_path and self.versions_installed.
|
|
|
|
"""
|
2019-07-18 11:20:40 -04:00
|
|
|
# this function can not be called for a different platform
|
|
|
|
assert self._platform == CURRENT_PLATFORM
|
2019-04-17 03:30:30 -04:00
|
|
|
# First check if the tool is in system PATH
|
|
|
|
try:
|
|
|
|
ver_str = self.check_version()
|
|
|
|
except ToolNotFound:
|
|
|
|
# not in PATH
|
|
|
|
pass
|
2021-04-12 04:06:56 -04:00
|
|
|
except ToolExecError as e:
|
|
|
|
warn('tool {} found in path, but {}'.format(
|
|
|
|
self.name, e))
|
2019-04-17 03:30:30 -04:00
|
|
|
else:
|
|
|
|
self.version_in_path = ver_str
|
|
|
|
|
|
|
|
# Now check all the versions installed in global_idf_tools_path
|
|
|
|
self.versions_installed = []
|
|
|
|
for version, version_obj in self.versions.items():
|
|
|
|
if not version_obj.compatible_with_platform():
|
|
|
|
continue
|
|
|
|
tool_path = self.get_path_for_version(version)
|
|
|
|
if not os.path.exists(tool_path):
|
|
|
|
# version not installed
|
|
|
|
continue
|
|
|
|
try:
|
|
|
|
ver_str = self.check_version(self.get_export_paths(version))
|
|
|
|
except ToolNotFound:
|
|
|
|
warn('directory for tool {} version {} is present, but tool was not found'.format(
|
|
|
|
self.name, version))
|
2021-04-12 04:06:56 -04:00
|
|
|
except ToolExecError as e:
|
|
|
|
warn('tool {} version {} is installed, but {}'.format(
|
|
|
|
self.name, version, e))
|
2019-04-17 03:30:30 -04:00
|
|
|
else:
|
|
|
|
if ver_str != version:
|
|
|
|
warn('tool {} version {} is installed, but has reported version {}'.format(
|
|
|
|
self.name, version, ver_str))
|
|
|
|
else:
|
|
|
|
self.versions_installed.append(version)
|
|
|
|
|
|
|
|
def download(self, version):
|
|
|
|
assert(version in self.versions)
|
2019-07-18 11:20:40 -04:00
|
|
|
download_obj = self.versions[version].get_download_for_platform(self._platform)
|
2019-04-17 03:30:30 -04:00
|
|
|
if not download_obj:
|
2019-07-18 11:20:40 -04:00
|
|
|
fatal('No packages for tool {} platform {}!'.format(self.name, self._platform))
|
2019-04-17 03:30:30 -04:00
|
|
|
raise DownloadError()
|
|
|
|
|
|
|
|
url = download_obj.url
|
|
|
|
archive_name = os.path.basename(url)
|
|
|
|
local_path = os.path.join(global_idf_tools_path, 'dist', archive_name)
|
|
|
|
mkdir_p(os.path.dirname(local_path))
|
|
|
|
|
|
|
|
if os.path.isfile(local_path):
|
|
|
|
if not self.check_download_file(download_obj, local_path):
|
|
|
|
warn('removing downloaded file {0} and downloading again'.format(archive_name))
|
|
|
|
os.unlink(local_path)
|
|
|
|
else:
|
|
|
|
info('file {0} is already downloaded'.format(archive_name))
|
|
|
|
return
|
|
|
|
|
|
|
|
downloaded = False
|
|
|
|
for retry in range(DOWNLOAD_RETRY_COUNT):
|
|
|
|
local_temp_path = local_path + '.tmp'
|
|
|
|
info('Downloading {} to {}'.format(archive_name, local_temp_path))
|
2019-11-15 08:31:19 -05:00
|
|
|
try:
|
2020-05-20 03:50:57 -04:00
|
|
|
ctx = None
|
|
|
|
# For dl.espressif.com, add the ISRG x1 root certificate.
|
|
|
|
# This works around the issue with outdated certificate stores in some installations.
|
2021-01-25 21:49:01 -05:00
|
|
|
if 'dl.espressif.com' in url:
|
2020-10-14 14:19:32 -04:00
|
|
|
try:
|
|
|
|
ctx = ssl.create_default_context()
|
|
|
|
ctx.load_verify_locations(cadata=ISRG_X1_ROOT_CERT)
|
|
|
|
except AttributeError:
|
|
|
|
# no ssl.create_default_context or load_verify_locations cadata argument
|
|
|
|
# in Python <=2.7.8
|
|
|
|
pass
|
2020-05-20 03:50:57 -04:00
|
|
|
|
|
|
|
urlretrieve_ctx(url, local_temp_path, report_progress if not global_non_interactive else None, context=ctx)
|
2021-01-25 21:49:01 -05:00
|
|
|
sys.stdout.write('\rDone\n')
|
2019-11-15 08:31:19 -05:00
|
|
|
except Exception as e:
|
|
|
|
# urlretrieve could throw different exceptions, e.g. IOError when the server is down
|
|
|
|
# Errors are ignored because the downloaded file is checked a couple of lines later.
|
|
|
|
warn('Download failure {}'.format(e))
|
2019-04-17 03:30:30 -04:00
|
|
|
sys.stdout.flush()
|
2019-11-15 08:31:19 -05:00
|
|
|
if not os.path.isfile(local_temp_path) or not self.check_download_file(download_obj, local_temp_path):
|
|
|
|
warn('Failed to download {} to {}'.format(url, local_temp_path))
|
2019-04-17 03:30:30 -04:00
|
|
|
continue
|
2019-09-23 00:01:09 -04:00
|
|
|
rename_with_retry(local_temp_path, local_path)
|
2019-04-17 03:30:30 -04:00
|
|
|
downloaded = True
|
|
|
|
break
|
|
|
|
if not downloaded:
|
|
|
|
fatal('Failed to download, and retry count has expired')
|
|
|
|
raise DownloadError()
|
|
|
|
|
|
|
|
def install(self, version):
|
|
|
|
# Currently this is called after calling 'download' method, so here are a few asserts
|
|
|
|
# for the conditions which should be true once that method is done.
|
|
|
|
assert (version in self.versions)
|
2019-07-18 11:20:40 -04:00
|
|
|
download_obj = self.versions[version].get_download_for_platform(self._platform)
|
2019-04-17 03:30:30 -04:00
|
|
|
assert (download_obj is not None)
|
|
|
|
archive_name = os.path.basename(download_obj.url)
|
|
|
|
archive_path = os.path.join(global_idf_tools_path, 'dist', archive_name)
|
|
|
|
assert (os.path.isfile(archive_path))
|
|
|
|
dest_dir = self.get_path_for_version(version)
|
|
|
|
if os.path.exists(dest_dir):
|
|
|
|
warn('destination path already exists, removing')
|
|
|
|
shutil.rmtree(dest_dir)
|
|
|
|
mkdir_p(dest_dir)
|
|
|
|
unpack(archive_path, dest_dir)
|
|
|
|
if self._current_options.strip_container_dirs:
|
|
|
|
strip_container_dirs(dest_dir, self._current_options.strip_container_dirs)
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def check_download_file(download_obj, local_path):
|
|
|
|
expected_sha256 = download_obj.sha256
|
|
|
|
expected_size = download_obj.size
|
|
|
|
file_size, file_sha256 = get_file_size_sha256(local_path)
|
|
|
|
if file_size != expected_size:
|
2019-06-28 08:23:21 -04:00
|
|
|
warn('file size mismatch for {}, expected {}, got {}'.format(local_path, expected_size, file_size))
|
2019-04-17 03:30:30 -04:00
|
|
|
return False
|
|
|
|
if file_sha256 != expected_sha256:
|
2019-06-28 08:23:21 -04:00
|
|
|
warn('hash mismatch for {}, expected {}, got {}'.format(local_path, expected_sha256, file_sha256))
|
2019-04-17 03:30:30 -04:00
|
|
|
return False
|
|
|
|
return True
|
|
|
|
|
|
|
|
@classmethod
|
|
|
|
def from_json(cls, tool_dict):
|
|
|
|
# json.load will return 'str' types in Python 3 and 'unicode' in Python 2
|
|
|
|
expected_str_type = type(u'')
|
|
|
|
|
|
|
|
# Validate json fields
|
|
|
|
tool_name = tool_dict.get('name')
|
|
|
|
if type(tool_name) is not expected_str_type:
|
|
|
|
raise RuntimeError('tool_name is not a string')
|
|
|
|
|
|
|
|
description = tool_dict.get('description')
|
|
|
|
if type(description) is not expected_str_type:
|
|
|
|
raise RuntimeError('description is not a string')
|
|
|
|
|
|
|
|
version_cmd = tool_dict.get('version_cmd')
|
|
|
|
if type(version_cmd) is not list:
|
|
|
|
raise RuntimeError('version_cmd for tool %s is not a list of strings' % tool_name)
|
|
|
|
|
|
|
|
version_regex = tool_dict.get('version_regex')
|
|
|
|
if type(version_regex) is not expected_str_type or not version_regex:
|
|
|
|
raise RuntimeError('version_regex for tool %s is not a non-empty string' % tool_name)
|
|
|
|
|
|
|
|
version_regex_replace = tool_dict.get('version_regex_replace')
|
|
|
|
if version_regex_replace and type(version_regex_replace) is not expected_str_type:
|
|
|
|
raise RuntimeError('version_regex_replace for tool %s is not a string' % tool_name)
|
|
|
|
|
|
|
|
export_paths = tool_dict.get('export_paths')
|
|
|
|
if type(export_paths) is not list:
|
|
|
|
raise RuntimeError('export_paths for tool %s is not a list' % tool_name)
|
|
|
|
|
|
|
|
export_vars = tool_dict.get('export_vars', {})
|
|
|
|
if type(export_vars) is not dict:
|
|
|
|
raise RuntimeError('export_vars for tool %s is not a mapping' % tool_name)
|
|
|
|
|
|
|
|
versions = tool_dict.get('versions')
|
|
|
|
if type(versions) is not list:
|
|
|
|
raise RuntimeError('versions for tool %s is not an array' % tool_name)
|
|
|
|
|
|
|
|
install = tool_dict.get('install', False)
|
|
|
|
if type(install) is not expected_str_type:
|
|
|
|
raise RuntimeError('install for tool %s is not a string' % tool_name)
|
|
|
|
|
|
|
|
info_url = tool_dict.get('info_url', False)
|
|
|
|
if type(info_url) is not expected_str_type:
|
|
|
|
raise RuntimeError('info_url for tool %s is not a string' % tool_name)
|
|
|
|
|
|
|
|
license = tool_dict.get('license', False)
|
|
|
|
if type(license) is not expected_str_type:
|
|
|
|
raise RuntimeError('license for tool %s is not a string' % tool_name)
|
|
|
|
|
|
|
|
strip_container_dirs = tool_dict.get('strip_container_dirs', 0)
|
|
|
|
if strip_container_dirs and type(strip_container_dirs) is not int:
|
|
|
|
raise RuntimeError('strip_container_dirs for tool %s is not an int' % tool_name)
|
|
|
|
|
|
|
|
overrides_list = tool_dict.get('platform_overrides', [])
|
|
|
|
if type(overrides_list) is not list:
|
|
|
|
raise RuntimeError('platform_overrides for tool %s is not a list' % tool_name)
|
|
|
|
|
|
|
|
# Create the object
|
|
|
|
tool_obj = cls(tool_name, description, install, info_url, license,
|
|
|
|
version_cmd, version_regex, version_regex_replace,
|
|
|
|
strip_container_dirs)
|
|
|
|
|
|
|
|
for path in export_paths:
|
|
|
|
tool_obj.options.export_paths.append(path)
|
|
|
|
|
|
|
|
for name, value in export_vars.items():
|
|
|
|
tool_obj.options.export_vars[name] = value
|
|
|
|
|
|
|
|
for index, override in enumerate(overrides_list):
|
|
|
|
platforms_list = override.get('platforms')
|
|
|
|
if type(platforms_list) is not list:
|
|
|
|
raise RuntimeError('platforms for override %d of tool %s is not a list' % (index, tool_name))
|
|
|
|
|
|
|
|
install = override.get('install')
|
|
|
|
if install is not None and type(install) is not expected_str_type:
|
|
|
|
raise RuntimeError('install for override %d of tool %s is not a string' % (index, tool_name))
|
|
|
|
|
|
|
|
version_cmd = override.get('version_cmd')
|
|
|
|
if version_cmd is not None and type(version_cmd) is not list:
|
|
|
|
raise RuntimeError('version_cmd for override %d of tool %s is not a list of strings' %
|
|
|
|
(index, tool_name))
|
|
|
|
|
|
|
|
version_regex = override.get('version_regex')
|
|
|
|
if version_regex is not None and (type(version_regex) is not expected_str_type or not version_regex):
|
|
|
|
raise RuntimeError('version_regex for override %d of tool %s is not a non-empty string' %
|
|
|
|
(index, tool_name))
|
|
|
|
|
|
|
|
version_regex_replace = override.get('version_regex_replace')
|
|
|
|
if version_regex_replace is not None and type(version_regex_replace) is not expected_str_type:
|
|
|
|
raise RuntimeError('version_regex_replace for override %d of tool %s is not a string' %
|
|
|
|
(index, tool_name))
|
|
|
|
|
|
|
|
export_paths = override.get('export_paths')
|
|
|
|
if export_paths is not None and type(export_paths) is not list:
|
|
|
|
raise RuntimeError('export_paths for override %d of tool %s is not a list' % (index, tool_name))
|
|
|
|
|
|
|
|
export_vars = override.get('export_vars')
|
|
|
|
if export_vars is not None and type(export_vars) is not dict:
|
|
|
|
raise RuntimeError('export_vars for override %d of tool %s is not a mapping' % (index, tool_name))
|
|
|
|
tool_obj.platform_overrides.append(override)
|
|
|
|
|
|
|
|
recommended_versions = {}
|
|
|
|
for version_dict in versions:
|
|
|
|
version = version_dict.get('name')
|
|
|
|
if type(version) is not expected_str_type:
|
|
|
|
raise RuntimeError('version name for tool {} is not a string'.format(tool_name))
|
|
|
|
|
|
|
|
version_status = version_dict.get('status')
|
|
|
|
if type(version_status) is not expected_str_type and version_status not in IDFToolVersion.STATUS_VALUES:
|
|
|
|
raise RuntimeError('tool {} version {} status is not one of {}', tool_name, version,
|
|
|
|
IDFToolVersion.STATUS_VALUES)
|
|
|
|
|
|
|
|
version_obj = IDFToolVersion(version, version_status)
|
|
|
|
for platform_id, platform_dict in version_dict.items():
|
|
|
|
if platform_id in ['name', 'status']:
|
|
|
|
continue
|
|
|
|
if platform_id not in PLATFORM_FROM_NAME.keys():
|
|
|
|
raise RuntimeError('invalid platform %s for tool %s version %s' %
|
|
|
|
(platform_id, tool_name, version))
|
|
|
|
|
|
|
|
version_obj.add_download(platform_id,
|
|
|
|
platform_dict['url'], platform_dict['size'], platform_dict['sha256'])
|
|
|
|
|
|
|
|
if version_status == IDFToolVersion.STATUS_RECOMMENDED:
|
|
|
|
if platform_id not in recommended_versions:
|
|
|
|
recommended_versions[platform_id] = []
|
|
|
|
recommended_versions[platform_id].append(version)
|
|
|
|
|
|
|
|
tool_obj.add_version(version_obj)
|
|
|
|
for platform_id, version_list in recommended_versions.items():
|
|
|
|
if len(version_list) > 1:
|
|
|
|
raise RuntimeError('tool {} for platform {} has {} recommended versions'.format(
|
|
|
|
tool_name, platform_id, len(recommended_versions)))
|
|
|
|
if install != IDFTool.INSTALL_NEVER and len(recommended_versions) == 0:
|
|
|
|
raise RuntimeError('required/optional tool {} for platform {} has no recommended versions'.format(
|
|
|
|
tool_name, platform_id))
|
|
|
|
|
|
|
|
tool_obj._update_current_options()
|
|
|
|
return tool_obj
|
|
|
|
|
|
|
|
def to_json(self):
|
|
|
|
versions_array = []
|
|
|
|
for version, version_obj in self.versions.items():
|
|
|
|
version_json = {
|
|
|
|
'name': version,
|
|
|
|
'status': version_obj.status
|
|
|
|
}
|
|
|
|
for platform_id, download in version_obj.downloads.items():
|
|
|
|
version_json[platform_id] = {
|
|
|
|
'url': download.url,
|
|
|
|
'size': download.size,
|
|
|
|
'sha256': download.sha256
|
|
|
|
}
|
|
|
|
versions_array.append(version_json)
|
|
|
|
overrides_array = self.platform_overrides
|
|
|
|
|
|
|
|
tool_json = {
|
|
|
|
'name': self.name,
|
|
|
|
'description': self.description,
|
|
|
|
'export_paths': self.options.export_paths,
|
|
|
|
'export_vars': self.options.export_vars,
|
|
|
|
'install': self.options.install,
|
|
|
|
'info_url': self.options.info_url,
|
|
|
|
'license': self.options.license,
|
|
|
|
'version_cmd': self.options.version_cmd,
|
|
|
|
'version_regex': self.options.version_regex,
|
|
|
|
'versions': versions_array,
|
|
|
|
}
|
|
|
|
if self.options.version_regex_replace != VERSION_REGEX_REPLACE_DEFAULT:
|
|
|
|
tool_json['version_regex_replace'] = self.options.version_regex_replace
|
|
|
|
if overrides_array:
|
|
|
|
tool_json['platform_overrides'] = overrides_array
|
|
|
|
if self.options.strip_container_dirs:
|
|
|
|
tool_json['strip_container_dirs'] = self.options.strip_container_dirs
|
|
|
|
return tool_json
|
|
|
|
|
|
|
|
|
2020-01-17 11:38:50 -05:00
|
|
|
def load_tools_info(): # type: () -> typing.Dict[str, IDFTool]
|
2019-04-17 03:30:30 -04:00
|
|
|
"""
|
|
|
|
Load tools metadata from tools.json, return a dictionary: tool name - tool info
|
|
|
|
"""
|
|
|
|
tool_versions_file_name = global_tools_json
|
|
|
|
|
|
|
|
with open(tool_versions_file_name, 'r') as f:
|
|
|
|
tools_info = json.load(f)
|
|
|
|
|
|
|
|
return parse_tools_info_json(tools_info)
|
|
|
|
|
|
|
|
|
|
|
|
def parse_tools_info_json(tools_info):
|
|
|
|
"""
|
|
|
|
Parse and validate the dictionary obtained by loading the tools.json file.
|
|
|
|
Returns a dictionary of tools (key: tool name, value: IDFTool object).
|
|
|
|
"""
|
|
|
|
if tools_info['version'] != TOOLS_FILE_VERSION:
|
|
|
|
raise RuntimeError('Invalid version')
|
|
|
|
|
|
|
|
tools_dict = OrderedDict()
|
|
|
|
|
|
|
|
tools_array = tools_info.get('tools')
|
|
|
|
if type(tools_array) is not list:
|
|
|
|
raise RuntimeError('tools property is missing or not an array')
|
|
|
|
|
|
|
|
for tool_dict in tools_array:
|
|
|
|
tool = IDFTool.from_json(tool_dict)
|
|
|
|
tools_dict[tool.name] = tool
|
|
|
|
|
|
|
|
return tools_dict
|
|
|
|
|
|
|
|
|
|
|
|
def dump_tools_json(tools_info):
|
|
|
|
tools_array = []
|
|
|
|
for tool_name, tool_obj in tools_info.items():
|
|
|
|
tool_json = tool_obj.to_json()
|
|
|
|
tools_array.append(tool_json)
|
|
|
|
file_json = {'version': TOOLS_FILE_VERSION, 'tools': tools_array}
|
|
|
|
return json.dumps(file_json, indent=2, separators=(',', ': '), sort_keys=True)
|
|
|
|
|
|
|
|
|
|
|
|
def get_python_env_path():
|
|
|
|
python_ver_major_minor = '{}.{}'.format(sys.version_info.major, sys.version_info.minor)
|
|
|
|
|
|
|
|
version_file_path = os.path.join(global_idf_path, 'version.txt')
|
|
|
|
if os.path.exists(version_file_path):
|
2021-01-25 21:49:01 -05:00
|
|
|
with open(version_file_path, 'r') as version_file:
|
2019-04-17 03:30:30 -04:00
|
|
|
idf_version_str = version_file.read()
|
|
|
|
else:
|
2019-11-18 04:12:05 -05:00
|
|
|
try:
|
2020-10-06 18:36:40 -04:00
|
|
|
idf_version_str = subprocess.check_output(['git', 'describe'],
|
2019-11-18 04:12:05 -05:00
|
|
|
cwd=global_idf_path, env=os.environ).decode()
|
|
|
|
except subprocess.CalledProcessError as e:
|
|
|
|
warn('Git describe was unsuccessul: {}'.format(e))
|
|
|
|
idf_version_str = ''
|
2019-04-17 03:30:30 -04:00
|
|
|
match = re.match(r'^v([0-9]+\.[0-9]+).*', idf_version_str)
|
2019-11-18 04:12:05 -05:00
|
|
|
if match:
|
|
|
|
idf_version = match.group(1)
|
|
|
|
else:
|
|
|
|
idf_version = None
|
|
|
|
# fallback when IDF is a shallow clone
|
|
|
|
try:
|
|
|
|
with open(os.path.join(global_idf_path, 'components', 'esp_common', 'include', 'esp_idf_version.h')) as f:
|
|
|
|
m = re.search(r'^#define\s+ESP_IDF_VERSION_MAJOR\s+(\d+).+?^#define\s+ESP_IDF_VERSION_MINOR\s+(\d+)',
|
|
|
|
f.read(), re.DOTALL | re.MULTILINE)
|
|
|
|
if m:
|
|
|
|
idf_version = '.'.join((m.group(1), m.group(2)))
|
|
|
|
else:
|
|
|
|
warn('Reading IDF version from C header file failed!')
|
|
|
|
except Exception as e:
|
|
|
|
warn('Is it not possible to determine the IDF version: {}'.format(e))
|
|
|
|
|
|
|
|
if idf_version is None:
|
|
|
|
fatal('IDF version cannot be determined')
|
|
|
|
raise SystemExit(1)
|
2019-04-17 03:30:30 -04:00
|
|
|
|
|
|
|
idf_python_env_path = os.path.join(global_idf_tools_path, 'python_env',
|
|
|
|
'idf{}_py{}_env'.format(idf_version, python_ver_major_minor))
|
|
|
|
|
|
|
|
if sys.platform == 'win32':
|
|
|
|
subdir = 'Scripts'
|
|
|
|
python_exe = 'python.exe'
|
|
|
|
else:
|
|
|
|
subdir = 'bin'
|
|
|
|
python_exe = 'python'
|
|
|
|
|
|
|
|
idf_python_export_path = os.path.join(idf_python_env_path, subdir)
|
|
|
|
virtualenv_python = os.path.join(idf_python_export_path, python_exe)
|
|
|
|
|
|
|
|
return idf_python_env_path, idf_python_export_path, virtualenv_python
|
|
|
|
|
|
|
|
|
|
|
|
def action_list(args):
|
|
|
|
tools_info = load_tools_info()
|
|
|
|
for name, tool in tools_info.items():
|
|
|
|
if tool.get_install_type() == IDFTool.INSTALL_NEVER:
|
|
|
|
continue
|
|
|
|
optional_str = ' (optional)' if tool.get_install_type() == IDFTool.INSTALL_ON_REQUEST else ''
|
|
|
|
info('* {}: {}{}'.format(name, tool.description, optional_str))
|
|
|
|
tool.find_installed_versions()
|
|
|
|
versions_for_platform = {k: v for k, v in tool.versions.items() if v.compatible_with_platform()}
|
|
|
|
if not versions_for_platform:
|
|
|
|
info(' (no versions compatible with platform {})'.format(PYTHON_PLATFORM))
|
|
|
|
continue
|
|
|
|
versions_sorted = sorted(versions_for_platform.keys(), key=tool.versions.get, reverse=True)
|
|
|
|
for version in versions_sorted:
|
|
|
|
version_obj = tool.versions[version]
|
|
|
|
info(' - {} ({}{})'.format(version, version_obj.status,
|
|
|
|
', installed' if version in tool.versions_installed else ''))
|
|
|
|
|
|
|
|
|
|
|
|
def action_check(args):
|
|
|
|
tools_info = load_tools_info()
|
|
|
|
not_found_list = []
|
|
|
|
info('Checking for installed tools...')
|
|
|
|
for name, tool in tools_info.items():
|
|
|
|
if tool.get_install_type() == IDFTool.INSTALL_NEVER:
|
|
|
|
continue
|
|
|
|
tool_found_somewhere = False
|
|
|
|
info('Checking tool %s' % name)
|
|
|
|
tool.find_installed_versions()
|
|
|
|
if tool.version_in_path:
|
|
|
|
info(' version found in PATH: %s' % tool.version_in_path)
|
|
|
|
tool_found_somewhere = True
|
|
|
|
else:
|
|
|
|
info(' no version found in PATH')
|
|
|
|
|
|
|
|
for version in tool.versions_installed:
|
|
|
|
info(' version installed in tools directory: %s' % version)
|
|
|
|
tool_found_somewhere = True
|
|
|
|
if not tool_found_somewhere and tool.get_install_type() == IDFTool.INSTALL_ALWAYS:
|
|
|
|
not_found_list.append(name)
|
|
|
|
if not_found_list:
|
|
|
|
fatal('The following required tools were not found: ' + ' '.join(not_found_list))
|
|
|
|
raise SystemExit(1)
|
|
|
|
|
|
|
|
|
|
|
|
def action_export(args):
|
|
|
|
tools_info = load_tools_info()
|
|
|
|
all_tools_found = True
|
|
|
|
export_vars = {}
|
|
|
|
paths_to_export = []
|
|
|
|
for name, tool in tools_info.items():
|
|
|
|
if tool.get_install_type() == IDFTool.INSTALL_NEVER:
|
|
|
|
continue
|
|
|
|
tool.find_installed_versions()
|
|
|
|
|
|
|
|
if tool.version_in_path:
|
|
|
|
if tool.version_in_path not in tool.versions:
|
|
|
|
# unsupported version
|
|
|
|
if args.prefer_system:
|
|
|
|
warn('using an unsupported version of tool {} found in PATH: {}'.format(
|
|
|
|
tool.name, tool.version_in_path))
|
|
|
|
continue
|
|
|
|
else:
|
|
|
|
# unsupported version in path
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
# supported/deprecated version in PATH, use it
|
|
|
|
version_obj = tool.versions[tool.version_in_path]
|
|
|
|
if version_obj.status == IDFToolVersion.STATUS_SUPPORTED:
|
|
|
|
info('Using a supported version of tool {} found in PATH: {}.'.format(name, tool.version_in_path),
|
|
|
|
f=sys.stderr)
|
|
|
|
info('However the recommended version is {}.'.format(tool.get_recommended_version()),
|
|
|
|
f=sys.stderr)
|
|
|
|
elif version_obj.status == IDFToolVersion.STATUS_DEPRECATED:
|
|
|
|
warn('using a deprecated version of tool {} found in PATH: {}'.format(name, tool.version_in_path))
|
|
|
|
continue
|
|
|
|
|
|
|
|
self_restart_cmd = '{} {}{}'.format(sys.executable, __file__,
|
|
|
|
(' --tools-json ' + args.tools_json) if args.tools_json else '')
|
|
|
|
self_restart_cmd = to_shell_specific_paths([self_restart_cmd])[0]
|
|
|
|
|
|
|
|
if IDF_TOOLS_EXPORT_CMD:
|
|
|
|
prefer_system_hint = ''
|
|
|
|
else:
|
|
|
|
prefer_system_hint = ' To use it, run \'{} export --prefer-system\''.format(self_restart_cmd)
|
|
|
|
|
|
|
|
if IDF_TOOLS_INSTALL_CMD:
|
|
|
|
install_cmd = to_shell_specific_paths([IDF_TOOLS_INSTALL_CMD])[0]
|
|
|
|
else:
|
|
|
|
install_cmd = self_restart_cmd + ' install'
|
|
|
|
|
|
|
|
if not tool.versions_installed:
|
|
|
|
if tool.get_install_type() == IDFTool.INSTALL_ALWAYS:
|
|
|
|
all_tools_found = False
|
|
|
|
fatal('tool {} has no installed versions. Please run \'{}\' to install it.'.format(
|
|
|
|
tool.name, install_cmd))
|
|
|
|
if tool.version_in_path and tool.version_in_path not in tool.versions:
|
|
|
|
info('An unsupported version of tool {} was found in PATH: {}. '.format(name, tool.version_in_path) +
|
|
|
|
prefer_system_hint, f=sys.stderr)
|
|
|
|
continue
|
|
|
|
else:
|
|
|
|
# tool is optional, and does not have versions installed
|
|
|
|
# use whatever is available in PATH
|
|
|
|
continue
|
|
|
|
|
|
|
|
if tool.version_in_path and tool.version_in_path not in tool.versions:
|
|
|
|
info('Not using an unsupported version of tool {} found in PATH: {}.'.format(
|
|
|
|
tool.name, tool.version_in_path) + prefer_system_hint, f=sys.stderr)
|
|
|
|
|
|
|
|
version_to_use = tool.get_preferred_installed_version()
|
|
|
|
export_paths = tool.get_export_paths(version_to_use)
|
|
|
|
if export_paths:
|
|
|
|
paths_to_export += export_paths
|
|
|
|
tool_export_vars = tool.get_export_vars(version_to_use)
|
|
|
|
for k, v in tool_export_vars.items():
|
|
|
|
old_v = os.environ.get(k)
|
|
|
|
if old_v is None or old_v != v:
|
|
|
|
export_vars[k] = v
|
|
|
|
|
|
|
|
current_path = os.getenv('PATH')
|
|
|
|
idf_python_env_path, idf_python_export_path, virtualenv_python = get_python_env_path()
|
|
|
|
if os.path.exists(virtualenv_python):
|
|
|
|
idf_python_env_path = to_shell_specific_paths([idf_python_env_path])[0]
|
|
|
|
if os.getenv('IDF_PYTHON_ENV_PATH') != idf_python_env_path:
|
|
|
|
export_vars['IDF_PYTHON_ENV_PATH'] = to_shell_specific_paths([idf_python_env_path])[0]
|
|
|
|
if idf_python_export_path not in current_path:
|
|
|
|
paths_to_export.append(idf_python_export_path)
|
|
|
|
|
|
|
|
idf_tools_dir = os.path.join(global_idf_path, 'tools')
|
|
|
|
idf_tools_dir = to_shell_specific_paths([idf_tools_dir])[0]
|
|
|
|
if idf_tools_dir not in current_path:
|
|
|
|
paths_to_export.append(idf_tools_dir)
|
|
|
|
|
|
|
|
if sys.platform == 'win32' and 'MSYSTEM' not in os.environ:
|
|
|
|
old_path = '%PATH%'
|
|
|
|
path_sep = ';'
|
|
|
|
else:
|
|
|
|
old_path = '$PATH'
|
|
|
|
# can't trust os.pathsep here, since for Windows Python started from MSYS shell,
|
|
|
|
# os.pathsep will be ';'
|
|
|
|
path_sep = ':'
|
|
|
|
|
|
|
|
if args.format == EXPORT_SHELL:
|
|
|
|
if sys.platform == 'win32' and 'MSYSTEM' not in os.environ:
|
|
|
|
export_format = 'SET "{}={}"'
|
|
|
|
export_sep = '\n'
|
|
|
|
else:
|
|
|
|
export_format = 'export {}="{}"'
|
|
|
|
export_sep = ';'
|
|
|
|
elif args.format == EXPORT_KEY_VALUE:
|
|
|
|
export_format = '{}={}'
|
|
|
|
export_sep = '\n'
|
|
|
|
else:
|
|
|
|
raise NotImplementedError('unsupported export format {}'.format(args.format))
|
|
|
|
|
|
|
|
if paths_to_export:
|
|
|
|
export_vars['PATH'] = path_sep.join(to_shell_specific_paths(paths_to_export) + [old_path])
|
|
|
|
|
|
|
|
export_statements = export_sep.join([export_format.format(k, v) for k, v in export_vars.items()])
|
|
|
|
|
|
|
|
if export_statements:
|
|
|
|
print(export_statements)
|
|
|
|
|
|
|
|
if not all_tools_found:
|
|
|
|
raise SystemExit(1)
|
|
|
|
|
|
|
|
|
2020-09-24 19:58:49 -04:00
|
|
|
def apply_url_mirrors(args, tool_download_obj):
|
|
|
|
apply_mirror_prefix_map(args, tool_download_obj)
|
|
|
|
apply_github_assets_option(tool_download_obj)
|
|
|
|
|
|
|
|
|
2019-07-18 11:20:40 -04:00
|
|
|
def apply_mirror_prefix_map(args, tool_download_obj):
|
2019-04-17 03:30:30 -04:00
|
|
|
"""Rewrite URL for given tool_obj, given tool_version, and current platform,
|
|
|
|
if --mirror-prefix-map flag or IDF_MIRROR_PREFIX_MAP environment variable is given.
|
|
|
|
"""
|
|
|
|
mirror_prefix_map = None
|
|
|
|
mirror_prefix_map_env = os.getenv('IDF_MIRROR_PREFIX_MAP')
|
|
|
|
if mirror_prefix_map_env:
|
|
|
|
mirror_prefix_map = mirror_prefix_map_env.split(';')
|
|
|
|
if IDF_MAINTAINER and args.mirror_prefix_map:
|
|
|
|
if mirror_prefix_map:
|
|
|
|
warn('Both IDF_MIRROR_PREFIX_MAP environment variable and --mirror-prefix-map flag are specified, ' +
|
|
|
|
'will use the value from the command line.')
|
|
|
|
mirror_prefix_map = args.mirror_prefix_map
|
2019-07-18 11:20:40 -04:00
|
|
|
if mirror_prefix_map and tool_download_obj:
|
2019-04-17 03:30:30 -04:00
|
|
|
for item in mirror_prefix_map:
|
|
|
|
if URL_PREFIX_MAP_SEPARATOR not in item:
|
|
|
|
warn('invalid mirror-prefix-map item (missing \'{}\') {}'.format(URL_PREFIX_MAP_SEPARATOR, item))
|
|
|
|
continue
|
|
|
|
search, replace = item.split(URL_PREFIX_MAP_SEPARATOR, 1)
|
2019-07-18 11:20:40 -04:00
|
|
|
old_url = tool_download_obj.url
|
2019-04-17 03:30:30 -04:00
|
|
|
new_url = re.sub(search, replace, old_url)
|
|
|
|
if new_url != old_url:
|
|
|
|
info('Changed download URL: {} => {}'.format(old_url, new_url))
|
2019-07-18 11:20:40 -04:00
|
|
|
tool_download_obj.url = new_url
|
2019-04-17 03:30:30 -04:00
|
|
|
break
|
|
|
|
|
|
|
|
|
2020-09-24 19:58:49 -04:00
|
|
|
def apply_github_assets_option(tool_download_obj):
|
|
|
|
""" Rewrite URL for given tool_obj if the download URL is an https://github.com/ URL and the variable
|
|
|
|
IDF_GITHUB_ASSETS is set. The github.com part of the URL will be replaced.
|
|
|
|
"""
|
|
|
|
try:
|
2021-01-25 21:49:01 -05:00
|
|
|
github_assets = os.environ['IDF_GITHUB_ASSETS'].strip()
|
2020-09-24 19:58:49 -04:00
|
|
|
except KeyError:
|
|
|
|
return # no IDF_GITHUB_ASSETS
|
|
|
|
if not github_assets: # variable exists but is empty
|
|
|
|
return
|
|
|
|
|
|
|
|
# check no URL qualifier in the mirror URL
|
|
|
|
if '://' in github_assets:
|
|
|
|
fatal("IDF_GITHUB_ASSETS shouldn't include any URL qualifier, https:// is assumed")
|
|
|
|
raise SystemExit(1)
|
|
|
|
|
|
|
|
# Strip any trailing / from the mirror URL
|
|
|
|
github_assets = github_assets.rstrip('/')
|
|
|
|
|
|
|
|
old_url = tool_download_obj.url
|
|
|
|
new_url = re.sub(r'^https://github.com/', 'https://{}/'.format(github_assets), old_url)
|
|
|
|
if new_url != old_url:
|
|
|
|
info('Using GitHub assets mirror for URL: {} => {}'.format(old_url, new_url))
|
|
|
|
tool_download_obj.url = new_url
|
|
|
|
|
|
|
|
|
2019-07-18 11:20:40 -04:00
|
|
|
def action_download(args):
|
|
|
|
tools_info = load_tools_info()
|
|
|
|
tools_spec = args.tools
|
|
|
|
|
|
|
|
if args.platform not in PLATFORM_FROM_NAME:
|
|
|
|
fatal('unknown platform: {}' % args.platform)
|
|
|
|
raise SystemExit(1)
|
|
|
|
platform = PLATFORM_FROM_NAME[args.platform]
|
|
|
|
|
|
|
|
tools_info_for_platform = OrderedDict()
|
|
|
|
for name, tool_obj in tools_info.items():
|
|
|
|
tool_for_platform = tool_obj.copy_for_platform(platform)
|
|
|
|
tools_info_for_platform[name] = tool_for_platform
|
|
|
|
|
2019-11-24 06:11:33 -05:00
|
|
|
if not tools_spec or 'required' in tools_spec:
|
|
|
|
tools_spec = [k for k, v in tools_info_for_platform.items() if v.get_install_type() == IDFTool.INSTALL_ALWAYS]
|
|
|
|
info('Downloading tools for {}: {}'.format(platform, ', '.join(tools_spec)))
|
|
|
|
elif 'all' in tools_spec:
|
2019-07-18 11:20:40 -04:00
|
|
|
tools_spec = [k for k, v in tools_info_for_platform.items() if v.get_install_type() != IDFTool.INSTALL_NEVER]
|
|
|
|
info('Downloading tools for {}: {}'.format(platform, ', '.join(tools_spec)))
|
|
|
|
|
|
|
|
for tool_spec in tools_spec:
|
|
|
|
if '@' not in tool_spec:
|
|
|
|
tool_name = tool_spec
|
|
|
|
tool_version = None
|
|
|
|
else:
|
|
|
|
tool_name, tool_version = tool_spec.split('@', 1)
|
|
|
|
if tool_name not in tools_info_for_platform:
|
|
|
|
fatal('unknown tool name: {}'.format(tool_name))
|
|
|
|
raise SystemExit(1)
|
|
|
|
tool_obj = tools_info_for_platform[tool_name]
|
|
|
|
if tool_version is not None and tool_version not in tool_obj.versions:
|
|
|
|
fatal('unknown version for tool {}: {}'.format(tool_name, tool_version))
|
|
|
|
raise SystemExit(1)
|
|
|
|
if tool_version is None:
|
|
|
|
tool_version = tool_obj.get_recommended_version()
|
2019-11-24 06:11:33 -05:00
|
|
|
if tool_version is None:
|
|
|
|
fatal('tool {} not found for {} platform'.format(tool_name, platform))
|
|
|
|
raise SystemExit(1)
|
2019-07-18 11:20:40 -04:00
|
|
|
tool_spec = '{}@{}'.format(tool_name, tool_version)
|
|
|
|
|
|
|
|
info('Downloading {}'.format(tool_spec))
|
2020-09-24 19:58:49 -04:00
|
|
|
apply_url_mirrors(args, tool_obj.versions[tool_version].get_download_for_platform(platform))
|
2019-07-18 11:20:40 -04:00
|
|
|
|
|
|
|
tool_obj.download(tool_version)
|
|
|
|
|
|
|
|
|
2019-04-17 03:30:30 -04:00
|
|
|
def action_install(args):
|
|
|
|
tools_info = load_tools_info()
|
|
|
|
tools_spec = args.tools
|
2019-11-24 06:11:33 -05:00
|
|
|
if not tools_spec or 'required' in tools_spec:
|
2019-04-17 03:30:30 -04:00
|
|
|
tools_spec = [k for k, v in tools_info.items() if v.get_install_type() == IDFTool.INSTALL_ALWAYS]
|
|
|
|
info('Installing tools: {}'.format(', '.join(tools_spec)))
|
|
|
|
elif 'all' in tools_spec:
|
|
|
|
tools_spec = [k for k, v in tools_info.items() if v.get_install_type() != IDFTool.INSTALL_NEVER]
|
|
|
|
info('Installing tools: {}'.format(', '.join(tools_spec)))
|
|
|
|
|
|
|
|
for tool_spec in tools_spec:
|
|
|
|
if '@' not in tool_spec:
|
|
|
|
tool_name = tool_spec
|
|
|
|
tool_version = None
|
|
|
|
else:
|
|
|
|
tool_name, tool_version = tool_spec.split('@', 1)
|
|
|
|
if tool_name not in tools_info:
|
|
|
|
fatal('unknown tool name: {}'.format(tool_name))
|
|
|
|
raise SystemExit(1)
|
|
|
|
tool_obj = tools_info[tool_name]
|
2019-07-23 00:55:30 -04:00
|
|
|
if not tool_obj.compatible_with_platform():
|
|
|
|
fatal('tool {} does not have versions compatible with platform {}'.format(tool_name, CURRENT_PLATFORM))
|
|
|
|
raise SystemExit(1)
|
2019-04-17 03:30:30 -04:00
|
|
|
if tool_version is not None and tool_version not in tool_obj.versions:
|
|
|
|
fatal('unknown version for tool {}: {}'.format(tool_name, tool_version))
|
|
|
|
raise SystemExit(1)
|
|
|
|
if tool_version is None:
|
|
|
|
tool_version = tool_obj.get_recommended_version()
|
|
|
|
assert tool_version is not None
|
|
|
|
tool_obj.find_installed_versions()
|
|
|
|
tool_spec = '{}@{}'.format(tool_name, tool_version)
|
|
|
|
if tool_version in tool_obj.versions_installed:
|
|
|
|
info('Skipping {} (already installed)'.format(tool_spec))
|
|
|
|
continue
|
|
|
|
|
|
|
|
info('Installing {}'.format(tool_spec))
|
2020-09-24 19:58:49 -04:00
|
|
|
apply_url_mirrors(args, tool_obj.versions[tool_version].get_download_for_platform(PYTHON_PLATFORM))
|
2019-04-17 03:30:30 -04:00
|
|
|
|
|
|
|
tool_obj.download(tool_version)
|
|
|
|
tool_obj.install(tool_version)
|
|
|
|
|
|
|
|
|
2020-12-16 11:40:13 -05:00
|
|
|
def get_wheels_dir():
|
|
|
|
tools_info = load_tools_info()
|
|
|
|
wheels_package_name = 'idf-python-wheels'
|
|
|
|
if wheels_package_name not in tools_info:
|
|
|
|
return None
|
|
|
|
wheels_package = tools_info[wheels_package_name]
|
|
|
|
recommended_version = wheels_package.get_recommended_version()
|
2021-01-18 11:26:20 -05:00
|
|
|
if recommended_version is None:
|
|
|
|
return None
|
2020-12-16 11:40:13 -05:00
|
|
|
wheels_dir = wheels_package.get_path_for_version(recommended_version)
|
|
|
|
if not os.path.exists(wheels_dir):
|
|
|
|
return None
|
|
|
|
return wheels_dir
|
|
|
|
|
|
|
|
|
2021-03-04 03:52:34 -05:00
|
|
|
def action_install_python_env(args): # type: ignore
|
|
|
|
reinstall = args.reinstall
|
2019-04-17 03:30:30 -04:00
|
|
|
idf_python_env_path, _, virtualenv_python = get_python_env_path()
|
|
|
|
|
2020-02-26 08:02:12 -05:00
|
|
|
is_virtualenv = hasattr(sys, 'real_prefix') or (hasattr(sys, 'base_prefix') and sys.base_prefix != sys.prefix)
|
2021-03-04 03:52:34 -05:00
|
|
|
if is_virtualenv and (not os.path.exists(idf_python_env_path) or reinstall):
|
2020-02-26 08:02:12 -05:00
|
|
|
fatal('This script was called from a virtual environment, can not create a virtual environment again')
|
|
|
|
raise SystemExit(1)
|
|
|
|
|
2021-03-04 03:52:34 -05:00
|
|
|
if os.path.exists(virtualenv_python):
|
|
|
|
try:
|
|
|
|
subprocess.check_call([virtualenv_python, '--version'], stdout=sys.stdout, stderr=sys.stderr)
|
|
|
|
except subprocess.CalledProcessError:
|
|
|
|
# At this point we can reinstall the virtual environment if it is non-functional. This can happen at least
|
|
|
|
# when the Python interpreter was removed which was used to create the virtual environment.
|
|
|
|
reinstall = True
|
|
|
|
|
|
|
|
try:
|
|
|
|
subprocess.check_call([virtualenv_python, '-m', 'pip', '--version'], stdout=sys.stdout, stderr=sys.stderr)
|
|
|
|
except subprocess.CalledProcessError:
|
|
|
|
warn('PIP is not available in the virtual environment.')
|
|
|
|
# Reinstallation of the virtual environment could help if PIP was installed for the main Python
|
|
|
|
reinstall = True
|
|
|
|
|
|
|
|
if reinstall and os.path.exists(idf_python_env_path):
|
2019-04-17 03:30:30 -04:00
|
|
|
warn('Removing the existing Python environment in {}'.format(idf_python_env_path))
|
|
|
|
shutil.rmtree(idf_python_env_path)
|
|
|
|
|
|
|
|
if not os.path.exists(virtualenv_python):
|
|
|
|
info('Creating a new Python environment in {}'.format(idf_python_env_path))
|
|
|
|
|
|
|
|
try:
|
2021-01-25 21:49:01 -05:00
|
|
|
import virtualenv # noqa: F401
|
2019-04-17 03:30:30 -04:00
|
|
|
except ImportError:
|
|
|
|
info('Installing virtualenv')
|
|
|
|
subprocess.check_call([sys.executable, '-m', 'pip', 'install', '--user', 'virtualenv'],
|
|
|
|
stdout=sys.stdout, stderr=sys.stderr)
|
|
|
|
|
2020-02-10 13:14:41 -05:00
|
|
|
subprocess.check_call([sys.executable, '-m', 'virtualenv', idf_python_env_path],
|
2019-04-17 03:30:30 -04:00
|
|
|
stdout=sys.stdout, stderr=sys.stderr)
|
|
|
|
run_args = [virtualenv_python, '-m', 'pip', 'install', '--no-warn-script-location']
|
|
|
|
requirements_txt = os.path.join(global_idf_path, 'requirements.txt')
|
|
|
|
run_args += ['-r', requirements_txt]
|
|
|
|
if args.extra_wheels_dir:
|
|
|
|
run_args += ['--find-links', args.extra_wheels_dir]
|
2020-12-16 11:40:13 -05:00
|
|
|
if args.no_index:
|
|
|
|
run_args += ['--no-index']
|
|
|
|
if args.extra_wheels_url:
|
|
|
|
run_args += ['--extra-index-url', args.extra_wheels_url]
|
|
|
|
|
|
|
|
wheels_dir = get_wheels_dir()
|
|
|
|
if wheels_dir is not None:
|
|
|
|
run_args += ['--find-links', wheels_dir]
|
|
|
|
|
2019-04-17 03:30:30 -04:00
|
|
|
info('Installing Python packages from {}'.format(requirements_txt))
|
|
|
|
subprocess.check_call(run_args, stdout=sys.stdout, stderr=sys.stderr)
|
|
|
|
|
|
|
|
|
|
|
|
def action_add_version(args):
|
|
|
|
tools_info = load_tools_info()
|
|
|
|
tool_name = args.tool
|
|
|
|
tool_obj = tools_info.get(tool_name)
|
|
|
|
if not tool_obj:
|
|
|
|
info('Creating new tool entry for {}'.format(tool_name))
|
|
|
|
tool_obj = IDFTool(tool_name, TODO_MESSAGE, IDFTool.INSTALL_ALWAYS,
|
|
|
|
TODO_MESSAGE, TODO_MESSAGE, [TODO_MESSAGE], TODO_MESSAGE)
|
|
|
|
tools_info[tool_name] = tool_obj
|
|
|
|
version = args.version
|
|
|
|
version_obj = tool_obj.versions.get(version)
|
|
|
|
if version not in tool_obj.versions:
|
|
|
|
info('Creating new version {}'.format(version))
|
|
|
|
version_obj = IDFToolVersion(version, IDFToolVersion.STATUS_SUPPORTED)
|
|
|
|
tool_obj.versions[version] = version_obj
|
|
|
|
url_prefix = args.url_prefix or 'https://%s/' % TODO_MESSAGE
|
|
|
|
for file_path in args.files:
|
|
|
|
file_name = os.path.basename(file_path)
|
|
|
|
# Guess which platform this file is for
|
|
|
|
found_platform = None
|
|
|
|
for platform_alias, platform_id in PLATFORM_FROM_NAME.items():
|
|
|
|
if platform_alias in file_name:
|
|
|
|
found_platform = platform_id
|
|
|
|
break
|
|
|
|
if found_platform is None:
|
|
|
|
info('Could not guess platform for file {}'.format(file_name))
|
|
|
|
found_platform = TODO_MESSAGE
|
|
|
|
# Get file size and calculate the SHA256
|
|
|
|
file_size, file_sha256 = get_file_size_sha256(file_path)
|
|
|
|
url = url_prefix + file_name
|
|
|
|
info('Adding download for platform {}'.format(found_platform))
|
|
|
|
info(' size: {}'.format(file_size))
|
|
|
|
info(' SHA256: {}'.format(file_sha256))
|
|
|
|
info(' URL: {}'.format(url))
|
|
|
|
version_obj.add_download(found_platform, url, file_size, file_sha256)
|
|
|
|
json_str = dump_tools_json(tools_info)
|
|
|
|
if not args.output:
|
|
|
|
args.output = os.path.join(global_idf_path, TOOLS_FILE_NEW)
|
|
|
|
with open(args.output, 'w') as f:
|
|
|
|
f.write(json_str)
|
|
|
|
f.write('\n')
|
|
|
|
info('Wrote output to {}'.format(args.output))
|
|
|
|
|
|
|
|
|
|
|
|
def action_rewrite(args):
|
|
|
|
tools_info = load_tools_info()
|
|
|
|
json_str = dump_tools_json(tools_info)
|
|
|
|
if not args.output:
|
|
|
|
args.output = os.path.join(global_idf_path, TOOLS_FILE_NEW)
|
|
|
|
with open(args.output, 'w') as f:
|
|
|
|
f.write(json_str)
|
|
|
|
f.write('\n')
|
|
|
|
info('Wrote output to {}'.format(args.output))
|
|
|
|
|
|
|
|
|
|
|
|
def action_validate(args):
|
|
|
|
try:
|
|
|
|
import jsonschema
|
|
|
|
except ImportError:
|
|
|
|
fatal('You need to install jsonschema package to use validate command')
|
|
|
|
raise SystemExit(1)
|
|
|
|
|
|
|
|
with open(os.path.join(global_idf_path, TOOLS_FILE), 'r') as tools_file:
|
|
|
|
tools_json = json.load(tools_file)
|
|
|
|
|
|
|
|
with open(os.path.join(global_idf_path, TOOLS_SCHEMA_FILE), 'r') as schema_file:
|
|
|
|
schema_json = json.load(schema_file)
|
|
|
|
jsonschema.validate(tools_json, schema_json)
|
|
|
|
# on failure, this will raise an exception with a fairly verbose diagnostic message
|
|
|
|
|
|
|
|
|
2020-01-17 11:38:50 -05:00
|
|
|
def action_gen_doc(args):
|
|
|
|
f = args.output
|
|
|
|
tools_info = load_tools_info()
|
|
|
|
|
|
|
|
def print_out(text):
|
|
|
|
f.write(text + '\n')
|
|
|
|
|
2021-01-25 21:49:01 -05:00
|
|
|
print_out('.. |zwsp| unicode:: U+200B')
|
|
|
|
print_out(' :trim:')
|
|
|
|
print_out('')
|
2020-01-17 11:38:50 -05:00
|
|
|
|
2021-01-25 21:49:01 -05:00
|
|
|
idf_gh_url = 'https://github.com/espressif/esp-idf'
|
2020-01-17 11:38:50 -05:00
|
|
|
for tool_name, tool_obj in tools_info.items():
|
|
|
|
info_url = tool_obj.options.info_url
|
2021-01-25 21:49:01 -05:00
|
|
|
if idf_gh_url + '/tree' in info_url:
|
|
|
|
info_url = re.sub(idf_gh_url + r'/tree/\w+/(.*)', r':idf:`\1`', info_url)
|
2020-01-17 11:38:50 -05:00
|
|
|
|
2021-01-25 21:49:01 -05:00
|
|
|
license_url = 'https://spdx.org/licenses/' + tool_obj.options.license
|
2020-01-17 11:38:50 -05:00
|
|
|
|
|
|
|
print_out("""
|
|
|
|
.. _tool-{name}:
|
|
|
|
|
|
|
|
{name}
|
|
|
|
{underline}
|
|
|
|
|
|
|
|
{description}
|
|
|
|
|
|
|
|
.. include:: idf-tools-notes.inc
|
|
|
|
:start-after: tool-{name}-notes
|
|
|
|
:end-before: ---
|
|
|
|
|
|
|
|
License: `{license} <{license_url}>`_
|
|
|
|
|
|
|
|
More info: {info_url}
|
|
|
|
|
|
|
|
.. list-table::
|
|
|
|
:widths: 10 10 80
|
|
|
|
:header-rows: 1
|
|
|
|
|
|
|
|
* - Platform
|
|
|
|
- Required
|
|
|
|
- Download
|
|
|
|
""".rstrip().format(name=tool_name,
|
|
|
|
underline=args.heading_underline_char * len(tool_name),
|
|
|
|
description=tool_obj.description,
|
|
|
|
license=tool_obj.options.license,
|
|
|
|
license_url=license_url,
|
|
|
|
info_url=info_url))
|
|
|
|
|
|
|
|
for platform_name in sorted(tool_obj.get_supported_platforms()):
|
|
|
|
platform_tool = tool_obj.copy_for_platform(platform_name)
|
|
|
|
install_type = platform_tool.get_install_type()
|
|
|
|
if install_type == IDFTool.INSTALL_NEVER:
|
|
|
|
continue
|
|
|
|
elif install_type == IDFTool.INSTALL_ALWAYS:
|
2021-01-25 21:49:01 -05:00
|
|
|
install_type_str = 'required'
|
2020-01-17 11:38:50 -05:00
|
|
|
elif install_type == IDFTool.INSTALL_ON_REQUEST:
|
2021-01-25 21:49:01 -05:00
|
|
|
install_type_str = 'optional'
|
2020-01-17 11:38:50 -05:00
|
|
|
else:
|
|
|
|
raise NotImplementedError()
|
|
|
|
|
|
|
|
version = platform_tool.get_recommended_version()
|
|
|
|
version_obj = platform_tool.versions[version]
|
|
|
|
download_obj = version_obj.get_download_for_platform(platform_name)
|
|
|
|
|
|
|
|
# Note: keep the list entries indented to the same number of columns
|
|
|
|
# as the list header above.
|
|
|
|
print_out("""
|
|
|
|
* - {}
|
|
|
|
- {}
|
|
|
|
- {}
|
|
|
|
|
|
|
|
.. rst-class:: tool-sha256
|
|
|
|
|
|
|
|
SHA256: {}
|
|
|
|
""".strip('\n').format(platform_name, install_type_str, download_obj.url, download_obj.sha256))
|
|
|
|
|
|
|
|
print_out('')
|
|
|
|
print_out('')
|
|
|
|
|
|
|
|
|
2019-04-17 03:30:30 -04:00
|
|
|
def main(argv):
|
|
|
|
parser = argparse.ArgumentParser()
|
|
|
|
|
|
|
|
parser.add_argument('--quiet', help='Don\'t output diagnostic messages to stdout/stderr', action='store_true')
|
2019-06-05 03:16:19 -04:00
|
|
|
parser.add_argument('--non-interactive', help='Don\'t output interactive messages and questions', action='store_true')
|
2019-04-17 03:30:30 -04:00
|
|
|
parser.add_argument('--tools-json', help='Path to the tools.json file to use')
|
|
|
|
parser.add_argument('--idf-path', help='ESP-IDF path to use')
|
|
|
|
|
|
|
|
subparsers = parser.add_subparsers(dest='action')
|
|
|
|
subparsers.add_parser('list', help='List tools and versions available')
|
|
|
|
subparsers.add_parser('check', help='Print summary of tools installed or found in PATH')
|
|
|
|
export = subparsers.add_parser('export', help='Output command for setting tool paths, suitable for shell')
|
|
|
|
export.add_argument('--format', choices=[EXPORT_SHELL, EXPORT_KEY_VALUE], default=EXPORT_SHELL,
|
|
|
|
help='Format of the output: shell (suitable for printing into shell), ' +
|
|
|
|
'or key-value (suitable for parsing by other tools')
|
|
|
|
export.add_argument('--prefer-system', help='Normally, if the tool is already present in PATH, ' +
|
|
|
|
'but has an unsupported version, a version from the tools directory ' +
|
|
|
|
'will be used instead. If this flag is given, the version in PATH ' +
|
|
|
|
'will be used.', action='store_true')
|
|
|
|
install = subparsers.add_parser('install', help='Download and install tools into the tools directory')
|
2019-11-24 06:11:33 -05:00
|
|
|
install.add_argument('tools', metavar='TOOL', nargs='*', default=['required'],
|
|
|
|
help='Tools to install. ' +
|
|
|
|
'To install a specific version use <tool_name>@<version> syntax. ' +
|
|
|
|
'Use empty or \'required\' to install required tools, not optional ones. ' +
|
|
|
|
'Use \'all\' to install all tools, including the optional ones.')
|
2019-04-17 03:30:30 -04:00
|
|
|
|
2019-07-18 11:20:40 -04:00
|
|
|
download = subparsers.add_parser('download', help='Download the tools into the dist directory')
|
|
|
|
download.add_argument('--platform', help='Platform to download the tools for')
|
2019-11-24 06:11:33 -05:00
|
|
|
download.add_argument('tools', metavar='TOOL', nargs='*', default=['required'],
|
|
|
|
help='Tools to download. ' +
|
|
|
|
'To download a specific version use <tool_name>@<version> syntax. ' +
|
|
|
|
'Use empty or \'required\' to download required tools, not optional ones. ' +
|
|
|
|
'Use \'all\' to download all tools, including the optional ones.')
|
2019-07-18 11:20:40 -04:00
|
|
|
|
|
|
|
if IDF_MAINTAINER:
|
|
|
|
for subparser in [download, install]:
|
|
|
|
subparser.add_argument('--mirror-prefix-map', nargs='*',
|
|
|
|
help='Pattern to rewrite download URLs, with source and replacement separated by comma.' +
|
|
|
|
' E.g. http://foo.com,http://test.foo.com')
|
|
|
|
|
2019-04-17 03:30:30 -04:00
|
|
|
install_python_env = subparsers.add_parser('install-python-env',
|
|
|
|
help='Create Python virtual environment and install the ' +
|
|
|
|
'required Python packages')
|
|
|
|
install_python_env.add_argument('--reinstall', help='Discard the previously installed environment',
|
|
|
|
action='store_true')
|
|
|
|
install_python_env.add_argument('--extra-wheels-dir', help='Additional directories with wheels ' +
|
|
|
|
'to use during installation')
|
2020-12-16 11:40:13 -05:00
|
|
|
install_python_env.add_argument('--extra-wheels-url', help='Additional URL with wheels', default='https://dl.espressif.com/pypi')
|
|
|
|
install_python_env.add_argument('--no-index', help='Work offline without retrieving wheels index')
|
2019-04-17 03:30:30 -04:00
|
|
|
|
|
|
|
if IDF_MAINTAINER:
|
|
|
|
add_version = subparsers.add_parser('add-version', help='Add or update download info for a version')
|
|
|
|
add_version.add_argument('--output', help='Save new tools.json into this file')
|
|
|
|
add_version.add_argument('--tool', help='Tool name to set add a version for', required=True)
|
|
|
|
add_version.add_argument('--version', help='Version identifier', required=True)
|
|
|
|
add_version.add_argument('--url-prefix', help='String to prepend to file names to obtain download URLs')
|
|
|
|
add_version.add_argument('files', help='File names of the download artifacts', nargs='*')
|
|
|
|
|
|
|
|
rewrite = subparsers.add_parser('rewrite', help='Load tools.json, validate, and save the result back into JSON')
|
|
|
|
rewrite.add_argument('--output', help='Save new tools.json into this file')
|
|
|
|
|
|
|
|
subparsers.add_parser('validate', help='Validate tools.json against schema file')
|
|
|
|
|
2020-01-17 11:38:50 -05:00
|
|
|
gen_doc = subparsers.add_parser('gen-doc', help='Write the list of tools as a documentation page')
|
|
|
|
gen_doc.add_argument('--output', type=argparse.FileType('w'), default=sys.stdout,
|
|
|
|
help='Output file name')
|
|
|
|
gen_doc.add_argument('--heading-underline-char', help='Character to use when generating RST sections', default='~')
|
|
|
|
|
2019-04-17 03:30:30 -04:00
|
|
|
args = parser.parse_args(argv)
|
|
|
|
|
|
|
|
if args.action is None:
|
|
|
|
parser.print_help()
|
|
|
|
parser.exit(1)
|
|
|
|
|
|
|
|
if args.quiet:
|
|
|
|
global global_quiet
|
|
|
|
global_quiet = True
|
|
|
|
|
2019-06-05 03:16:19 -04:00
|
|
|
if args.non_interactive:
|
|
|
|
global global_non_interactive
|
|
|
|
global_non_interactive = True
|
|
|
|
|
2019-04-17 03:30:30 -04:00
|
|
|
global global_idf_path
|
|
|
|
global_idf_path = os.environ.get('IDF_PATH')
|
|
|
|
if args.idf_path:
|
|
|
|
global_idf_path = args.idf_path
|
|
|
|
if not global_idf_path:
|
2021-01-25 21:49:01 -05:00
|
|
|
global_idf_path = os.path.realpath(os.path.join(os.path.dirname(__file__), '..'))
|
2019-11-18 10:17:30 -05:00
|
|
|
os.environ['IDF_PATH'] = global_idf_path
|
2019-04-17 03:30:30 -04:00
|
|
|
|
|
|
|
global global_idf_tools_path
|
|
|
|
global_idf_tools_path = os.environ.get('IDF_TOOLS_PATH') or os.path.expanduser(IDF_TOOLS_PATH_DEFAULT)
|
|
|
|
|
2019-10-30 11:19:17 -04:00
|
|
|
# On macOS, unset __PYVENV_LAUNCHER__ variable if it is set.
|
|
|
|
# Otherwise sys.executable keeps pointing to the system Python, even when a python binary from a virtualenv is invoked.
|
|
|
|
# See https://bugs.python.org/issue22490#msg283859.
|
2020-01-21 04:48:24 -05:00
|
|
|
os.environ.pop('__PYVENV_LAUNCHER__', None)
|
2019-10-30 11:19:17 -04:00
|
|
|
|
2019-04-17 03:30:30 -04:00
|
|
|
if sys.version_info.major == 2:
|
|
|
|
try:
|
|
|
|
global_idf_tools_path.decode('ascii')
|
|
|
|
except UnicodeDecodeError:
|
|
|
|
fatal('IDF_TOOLS_PATH contains non-ASCII characters: {}'.format(global_idf_tools_path) +
|
|
|
|
'\nThis is not supported yet with Python 2. ' +
|
|
|
|
'Please set IDF_TOOLS_PATH to a directory with an ASCII name, or switch to Python 3.')
|
|
|
|
raise SystemExit(1)
|
|
|
|
|
|
|
|
if CURRENT_PLATFORM == UNKNOWN_PLATFORM:
|
|
|
|
fatal('Platform {} appears to be unsupported'.format(PYTHON_PLATFORM))
|
|
|
|
raise SystemExit(1)
|
|
|
|
|
|
|
|
global global_tools_json
|
|
|
|
if args.tools_json:
|
|
|
|
global_tools_json = args.tools_json
|
|
|
|
else:
|
|
|
|
global_tools_json = os.path.join(global_idf_path, TOOLS_FILE)
|
|
|
|
|
|
|
|
action_func_name = 'action_' + args.action.replace('-', '_')
|
|
|
|
action_func = globals()[action_func_name]
|
|
|
|
|
|
|
|
action_func(args)
|
|
|
|
|
|
|
|
|
|
|
|
if __name__ == '__main__':
|
|
|
|
main(sys.argv[1:])
|