2019-04-17 03:30:30 -04:00
#!/usr/bin/env python
# coding=utf-8
#
2023-06-28 04:44:27 -04:00
# SPDX-FileCopyrightText: 2019-2023 Espressif Systems (Shanghai) CO LTD
2021-05-26 04:44:20 -04:00
#
# SPDX-License-Identifier: Apache-2.0
#
2019-04-17 03:30:30 -04:00
# This script helps installing tools required to use the ESP-IDF, and updating PATH
# to use the installed tools. It can also create a Python virtual environment,
# and install Python requirements into it.
# It does not install OS dependencies. It does install tools such as the Xtensa
# GCC toolchain and ESP32 ULP coprocessor toolchain.
#
# By default, downloaded tools will be installed under $HOME/.espressif directory
# (%USERPROFILE%/.espressif on Windows). This path can be modified by setting
# IDF_TOOLS_PATH variable prior to running this tool.
#
# Users do not need to interact with this script directly. In IDF root directory,
# install.sh (.bat) and export.sh (.bat) scripts are provided to invoke this script.
#
# Usage:
#
# * To install the tools, run `idf_tools.py install`.
#
# * To install the Python environment, run `idf_tools.py install-python-env`.
#
# * To start using the tools, run `eval "$(idf_tools.py export)"` — this will update
# the PATH to point to the installed tools and set up other environment variables
# needed by the tools.
2021-01-25 21:49:01 -05:00
import argparse
import contextlib
import copy
2022-01-18 08:49:53 -05:00
import datetime
2021-01-25 21:49:01 -05:00
import errno
2023-09-14 01:55:24 -04:00
import fnmatch
2021-01-25 21:49:01 -05:00
import functools
import hashlib
2019-04-17 03:30:30 -04:00
import json
import os
2021-01-25 21:49:01 -05:00
import platform
import re
import shutil
import ssl
2019-04-17 03:30:30 -04:00
import subprocess
import sys
import tarfile
2022-08-09 12:13:29 -04:00
import tempfile
2022-02-07 09:18:14 -05:00
import time
2019-04-17 03:30:30 -04:00
from collections import OrderedDict , namedtuple
2022-03-19 16:37:16 -04:00
from json import JSONEncoder
2021-03-04 03:53:47 -05:00
from ssl import SSLContext # noqa: F401
from tarfile import TarFile # noqa: F401
from zipfile import ZipFile
2019-04-17 03:30:30 -04:00
2021-05-26 04:44:20 -04:00
# Important notice: Please keep the lines above compatible with old Pythons so it won't fail with ImportError but with
# a nice message printed by python_version_checker.check()
2020-01-17 11:38:50 -05:00
try :
2021-05-26 04:44:20 -04:00
import python_version_checker
2020-01-17 11:38:50 -05:00
2021-05-26 04:44:20 -04:00
# check the Python version before it will fail with an exception on syntax or package incompatibility.
python_version_checker . check ( )
except RuntimeError as e :
print ( e )
raise SystemExit ( 1 )
2022-03-15 15:02:37 -04:00
from typing import IO , Any , Callable , Dict , Iterator , List , Optional , Set , Tuple , Union # noqa: F401
2021-05-26 04:44:20 -04:00
from urllib . error import ContentTooShortError
2022-03-15 15:02:37 -04:00
from urllib . parse import urljoin , urlparse
2021-05-26 04:44:20 -04:00
from urllib . request import urlopen
# the following is only for typing annotation
from urllib . response import addinfourl # noqa: F401
2019-04-17 03:30:30 -04:00
2019-09-23 00:01:09 -04:00
try :
from exceptions import WindowsError
except ImportError :
2021-03-04 03:53:47 -05:00
# Unix
class WindowsError ( OSError ) : # type: ignore
2019-09-23 00:01:09 -04:00
pass
2019-04-17 03:30:30 -04:00
TOOLS_FILE = ' tools/tools.json '
TOOLS_SCHEMA_FILE = ' tools/tools_schema.json '
TOOLS_FILE_NEW = ' tools/tools.new.json '
2021-04-26 15:34:48 -04:00
IDF_ENV_FILE = ' idf-env.json '
2019-04-17 03:30:30 -04:00
TOOLS_FILE_VERSION = 1
IDF_TOOLS_PATH_DEFAULT = os . path . join ( ' ~ ' , ' .espressif ' )
UNKNOWN_VERSION = ' unknown '
SUBST_TOOL_PATH_REGEX = re . compile ( r ' \ $ {TOOL_PATH} ' )
VERSION_REGEX_REPLACE_DEFAULT = r ' \ 1 '
IDF_MAINTAINER = os . environ . get ( ' IDF_MAINTAINER ' ) or False
TODO_MESSAGE = ' TODO '
DOWNLOAD_RETRY_COUNT = 3
URL_PREFIX_MAP_SEPARATOR = ' , '
IDF_TOOLS_INSTALL_CMD = os . environ . get ( ' IDF_TOOLS_INSTALL_CMD ' )
IDF_TOOLS_EXPORT_CMD = os . environ . get ( ' IDF_TOOLS_INSTALL_CMD ' )
2021-09-16 10:48:03 -04:00
IDF_DL_URL = ' https://dl.espressif.com/dl/esp-idf '
2022-11-10 23:00:56 -05:00
IDF_PIP_WHEELS_URL = os . environ . get ( ' IDF_PIP_WHEELS_URL ' , ' https://dl.espressif.com/pypi ' )
2019-04-17 03:30:30 -04:00
PYTHON_PLATFORM = platform . system ( ) + ' - ' + platform . machine ( )
# Identifiers used in tools.json for different platforms.
PLATFORM_WIN32 = ' win32 '
PLATFORM_WIN64 = ' win64 '
PLATFORM_MACOS = ' macos '
2022-04-13 16:11:28 -04:00
PLATFORM_MACOS_ARM64 = ' macos-arm64 '
2019-04-17 03:30:30 -04:00
PLATFORM_LINUX32 = ' linux-i686 '
PLATFORM_LINUX64 = ' linux-amd64 '
PLATFORM_LINUX_ARM32 = ' linux-armel '
2021-11-11 07:42:55 -05:00
PLATFORM_LINUX_ARMHF = ' linux-armhf '
2019-04-17 03:30:30 -04:00
PLATFORM_LINUX_ARM64 = ' linux-arm64 '
2022-05-16 16:11:01 -04:00
class Platforms :
# Mappings from various other names these platforms are known as, to the identifiers above.
# This includes strings produced from "platform.system() + '-' + platform.machine()", see PYTHON_PLATFORM
# definition above.
# This list also includes various strings used in release archives of xtensa-esp32-elf-gcc, OpenOCD, etc.
PLATFORM_FROM_NAME = {
# Windows
PLATFORM_WIN32 : PLATFORM_WIN32 ,
' Windows-i686 ' : PLATFORM_WIN32 ,
' Windows-x86 ' : PLATFORM_WIN32 ,
' i686-w64-mingw32 ' : PLATFORM_WIN32 ,
PLATFORM_WIN64 : PLATFORM_WIN64 ,
' Windows-x86_64 ' : PLATFORM_WIN64 ,
' Windows-AMD64 ' : PLATFORM_WIN64 ,
' x86_64-w64-mingw32 ' : PLATFORM_WIN64 ,
2023-08-24 10:18:13 -04:00
' Windows-ARM64 ' : PLATFORM_WIN64 ,
2022-05-16 16:11:01 -04:00
# macOS
PLATFORM_MACOS : PLATFORM_MACOS ,
' osx ' : PLATFORM_MACOS ,
' darwin ' : PLATFORM_MACOS ,
' Darwin-x86_64 ' : PLATFORM_MACOS ,
2022-08-23 04:17:24 -04:00
' x86_64-apple-darwin ' : PLATFORM_MACOS ,
2022-05-16 16:11:01 -04:00
PLATFORM_MACOS_ARM64 : PLATFORM_MACOS_ARM64 ,
' Darwin-arm64 ' : PLATFORM_MACOS_ARM64 ,
2022-08-23 04:17:24 -04:00
' aarch64-apple-darwin ' : PLATFORM_MACOS_ARM64 ,
' arm64-apple-darwin ' : PLATFORM_MACOS_ARM64 ,
2022-05-16 16:11:01 -04:00
# Linux
PLATFORM_LINUX64 : PLATFORM_LINUX64 ,
' linux64 ' : PLATFORM_LINUX64 ,
' Linux-x86_64 ' : PLATFORM_LINUX64 ,
' FreeBSD-amd64 ' : PLATFORM_LINUX64 ,
' x86_64-linux-gnu ' : PLATFORM_LINUX64 ,
PLATFORM_LINUX32 : PLATFORM_LINUX32 ,
' linux32 ' : PLATFORM_LINUX32 ,
' Linux-i686 ' : PLATFORM_LINUX32 ,
' FreeBSD-i386 ' : PLATFORM_LINUX32 ,
' i586-linux-gnu ' : PLATFORM_LINUX32 ,
2023-01-19 08:30:17 -05:00
' i686-linux-gnu ' : PLATFORM_LINUX32 ,
2022-05-16 16:11:01 -04:00
PLATFORM_LINUX_ARM64 : PLATFORM_LINUX_ARM64 ,
' Linux-arm64 ' : PLATFORM_LINUX_ARM64 ,
' Linux-aarch64 ' : PLATFORM_LINUX_ARM64 ,
' Linux-armv8l ' : PLATFORM_LINUX_ARM64 ,
' aarch64 ' : PLATFORM_LINUX_ARM64 ,
PLATFORM_LINUX_ARMHF : PLATFORM_LINUX_ARMHF ,
' arm-linux-gnueabihf ' : PLATFORM_LINUX_ARMHF ,
PLATFORM_LINUX_ARM32 : PLATFORM_LINUX_ARM32 ,
' arm-linux-gnueabi ' : PLATFORM_LINUX_ARM32 ,
' Linux-armv7l ' : PLATFORM_LINUX_ARM32 ,
' Linux-arm ' : PLATFORM_LINUX_ARM32 ,
}
@staticmethod
def get ( platform_alias ) : # type: (Optional[str]) -> Optional[str]
if platform_alias is None :
return None
2022-08-15 10:47:24 -04:00
if platform_alias == ' any ' and CURRENT_PLATFORM :
platform_alias = CURRENT_PLATFORM
2022-05-16 16:11:01 -04:00
platform_name = Platforms . PLATFORM_FROM_NAME . get ( platform_alias , None )
# ARM platform may run on armhf hardware but having armel installed packages.
# To avoid possible armel/armhf libraries mixing need to define user's
# packages architecture to use the same
# See note section in https://gcc.gnu.org/onlinedocs/gcc/ARM-Options.html#index-mfloat-abi
if platform_name in ( PLATFORM_LINUX_ARM32 , PLATFORM_LINUX_ARMHF ) and ' arm ' in platform . machine ( ) :
# suppose that installed python was built with a right ABI
with open ( sys . executable , ' rb ' ) as f :
if int . from_bytes ( f . read ( 4 ) , sys . byteorder ) != int . from_bytes ( b ' \x7f ELF ' , sys . byteorder ) :
return platform_name # ELF magic not found. Use default platform name from PLATFORM_FROM_NAME
f . seek ( 36 ) # seek to e_flags (https://man7.org/linux/man-pages/man5/elf.5.html)
e_flags = int . from_bytes ( f . read ( 4 ) , sys . byteorder )
platform_name = PLATFORM_LINUX_ARMHF if e_flags & 0x400 else PLATFORM_LINUX_ARM32
return platform_name
@staticmethod
def get_by_filename ( file_name ) : # type: (str) -> Optional[str]
found_alias = ' '
for platform_alias in Platforms . PLATFORM_FROM_NAME :
# Find the longest alias which matches with file name to avoid mismatching
if platform_alias in file_name and len ( found_alias ) < len ( platform_alias ) :
found_alias = platform_alias
return Platforms . get ( found_alias )
2023-09-14 01:55:24 -04:00
def parse_platform_arg ( platform_str ) : # type: (str) -> str
platform = Platforms . get ( platform_str )
if platform is None :
fatal ( f ' unknown platform: { platform } ' )
raise SystemExit ( 1 )
return platform
CURRENT_PLATFORM = parse_platform_arg ( PYTHON_PLATFORM )
2019-04-17 03:30:30 -04:00
EXPORT_SHELL = ' shell '
EXPORT_KEY_VALUE = ' key-value '
2023-05-18 08:46:17 -04:00
# the older "DigiCert Global Root CA" certificate used with github.com
DIGICERT_ROOT_CA_CERT = """
2020-05-20 03:50:57 -04:00
- - - - - BEGIN CERTIFICATE - - - - -
2022-09-09 11:33:33 -04:00
MIIDrzCCApegAwIBAgIQCDvgVpBCRrGhdWrJWZHHSjANBgkqhkiG9w0BAQUFADBh
MQswCQYDVQQGEwJVUzEVMBMGA1UEChMMRGlnaUNlcnQgSW5jMRkwFwYDVQQLExB3
d3cuZGlnaWNlcnQuY29tMSAwHgYDVQQDExdEaWdpQ2VydCBHbG9iYWwgUm9vdCBD
QTAeFw0wNjExMTAwMDAwMDBaFw0zMTExMTAwMDAwMDBaMGExCzAJBgNVBAYTAlVT
MRUwEwYDVQQKEwxEaWdpQ2VydCBJbmMxGTAXBgNVBAsTEHd3dy5kaWdpY2VydC5j
b20xIDAeBgNVBAMTF0RpZ2lDZXJ0IEdsb2JhbCBSb290IENBMIIBIjANBgkqhkiG
9 w0BAQEFAAOCAQ8AMIIBCgKCAQEA4jvhEXLeqKTTo1eqUKKPC3eQyaKl7hLOllsB
CSDMAZOnTjC3U / dDxGkAV53ijSLdhwZAAIEJzs4bg7 / fzTtxRuLWZscFs3YnFo97
nh6Vfe63SKMI2tavegw5BmV / Sl0fvBf4q77uKNd0f3p4mVmFaG5cIzJLv07A6Fpt
43 C / dxC / / AH2hdmoRBBYMql1GNXRor5H4idq9Joz + EkIYIvUX7Q6hL + hqkpMfT7P
T19sdl6gSzeRntwi5m3OFBqOasv + zbMUZBfHWymeMr / y7vrTC0LUq7dBMtoM1O / 4
gdW7jVg / tRvoSSiicNoxBN33shbyTApOB6jtSj1etX + jkMOvJwIDAQABo2MwYTAO
BgNVHQ8BAf8EBAMCAYYwDwYDVR0TAQH / BAUwAwEB / zAdBgNVHQ4EFgQUA95QNVbR
TLtm8KPiGxvDl7I90VUwHwYDVR0jBBgwFoAUA95QNVbRTLtm8KPiGxvDl7I90VUw
DQYJKoZIhvcNAQEFBQADggEBAMucN6pIExIK + t1EnE9SsPTfrgT1eXkIoyQY / Esr
hMAtudXH / vTBH1jLuG2cenTnmCmrEbXjcKChzUyImZOMkXDiqw8cvpOp / 2 PV5Adg
06 O / nVsJ8dWO41P0jmP6P6fbtGbfYmbW0W5BjfIttep3Sp + dWOIrWcBAI + 0 tKIJF
PnlUkiaY4IBIqDfv8NZ5YBberOgOzW6sRBc4L0na4UU + Krk2U886UAb3LujEV0ls
YSEY1QSteDwsOoBrp + uvFRTp2InBuThs4pFsiv9kuXclVzDAGySj4dzp30d8tbQk
CAUw7C29C79Fv1C5qfPrmAESrciIxpg0X40KPMbp1ZWVbd4 =
2020-05-20 03:50:57 -04:00
- - - - - END CERTIFICATE - - - - -
"""
2023-05-18 08:46:17 -04:00
# the newer "DigiCert Global Root G2" certificate used with dl.espressif.com
DIGICERT_ROOT_G2_CERT = """
- - - - - BEGIN CERTIFICATE - - - - -
MIIDjjCCAnagAwIBAgIQAzrx5qcRqaC7KGSxHQn65TANBgkqhkiG9w0BAQsFADBh
MQswCQYDVQQGEwJVUzEVMBMGA1UEChMMRGlnaUNlcnQgSW5jMRkwFwYDVQQLExB3
d3cuZGlnaWNlcnQuY29tMSAwHgYDVQQDExdEaWdpQ2VydCBHbG9iYWwgUm9vdCBH
MjAeFw0xMzA4MDExMjAwMDBaFw0zODAxMTUxMjAwMDBaMGExCzAJBgNVBAYTAlVT
MRUwEwYDVQQKEwxEaWdpQ2VydCBJbmMxGTAXBgNVBAsTEHd3dy5kaWdpY2VydC5j
b20xIDAeBgNVBAMTF0RpZ2lDZXJ0IEdsb2JhbCBSb290IEcyMIIBIjANBgkqhkiG
9 w0BAQEFAAOCAQ8AMIIBCgKCAQEAuzfNNNx7a8myaJCtSnX / RrohCgiN9RlUyfuI
2 / Ou8jqJkTx65qsGGmvPrC3oXgkkRLpimn7Wo6h + 4 FR1IAWsULecYxpsMNzaHxmx
1 x7e / dfgy5SDN67sH0NO3Xss0r0upS / kqbitOtSZpLYl6ZtrAGCSYP9PIUkY92eQ
q2EGnI / yuum06ZIya7XzV + hdG82MHauVBJVJ8zUtluNJbd134 / tJS7SsVQepj5Wz
tCO7TG1F8PapspUwtP1MVYwnSlcUfIKdzXOS0xZKBgyMUNGPHgm + F6HmIcr9g + UQ
vIOlCsRnKPZzFBQ9RnbDhxSJITRNrw9FDKZJobq7nMWxM4MphQIDAQABo0IwQDAP
BgNVHRMBAf8EBTADAQH / MA4GA1UdDwEB / wQEAwIBhjAdBgNVHQ4EFgQUTiJUIBiV
5 uNu5g / 6 + rkS7QYXjzkwDQYJKoZIhvcNAQELBQADggEBAGBnKJRvDkhj6zHd6mcY
1 Yl9PMWLSn / pvtsrF9 + wX3N3KjITOYFnQoQj8kVnNeyIv / iPsGEMNKSuIEyExtv4
NeF22d + mQrvHRAiGfzZ0JFrabA0UWTW98kndth / Jsw1HKj2ZL7tcu7XUIOGZX1NG
Fdtom / DzMNU + MeKNhJ7jitralj41E6Vf8PlwUHBHQRFXGU7Aj64GxJUTFy8bJZ91
8 rGOmaFvE7FBcf6IKshPECBV1 / MUReXgRPTqh5Uykw7 + U0b6LJ3 / iyK5S9kJRaTe
pLiaWN0bfVKfjllDiIGknibVb63dDcY3fe0Dkhvld1927jyNxF1WW6LZZm6zNTfl
MrY =
- - - - - END CERTIFICATE - - - - -
"""
DL_CERT_DICT = { ' dl.espressif.com ' : DIGICERT_ROOT_G2_CERT ,
' github.com ' : DIGICERT_ROOT_CA_CERT }
2019-04-17 03:30:30 -04:00
global_quiet = False
2019-06-05 03:16:19 -04:00
global_non_interactive = False
2021-03-04 03:53:47 -05:00
global_idf_path = None # type: Optional[str]
global_idf_tools_path = None # type: Optional[str]
global_tools_json = None # type: Optional[str]
2019-04-17 03:30:30 -04:00
2021-03-04 03:53:47 -05:00
def fatal ( text , * args ) : # type: (str, str) -> None
2019-04-17 03:30:30 -04:00
if not global_quiet :
sys . stderr . write ( ' ERROR: ' + text + ' \n ' , * args )
2021-03-04 03:53:47 -05:00
def warn ( text , * args ) : # type: (str, str) -> None
2019-04-17 03:30:30 -04:00
if not global_quiet :
sys . stderr . write ( ' WARNING: ' + text + ' \n ' , * args )
2021-03-04 03:53:47 -05:00
def info ( text , f = None , * args ) : # type: (str, Optional[IO[str]], str) -> None
2019-04-17 03:30:30 -04:00
if not global_quiet :
if f is None :
f = sys . stdout
f . write ( text + ' \n ' , * args )
2022-09-22 10:14:17 -04:00
def print_hints_on_download_error ( err ) : # type: (str) -> None
info ( ' Please make sure you have a working Internet connection. ' )
if ' CERTIFICATE ' in err :
info ( ' Certificate issues are usually caused by an outdated certificate database on your computer. ' )
info ( ' Please check the documentation of your operating system for how to upgrade it. ' )
if sys . platform == ' darwin ' :
info ( ' Running " ./Install \\ Certificates.command " might be able to fix this issue. ' )
info ( ' Running " {} -m pip install --upgrade certifi " can also resolve this issue in some cases. ' . format ( sys . executable ) )
# Certificate issue on Windows can be hidden under different errors which might be even translated,
# e.g. "[WinError -2146881269] ASN1 valor de tag inválido encontrado"
if sys . platform == ' win32 ' :
info ( ' By downloading and using the offline installer from https://dl.espressif.com/dl/esp-idf '
' you might be able to work around this issue. ' )
2019-04-17 03:30:30 -04:00
def run_cmd_check_output ( cmd , input_text = None , extra_paths = None ) :
2021-06-21 04:22:54 -04:00
# type: (List[str], Optional[str], Optional[List[str]]) -> bytes
2019-04-17 03:30:30 -04:00
# If extra_paths is given, locate the executable in one of these directories.
# Note: it would seem logical to add extra_paths to env[PATH], instead, and let OS do the job of finding the
# executable for us. However this does not work on Windows: https://bugs.python.org/issue8557.
if extra_paths :
found = False
extensions = [ ' ' ]
if sys . platform == ' win32 ' :
extensions . append ( ' .exe ' )
for path in extra_paths :
for ext in extensions :
fullpath = os . path . join ( path , cmd [ 0 ] + ext )
if os . path . exists ( fullpath ) :
cmd [ 0 ] = fullpath
found = True
break
if found :
break
try :
2021-03-04 03:53:47 -05:00
input_bytes = None
2019-04-17 03:30:30 -04:00
if input_text :
2021-03-04 03:53:47 -05:00
input_bytes = input_text . encode ( )
result = subprocess . run ( cmd , stdout = subprocess . PIPE , stderr = subprocess . PIPE , check = True , input = input_bytes )
2019-04-17 03:30:30 -04:00
return result . stdout + result . stderr
tools: fix idf_tools.py exception with python3
use `subprocess.Popen` when catch TypeError:
```
Traceback (most recent call last):
File "tools/idf_tools.py", line 1249, in <module>
main(sys.argv[1:])
File "tools/idf_tools.py", line 1245, in main
action_func(args)
File "tools/idf_tools.py", line 1038, in action_install
tool_obj.find_installed_versions()
File "tools/idf_tools.py", line 468, in find_installed_versions
ver_str = self.check_version()
File "tools/idf_tools.py", line 426, in check_version
version_cmd_result = run_cmd_check_output(cmd, None, extra_paths)
File "tools/idf_tools.py", line 176, in run_cmd_check_output
result = subprocess.run(cmd, capture_output=True, check=True, input=input_text)
File "/opt/pyenv/pyenv-1.2.6/versions/3.5.5/lib/python3.5/subprocess.py", line 383, in run
with Popen(*popenargs, **kwargs) as process:
TypeError: __init__() got an unexpected keyword argument 'capture_output'
```
2019-07-04 03:06:00 -04:00
except ( AttributeError , TypeError ) :
2019-04-17 03:30:30 -04:00
p = subprocess . Popen ( cmd , stdout = subprocess . PIPE , stdin = subprocess . PIPE , stderr = subprocess . PIPE )
2021-03-04 03:53:47 -05:00
stdout , stderr = p . communicate ( input_bytes )
2019-04-17 03:30:30 -04:00
if p . returncode != 0 :
try :
raise subprocess . CalledProcessError ( p . returncode , cmd , stdout , stderr )
except TypeError :
raise subprocess . CalledProcessError ( p . returncode , cmd , stdout )
return stdout + stderr
2021-06-21 04:22:54 -04:00
def to_shell_specific_paths ( paths_list ) : # type: (List[str]) -> List[str]
2019-04-17 03:30:30 -04:00
if sys . platform == ' win32 ' :
paths_list = [ p . replace ( ' / ' , os . path . sep ) if os . path . sep in p else p for p in paths_list ]
return paths_list
2021-06-21 04:22:54 -04:00
def get_env_for_extra_paths ( extra_paths ) : # type: (List[str]) -> Dict[str, str]
2019-04-17 03:30:30 -04:00
"""
Return a copy of environment variables dict , prepending paths listed in extra_paths
to the PATH environment variable .
"""
env_arg = os . environ . copy ( )
new_path = os . pathsep . join ( extra_paths ) + os . pathsep + env_arg [ ' PATH ' ]
if sys . version_info . major == 2 :
2021-03-04 03:53:47 -05:00
env_arg [ ' PATH ' ] = new_path . encode ( ' utf8 ' ) # type: ignore
2019-04-17 03:30:30 -04:00
else :
env_arg [ ' PATH ' ] = new_path
return env_arg
2021-03-04 03:53:47 -05:00
def get_file_size_sha256 ( filename , block_size = 65536 ) : # type: (str, int) -> Tuple[int, str]
2019-04-17 03:30:30 -04:00
sha256 = hashlib . sha256 ( )
size = 0
with open ( filename , ' rb ' ) as f :
for block in iter ( lambda : f . read ( block_size ) , b ' ' ) :
sha256 . update ( block )
size + = len ( block )
return size , sha256 . hexdigest ( )
2021-03-04 03:53:47 -05:00
def report_progress ( count , block_size , total_size ) : # type: (int, int, int) -> None
2019-04-17 03:30:30 -04:00
percent = int ( count * block_size * 100 / total_size )
percent = min ( 100 , percent )
2021-01-25 21:49:01 -05:00
sys . stdout . write ( ' \r %d %% ' % percent )
2019-04-17 03:30:30 -04:00
sys . stdout . flush ( )
2021-03-04 03:53:47 -05:00
def mkdir_p ( path ) : # type: (str) -> None
2019-04-17 03:30:30 -04:00
try :
os . makedirs ( path )
except OSError as exc :
if exc . errno != errno . EEXIST or not os . path . isdir ( path ) :
raise
2021-03-04 03:53:47 -05:00
def unpack ( filename , destination ) : # type: (str, str) -> None
2019-04-17 03:30:30 -04:00
info ( ' Extracting {0} to {1} ' . format ( filename , destination ) )
2020-06-08 09:34:05 -04:00
if filename . endswith ( ( ' .tar.gz ' , ' .tgz ' ) ) :
2021-03-04 03:53:47 -05:00
archive_obj = tarfile . open ( filename , ' r:gz ' ) # type: Union[TarFile, ZipFile]
2021-09-10 07:23:00 -04:00
elif filename . endswith ( ( ' .tar.xz ' ) ) :
archive_obj = tarfile . open ( filename , ' r:xz ' )
2023-09-14 01:55:24 -04:00
elif filename . endswith ( ( ' .tar.bz2 ' ) ) :
archive_obj = tarfile . open ( filename , ' r:bz2 ' )
2019-04-17 03:30:30 -04:00
elif filename . endswith ( ' zip ' ) :
2021-03-04 03:53:47 -05:00
archive_obj = ZipFile ( filename )
2019-04-17 03:30:30 -04:00
else :
raise NotImplementedError ( ' Unsupported archive type ' )
if sys . version_info . major == 2 :
# This is a workaround for the issue that unicode destination is not handled:
# https://bugs.python.org/issue17153
destination = str ( destination )
archive_obj . extractall ( destination )
2021-03-04 03:53:47 -05:00
def splittype ( url ) : # type: (str) -> Tuple[Optional[str], str]
2021-01-11 10:58:21 -05:00
match = re . match ( ' ([^/:]+):(.*) ' , url , re . DOTALL )
if match :
scheme , data = match . groups ( )
return scheme . lower ( ) , data
return None , url
2020-05-20 03:50:57 -04:00
# An alternative version of urlretrieve which takes SSL context as an argument
def urlretrieve_ctx ( url , filename , reporthook = None , data = None , context = None ) :
2021-03-04 03:53:47 -05:00
# type: (str, str, Optional[Callable[[int, int, int], None]], Optional[bytes], Optional[SSLContext]) -> Tuple[str, addinfourl]
2020-05-20 03:50:57 -04:00
url_type , path = splittype ( url )
2020-10-14 14:19:32 -04:00
# urlopen doesn't have context argument in Python <=2.7.9
extra_urlopen_args = { }
if context :
2021-01-25 21:49:01 -05:00
extra_urlopen_args [ ' context ' ] = context
2021-03-04 03:53:47 -05:00
with contextlib . closing ( urlopen ( url , data , * * extra_urlopen_args ) ) as fp : # type: ignore
2020-05-20 03:50:57 -04:00
headers = fp . info ( )
# Just return the local path and the "headers" for file://
# URLs. No sense in performing a copy unless requested.
2021-01-25 21:49:01 -05:00
if url_type == ' file ' and not filename :
2020-05-20 03:50:57 -04:00
return os . path . normpath ( path ) , headers
# Handle temporary file setup.
tfp = open ( filename , ' wb ' )
with tfp :
result = filename , headers
bs = 1024 * 8
2021-01-25 21:49:01 -05:00
size = int ( headers . get ( ' content-length ' , - 1 ) )
2020-05-20 03:50:57 -04:00
read = 0
blocknum = 0
if reporthook :
reporthook ( blocknum , bs , size )
while True :
block = fp . read ( bs )
if not block :
break
read + = len ( block )
tfp . write ( block )
blocknum + = 1
if reporthook :
reporthook ( blocknum , bs , size )
if size > = 0 and read < size :
raise ContentTooShortError (
2021-01-25 21:49:01 -05:00
' retrieval incomplete: got only %i out of %i bytes '
2020-05-20 03:50:57 -04:00
% ( read , size ) , result )
return result
2022-09-22 10:14:17 -04:00
def download ( url , destination ) : # type: (str, str) -> Optional[Exception]
2022-03-30 02:11:18 -04:00
info ( f ' Downloading { url } ' )
info ( f ' Destination: { destination } ' )
2021-09-16 10:48:03 -04:00
try :
2023-05-18 08:46:17 -04:00
for site , cert in DL_CERT_DICT . items ( ) :
# For dl.espressif.com and github.com, add the DigiCert root certificate.
# This works around the issue with outdated certificate stores in some installations.
if site in url :
2021-09-16 10:48:03 -04:00
ctx = ssl . create_default_context ( )
2023-05-18 08:46:17 -04:00
ctx . load_verify_locations ( cadata = cert )
break
else :
ctx = None
2021-09-16 10:48:03 -04:00
urlretrieve_ctx ( url , destination , report_progress if not global_non_interactive else None , context = ctx )
sys . stdout . write ( ' \r Done \n ' )
2022-09-22 10:14:17 -04:00
return None
2021-09-16 10:48:03 -04:00
except Exception as e :
# urlretrieve could throw different exceptions, e.g. IOError when the server is down
2022-09-22 10:14:17 -04:00
return e
2021-09-16 10:48:03 -04:00
finally :
sys . stdout . flush ( )
2019-09-23 00:01:09 -04:00
# Sometimes renaming a directory on Windows (randomly?) causes a PermissionError.
# This is confirmed to be a workaround:
# https://github.com/espressif/esp-idf/issues/3819#issuecomment-515167118
# https://github.com/espressif/esp-idf/issues/4063#issuecomment-531490140
# https://stackoverflow.com/a/43046729
2021-03-04 03:53:47 -05:00
def rename_with_retry ( path_from , path_to ) : # type: (str, str) -> None
2022-02-07 09:18:14 -05:00
retry_count = 20 if sys . platform . startswith ( ' win ' ) else 1
2019-09-23 00:01:09 -04:00
for retry in range ( retry_count ) :
try :
os . rename ( path_from , path_to )
return
2022-02-07 09:18:14 -05:00
except OSError :
msg = f ' Rename { path_from } to { path_to } failed '
2019-09-23 00:01:09 -04:00
if retry == retry_count - 1 :
2022-02-07 09:18:14 -05:00
fatal ( msg + ' . Antivirus software might be causing this. Disabling it temporarily could solve the issue. ' )
2019-09-23 00:01:09 -04:00
raise
2022-02-07 09:18:14 -05:00
warn ( msg + ' , retrying... ' )
# Sleep before the next try in order to pass the antivirus check on Windows
time . sleep ( 0.5 )
2019-09-23 00:01:09 -04:00
2022-08-15 10:47:24 -04:00
def do_strip_container_dirs ( path , levels ) : # type: (str, int) -> None
2019-04-17 03:30:30 -04:00
assert levels > 0
# move the original directory out of the way (add a .tmp suffix)
tmp_path = path + ' .tmp '
if os . path . exists ( tmp_path ) :
shutil . rmtree ( tmp_path )
2019-09-23 00:01:09 -04:00
rename_with_retry ( path , tmp_path )
2019-04-17 03:30:30 -04:00
os . mkdir ( path )
base_path = tmp_path
# walk given number of levels down
for level in range ( levels ) :
contents = os . listdir ( base_path )
if len ( contents ) > 1 :
raise RuntimeError ( ' at level {} , expected 1 entry, got {} ' . format ( level , contents ) )
base_path = os . path . join ( base_path , contents [ 0 ] )
if not os . path . isdir ( base_path ) :
raise RuntimeError ( ' at level {} , {} is not a directory ' . format ( level , contents [ 0 ] ) )
# get the list of directories/files to move
contents = os . listdir ( base_path )
for name in contents :
move_from = os . path . join ( base_path , name )
move_to = os . path . join ( path , name )
2019-09-23 00:01:09 -04:00
rename_with_retry ( move_from , move_to )
2019-04-17 03:30:30 -04:00
shutil . rmtree ( tmp_path )
class ToolNotFound ( RuntimeError ) :
pass
class ToolExecError ( RuntimeError ) :
pass
class IDFToolDownload ( object ) :
2021-03-04 03:53:47 -05:00
def __init__ ( self , platform_name , url , size , sha256 ) : # type: (str, str, int, str) -> None
2019-04-17 03:30:30 -04:00
self . platform_name = platform_name
self . url = url
self . size = size
self . sha256 = sha256
self . platform_name = platform_name
@functools.total_ordering
class IDFToolVersion ( object ) :
STATUS_RECOMMENDED = ' recommended '
STATUS_SUPPORTED = ' supported '
STATUS_DEPRECATED = ' deprecated '
STATUS_VALUES = [ STATUS_RECOMMENDED , STATUS_SUPPORTED , STATUS_DEPRECATED ]
2021-03-04 03:53:47 -05:00
def __init__ ( self , version , status ) : # type: (str, str) -> None
2019-04-17 03:30:30 -04:00
self . version = version
self . status = status
2021-03-04 03:53:47 -05:00
self . downloads = OrderedDict ( ) # type: OrderedDict[str, IDFToolDownload]
2019-04-17 03:30:30 -04:00
self . latest = False
2021-03-04 03:53:47 -05:00
def __lt__ ( self , other ) : # type: (IDFToolVersion) -> bool
2019-04-17 03:30:30 -04:00
if self . status != other . status :
return self . status > other . status
else :
assert not ( self . status == IDFToolVersion . STATUS_RECOMMENDED
and other . status == IDFToolVersion . STATUS_RECOMMENDED )
return self . version < other . version
2021-03-04 03:53:47 -05:00
def __eq__ ( self , other ) : # type: (object) -> bool
if not isinstance ( other , IDFToolVersion ) :
return NotImplemented
2019-04-17 03:30:30 -04:00
return self . status == other . status and self . version == other . version
2021-03-04 03:53:47 -05:00
def add_download ( self , platform_name , url , size , sha256 ) : # type: (str, str, int, str) -> None
2019-04-17 03:30:30 -04:00
self . downloads [ platform_name ] = IDFToolDownload ( platform_name , url , size , sha256 )
2022-05-16 16:11:01 -04:00
def get_download_for_platform ( self , platform_name ) : # type: (Optional[str]) -> Optional[IDFToolDownload]
platform_name = Platforms . get ( platform_name )
if platform_name and platform_name in self . downloads . keys ( ) :
2019-04-17 03:30:30 -04:00
return self . downloads [ platform_name ]
if ' any ' in self . downloads . keys ( ) :
return self . downloads [ ' any ' ]
return None
def compatible_with_platform ( self , platform_name = PYTHON_PLATFORM ) :
2022-05-16 16:11:01 -04:00
# type: (Optional[str]) -> bool
2019-04-17 03:30:30 -04:00
return self . get_download_for_platform ( platform_name ) is not None
2021-03-04 03:53:47 -05:00
def get_supported_platforms ( self ) : # type: () -> set[str]
2020-01-17 11:38:50 -05:00
return set ( self . downloads . keys ( ) )
2019-04-17 03:30:30 -04:00
2021-06-21 04:22:54 -04:00
IDFToolOptions = namedtuple ( ' IDFToolOptions ' , [
' version_cmd ' ,
' version_regex ' ,
' version_regex_replace ' ,
2022-08-15 10:47:24 -04:00
' is_executable ' ,
2021-06-21 04:22:54 -04:00
' export_paths ' ,
' export_vars ' ,
' install ' ,
' info_url ' ,
' license ' ,
' strip_container_dirs ' ,
' supported_targets ' ] )
2019-04-17 03:30:30 -04:00
class IDFTool ( object ) :
# possible values of 'install' field
INSTALL_ALWAYS = ' always '
INSTALL_ON_REQUEST = ' on_request '
INSTALL_NEVER = ' never '
2021-04-26 15:34:48 -04:00
def __init__ ( self , name , description , install , info_url , license , version_cmd , version_regex , supported_targets , version_regex_replace = None ,
2022-08-15 10:47:24 -04:00
strip_container_dirs = 0 , is_executable = True ) :
# type: (str, str, str, str, str, List[str], str, List[str], Optional[str], int, bool) -> None
2019-04-17 03:30:30 -04:00
self . name = name
self . description = description
2022-03-15 15:02:37 -04:00
self . drop_versions ( )
2021-03-04 03:53:47 -05:00
self . version_in_path = None # type: Optional[str]
2021-06-21 04:22:54 -04:00
self . versions_installed = [ ] # type: List[str]
2019-04-17 03:30:30 -04:00
if version_regex_replace is None :
version_regex_replace = VERSION_REGEX_REPLACE_DEFAULT
2022-08-15 10:47:24 -04:00
self . options = IDFToolOptions ( version_cmd , version_regex , version_regex_replace , is_executable ,
2021-04-26 15:34:48 -04:00
[ ] , OrderedDict ( ) , install , info_url , license , strip_container_dirs , supported_targets ) # type: ignore
2021-06-21 04:22:54 -04:00
self . platform_overrides = [ ] # type: List[Dict[str, str]]
2019-07-18 11:20:40 -04:00
self . _platform = CURRENT_PLATFORM
2019-04-17 03:30:30 -04:00
self . _update_current_options ( )
2022-08-15 10:47:24 -04:00
self . is_executable = is_executable
2019-04-17 03:30:30 -04:00
2020-01-17 11:38:50 -05:00
def copy_for_platform ( self , platform ) : # type: (str) -> IDFTool
2019-07-18 11:20:40 -04:00
result = copy . deepcopy ( self )
result . _platform = platform
result . _update_current_options ( )
return result
2021-03-04 03:53:47 -05:00
def _update_current_options ( self ) : # type: () -> None
2019-04-17 03:30:30 -04:00
self . _current_options = IDFToolOptions ( * self . options )
for override in self . platform_overrides :
2022-05-16 16:11:01 -04:00
if self . _platform and self . _platform not in override [ ' platforms ' ] :
2019-04-17 03:30:30 -04:00
continue
override_dict = override . copy ( )
del override_dict [ ' platforms ' ]
2021-03-04 03:53:47 -05:00
self . _current_options = self . _current_options . _replace ( * * override_dict ) # type: ignore
2019-04-17 03:30:30 -04:00
2022-03-15 15:02:37 -04:00
def drop_versions ( self ) : # type: () -> None
self . versions = OrderedDict ( ) # type: Dict[str, IDFToolVersion]
2021-03-04 03:53:47 -05:00
def add_version ( self , version ) : # type: (IDFToolVersion) -> None
2022-08-10 03:01:57 -04:00
assert type ( version ) is IDFToolVersion
2019-04-17 03:30:30 -04:00
self . versions [ version . version ] = version
2020-01-17 11:38:50 -05:00
def get_path ( self ) : # type: () -> str
2021-11-08 11:55:03 -05:00
return os . path . join ( global_idf_tools_path or ' ' , ' tools ' , self . name )
2019-04-17 03:30:30 -04:00
2020-01-17 11:38:50 -05:00
def get_path_for_version ( self , version ) : # type: (str) -> str
2022-08-10 03:01:57 -04:00
assert version in self . versions
2019-04-17 03:30:30 -04:00
return os . path . join ( self . get_path ( ) , version )
2021-06-21 04:22:54 -04:00
def get_export_paths ( self , version ) : # type: (str) -> List[str]
2019-04-17 03:30:30 -04:00
tool_path = self . get_path_for_version ( version )
2021-03-04 03:53:47 -05:00
return [ os . path . join ( tool_path , * p ) for p in self . _current_options . export_paths ] # type: ignore
2019-04-17 03:30:30 -04:00
2021-06-21 04:22:54 -04:00
def get_export_vars ( self , version ) : # type: (str) -> Dict[str, str]
2019-04-17 03:30:30 -04:00
"""
Get the dictionary of environment variables to be exported , for the given version .
Expands :
- $ { TOOL_PATH } = > the actual path where the version is installed
"""
result = { }
2021-03-04 03:53:47 -05:00
for k , v in self . _current_options . export_vars . items ( ) : # type: ignore
2019-04-17 03:30:30 -04:00
replace_path = self . get_path_for_version ( version ) . replace ( ' \\ ' , ' \\ \\ ' )
v_repl = re . sub ( SUBST_TOOL_PATH_REGEX , replace_path , v )
if v_repl != v :
v_repl = to_shell_specific_paths ( [ v_repl ] ) [ 0 ]
2022-08-15 10:47:24 -04:00
old_v = os . environ . get ( k )
if old_v is None or old_v != v_repl :
result [ k ] = v_repl
2019-04-17 03:30:30 -04:00
return result
2022-08-31 16:59:15 -04:00
def get_version ( self , extra_paths = None , executable_path = None ) : # type: (Optional[List[str]], Optional[str]) -> str
2019-04-17 03:30:30 -04:00
"""
Execute the tool , optionally prepending extra_paths to PATH ,
extract the version string and return it as a result .
Raises ToolNotFound if the tool is not found ( not present in the paths ) .
Raises ToolExecError if the tool returns with a non - zero exit code .
Returns ' unknown ' if tool returns something from which version string
can not be extracted .
"""
2019-07-18 11:20:40 -04:00
# this function can not be called for a different platform
assert self . _platform == CURRENT_PLATFORM
2021-03-04 03:53:47 -05:00
cmd = self . _current_options . version_cmd # type: ignore
2022-08-31 16:59:15 -04:00
if executable_path :
cmd [ 0 ] = executable_path
2023-07-24 10:19:30 -04:00
if not cmd [ 0 ] :
# There is no command available, so return early. It seems that
# within some very strange context empty [''] may actually execute
# something https://github.com/espressif/esp-idf/issues/11880
raise ToolNotFound ( ' Tool {} not found ' . format ( self . name ) )
2019-04-17 03:30:30 -04:00
try :
version_cmd_result = run_cmd_check_output ( cmd , None , extra_paths )
except OSError :
# tool is not on the path
raise ToolNotFound ( ' Tool {} not found ' . format ( self . name ) )
except subprocess . CalledProcessError as e :
2021-04-12 04:06:56 -04:00
raise ToolExecError ( ' returned non-zero exit code ( {} ) with error message: \n {} ' . format (
e . returncode , e . stderr . decode ( ' utf-8 ' , errors = ' ignore ' ) ) ) # type: ignore
2019-04-17 03:30:30 -04:00
2021-01-25 21:49:01 -05:00
in_str = version_cmd_result . decode ( ' utf-8 ' )
2021-03-04 03:53:47 -05:00
match = re . search ( self . _current_options . version_regex , in_str ) # type: ignore
2019-04-17 03:30:30 -04:00
if not match :
return UNKNOWN_VERSION
2021-03-04 03:53:47 -05:00
return re . sub ( self . _current_options . version_regex , self . _current_options . version_regex_replace , match . group ( 0 ) ) # type: ignore
2019-04-17 03:30:30 -04:00
2022-08-31 16:59:15 -04:00
def check_version ( self , executable_path ) : # type: (Optional[str]) -> bool
version = self . get_version ( executable_path = executable_path )
return version in self . versions
2021-03-04 03:53:47 -05:00
def get_install_type ( self ) : # type: () -> Callable[[str], None]
return self . _current_options . install # type: ignore
2019-04-17 03:30:30 -04:00
2021-04-26 15:34:48 -04:00
def get_supported_targets ( self ) : # type: () -> list[str]
return self . _current_options . supported_targets # type: ignore
2023-09-14 01:55:24 -04:00
def is_supported_for_any_of_targets ( self , targets ) : # type: (list[str]) -> bool
"""
Checks whether the tool is suitable for at least one of the specified targets .
"""
supported_targets = self . get_supported_targets ( )
return ( any ( item in targets for item in supported_targets ) or supported_targets == [ ' all ' ] )
2021-03-04 03:53:47 -05:00
def compatible_with_platform ( self ) : # type: () -> bool
2019-07-23 00:55:30 -04:00
return any ( [ v . compatible_with_platform ( ) for v in self . versions . values ( ) ] )
2021-06-21 04:22:54 -04:00
def get_supported_platforms ( self ) : # type: () -> Set[str]
2020-01-17 11:38:50 -05:00
result = set ( )
for v in self . versions . values ( ) :
result . update ( v . get_supported_platforms ( ) )
return result
2021-03-04 03:53:47 -05:00
def get_recommended_version ( self ) : # type: () -> Optional[str]
2019-04-17 03:30:30 -04:00
recommended_versions = [ k for k , v in self . versions . items ( )
if v . status == IDFToolVersion . STATUS_RECOMMENDED
2019-07-18 11:20:40 -04:00
and v . compatible_with_platform ( self . _platform ) ]
2019-04-17 03:30:30 -04:00
assert len ( recommended_versions ) < = 1
if recommended_versions :
return recommended_versions [ 0 ]
return None
2021-03-04 03:53:47 -05:00
def get_preferred_installed_version ( self ) : # type: () -> Optional[str]
2019-04-17 03:30:30 -04:00
recommended_versions = [ k for k in self . versions_installed
if self . versions [ k ] . status == IDFToolVersion . STATUS_RECOMMENDED
2019-07-18 11:20:40 -04:00
and self . versions [ k ] . compatible_with_platform ( self . _platform ) ]
2019-04-17 03:30:30 -04:00
assert len ( recommended_versions ) < = 1
if recommended_versions :
return recommended_versions [ 0 ]
return None
2021-03-04 03:53:47 -05:00
def find_installed_versions ( self ) : # type: () -> None
2019-04-17 03:30:30 -04:00
"""
Checks whether the tool can be found in PATH and in global_idf_tools_path .
Writes results to self . version_in_path and self . versions_installed .
"""
2019-07-18 11:20:40 -04:00
# this function can not be called for a different platform
assert self . _platform == CURRENT_PLATFORM
2019-04-17 03:30:30 -04:00
# First check if the tool is in system PATH
try :
2022-08-31 16:59:15 -04:00
ver_str = self . get_version ( )
2019-04-17 03:30:30 -04:00
except ToolNotFound :
# not in PATH
pass
2021-04-12 04:06:56 -04:00
except ToolExecError as e :
warn ( ' tool {} found in path, but {} ' . format (
self . name , e ) )
2019-04-17 03:30:30 -04:00
else :
self . version_in_path = ver_str
# Now check all the versions installed in global_idf_tools_path
self . versions_installed = [ ]
for version , version_obj in self . versions . items ( ) :
if not version_obj . compatible_with_platform ( ) :
continue
tool_path = self . get_path_for_version ( version )
if not os . path . exists ( tool_path ) :
# version not installed
continue
2022-08-15 10:47:24 -04:00
if not self . is_executable :
self . versions_installed . append ( version )
continue
2019-04-17 03:30:30 -04:00
try :
2022-08-31 16:59:15 -04:00
ver_str = self . get_version ( self . get_export_paths ( version ) )
2019-04-17 03:30:30 -04:00
except ToolNotFound :
warn ( ' directory for tool {} version {} is present, but tool was not found ' . format (
self . name , version ) )
2021-04-12 04:06:56 -04:00
except ToolExecError as e :
warn ( ' tool {} version {} is installed, but {} ' . format (
self . name , version , e ) )
2019-04-17 03:30:30 -04:00
else :
if ver_str != version :
warn ( ' tool {} version {} is installed, but has reported version {} ' . format (
self . name , version , ver_str ) )
else :
self . versions_installed . append ( version )
2023-02-05 10:29:03 -05:00
def latest_installed_version ( self ) : # type: () -> Optional[str]
"""
Get the latest installed tool version by directly checking the
tool ' s version directories.
"""
tool_path = self . get_path ( )
if not os . path . exists ( tool_path ) :
return None
dentries = os . listdir ( tool_path )
dirs = [ d for d in dentries if os . path . isdir ( os . path . join ( tool_path , d ) ) ]
for version in sorted ( dirs , reverse = True ) :
# get_path_for_version() has assert to check if version is in versions
# dict, so get_export_paths() cannot be used. Let's just create the
# export paths list directly here.
paths = [ os . path . join ( tool_path , version , * p ) for p in self . _current_options . export_paths ]
try :
ver_str = self . get_version ( paths )
except ( ToolNotFound , ToolExecError ) :
continue
if ver_str != version :
continue
return version
return None
2021-03-04 03:53:47 -05:00
def download ( self , version ) : # type: (str) -> None
2022-08-10 03:01:57 -04:00
assert version in self . versions
2019-07-18 11:20:40 -04:00
download_obj = self . versions [ version ] . get_download_for_platform ( self . _platform )
2019-04-17 03:30:30 -04:00
if not download_obj :
2019-07-18 11:20:40 -04:00
fatal ( ' No packages for tool {} platform {} ! ' . format ( self . name , self . _platform ) )
2022-09-22 10:14:17 -04:00
raise SystemExit ( 1 )
2019-04-17 03:30:30 -04:00
url = download_obj . url
archive_name = os . path . basename ( url )
2021-11-08 11:55:03 -05:00
local_path = os . path . join ( global_idf_tools_path or ' ' , ' dist ' , archive_name )
2019-04-17 03:30:30 -04:00
mkdir_p ( os . path . dirname ( local_path ) )
if os . path . isfile ( local_path ) :
if not self . check_download_file ( download_obj , local_path ) :
warn ( ' removing downloaded file {0} and downloading again ' . format ( archive_name ) )
os . unlink ( local_path )
else :
info ( ' file {0} is already downloaded ' . format ( archive_name ) )
return
downloaded = False
2021-09-16 10:48:03 -04:00
local_temp_path = local_path + ' .tmp '
2019-04-17 03:30:30 -04:00
for retry in range ( DOWNLOAD_RETRY_COUNT ) :
2022-09-22 10:14:17 -04:00
err = download ( url , local_temp_path )
2019-11-15 08:31:19 -05:00
if not os . path . isfile ( local_temp_path ) or not self . check_download_file ( download_obj , local_temp_path ) :
2022-09-22 10:14:17 -04:00
warn ( ' Download failure: {} ' . format ( err ) )
2019-11-15 08:31:19 -05:00
warn ( ' Failed to download {} to {} ' . format ( url , local_temp_path ) )
2019-04-17 03:30:30 -04:00
continue
2019-09-23 00:01:09 -04:00
rename_with_retry ( local_temp_path , local_path )
2019-04-17 03:30:30 -04:00
downloaded = True
break
if not downloaded :
fatal ( ' Failed to download, and retry count has expired ' )
2022-09-22 10:14:17 -04:00
print_hints_on_download_error ( str ( err ) )
raise SystemExit ( 1 )
2019-04-17 03:30:30 -04:00
2021-03-04 03:53:47 -05:00
def install ( self , version ) : # type: (str) -> None
2019-04-17 03:30:30 -04:00
# Currently this is called after calling 'download' method, so here are a few asserts
# for the conditions which should be true once that method is done.
2022-08-10 03:01:57 -04:00
assert version in self . versions
2019-07-18 11:20:40 -04:00
download_obj = self . versions [ version ] . get_download_for_platform ( self . _platform )
2022-08-10 03:01:57 -04:00
assert download_obj is not None
2019-04-17 03:30:30 -04:00
archive_name = os . path . basename ( download_obj . url )
2021-11-08 11:55:03 -05:00
archive_path = os . path . join ( global_idf_tools_path or ' ' , ' dist ' , archive_name )
2022-08-10 03:01:57 -04:00
assert os . path . isfile ( archive_path )
2019-04-17 03:30:30 -04:00
dest_dir = self . get_path_for_version ( version )
if os . path . exists ( dest_dir ) :
warn ( ' destination path already exists, removing ' )
shutil . rmtree ( dest_dir )
mkdir_p ( dest_dir )
unpack ( archive_path , dest_dir )
2021-03-04 03:53:47 -05:00
if self . _current_options . strip_container_dirs : # type: ignore
2022-08-15 10:47:24 -04:00
do_strip_container_dirs ( dest_dir , self . _current_options . strip_container_dirs ) # type: ignore
2019-04-17 03:30:30 -04:00
@staticmethod
2021-03-04 03:53:47 -05:00
def check_download_file ( download_obj , local_path ) : # type: (IDFToolDownload, str) -> bool
2019-04-17 03:30:30 -04:00
expected_sha256 = download_obj . sha256
expected_size = download_obj . size
file_size , file_sha256 = get_file_size_sha256 ( local_path )
if file_size != expected_size :
2019-06-28 08:23:21 -04:00
warn ( ' file size mismatch for {} , expected {} , got {} ' . format ( local_path , expected_size , file_size ) )
2019-04-17 03:30:30 -04:00
return False
if file_sha256 != expected_sha256 :
2019-06-28 08:23:21 -04:00
warn ( ' hash mismatch for {} , expected {} , got {} ' . format ( local_path , expected_sha256 , file_sha256 ) )
2019-04-17 03:30:30 -04:00
return False
return True
@classmethod
2021-06-21 04:22:54 -04:00
def from_json ( cls , tool_dict ) : # type: (Dict[str, Union[str, List[str], Dict[str, str]]]) -> IDFTool
2019-04-17 03:30:30 -04:00
# Validate json fields
2021-03-04 03:53:47 -05:00
tool_name = tool_dict . get ( ' name ' ) # type: ignore
2022-08-15 10:47:24 -04:00
if not isinstance ( tool_name , str ) :
2019-04-17 03:30:30 -04:00
raise RuntimeError ( ' tool_name is not a string ' )
2021-03-04 03:53:47 -05:00
description = tool_dict . get ( ' description ' ) # type: ignore
2022-08-15 10:47:24 -04:00
if not isinstance ( description , str ) :
2019-04-17 03:30:30 -04:00
raise RuntimeError ( ' description is not a string ' )
2022-08-15 10:47:24 -04:00
is_executable = tool_dict . get ( ' is_executable ' , True ) # type: ignore
if not isinstance ( is_executable , bool ) :
raise RuntimeError ( ' is_executable for tool %s is not a bool ' % tool_name )
2019-04-17 03:30:30 -04:00
version_cmd = tool_dict . get ( ' version_cmd ' )
if type ( version_cmd ) is not list :
raise RuntimeError ( ' version_cmd for tool %s is not a list of strings ' % tool_name )
version_regex = tool_dict . get ( ' version_regex ' )
2022-08-15 10:47:24 -04:00
if not isinstance ( version_regex , str ) or ( not version_regex and is_executable ) :
2019-04-17 03:30:30 -04:00
raise RuntimeError ( ' version_regex for tool %s is not a non-empty string ' % tool_name )
version_regex_replace = tool_dict . get ( ' version_regex_replace ' )
2022-08-15 10:47:24 -04:00
if version_regex_replace and not isinstance ( version_regex_replace , str ) :
2019-04-17 03:30:30 -04:00
raise RuntimeError ( ' version_regex_replace for tool %s is not a string ' % tool_name )
export_paths = tool_dict . get ( ' export_paths ' )
if type ( export_paths ) is not list :
raise RuntimeError ( ' export_paths for tool %s is not a list ' % tool_name )
2021-03-04 03:53:47 -05:00
export_vars = tool_dict . get ( ' export_vars ' , { } ) # type: ignore
2019-04-17 03:30:30 -04:00
if type ( export_vars ) is not dict :
raise RuntimeError ( ' export_vars for tool %s is not a mapping ' % tool_name )
versions = tool_dict . get ( ' versions ' )
if type ( versions ) is not list :
raise RuntimeError ( ' versions for tool %s is not an array ' % tool_name )
2021-03-04 03:53:47 -05:00
install = tool_dict . get ( ' install ' , False ) # type: ignore
2022-08-15 10:47:24 -04:00
if not isinstance ( install , str ) :
2019-04-17 03:30:30 -04:00
raise RuntimeError ( ' install for tool %s is not a string ' % tool_name )
2021-03-04 03:53:47 -05:00
info_url = tool_dict . get ( ' info_url ' , False ) # type: ignore
2022-08-15 10:47:24 -04:00
if not isinstance ( info_url , str ) :
2019-04-17 03:30:30 -04:00
raise RuntimeError ( ' info_url for tool %s is not a string ' % tool_name )
2021-03-04 03:53:47 -05:00
license = tool_dict . get ( ' license ' , False ) # type: ignore
2022-08-15 10:47:24 -04:00
if not isinstance ( license , str ) :
2019-04-17 03:30:30 -04:00
raise RuntimeError ( ' license for tool %s is not a string ' % tool_name )
strip_container_dirs = tool_dict . get ( ' strip_container_dirs ' , 0 )
if strip_container_dirs and type ( strip_container_dirs ) is not int :
raise RuntimeError ( ' strip_container_dirs for tool %s is not an int ' % tool_name )
2021-03-04 03:53:47 -05:00
overrides_list = tool_dict . get ( ' platform_overrides ' , [ ] ) # type: ignore
2019-04-17 03:30:30 -04:00
if type ( overrides_list ) is not list :
raise RuntimeError ( ' platform_overrides for tool %s is not a list ' % tool_name )
2021-04-26 15:34:48 -04:00
supported_targets = tool_dict . get ( ' supported_targets ' )
if not isinstance ( supported_targets , list ) :
raise RuntimeError ( ' supported_targets for tool %s is not a list of strings ' % tool_name )
2019-04-17 03:30:30 -04:00
# Create the object
2021-03-04 03:53:47 -05:00
tool_obj = cls ( tool_name , description , install , info_url , license , # type: ignore
2021-04-26 15:34:48 -04:00
version_cmd , version_regex , supported_targets , version_regex_replace , # type: ignore
2022-08-15 10:47:24 -04:00
strip_container_dirs , is_executable ) # type: ignore
2019-04-17 03:30:30 -04:00
2021-03-04 03:53:47 -05:00
for path in export_paths : # type: ignore
tool_obj . options . export_paths . append ( path ) # type: ignore
2019-04-17 03:30:30 -04:00
2021-03-04 03:53:47 -05:00
for name , value in export_vars . items ( ) : # type: ignore
tool_obj . options . export_vars [ name ] = value # type: ignore
2019-04-17 03:30:30 -04:00
for index , override in enumerate ( overrides_list ) :
2021-03-04 03:53:47 -05:00
platforms_list = override . get ( ' platforms ' ) # type: ignore
2019-04-17 03:30:30 -04:00
if type ( platforms_list ) is not list :
raise RuntimeError ( ' platforms for override %d of tool %s is not a list ' % ( index , tool_name ) )
2021-03-04 03:53:47 -05:00
install = override . get ( ' install ' ) # type: ignore
2022-08-15 10:47:24 -04:00
if install is not None and not isinstance ( install , str ) :
2019-04-17 03:30:30 -04:00
raise RuntimeError ( ' install for override %d of tool %s is not a string ' % ( index , tool_name ) )
2021-03-04 03:53:47 -05:00
version_cmd = override . get ( ' version_cmd ' ) # type: ignore
2019-04-17 03:30:30 -04:00
if version_cmd is not None and type ( version_cmd ) is not list :
raise RuntimeError ( ' version_cmd for override %d of tool %s is not a list of strings ' %
( index , tool_name ) )
2021-03-04 03:53:47 -05:00
version_regex = override . get ( ' version_regex ' ) # type: ignore
2022-08-15 10:47:24 -04:00
if version_regex is not None and ( not isinstance ( version_regex , str ) or not version_regex ) :
2019-04-17 03:30:30 -04:00
raise RuntimeError ( ' version_regex for override %d of tool %s is not a non-empty string ' %
( index , tool_name ) )
2021-03-04 03:53:47 -05:00
version_regex_replace = override . get ( ' version_regex_replace ' ) # type: ignore
2022-08-15 10:47:24 -04:00
if version_regex_replace is not None and not isinstance ( version_regex_replace , str ) :
2019-04-17 03:30:30 -04:00
raise RuntimeError ( ' version_regex_replace for override %d of tool %s is not a string ' %
( index , tool_name ) )
2021-03-04 03:53:47 -05:00
export_paths = override . get ( ' export_paths ' ) # type: ignore
2019-04-17 03:30:30 -04:00
if export_paths is not None and type ( export_paths ) is not list :
raise RuntimeError ( ' export_paths for override %d of tool %s is not a list ' % ( index , tool_name ) )
2021-03-04 03:53:47 -05:00
export_vars = override . get ( ' export_vars ' ) # type: ignore
2019-04-17 03:30:30 -04:00
if export_vars is not None and type ( export_vars ) is not dict :
raise RuntimeError ( ' export_vars for override %d of tool %s is not a mapping ' % ( index , tool_name ) )
2021-03-04 03:53:47 -05:00
tool_obj . platform_overrides . append ( override ) # type: ignore
2019-04-17 03:30:30 -04:00
2021-03-04 03:53:47 -05:00
recommended_versions = { } # type: dict[str, list[str]]
for version_dict in versions : # type: ignore
version = version_dict . get ( ' name ' ) # type: ignore
2022-08-15 10:47:24 -04:00
if not isinstance ( version , str ) :
2019-04-17 03:30:30 -04:00
raise RuntimeError ( ' version name for tool {} is not a string ' . format ( tool_name ) )
2021-03-04 03:53:47 -05:00
version_status = version_dict . get ( ' status ' ) # type: ignore
2022-08-15 10:47:24 -04:00
if not isinstance ( version_status , str ) and version_status not in IDFToolVersion . STATUS_VALUES :
2019-04-17 03:30:30 -04:00
raise RuntimeError ( ' tool {} version {} status is not one of {} ' , tool_name , version ,
IDFToolVersion . STATUS_VALUES )
version_obj = IDFToolVersion ( version , version_status )
2021-03-04 03:53:47 -05:00
for platform_id , platform_dict in version_dict . items ( ) : # type: ignore
2019-04-17 03:30:30 -04:00
if platform_id in [ ' name ' , ' status ' ] :
continue
2022-05-16 16:11:01 -04:00
if Platforms . get ( platform_id ) is None :
2019-04-17 03:30:30 -04:00
raise RuntimeError ( ' invalid platform %s for tool %s version %s ' %
( platform_id , tool_name , version ) )
version_obj . add_download ( platform_id ,
platform_dict [ ' url ' ] , platform_dict [ ' size ' ] , platform_dict [ ' sha256 ' ] )
if version_status == IDFToolVersion . STATUS_RECOMMENDED :
if platform_id not in recommended_versions :
recommended_versions [ platform_id ] = [ ]
recommended_versions [ platform_id ] . append ( version )
tool_obj . add_version ( version_obj )
for platform_id , version_list in recommended_versions . items ( ) :
if len ( version_list ) > 1 :
raise RuntimeError ( ' tool {} for platform {} has {} recommended versions ' . format (
tool_name , platform_id , len ( recommended_versions ) ) )
if install != IDFTool . INSTALL_NEVER and len ( recommended_versions ) == 0 :
raise RuntimeError ( ' required/optional tool {} for platform {} has no recommended versions ' . format (
tool_name , platform_id ) )
tool_obj . _update_current_options ( )
return tool_obj
2021-03-04 03:53:47 -05:00
def to_json ( self ) : # type: ignore
2019-04-17 03:30:30 -04:00
versions_array = [ ]
for version , version_obj in self . versions . items ( ) :
version_json = {
' name ' : version ,
' status ' : version_obj . status
}
for platform_id , download in version_obj . downloads . items ( ) :
version_json [ platform_id ] = {
' url ' : download . url ,
' size ' : download . size ,
' sha256 ' : download . sha256
}
versions_array . append ( version_json )
overrides_array = self . platform_overrides
tool_json = {
' name ' : self . name ,
' description ' : self . description ,
' export_paths ' : self . options . export_paths ,
' export_vars ' : self . options . export_vars ,
' install ' : self . options . install ,
' info_url ' : self . options . info_url ,
' license ' : self . options . license ,
' version_cmd ' : self . options . version_cmd ,
' version_regex ' : self . options . version_regex ,
2021-04-26 15:34:48 -04:00
' supported_targets ' : self . options . supported_targets ,
2019-04-17 03:30:30 -04:00
' versions ' : versions_array ,
}
if self . options . version_regex_replace != VERSION_REGEX_REPLACE_DEFAULT :
tool_json [ ' version_regex_replace ' ] = self . options . version_regex_replace
if overrides_array :
tool_json [ ' platform_overrides ' ] = overrides_array
if self . options . strip_container_dirs :
tool_json [ ' strip_container_dirs ' ] = self . options . strip_container_dirs
2022-08-15 10:47:24 -04:00
if self . options . is_executable is False :
tool_json [ ' is_executable ' ] = self . options . is_executable
2019-04-17 03:30:30 -04:00
return tool_json
2022-03-19 16:37:16 -04:00
class IDFEnvEncoder ( JSONEncoder ) :
"""
IDFEnvEncoder is used for encoding IDFEnv , IDFRecord , SelectedIDFRecord classes to JSON in readable format . Not as ( __main__ . IDFRecord object at ' 0x7fcxx ' )
Additionally remove first underscore with private properties when processing
"""
def default ( self , obj ) : # type: ignore
return { k . lstrip ( ' _ ' ) : v for k , v in vars ( obj ) . items ( ) }
class IDFRecord :
"""
IDFRecord represents one record of installed ESP - IDF on system .
Contains :
* version - actual version of ESP - IDF ( example ' 5.0 ' )
* path - absolute path to the ESP - IDF
* features - features using ESP - IDF
* targets - ESP chips for which are installed needed toolchains ( example [ ' esp32 ' , ' esp32s2 ' ] )
- Default value is [ ] , since user didn ' t define any targets yet
"""
def __init__ ( self ) - > None :
self . version = ' ' # type: str
self . path = ' ' # type: str
self . _features = [ ' core ' ] # type: list[str]
self . _targets = [ ] # type: list[str]
def __iter__ ( self ) : # type: ignore
yield from {
' version ' : self . version ,
' path ' : self . path ,
' features ' : self . _features ,
' targets ' : self . _targets
} . items ( )
def __str__ ( self ) - > str :
return json . dumps ( dict ( self ) , ensure_ascii = False , indent = 4 ) # type: ignore
def __repr__ ( self ) - > str :
return self . __str__ ( )
2022-08-09 12:13:29 -04:00
def __eq__ ( self , other : object ) - > bool :
if not isinstance ( other , IDFRecord ) :
return False
return all ( getattr ( self , x ) == getattr ( other , x ) for x in ( ' version ' , ' path ' , ' features ' , ' targets ' ) )
def __ne__ ( self , other : object ) - > bool :
if not isinstance ( other , IDFRecord ) :
return False
return not self . __eq__ ( other )
2022-03-19 16:37:16 -04:00
@property
def features ( self ) - > List [ str ] :
return self . _features
2022-05-20 09:15:52 -04:00
def update_features ( self , add : Tuple [ str , . . . ] = ( ) , remove : Tuple [ str , . . . ] = ( ) ) - > None :
# Update features, but maintain required feature 'core'
# If the same feature is present in both argument's tuples, do not update this feature
add_set = set ( add )
remove_set = set ( remove )
# Remove duplicates
features_to_add = add_set . difference ( remove_set )
features_to_remove = remove_set . difference ( add_set )
features = set ( self . _features )
features . update ( features_to_add )
features . difference_update ( features_to_remove )
features . add ( ' core ' )
self . _features = list ( features )
2022-03-19 16:37:16 -04:00
@property
def targets ( self ) - > List [ str ] :
return self . _targets
def extend_targets ( self , targets : List [ str ] ) - > None :
# Targets can be only updated, but always maintain existing targets.
self . _targets = list ( set ( targets + self . _targets ) )
@classmethod
def get_active_idf_record ( cls ) : # type: () -> IDFRecord
idf_record_obj = cls ( )
idf_record_obj . version = get_idf_version ( )
idf_record_obj . path = global_idf_path or ' '
return idf_record_obj
@classmethod
def get_idf_record_from_dict ( cls , record_dict ) : # type: (Dict[str, Any]) -> IDFRecord
idf_record_obj = cls ( )
try :
idf_record_obj . version = record_dict [ ' version ' ]
idf_record_obj . path = record_dict [ ' path ' ]
except KeyError :
# When some of these key attributes, which are irreplaceable with default values, are not found, raise VallueError
raise ValueError ( ' Inconsistent record ' )
2022-05-20 09:15:52 -04:00
idf_record_obj . update_features ( record_dict . get ( ' features ' , [ ] ) )
2022-03-19 16:37:16 -04:00
idf_record_obj . extend_targets ( record_dict . get ( ' targets ' , [ ] ) )
return idf_record_obj
class IDFEnv :
"""
2022-08-09 12:13:29 -04:00
IDFEnv represents ESP - IDF Environments installed on system and is responsible for loading and saving structured data
All information is saved and loaded from IDF_ENV_FILE
2022-03-19 16:37:16 -04:00
Contains :
* idf_installed - all installed environments of ESP - IDF on system
"""
def __init__ ( self ) - > None :
active_idf_id = active_repo_id ( )
self . idf_installed = { active_idf_id : IDFRecord . get_active_idf_record ( ) } # type: Dict[str, IDFRecord]
def __iter__ ( self ) : # type: ignore
yield from {
' idfInstalled ' : self . idf_installed ,
} . items ( )
def __str__ ( self ) - > str :
return json . dumps ( dict ( self ) , cls = IDFEnvEncoder , ensure_ascii = False , indent = 4 ) # type: ignore
def __repr__ ( self ) - > str :
return self . __str__ ( )
def save ( self ) - > None :
2022-08-09 12:13:29 -04:00
"""
Diff current class instance with instance loaded from IDF_ENV_FILE and save only if are different
"""
# It is enough to compare just active records because others can't be touched by the running script
if self . get_active_idf_record ( ) != self . get_idf_env ( ) . get_active_idf_record ( ) :
idf_env_file_path = os . path . join ( global_idf_tools_path or ' ' , IDF_ENV_FILE )
try :
if global_idf_tools_path : # mypy fix for Optional[str] in the next call
# the directory doesn't exist if this is run on a clean system the first time
mkdir_p ( global_idf_tools_path )
2023-06-01 04:44:54 -04:00
with open ( idf_env_file_path , ' w ' , encoding = ' utf-8 ' ) as w :
2022-08-09 12:13:29 -04:00
info ( ' Updating {} ' . format ( idf_env_file_path ) )
json . dump ( dict ( self ) , w , cls = IDFEnvEncoder , ensure_ascii = False , indent = 4 ) # type: ignore
except ( IOError , OSError ) :
if not os . access ( global_idf_tools_path or ' ' , os . W_OK ) :
raise OSError ( ' IDF_TOOLS_PATH {} is not accessible to write. Required changes have not been saved ' . format ( global_idf_tools_path or ' ' ) )
raise OSError ( ' File {} is not accessible to write or corrupted. Required changes have not been saved ' . format ( idf_env_file_path ) )
2022-03-19 16:37:16 -04:00
def get_active_idf_record ( self ) - > IDFRecord :
return self . idf_installed [ active_repo_id ( ) ]
@classmethod
def get_idf_env ( cls ) : # type: () -> IDFEnv
# IDFEnv class is used to process IDF_ENV_FILE file. The constructor is therefore called only in this method that loads the file and checks its contents
idf_env_obj = cls ( )
try :
idf_env_file_path = os . path . join ( global_idf_tools_path or ' ' , IDF_ENV_FILE )
2023-06-01 04:44:54 -04:00
with open ( idf_env_file_path , ' r ' , encoding = ' utf-8 ' ) as idf_env_file :
2022-03-19 16:37:16 -04:00
idf_env_json = json . load ( idf_env_file )
try :
idf_installed = idf_env_json [ ' idfInstalled ' ]
except KeyError :
# If no ESP-IDF record is found in loaded file, do not update and keep default value from constructor
pass
else :
# Load and verify ESP-IDF records found in IDF_ENV_FILE
idf_installed . pop ( ' sha ' , None )
idf_installed_verified = { } # type: dict[str, IDFRecord]
for idf in idf_installed :
try :
idf_installed_verified [ idf ] = IDFRecord . get_idf_record_from_dict ( idf_installed [ idf ] )
except ValueError as err :
warn ( ' {} " {} " found in {} , removing this record. ' . format ( err , idf , idf_env_file_path ) )
# Combine ESP-IDF loaded records with the one in constructor, to be sure that there is an active ESP-IDF record in the idf_installed
# If the active record is already in idf_installed, it is not overwritten
idf_env_obj . idf_installed = dict ( idf_env_obj . idf_installed , * * idf_installed_verified )
except ( IOError , OSError , ValueError ) :
# If no, empty or not-accessible to read IDF_ENV_FILE found, use default values from constructor
pass
return idf_env_obj
2022-08-09 12:13:29 -04:00
class ENVState :
"""
ENVState is used to handle IDF global variables that are set in environment and need to be removed when switching between ESP - IDF versions in opened shell
Every opened shell / terminal has it ' s own temporary file to store these variables
The temporary file ' s name is generated automatically with suffix ' idf_ + opened shell ID ' . Path to this tmp file is stored as env global variable (env_key)
The shell ID is crucial , since in one terminal can be opened more shells
* env_key - global variable name / key
* deactivate_file_path - global variable value ( generated tmp file name )
* idf_variables - loaded IDF variables from file
"""
env_key = ' IDF_DEACTIVATE_FILE_PATH '
deactivate_file_path = os . environ . get ( env_key , ' ' )
def __init__ ( self ) - > None :
self . idf_variables = { } # type: Dict[str, Any]
@classmethod
def get_env_state ( cls ) : # type: () -> ENVState
env_state_obj = cls ( )
if cls . deactivate_file_path :
try :
with open ( cls . deactivate_file_path , ' r ' ) as fp :
env_state_obj . idf_variables = json . load ( fp )
except ( IOError , OSError , ValueError ) :
pass
return env_state_obj
def save ( self ) - > str :
try :
if self . deactivate_file_path and os . path . basename ( self . deactivate_file_path ) . endswith ( ' idf_ ' + str ( os . getppid ( ) ) ) :
# If exported file path/name exists and belongs to actual opened shell
with open ( self . deactivate_file_path , ' w ' ) as w :
json . dump ( self . idf_variables , w , ensure_ascii = False , indent = 4 ) # type: ignore
else :
with tempfile . NamedTemporaryFile ( delete = False , suffix = ' idf_ ' + str ( os . getppid ( ) ) ) as fp :
self . deactivate_file_path = fp . name
fp . write ( json . dumps ( self . idf_variables , ensure_ascii = False , indent = 4 ) . encode ( ' utf-8 ' ) )
except ( IOError , OSError ) :
warn ( ' File storing IDF env variables {} is not accessible to write. '
' Potentional switching ESP-IDF versions may cause problems ' . format ( self . deactivate_file_path ) )
return self . deactivate_file_path
2021-03-04 03:53:47 -05:00
def load_tools_info ( ) : # type: () -> dict[str, IDFTool]
2019-04-17 03:30:30 -04:00
"""
Load tools metadata from tools . json , return a dictionary : tool name - tool info
"""
tool_versions_file_name = global_tools_json
2021-03-04 03:53:47 -05:00
with open ( tool_versions_file_name , ' r ' ) as f : # type: ignore
2019-04-17 03:30:30 -04:00
tools_info = json . load ( f )
2021-03-04 03:53:47 -05:00
return parse_tools_info_json ( tools_info ) # type: ignore
2019-04-17 03:30:30 -04:00
2021-03-04 03:53:47 -05:00
def parse_tools_info_json ( tools_info ) : # type: ignore
2019-04-17 03:30:30 -04:00
"""
Parse and validate the dictionary obtained by loading the tools . json file .
Returns a dictionary of tools ( key : tool name , value : IDFTool object ) .
"""
if tools_info [ ' version ' ] != TOOLS_FILE_VERSION :
raise RuntimeError ( ' Invalid version ' )
tools_dict = OrderedDict ( )
tools_array = tools_info . get ( ' tools ' )
if type ( tools_array ) is not list :
raise RuntimeError ( ' tools property is missing or not an array ' )
for tool_dict in tools_array :
tool = IDFTool . from_json ( tool_dict )
tools_dict [ tool . name ] = tool
return tools_dict
2021-03-04 03:53:47 -05:00
def dump_tools_json ( tools_info ) : # type: ignore
2019-04-17 03:30:30 -04:00
tools_array = [ ]
for tool_name , tool_obj in tools_info . items ( ) :
tool_json = tool_obj . to_json ( )
tools_array . append ( tool_json )
file_json = { ' version ' : TOOLS_FILE_VERSION , ' tools ' : tools_array }
return json . dumps ( file_json , indent = 2 , separators = ( ' , ' , ' : ' ) , sort_keys = True )
2021-12-13 10:45:11 -05:00
def get_python_exe_and_subdir ( ) - > Tuple [ str , str ] :
if sys . platform == ' win32 ' :
subdir = ' Scripts '
python_exe = ' python.exe '
else :
subdir = ' bin '
python_exe = ' python '
return python_exe , subdir
2021-11-08 11:55:03 -05:00
def get_idf_version ( ) - > str :
2021-03-04 03:53:47 -05:00
version_file_path = os . path . join ( global_idf_path , ' version.txt ' ) # type: ignore
2019-04-17 03:30:30 -04:00
if os . path . exists ( version_file_path ) :
2021-01-25 21:49:01 -05:00
with open ( version_file_path , ' r ' ) as version_file :
2019-04-17 03:30:30 -04:00
idf_version_str = version_file . read ( )
else :
2021-06-23 05:46:00 -04:00
idf_version_str = ' '
2019-11-18 04:12:05 -05:00
try :
2020-10-06 18:36:40 -04:00
idf_version_str = subprocess . check_output ( [ ' git ' , ' describe ' ] ,
2022-02-10 11:29:41 -05:00
cwd = global_idf_path , env = os . environ ,
stderr = subprocess . DEVNULL ) . decode ( )
2021-06-23 05:46:00 -04:00
except OSError :
# OSError should cover FileNotFoundError and WindowsError
warn ( ' Git was not found ' )
2022-02-10 11:29:41 -05:00
except subprocess . CalledProcessError :
# This happens quite often when the repo is shallow. Don't print a warning because there are other
# possibilities for version detection.
pass
2019-04-17 03:30:30 -04:00
match = re . match ( r ' ^v([0-9]+ \ .[0-9]+).* ' , idf_version_str )
2019-11-18 04:12:05 -05:00
if match :
2021-03-04 03:53:47 -05:00
idf_version = match . group ( 1 ) # type: Optional[str]
2019-11-18 04:12:05 -05:00
else :
idf_version = None
# fallback when IDF is a shallow clone
try :
2021-03-04 03:53:47 -05:00
with open ( os . path . join ( global_idf_path , ' components ' , ' esp_common ' , ' include ' , ' esp_idf_version.h ' ) ) as f : # type: ignore
2019-11-18 04:12:05 -05:00
m = re . search ( r ' ^#define \ s+ESP_IDF_VERSION_MAJOR \ s+( \ d+).+?^#define \ s+ESP_IDF_VERSION_MINOR \ s+( \ d+) ' ,
f . read ( ) , re . DOTALL | re . MULTILINE )
if m :
idf_version = ' . ' . join ( ( m . group ( 1 ) , m . group ( 2 ) ) )
else :
warn ( ' Reading IDF version from C header file failed! ' )
except Exception as e :
warn ( ' Is it not possible to determine the IDF version: {} ' . format ( e ) )
if idf_version is None :
fatal ( ' IDF version cannot be determined ' )
raise SystemExit ( 1 )
2019-04-17 03:30:30 -04:00
2021-11-08 11:55:03 -05:00
return idf_version
def get_python_env_path ( ) - > Tuple [ str , str , str , str ] :
python_ver_major_minor = ' {} . {} ' . format ( sys . version_info . major , sys . version_info . minor )
idf_version = get_idf_version ( )
2023-02-14 10:14:54 -05:00
idf_python_env_path = os . getenv ( ' IDF_PYTHON_ENV_PATH ' ) or os . path . join ( global_idf_tools_path or ' ' , ' python_env ' ,
' idf {} _py {} _env ' . format ( idf_version , python_ver_major_minor ) )
2019-04-17 03:30:30 -04:00
2021-12-13 10:45:11 -05:00
python_exe , subdir = get_python_exe_and_subdir ( )
2019-04-17 03:30:30 -04:00
idf_python_export_path = os . path . join ( idf_python_env_path , subdir )
virtualenv_python = os . path . join ( idf_python_export_path , python_exe )
2021-09-16 10:48:03 -04:00
return idf_python_env_path , idf_python_export_path , virtualenv_python , idf_version
2019-04-17 03:30:30 -04:00
2023-09-14 01:55:24 -04:00
def parse_tools_arg ( tools_str ) : # type: (List[str]) -> List[str]
2021-11-08 11:55:03 -05:00
"""
2023-09-14 01:55:24 -04:00
Base parsing " tools " argumets : all , required , etc
"""
if not tools_str :
return [ ' required ' ]
else :
return tools_str
def expand_tools_arg ( tools_spec , overall_tools , targets ) : # type: (list[str], OrderedDict, list[str]) -> list[str]
""" Expand list of tools ' tools_spec ' in according:
- a tool is in the ' overall_tools ' list
- consider metapackages like " required " and " all "
- process wildcards in tool names
- a tool supports chips from ' targets '
"""
tools = [ ]
# Filtering tools if they are in overall_tools
# Processing wildcards if possible
for tool_pattern in tools_spec :
tools . extend ( [ k for k , _ in overall_tools . items ( ) if fnmatch . fnmatch ( k , tool_pattern ) and k not in tools ] )
# Processing "metapackage"
if ' required ' in tools_spec :
tools . extend ( [ k for k , v in overall_tools . items ( ) if v . get_install_type ( ) == IDFTool . INSTALL_ALWAYS and k not in tools ] )
elif ' all ' in tools_spec :
tools . extend ( [ k for k , v in overall_tools . items ( ) if v . get_install_type ( ) != IDFTool . INSTALL_NEVER and k not in tools ] )
# Filtering by ESP_targets
tools = [ k for k in tools if overall_tools [ k ] . is_supported_for_any_of_targets ( targets ) ]
return tools
def parse_targets_arg ( targets_str ) : # type: (str) -> List[str]
"""
Parse and check if targets_str is a valid list of targets and return a target list
2021-11-08 11:55:03 -05:00
"""
2021-04-26 15:34:48 -04:00
targets_from_tools_json = get_all_targets_from_tools_json ( )
invalid_targets = [ ]
targets_str = targets_str . lower ( )
targets = targets_str . replace ( ' - ' , ' ' ) . split ( ' , ' )
2023-09-14 01:55:24 -04:00
if targets == [ ' all ' ] :
return targets_from_tools_json
else :
2021-04-26 15:34:48 -04:00
invalid_targets = [ t for t in targets if t not in targets_from_tools_json ]
if invalid_targets :
warn ( ' Targets: " {} " are not supported. Only allowed options are: {} . ' . format ( ' , ' . join ( invalid_targets ) , ' , ' . join ( targets_from_tools_json ) ) )
raise SystemExit ( 1 )
2023-09-14 01:55:24 -04:00
return targets
2021-11-08 11:55:03 -05:00
2023-09-14 01:55:24 -04:00
def add_and_check_targets ( idf_env_obj , targets_str ) : # type: (IDFEnv, str) -> list[str]
"""
Define targets from targets_str , check that the target names are valid and add them to idf_env_obj
"""
targets = parse_targets_arg ( targets_str )
idf_env_obj . get_active_idf_record ( ) . extend_targets ( targets )
2022-03-19 16:37:16 -04:00
return idf_env_obj . get_active_idf_record ( ) . targets
2021-04-26 15:34:48 -04:00
2021-09-16 10:48:03 -04:00
def feature_to_requirements_path ( feature ) : # type: (str) -> str
2022-02-01 13:11:46 -05:00
return os . path . join ( global_idf_path or ' ' , ' tools ' , ' requirements ' , ' requirements. {} .txt ' . format ( feature ) )
2021-09-16 10:48:03 -04:00
2022-05-20 09:15:52 -04:00
def process_and_check_features ( idf_env_obj , features_str ) : # type: (IDFEnv, str) -> list[str]
2022-03-19 16:37:16 -04:00
new_features = [ ]
2022-05-20 09:15:52 -04:00
remove_features = [ ]
2021-09-16 10:48:03 -04:00
for new_feature_candidate in features_str . split ( ' , ' ) :
2022-05-20 09:15:52 -04:00
if new_feature_candidate . startswith ( ' - ' ) :
remove_features + = [ new_feature_candidate . lstrip ( ' - ' ) ]
else :
new_feature_candidate = new_feature_candidate . lstrip ( ' + ' )
# Feature to be added needs to be checked if is valid
if os . path . isfile ( feature_to_requirements_path ( new_feature_candidate ) ) :
new_features + = [ new_feature_candidate ]
idf_env_obj . get_active_idf_record ( ) . update_features ( tuple ( new_features ) , tuple ( remove_features ) )
2022-03-19 16:37:16 -04:00
return idf_env_obj . get_active_idf_record ( ) . features
2021-04-26 15:34:48 -04:00
def get_all_targets_from_tools_json ( ) : # type: () -> list[str]
tools_info = load_tools_info ( )
targets_from_tools_json = [ ] # type: list[str]
for _ , v in tools_info . items ( ) :
targets_from_tools_json . extend ( v . get_supported_targets ( ) )
# remove duplicates
targets_from_tools_json = list ( set ( targets_from_tools_json ) )
if ' all ' in targets_from_tools_json :
targets_from_tools_json . remove ( ' all ' )
return sorted ( targets_from_tools_json )
2022-03-19 16:37:16 -04:00
def filter_tools_info ( idf_env_obj , tools_info ) : # type: (IDFEnv, OrderedDict[str, IDFTool]) -> OrderedDict[str,IDFTool]
targets = idf_env_obj . get_active_idf_record ( ) . targets
2021-04-26 15:34:48 -04:00
if not targets :
return tools_info
else :
filtered_tools_spec = { k : v for k , v in tools_info . items ( ) if
( v . get_install_type ( ) == IDFTool . INSTALL_ALWAYS or v . get_install_type ( ) == IDFTool . INSTALL_ON_REQUEST ) and
( any ( item in targets for item in v . get_supported_targets ( ) ) or v . get_supported_targets ( ) == [ ' all ' ] ) }
return OrderedDict ( filtered_tools_spec )
2022-08-09 12:13:29 -04:00
def add_variables_to_deactivate_file ( args , new_idf_vars ) : # type: (list[str], dict[str, Any]) -> str
2021-11-08 11:55:03 -05:00
"""
2022-08-09 12:13:29 -04:00
Add IDF global variables that need to be removed when the active esp - idf environment is deactivated .
2021-11-08 11:55:03 -05:00
"""
2022-08-09 12:13:29 -04:00
if ' PATH ' in new_idf_vars :
new_idf_vars [ ' PATH ' ] = new_idf_vars [ ' PATH ' ] . split ( ' : ' ) [ : - 1 ] # PATH is stored as list of sub-paths without '$PATH'
2022-03-19 16:37:16 -04:00
2022-08-09 12:13:29 -04:00
new_idf_vars [ ' PATH ' ] = new_idf_vars . get ( ' PATH ' , [ ] )
2022-03-19 16:37:16 -04:00
args_add_paths_extras = vars ( args ) . get ( ' add_paths_extras ' ) # remove mypy error with args
2022-08-09 12:13:29 -04:00
new_idf_vars [ ' PATH ' ] = new_idf_vars [ ' PATH ' ] + args_add_paths_extras . split ( ' : ' ) if args_add_paths_extras else new_idf_vars [ ' PATH ' ]
2022-03-19 16:37:16 -04:00
2022-08-09 12:13:29 -04:00
env_state_obj = ENVState . get_env_state ( )
2022-03-19 16:37:16 -04:00
2022-08-09 12:13:29 -04:00
if env_state_obj . idf_variables :
exported_idf_vars = env_state_obj . idf_variables
new_idf_vars [ ' PATH ' ] = list ( set ( new_idf_vars [ ' PATH ' ] + exported_idf_vars . get ( ' PATH ' , [ ] ) ) ) # remove duplicates
env_state_obj . idf_variables = dict ( exported_idf_vars , * * new_idf_vars ) # merge two dicts
2022-03-19 16:37:16 -04:00
else :
2022-08-09 12:13:29 -04:00
env_state_obj . idf_variables = new_idf_vars
deactivate_file_path = env_state_obj . save ( )
2021-11-08 11:55:03 -05:00
2022-08-09 12:13:29 -04:00
return deactivate_file_path
2021-11-08 11:55:03 -05:00
2022-08-09 12:13:29 -04:00
def deactivate_statement ( args ) : # type: (list[str]) -> None
2021-11-08 11:55:03 -05:00
"""
2022-08-09 12:13:29 -04:00
Deactivate statement is sequence of commands , that remove IDF global variables from enviroment ,
2021-11-08 11:55:03 -05:00
so the environment gets to the state it was before calling export . { sh / fish } script .
"""
2022-08-09 12:13:29 -04:00
env_state_obj = ENVState . get_env_state ( )
if not env_state_obj . idf_variables :
warn ( ' No IDF variables to remove from environment found. Deactivation of previous esp-idf version was not successful. ' )
2021-11-08 11:55:03 -05:00
return
2022-08-09 12:13:29 -04:00
unset_vars = env_state_obj . idf_variables
2021-11-08 11:55:03 -05:00
env_path = os . getenv ( ' PATH ' ) # type: Optional[str]
if env_path :
2022-08-09 12:13:29 -04:00
cleared_env_path = ' : ' . join ( [ k for k in env_path . split ( ' : ' ) if k not in unset_vars [ ' PATH ' ] ] )
2021-11-08 11:55:03 -05:00
2022-08-09 12:13:29 -04:00
unset_list = [ k for k in unset_vars . keys ( ) if k != ' PATH ' ]
2021-11-08 11:55:03 -05:00
unset_format , sep = get_unset_format_and_separator ( args )
unset_statement = sep . join ( [ unset_format . format ( k ) for k in unset_list ] )
export_format , sep = get_export_format_and_separator ( args )
export_statement = export_format . format ( ' PATH ' , cleared_env_path )
deactivate_statement_str = sep . join ( [ unset_statement , export_statement ] )
print ( deactivate_statement_str )
2022-08-09 12:13:29 -04:00
# After deactivation clear old variables
env_state_obj . idf_variables . clear ( )
env_state_obj . save ( )
2021-11-08 11:55:03 -05:00
return
def get_export_format_and_separator ( args ) : # type: (list[str]) -> Tuple[str, str]
return { EXPORT_SHELL : ( ' export {} = " {} " ' , ' ; ' ) , EXPORT_KEY_VALUE : ( ' {} = {} ' , ' \n ' ) } [ args . format ] # type: ignore
def get_unset_format_and_separator ( args ) : # type: (list[str]) -> Tuple[str, str]
return { EXPORT_SHELL : ( ' unset {} ' , ' ; ' ) , EXPORT_KEY_VALUE : ( ' {} ' , ' \n ' ) } [ args . format ] # type: ignore
2022-03-19 16:37:16 -04:00
def different_idf_detected ( ) - > bool :
# If IDF global variable found, test if belong to different ESP-IDF version
if ' IDF_TOOLS_EXPORT_CMD ' in os . environ :
if global_idf_path != os . path . dirname ( os . environ [ ' IDF_TOOLS_EXPORT_CMD ' ] ) :
return True
2021-11-08 11:55:03 -05:00
# No previous ESP-IDF export detected, nothing to be unset
2022-03-19 16:37:16 -04:00
if all ( s not in os . environ for s in [ ' IDF_PYTHON_ENV_PATH ' , ' OPENOCD_SCRIPTS ' , ' ESP_IDF_VERSION ' ] ) :
2021-11-08 11:55:03 -05:00
return False
# User is exporting the same version as is in env
if os . getenv ( ' ESP_IDF_VERSION ' ) == get_idf_version ( ) :
return False
# Different version detected
return True
# Function returns unique id of running ESP-IDF combining current idfpath with version.
# The id is unique with same version & different path or same path & different version.
def active_repo_id ( ) - > str :
if global_idf_path is None :
return ' UNKNOWN_PATH ' + ' -v ' + get_idf_version ( )
return global_idf_path + ' -v ' + get_idf_version ( )
2023-02-05 10:29:03 -05:00
def list_default ( args ) : # type: ignore
2019-04-17 03:30:30 -04:00
tools_info = load_tools_info ( )
for name , tool in tools_info . items ( ) :
if tool . get_install_type ( ) == IDFTool . INSTALL_NEVER :
continue
optional_str = ' (optional) ' if tool . get_install_type ( ) == IDFTool . INSTALL_ON_REQUEST else ' '
info ( ' * {} : {} {} ' . format ( name , tool . description , optional_str ) )
tool . find_installed_versions ( )
versions_for_platform = { k : v for k , v in tool . versions . items ( ) if v . compatible_with_platform ( ) }
if not versions_for_platform :
info ( ' (no versions compatible with platform {} ) ' . format ( PYTHON_PLATFORM ) )
continue
2021-03-04 03:53:47 -05:00
versions_sorted = sorted ( versions_for_platform . keys ( ) , key = tool . versions . get , reverse = True ) # type: ignore
2019-04-17 03:30:30 -04:00
for version in versions_sorted :
version_obj = tool . versions [ version ]
info ( ' - {} ( {} {} ) ' . format ( version , version_obj . status ,
' , installed ' if version in tool . versions_installed else ' ' ) )
2023-02-05 10:29:03 -05:00
def list_outdated ( args ) : # type: ignore
tools_info = load_tools_info ( )
for name , tool in tools_info . items ( ) :
if tool . get_install_type ( ) == IDFTool . INSTALL_NEVER :
continue
versions_for_platform = { k : v for k , v in tool . versions . items ( ) if v . compatible_with_platform ( ) }
if not versions_for_platform :
continue
version_installed = tool . latest_installed_version ( )
if not version_installed :
continue
version_available = sorted ( versions_for_platform . keys ( ) , key = tool . versions . get , reverse = True ) [ 0 ]
if version_installed < version_available :
info ( f ' { name } : version { version_installed } is outdated by { version_available } ' )
def action_list ( args ) : # type: ignore
if args . outdated :
list_outdated ( args )
else :
list_default ( args )
2021-03-04 03:53:47 -05:00
def action_check ( args ) : # type: ignore
2019-04-17 03:30:30 -04:00
tools_info = load_tools_info ( )
2022-03-19 16:37:16 -04:00
tools_info = filter_tools_info ( IDFEnv . get_idf_env ( ) , tools_info )
2019-04-17 03:30:30 -04:00
not_found_list = [ ]
info ( ' Checking for installed tools... ' )
for name , tool in tools_info . items ( ) :
if tool . get_install_type ( ) == IDFTool . INSTALL_NEVER :
continue
tool_found_somewhere = False
info ( ' Checking tool %s ' % name )
tool . find_installed_versions ( )
if tool . version_in_path :
info ( ' version found in PATH: %s ' % tool . version_in_path )
tool_found_somewhere = True
else :
info ( ' no version found in PATH ' )
for version in tool . versions_installed :
info ( ' version installed in tools directory: %s ' % version )
tool_found_somewhere = True
if not tool_found_somewhere and tool . get_install_type ( ) == IDFTool . INSTALL_ALWAYS :
not_found_list . append ( name )
if not_found_list :
fatal ( ' The following required tools were not found: ' + ' ' . join ( not_found_list ) )
raise SystemExit ( 1 )
2021-03-04 03:53:47 -05:00
def action_export ( args ) : # type: ignore
2022-08-09 12:13:29 -04:00
if args . deactivate and different_idf_detected ( ) :
deactivate_statement ( args )
2021-11-08 11:55:03 -05:00
return
2019-04-17 03:30:30 -04:00
tools_info = load_tools_info ( )
2022-08-09 12:13:29 -04:00
tools_info = filter_tools_info ( IDFEnv . get_idf_env ( ) , tools_info )
2019-04-17 03:30:30 -04:00
all_tools_found = True
export_vars = { }
paths_to_export = [ ]
2022-08-15 10:47:24 -04:00
self_restart_cmd = f ' { sys . executable } { __file__ } { ( " --tools-json " + args . tools_json ) if args . tools_json else " " } '
self_restart_cmd = to_shell_specific_paths ( [ self_restart_cmd ] ) [ 0 ]
prefer_system_hint = ' ' if IDF_TOOLS_EXPORT_CMD else f ' To use it, run \' { self_restart_cmd } export --prefer-system \' '
install_cmd = to_shell_specific_paths ( [ IDF_TOOLS_INSTALL_CMD ] ) [ 0 ] if IDF_TOOLS_INSTALL_CMD else self_restart_cmd + ' install '
2019-04-17 03:30:30 -04:00
for name , tool in tools_info . items ( ) :
if tool . get_install_type ( ) == IDFTool . INSTALL_NEVER :
continue
tool . find_installed_versions ( )
2022-08-15 10:47:24 -04:00
version_to_use = tool . get_preferred_installed_version ( )
if not tool . is_executable and version_to_use :
tool_export_vars = tool . get_export_vars ( version_to_use )
export_vars = { * * export_vars , * * tool_export_vars }
continue
2019-04-17 03:30:30 -04:00
if tool . version_in_path :
if tool . version_in_path not in tool . versions :
# unsupported version
2021-03-04 03:53:47 -05:00
if args . prefer_system : # type: ignore
2019-04-17 03:30:30 -04:00
warn ( ' using an unsupported version of tool {} found in PATH: {} ' . format (
tool . name , tool . version_in_path ) )
continue
else :
# unsupported version in path
pass
else :
# supported/deprecated version in PATH, use it
version_obj = tool . versions [ tool . version_in_path ]
if version_obj . status == IDFToolVersion . STATUS_SUPPORTED :
info ( ' Using a supported version of tool {} found in PATH: {} . ' . format ( name , tool . version_in_path ) ,
f = sys . stderr )
info ( ' However the recommended version is {} . ' . format ( tool . get_recommended_version ( ) ) ,
f = sys . stderr )
elif version_obj . status == IDFToolVersion . STATUS_DEPRECATED :
warn ( ' using a deprecated version of tool {} found in PATH: {} ' . format ( name , tool . version_in_path ) )
continue
if not tool . versions_installed :
if tool . get_install_type ( ) == IDFTool . INSTALL_ALWAYS :
all_tools_found = False
fatal ( ' tool {} has no installed versions. Please run \' {} \' to install it. ' . format (
tool . name , install_cmd ) )
if tool . version_in_path and tool . version_in_path not in tool . versions :
info ( ' An unsupported version of tool {} was found in PATH: {} . ' . format ( name , tool . version_in_path ) +
prefer_system_hint , f = sys . stderr )
continue
else :
# tool is optional, and does not have versions installed
# use whatever is available in PATH
continue
if tool . version_in_path and tool . version_in_path not in tool . versions :
info ( ' Not using an unsupported version of tool {} found in PATH: {} . ' . format (
tool . name , tool . version_in_path ) + prefer_system_hint , f = sys . stderr )
export_paths = tool . get_export_paths ( version_to_use )
if export_paths :
paths_to_export + = export_paths
tool_export_vars = tool . get_export_vars ( version_to_use )
2022-08-15 10:47:24 -04:00
export_vars = { * * export_vars , * * tool_export_vars }
2019-04-17 03:30:30 -04:00
current_path = os . getenv ( ' PATH ' )
2021-09-16 10:48:03 -04:00
idf_python_env_path , idf_python_export_path , virtualenv_python , _ = get_python_env_path ( )
2019-04-17 03:30:30 -04:00
if os . path . exists ( virtualenv_python ) :
idf_python_env_path = to_shell_specific_paths ( [ idf_python_env_path ] ) [ 0 ]
if os . getenv ( ' IDF_PYTHON_ENV_PATH ' ) != idf_python_env_path :
export_vars [ ' IDF_PYTHON_ENV_PATH ' ] = to_shell_specific_paths ( [ idf_python_env_path ] ) [ 0 ]
if idf_python_export_path not in current_path :
paths_to_export . append ( idf_python_export_path )
2022-03-19 16:37:16 -04:00
idf_version = get_idf_version ( )
if os . getenv ( ' ESP_IDF_VERSION ' ) != idf_version :
export_vars [ ' ESP_IDF_VERSION ' ] = idf_version
2021-11-08 11:55:03 -05:00
2019-04-17 03:30:30 -04:00
idf_tools_dir = os . path . join ( global_idf_path , ' tools ' )
idf_tools_dir = to_shell_specific_paths ( [ idf_tools_dir ] ) [ 0 ]
if idf_tools_dir not in current_path :
paths_to_export . append ( idf_tools_dir )
2021-11-09 07:10:56 -05:00
if sys . platform == ' win32 ' :
2019-04-17 03:30:30 -04:00
old_path = ' % PATH % '
path_sep = ' ; '
else :
old_path = ' $PATH '
path_sep = ' : '
2021-11-08 11:55:03 -05:00
export_format , export_sep = get_export_format_and_separator ( args )
2019-04-17 03:30:30 -04:00
if paths_to_export :
export_vars [ ' PATH ' ] = path_sep . join ( to_shell_specific_paths ( paths_to_export ) + [ old_path ] )
2022-08-09 12:13:29 -04:00
if export_vars :
# if not copy of export_vars is given to function, it brekas the formatting string for 'export_statements'
deactivate_file_path = add_variables_to_deactivate_file ( args , export_vars . copy ( ) )
export_vars [ ENVState . env_key ] = deactivate_file_path
export_statements = export_sep . join ( [ export_format . format ( k , v ) for k , v in export_vars . items ( ) ] )
2019-04-17 03:30:30 -04:00
print ( export_statements )
if not all_tools_found :
raise SystemExit ( 1 )
2022-11-09 11:51:07 -05:00
def get_idf_download_url_apply_mirrors ( args = None , download_url = IDF_DL_URL ) : # type: (Any, str) -> str
url = apply_mirror_prefix_map ( args , download_url )
url = apply_github_assets_option ( url )
return url
2020-09-24 19:58:49 -04:00
2022-11-09 11:51:07 -05:00
def apply_mirror_prefix_map ( args , idf_download_url ) : # type: (Any, str) -> str
""" Rewrite URL for given idf_download_url.
2019-04-17 03:30:30 -04:00
if - - mirror - prefix - map flag or IDF_MIRROR_PREFIX_MAP environment variable is given .
"""
2022-11-09 11:51:07 -05:00
new_url = idf_download_url
2019-04-17 03:30:30 -04:00
mirror_prefix_map = None
mirror_prefix_map_env = os . getenv ( ' IDF_MIRROR_PREFIX_MAP ' )
if mirror_prefix_map_env :
mirror_prefix_map = mirror_prefix_map_env . split ( ' ; ' )
2022-11-09 11:51:07 -05:00
if IDF_MAINTAINER and args and args . mirror_prefix_map :
2019-04-17 03:30:30 -04:00
if mirror_prefix_map :
warn ( ' Both IDF_MIRROR_PREFIX_MAP environment variable and --mirror-prefix-map flag are specified, ' +
' will use the value from the command line. ' )
mirror_prefix_map = args . mirror_prefix_map
2022-11-09 11:51:07 -05:00
if mirror_prefix_map :
2019-04-17 03:30:30 -04:00
for item in mirror_prefix_map :
if URL_PREFIX_MAP_SEPARATOR not in item :
warn ( ' invalid mirror-prefix-map item (missing \' {} \' ) {} ' . format ( URL_PREFIX_MAP_SEPARATOR , item ) )
continue
search , replace = item . split ( URL_PREFIX_MAP_SEPARATOR , 1 )
2022-11-09 11:51:07 -05:00
new_url = re . sub ( search , replace , idf_download_url )
if new_url != idf_download_url :
info ( ' Changed download URL: {} => {} ' . format ( idf_download_url , new_url ) )
2019-04-17 03:30:30 -04:00
break
2022-11-09 11:51:07 -05:00
return new_url
2019-04-17 03:30:30 -04:00
2022-11-09 11:51:07 -05:00
def apply_github_assets_option ( idf_download_url ) : # type: (str) -> str
""" Rewrite URL for given idf_download_url if the download URL is an https://github.com/ URL and the variable
2020-09-24 19:58:49 -04:00
IDF_GITHUB_ASSETS is set . The github . com part of the URL will be replaced .
"""
2022-11-09 11:51:07 -05:00
new_url = idf_download_url
github_assets = os . environ . get ( ' IDF_GITHUB_ASSETS ' , ' ' ) . strip ( )
if not github_assets :
# no IDF_GITHUB_ASSETS or variable exists but is empty
return new_url
2020-09-24 19:58:49 -04:00
# check no URL qualifier in the mirror URL
if ' :// ' in github_assets :
fatal ( " IDF_GITHUB_ASSETS shouldn ' t include any URL qualifier, https:// is assumed " )
raise SystemExit ( 1 )
# Strip any trailing / from the mirror URL
github_assets = github_assets . rstrip ( ' / ' )
2022-11-09 11:51:07 -05:00
new_url = re . sub ( r ' ^https://github.com/ ' , ' https:// {} / ' . format ( github_assets ) , idf_download_url )
if new_url != idf_download_url :
info ( ' Using GitHub assets mirror for URL: {} => {} ' . format ( idf_download_url , new_url ) )
return new_url
2020-09-24 19:58:49 -04:00
2021-12-13 10:45:11 -05:00
def get_tools_spec_and_platform_info ( selected_platform , targets , tools_spec ,
2023-09-14 01:55:24 -04:00
quiet = False ) : # type: (str, list[str], list[str], bool) -> Tuple[list[str], Dict[str, IDFTool]]
2021-12-13 10:45:11 -05:00
# If this function is not called from action_download, but is used just for detecting active tools, info about downloading is unwanted.
global global_quiet
try :
old_global_quiet = global_quiet
global_quiet = quiet
tools_info = load_tools_info ( )
tools_info_for_platform = OrderedDict ( )
for name , tool_obj in tools_info . items ( ) :
tool_for_platform = tool_obj . copy_for_platform ( selected_platform )
tools_info_for_platform [ name ] = tool_for_platform
2023-09-14 01:55:24 -04:00
tools_spec = expand_tools_arg ( tools_spec , tools_info_for_platform , targets )
info ( ' Downloading tools for {} : {} ' . format ( selected_platform , ' , ' . join ( tools_spec ) ) )
2021-12-13 10:45:11 -05:00
finally :
global_quiet = old_global_quiet
return tools_spec , tools_info_for_platform
2021-03-04 03:53:47 -05:00
def action_download ( args ) : # type: ignore
2023-09-14 01:55:24 -04:00
tools_spec = parse_tools_arg ( args . tools )
2021-04-26 15:34:48 -04:00
targets = [ ] # type: list[str]
2023-09-14 01:55:24 -04:00
# Saving IDFEnv::targets for selected ESP_targets if all tools have been specified
if ' required ' in tools_spec or ' all ' in tools_spec :
2022-03-19 16:37:16 -04:00
idf_env_obj = IDFEnv . get_idf_env ( )
targets = add_and_check_targets ( idf_env_obj , args . targets )
2022-08-09 12:13:29 -04:00
try :
idf_env_obj . save ( )
except OSError as err :
if args . targets in targets :
targets . remove ( args . targets )
warn ( ' Downloading tools for targets was not successful with error: {} ' . format ( err ) )
2023-09-14 01:55:24 -04:00
# Taking into account ESP_targets but not saving them for individual tools (specified list of tools)
else :
targets = parse_targets_arg ( args . targets )
2019-07-18 11:20:40 -04:00
2023-09-14 01:55:24 -04:00
platform = parse_platform_arg ( args . platform )
2019-07-18 11:20:40 -04:00
2023-09-14 01:55:24 -04:00
tools_spec , tools_info_for_platform = get_tools_spec_and_platform_info ( platform , targets , tools_spec )
2019-07-18 11:20:40 -04:00
for tool_spec in tools_spec :
if ' @ ' not in tool_spec :
tool_name = tool_spec
tool_version = None
else :
tool_name , tool_version = tool_spec . split ( ' @ ' , 1 )
if tool_name not in tools_info_for_platform :
fatal ( ' unknown tool name: {} ' . format ( tool_name ) )
raise SystemExit ( 1 )
tool_obj = tools_info_for_platform [ tool_name ]
if tool_version is not None and tool_version not in tool_obj . versions :
fatal ( ' unknown version for tool {} : {} ' . format ( tool_name , tool_version ) )
raise SystemExit ( 1 )
if tool_version is None :
tool_version = tool_obj . get_recommended_version ( )
2019-11-24 06:11:33 -05:00
if tool_version is None :
fatal ( ' tool {} not found for {} platform ' . format ( tool_name , platform ) )
raise SystemExit ( 1 )
2019-07-18 11:20:40 -04:00
tool_spec = ' {} @ {} ' . format ( tool_name , tool_version )
info ( ' Downloading {} ' . format ( tool_spec ) )
2023-09-14 01:55:24 -04:00
_idf_tool_obj = tool_obj . versions [ tool_version ] . get_download_for_platform ( platform )
2022-11-09 11:51:07 -05:00
_idf_tool_obj . url = get_idf_download_url_apply_mirrors ( args , _idf_tool_obj . url )
2019-07-18 11:20:40 -04:00
tool_obj . download ( tool_version )
2021-03-04 03:53:47 -05:00
def action_install ( args ) : # type: ignore
2023-09-14 01:55:24 -04:00
tools_spec = parse_tools_arg ( args . tools )
2021-04-26 15:34:48 -04:00
targets = [ ] # type: list[str]
2023-09-14 01:55:24 -04:00
# Saving IDFEnv::targets for selected ESP_targets if all tools have been specified
2022-03-19 16:37:16 -04:00
if ' required ' in tools_spec or ' all ' in tools_spec :
idf_env_obj = IDFEnv . get_idf_env ( )
targets = add_and_check_targets ( idf_env_obj , args . targets )
2022-08-09 12:13:29 -04:00
try :
idf_env_obj . save ( )
except OSError as err :
if args . targets in targets :
targets . remove ( args . targets )
warn ( ' Installing targets was not successful with error: {} ' . format ( err ) )
2022-04-14 02:39:16 -04:00
info ( ' Selected targets are: {} ' . format ( ' , ' . join ( targets ) ) )
2023-09-14 01:55:24 -04:00
# Taking into account ESP_targets but not saving them for individual tools (specified list of tools)
else :
targets = parse_targets_arg ( args . targets )
2021-04-26 15:34:48 -04:00
2023-09-14 01:55:24 -04:00
info ( ' Current system platform: {} ' . format ( CURRENT_PLATFORM ) )
tools_info = load_tools_info ( )
tools_spec = expand_tools_arg ( tools_spec , tools_info , targets )
info ( ' Installing tools: {} ' . format ( ' , ' . join ( tools_spec ) ) )
2019-04-17 03:30:30 -04:00
for tool_spec in tools_spec :
if ' @ ' not in tool_spec :
tool_name = tool_spec
tool_version = None
else :
tool_name , tool_version = tool_spec . split ( ' @ ' , 1 )
if tool_name not in tools_info :
fatal ( ' unknown tool name: {} ' . format ( tool_name ) )
raise SystemExit ( 1 )
tool_obj = tools_info [ tool_name ]
2019-07-23 00:55:30 -04:00
if not tool_obj . compatible_with_platform ( ) :
fatal ( ' tool {} does not have versions compatible with platform {} ' . format ( tool_name , CURRENT_PLATFORM ) )
raise SystemExit ( 1 )
2019-04-17 03:30:30 -04:00
if tool_version is not None and tool_version not in tool_obj . versions :
fatal ( ' unknown version for tool {} : {} ' . format ( tool_name , tool_version ) )
raise SystemExit ( 1 )
if tool_version is None :
tool_version = tool_obj . get_recommended_version ( )
assert tool_version is not None
tool_obj . find_installed_versions ( )
tool_spec = ' {} @ {} ' . format ( tool_name , tool_version )
if tool_version in tool_obj . versions_installed :
info ( ' Skipping {} (already installed) ' . format ( tool_spec ) )
continue
info ( ' Installing {} ' . format ( tool_spec ) )
2022-11-09 11:51:07 -05:00
_idf_tool_obj = tool_obj . versions [ tool_version ] . get_download_for_platform ( PYTHON_PLATFORM )
_idf_tool_obj . url = get_idf_download_url_apply_mirrors ( args , _idf_tool_obj . url )
2019-04-17 03:30:30 -04:00
tool_obj . download ( tool_version )
tool_obj . install ( tool_version )
2021-03-04 03:53:47 -05:00
def get_wheels_dir ( ) : # type: () -> Optional[str]
2020-12-16 11:40:13 -05:00
tools_info = load_tools_info ( )
wheels_package_name = ' idf-python-wheels '
if wheels_package_name not in tools_info :
return None
wheels_package = tools_info [ wheels_package_name ]
recommended_version = wheels_package . get_recommended_version ( )
2021-01-18 11:26:20 -05:00
if recommended_version is None :
return None
2020-12-16 11:40:13 -05:00
wheels_dir = wheels_package . get_path_for_version ( recommended_version )
if not os . path . exists ( wheels_dir ) :
return None
return wheels_dir
2021-09-16 10:48:03 -04:00
def get_requirements ( new_features ) : # type: (str) -> list[str]
2022-03-19 16:37:16 -04:00
idf_env_obj = IDFEnv . get_idf_env ( )
2022-05-20 09:15:52 -04:00
features = process_and_check_features ( idf_env_obj , new_features )
2022-08-09 12:13:29 -04:00
try :
idf_env_obj . save ( )
except OSError as err :
if new_features in features :
features . remove ( new_features )
warn ( ' Updating features was not successful with error: {} ' . format ( err ) )
2021-09-16 10:48:03 -04:00
return [ feature_to_requirements_path ( feature ) for feature in features ]
2022-10-03 04:59:31 -04:00
def get_constraints ( idf_version , online = True ) : # type: (str, bool) -> str
2022-11-09 11:51:07 -05:00
idf_download_url = get_idf_download_url_apply_mirrors ( )
2021-09-16 10:48:03 -04:00
constraint_file = ' espidf.constraints.v {} .txt ' . format ( idf_version )
2022-04-14 05:08:20 -04:00
constraint_path = os . path . join ( global_idf_tools_path or ' ' , constraint_file )
2022-11-09 11:51:07 -05:00
constraint_url = ' / ' . join ( [ idf_download_url , constraint_file ] )
2021-09-16 10:48:03 -04:00
temp_path = constraint_path + ' .tmp '
2022-10-03 04:59:31 -04:00
if not online :
if os . path . isfile ( constraint_path ) :
return constraint_path
else :
fatal ( f ' { constraint_path } doesn \' t exist. Perhaps you \' ve forgotten to run the install scripts. '
f ' Please check the installation guide for more information. ' )
raise SystemExit ( 1 )
2021-09-16 10:48:03 -04:00
mkdir_p ( os . path . dirname ( temp_path ) )
2022-01-18 08:49:53 -05:00
try :
age = datetime . date . today ( ) - datetime . date . fromtimestamp ( os . path . getmtime ( constraint_path ) )
if age < datetime . timedelta ( days = 1 ) :
2022-07-01 08:44:12 -04:00
info ( f ' Skipping the download of { constraint_path } because it was downloaded recently. ' )
2022-01-18 08:49:53 -05:00
return constraint_path
except OSError :
# doesn't exist or inaccessible
pass
2021-09-16 10:48:03 -04:00
for _ in range ( DOWNLOAD_RETRY_COUNT ) :
2022-09-22 10:14:17 -04:00
err = download ( constraint_url , temp_path )
2021-09-16 10:48:03 -04:00
if not os . path . isfile ( temp_path ) :
2022-09-22 10:14:17 -04:00
warn ( ' Download failure: {} ' . format ( err ) )
2021-09-16 10:48:03 -04:00
warn ( ' Failed to download {} to {} ' . format ( constraint_url , temp_path ) )
continue
if os . path . isfile ( constraint_path ) :
# Windows cannot rename to existing file. It needs to be deleted.
os . remove ( constraint_path )
rename_with_retry ( temp_path , constraint_path )
return constraint_path
if os . path . isfile ( constraint_path ) :
warn ( ' Failed to download, retry count has expired, using a previously downloaded version ' )
return constraint_path
else :
fatal ( ' Failed to download, and retry count has expired ' )
2022-09-22 10:14:17 -04:00
print_hints_on_download_error ( str ( err ) )
2022-07-01 08:44:12 -04:00
info ( ' See the help on how to disable constraints in order to work around this issue. ' )
2022-09-22 10:14:17 -04:00
raise SystemExit ( 1 )
2021-09-16 10:48:03 -04:00
2022-05-25 09:06:13 -04:00
def install_legacy_python_virtualenv ( path ) : # type: (str) -> None
# Before creating the virtual environment, check if pip is installed.
try :
subprocess . check_call ( [ sys . executable , ' -m ' , ' pip ' , ' --version ' ] )
except subprocess . CalledProcessError :
fatal ( ' Python interpreter at {} doesn \' t have pip installed. '
' Please check the Getting Started Guides for the steps to install prerequisites for your OS. ' . format ( sys . executable ) )
raise SystemExit ( 1 )
virtualenv_installed_via_pip = False
try :
import virtualenv # noqa: F401
except ImportError :
info ( ' Installing virtualenv ' )
subprocess . check_call ( [ sys . executable , ' -m ' , ' pip ' , ' install ' , ' --user ' , ' virtualenv ' ] ,
stdout = sys . stdout , stderr = sys . stderr )
virtualenv_installed_via_pip = True
# since we just installed virtualenv via pip, we know that version is recent enough
# so the version check below is not necessary.
with_seeder_option = True
if not virtualenv_installed_via_pip :
# virtualenv is already present in the system and may have been installed via OS package manager
# check the version to determine if we should add --seeder option
try :
major_ver = int ( virtualenv . __version__ . split ( ' . ' ) [ 0 ] )
if major_ver < 20 :
warn ( ' Virtualenv version {} is old, please consider upgrading it ' . format ( virtualenv . __version__ ) )
with_seeder_option = False
except ( ValueError , NameError , AttributeError , IndexError ) :
pass
info ( f ' Creating a new Python environment using virtualenv in { path } ' )
virtualenv_options = [ ' --python ' , sys . executable ]
if with_seeder_option :
virtualenv_options + = [ ' --seeder ' , ' pip ' ]
subprocess . check_call ( [ sys . executable , ' -m ' , ' virtualenv ' ,
* virtualenv_options ,
path ] ,
stdout = sys . stdout , stderr = sys . stderr )
2021-03-04 03:53:47 -05:00
def action_install_python_env ( args ) : # type: ignore
2021-09-16 10:48:03 -04:00
use_constraints = not args . no_constraints
2021-03-04 03:52:34 -05:00
reinstall = args . reinstall
2021-09-16 10:48:03 -04:00
idf_python_env_path , _ , virtualenv_python , idf_version = get_python_env_path ( )
2019-04-17 03:30:30 -04:00
2020-02-26 08:02:12 -05:00
is_virtualenv = hasattr ( sys , ' real_prefix ' ) or ( hasattr ( sys , ' base_prefix ' ) and sys . base_prefix != sys . prefix )
2021-03-04 03:52:34 -05:00
if is_virtualenv and ( not os . path . exists ( idf_python_env_path ) or reinstall ) :
2020-02-26 08:02:12 -05:00
fatal ( ' This script was called from a virtual environment, can not create a virtual environment again ' )
raise SystemExit ( 1 )
2021-03-04 03:52:34 -05:00
if os . path . exists ( virtualenv_python ) :
try :
subprocess . check_call ( [ virtualenv_python , ' --version ' ] , stdout = sys . stdout , stderr = sys . stderr )
2021-06-23 05:46:00 -04:00
except ( OSError , subprocess . CalledProcessError ) :
2021-03-04 03:52:34 -05:00
# At this point we can reinstall the virtual environment if it is non-functional. This can happen at least
# when the Python interpreter was removed which was used to create the virtual environment.
reinstall = True
try :
subprocess . check_call ( [ virtualenv_python , ' -m ' , ' pip ' , ' --version ' ] , stdout = sys . stdout , stderr = sys . stderr )
except subprocess . CalledProcessError :
2022-01-04 14:01:35 -05:00
warn ( ' pip is not available in the existing virtual environment, new virtual environment will be created. ' )
# Reinstallation of the virtual environment could help if pip was installed for the main Python
2021-03-04 03:52:34 -05:00
reinstall = True
2023-06-28 04:44:27 -04:00
if sys . platform != ' win32 ' :
try :
subprocess . check_call ( [ virtualenv_python , ' -c ' , ' import curses ' ] , stdout = sys . stdout , stderr = sys . stderr )
except subprocess . CalledProcessError :
warn ( ' curses can not be imported, new virtual environment will be created. ' )
reinstall = True
2021-03-04 03:52:34 -05:00
if reinstall and os . path . exists ( idf_python_env_path ) :
2019-04-17 03:30:30 -04:00
warn ( ' Removing the existing Python environment in {} ' . format ( idf_python_env_path ) )
shutil . rmtree ( idf_python_env_path )
2022-08-10 07:45:32 -04:00
venv_can_upgrade = False
2019-04-17 03:30:30 -04:00
if not os . path . exists ( virtualenv_python ) :
2022-10-10 06:20:46 -04:00
if subprocess . run ( [ sys . executable , ' -m ' , ' venv ' , ' -h ' ] , check = False , stdout = subprocess . DEVNULL , stderr = subprocess . DEVNULL ) . returncode == 0 :
2022-05-25 09:06:13 -04:00
# venv available
virtualenv_options = [ ' --clear ' ] # delete environment if already exists
if sys . version_info [ : 2 ] > = ( 3 , 9 ) :
# upgrade pip & setuptools
virtualenv_options + = [ ' --upgrade-deps ' ]
2022-08-10 07:45:32 -04:00
venv_can_upgrade = True
2022-05-25 09:06:13 -04:00
info ( ' Creating a new Python environment in {} ' . format ( idf_python_env_path ) )
subprocess . check_call ( [ sys . executable , ' -m ' , ' venv ' ,
* virtualenv_options ,
idf_python_env_path ] ,
2019-04-17 03:30:30 -04:00
stdout = sys . stdout , stderr = sys . stderr )
2022-10-10 06:20:46 -04:00
else :
2022-05-25 09:06:13 -04:00
# The embeddable Python for Windows doesn't have the built-in venv module
install_legacy_python_virtualenv ( idf_python_env_path )
2019-04-17 03:30:30 -04:00
2021-11-29 08:03:21 -05:00
env_copy = os . environ . copy ( )
if env_copy . get ( ' PIP_USER ' ) == ' yes ' :
warn ( ' Found PIP_USER= " yes " in the environment. Disabling PIP_USER in this shell to install packages into a virtual environment. ' )
env_copy [ ' PIP_USER ' ] = ' no '
2022-08-10 07:45:32 -04:00
if not venv_can_upgrade :
info ( ' Upgrading pip and setuptools... ' )
subprocess . check_call ( [ virtualenv_python , ' -m ' , ' pip ' , ' install ' , ' --upgrade ' , ' pip ' , ' setuptools ' ] ,
stdout = sys . stdout , stderr = sys . stderr , env = env_copy )
2019-04-17 03:30:30 -04:00
run_args = [ virtualenv_python , ' -m ' , ' pip ' , ' install ' , ' --no-warn-script-location ' ]
2021-09-16 10:48:03 -04:00
requirements_file_list = get_requirements ( args . features )
for requirement_file in requirements_file_list :
run_args + = [ ' -r ' , requirement_file ]
if use_constraints :
constraint_file = get_constraints ( idf_version )
run_args + = [ ' --upgrade ' , ' --constraint ' , constraint_file ]
2019-04-17 03:30:30 -04:00
if args . extra_wheels_dir :
run_args + = [ ' --find-links ' , args . extra_wheels_dir ]
2020-12-16 11:40:13 -05:00
if args . no_index :
run_args + = [ ' --no-index ' ]
if args . extra_wheels_url :
run_args + = [ ' --extra-index-url ' , args . extra_wheels_url ]
wheels_dir = get_wheels_dir ( )
if wheels_dir is not None :
run_args + = [ ' --find-links ' , wheels_dir ]
2021-09-16 10:48:03 -04:00
info ( ' Installing Python packages ' )
if use_constraints :
info ( ' Constraint file: {} ' . format ( constraint_file ) )
info ( ' Requirement files: ' )
info ( os . linesep . join ( ' - {} ' . format ( path ) for path in requirements_file_list ) )
2021-11-29 08:03:21 -05:00
subprocess . check_call ( run_args , stdout = sys . stdout , stderr = sys . stderr , env = env_copy )
2019-04-17 03:30:30 -04:00
2021-09-16 10:48:03 -04:00
def action_check_python_dependencies ( args ) : # type: ignore
use_constraints = not args . no_constraints
req_paths = get_requirements ( ' ' ) # no new features -> just detect the existing ones
_ , _ , virtualenv_python , idf_version = get_python_env_path ( )
if not os . path . isfile ( virtualenv_python ) :
fatal ( ' {} doesn \' t exist! Please run the install script or " idf_tools.py install-python-env " in order to '
' create it ' . format ( virtualenv_python ) )
raise SystemExit ( 1 )
if use_constraints :
2022-10-03 04:59:31 -04:00
constr_path = get_constraints ( idf_version , online = False ) # keep offline for checking
2021-09-16 10:48:03 -04:00
info ( ' Constraint file: {} ' . format ( constr_path ) )
info ( ' Requirement files: ' )
info ( os . linesep . join ( ' - {} ' . format ( path ) for path in req_paths ) )
info ( ' Python being checked: {} ' . format ( virtualenv_python ) )
# The dependency checker will be invoked with virtualenv_python. idf_tools.py could have been invoked with a
# different one, therefore, importing is not a suitable option.
dep_check_cmd = [ virtualenv_python ,
os . path . join ( global_idf_path ,
' tools ' ,
' check_python_dependencies.py ' ) ]
if use_constraints :
dep_check_cmd + = [ ' -c ' , constr_path ]
for req_path in req_paths :
dep_check_cmd + = [ ' -r ' , req_path ]
try :
ret = subprocess . run ( dep_check_cmd )
if ret and ret . returncode :
# returncode is a negative number and system exit output is usually expected be positive.
raise SystemExit ( - ret . returncode )
except FileNotFoundError :
# Python environment not yet created
fatal ( ' Requirements are not satisfied! ' )
raise SystemExit ( 1 )
2022-03-15 15:02:37 -04:00
class ChecksumCalculator ( ) :
"""
A class used to get size / checksum / basename of local artifact files .
"""
def __init__ ( self , files ) : # type: (list[str]) -> None
self . files = files
def __iter__ ( self ) : # type: () -> Iterator[Tuple[int, str, str]]
for f in self . files :
yield ( * get_file_size_sha256 ( f ) , os . path . basename ( f ) )
class ChecksumParsingError ( RuntimeError ) :
pass
class ChecksumFileParser ( ) :
"""
A class used to get size / sha256 / filename of artifact using checksum - file with format :
# <artifact-filename>: <size> bytes
< sha256sum - string > * < artifact - filename >
. . . ( 2 lines for every artifact ) . . .
"""
def __init__ ( self , tool_name , url ) : # type: (str, str) -> None
self . tool_name = tool_name
sha256_file_tmp = os . path . join ( global_idf_tools_path or ' ' , ' tools ' , ' add-version.sha256.tmp ' )
sha256_file = os . path . abspath ( url )
# download sha256 file if URL presented
if urlparse ( url ) . scheme :
sha256_file = sha256_file_tmp
download ( url , sha256_file )
with open ( sha256_file , ' r ' ) as f :
self . checksum = f . read ( ) . splitlines ( )
# remove temp file
if os . path . isfile ( sha256_file_tmp ) :
os . remove ( sha256_file_tmp )
def parseLine ( self , regex , line ) : # type: (str, str) -> str
match = re . search ( regex , line )
if not match :
raise ChecksumParsingError ( f ' Can not parse line " { line } " with regex " { regex } " ' )
return match . group ( 1 )
# parse checksum file with formatting used by crosstool-ng, gdb, ... releases
# e.g. https://github.com/espressif/crosstool-NG/releases/download/esp-2021r2/crosstool-NG-esp-2021r2-checksum.sha256
def __iter__ ( self ) : # type: () -> Iterator[Tuple[int, str, str]]
try :
for bytes_str , hash_str in zip ( self . checksum [ 0 : : 2 ] , self . checksum [ 1 : : 2 ] ) :
bytes_filename = self . parseLine ( r ' ^# ( \ S*): ' , bytes_str )
hash_filename = self . parseLine ( r ' ^ \ S* \ *( \ S*) ' , hash_str )
if hash_filename != bytes_filename :
fatal ( ' filename in hash-line and in bytes-line are not the same ' )
raise SystemExit ( 1 )
# crosstool-ng checksum file contains info about few tools
# e.g.: "xtensa-esp32-elf", "xtensa-esp32s2-elf"
# filter records for file by tool_name to avoid mismatch
if not hash_filename . startswith ( self . tool_name ) :
continue
size = self . parseLine ( r ' ^# \ S*: ( \ d*) bytes ' , bytes_str )
sha256 = self . parseLine ( r ' ^( \ S*) ' , hash_str )
yield int ( size ) , sha256 , hash_filename
except ( TypeError , AttributeError ) as err :
fatal ( f ' Error while parsing, check checksum file ( { err } ) ' )
raise SystemExit ( 1 )
2021-03-04 03:53:47 -05:00
def action_add_version ( args ) : # type: ignore
2019-04-17 03:30:30 -04:00
tools_info = load_tools_info ( )
tool_name = args . tool
tool_obj = tools_info . get ( tool_name )
if not tool_obj :
info ( ' Creating new tool entry for {} ' . format ( tool_name ) )
tool_obj = IDFTool ( tool_name , TODO_MESSAGE , IDFTool . INSTALL_ALWAYS ,
TODO_MESSAGE , TODO_MESSAGE , [ TODO_MESSAGE ] , TODO_MESSAGE )
tools_info [ tool_name ] = tool_obj
version = args . version
2022-03-15 15:02:37 -04:00
version_status = IDFToolVersion . STATUS_SUPPORTED
if args . override and len ( tool_obj . versions ) :
tool_obj . drop_versions ( )
version_status = IDFToolVersion . STATUS_RECOMMENDED
2019-04-17 03:30:30 -04:00
version_obj = tool_obj . versions . get ( version )
2022-03-15 15:02:37 -04:00
if not version_obj :
2019-04-17 03:30:30 -04:00
info ( ' Creating new version {} ' . format ( version ) )
2022-03-15 15:02:37 -04:00
version_obj = IDFToolVersion ( version , version_status )
2019-04-17 03:30:30 -04:00
tool_obj . versions [ version ] = version_obj
url_prefix = args . url_prefix or ' https:// %s / ' % TODO_MESSAGE
2022-03-15 15:02:37 -04:00
checksum_info = ChecksumFileParser ( tool_name , args . checksum_file ) if args . checksum_file else ChecksumCalculator ( args . artifact_file )
for file_size , file_sha256 , file_name in checksum_info :
2019-04-17 03:30:30 -04:00
# Guess which platform this file is for
2022-05-16 16:11:01 -04:00
found_platform = Platforms . get_by_filename ( file_name )
2019-04-17 03:30:30 -04:00
if found_platform is None :
info ( ' Could not guess platform for file {} ' . format ( file_name ) )
found_platform = TODO_MESSAGE
2022-03-15 15:02:37 -04:00
url = urljoin ( url_prefix , file_name )
2019-04-17 03:30:30 -04:00
info ( ' Adding download for platform {} ' . format ( found_platform ) )
info ( ' size: {} ' . format ( file_size ) )
info ( ' SHA256: {} ' . format ( file_sha256 ) )
info ( ' URL: {} ' . format ( url ) )
version_obj . add_download ( found_platform , url , file_size , file_sha256 )
json_str = dump_tools_json ( tools_info )
if not args . output :
args . output = os . path . join ( global_idf_path , TOOLS_FILE_NEW )
with open ( args . output , ' w ' ) as f :
f . write ( json_str )
f . write ( ' \n ' )
info ( ' Wrote output to {} ' . format ( args . output ) )
2021-03-04 03:53:47 -05:00
def action_rewrite ( args ) : # type: ignore
2019-04-17 03:30:30 -04:00
tools_info = load_tools_info ( )
json_str = dump_tools_json ( tools_info )
if not args . output :
args . output = os . path . join ( global_idf_path , TOOLS_FILE_NEW )
with open ( args . output , ' w ' ) as f :
f . write ( json_str )
f . write ( ' \n ' )
info ( ' Wrote output to {} ' . format ( args . output ) )
2021-12-13 10:45:11 -05:00
def action_uninstall ( args ) : # type: (Any) -> None
2022-10-20 10:10:38 -04:00
""" Print or remove installed tools versions, that are not used by active ESP-IDF version anymore.
2021-12-13 10:45:11 -05:00
Additionally remove all older versions of previously downloaded archives .
"""
tools_info = load_tools_info ( )
tools_path = os . path . join ( global_idf_tools_path or ' ' , ' tools ' )
dist_path = os . path . join ( global_idf_tools_path or ' ' , ' dist ' )
installed_tools = os . listdir ( tools_path ) if os . path . isdir ( tools_path ) else [ ]
2022-10-20 10:10:38 -04:00
unused_tools_versions = { }
for tool in installed_tools :
tool_versions = os . listdir ( os . path . join ( tools_path , tool ) ) if os . path . isdir ( os . path . join ( tools_path , tool ) ) else [ ]
try :
unused_versions = ( [ x for x in tool_versions if x != tools_info [ tool ] . get_recommended_version ( ) ] )
except KeyError : # When tool that is not supported by tools_info (tools.json) anymore, remove the whole tool file
unused_versions = [ ' ' ]
if unused_versions :
unused_tools_versions [ tool ] = unused_versions
2021-12-13 10:45:11 -05:00
# Keeping tools added by windows installer
KEEP_WIN_TOOLS = [ ' idf-git ' , ' idf-python ' ]
for tool in KEEP_WIN_TOOLS :
2022-10-20 10:10:38 -04:00
if tool in unused_tools_versions :
unused_tools_versions . pop ( tool )
2021-12-13 10:45:11 -05:00
# Print unused tools.
if args . dry_run :
2022-10-20 10:10:38 -04:00
if unused_tools_versions :
print ( ' For removing old versions of {} use command \' {} {} {} \' ' . format ( ' , ' . join ( unused_tools_versions ) , get_python_exe_and_subdir ( ) [ 0 ] ,
2021-12-13 10:45:11 -05:00
os . path . join ( global_idf_path or ' ' , ' tools ' , ' idf_tools.py ' ) , ' uninstall ' ) )
return
# Remove installed tools that are not used by current ESP-IDF version.
2022-10-20 10:10:38 -04:00
for tool in unused_tools_versions :
for version in unused_tools_versions [ tool ] :
try :
if version :
path_to_remove = os . path . join ( tools_path , tool , version )
else :
path_to_remove = os . path . join ( tools_path , tool )
shutil . rmtree ( path_to_remove )
info ( path_to_remove + ' was removed. ' )
except OSError as error :
warn ( f ' { error . filename } can not be removed because { error . strerror } . ' )
2021-12-13 10:45:11 -05:00
# Remove old archives versions and archives that are not used by the current ESP-IDF version.
if args . remove_archives :
2022-10-20 10:10:38 -04:00
tools_spec , tools_info_for_platform = get_tools_spec_and_platform_info ( CURRENT_PLATFORM , [ ' all ' ] , [ ' all ' ] , quiet = True )
2021-12-13 10:45:11 -05:00
used_archives = [ ]
# Detect used active archives
for tool_spec in tools_spec :
if ' @ ' not in tool_spec :
tool_name = tool_spec
tool_version = None
else :
tool_name , tool_version = tool_spec . split ( ' @ ' , 1 )
tool_obj = tools_info_for_platform [ tool_name ]
if tool_version is None :
tool_version = tool_obj . get_recommended_version ( )
# mypy-checks
if tool_version is not None :
archive_version = tool_obj . versions [ tool_version ] . get_download_for_platform ( CURRENT_PLATFORM )
if archive_version is not None :
archive_version_url = archive_version . url
archive = os . path . basename ( archive_version_url )
used_archives . append ( archive )
downloaded_archives = os . listdir ( dist_path )
for archive in downloaded_archives :
if archive not in used_archives :
os . remove ( os . path . join ( dist_path , archive ) )
info ( os . path . join ( dist_path , archive ) + ' was removed. ' )
2021-03-04 03:53:47 -05:00
def action_validate ( args ) : # type: ignore
2019-04-17 03:30:30 -04:00
try :
import jsonschema
except ImportError :
fatal ( ' You need to install jsonschema package to use validate command ' )
raise SystemExit ( 1 )
with open ( os . path . join ( global_idf_path , TOOLS_FILE ) , ' r ' ) as tools_file :
tools_json = json . load ( tools_file )
with open ( os . path . join ( global_idf_path , TOOLS_SCHEMA_FILE ) , ' r ' ) as schema_file :
schema_json = json . load ( schema_file )
jsonschema . validate ( tools_json , schema_json )
# on failure, this will raise an exception with a fairly verbose diagnostic message
2021-03-04 03:53:47 -05:00
def action_gen_doc ( args ) : # type: ignore
2020-01-17 11:38:50 -05:00
f = args . output
tools_info = load_tools_info ( )
2021-03-04 03:53:47 -05:00
def print_out ( text ) : # type: (str) -> None
2020-01-17 11:38:50 -05:00
f . write ( text + ' \n ' )
2021-01-25 21:49:01 -05:00
print_out ( ' .. |zwsp| unicode:: U+200B ' )
print_out ( ' :trim: ' )
print_out ( ' ' )
2020-01-17 11:38:50 -05:00
2021-01-25 21:49:01 -05:00
idf_gh_url = ' https://github.com/espressif/esp-idf '
2020-01-17 11:38:50 -05:00
for tool_name , tool_obj in tools_info . items ( ) :
info_url = tool_obj . options . info_url
2021-01-25 21:49:01 -05:00
if idf_gh_url + ' /tree ' in info_url :
info_url = re . sub ( idf_gh_url + r ' /tree/ \ w+/(.*) ' , r ' :idf:` \ 1` ' , info_url )
2020-01-17 11:38:50 -05:00
2021-01-25 21:49:01 -05:00
license_url = ' https://spdx.org/licenses/ ' + tool_obj . options . license
2020-01-17 11:38:50 -05:00
print_out ( """
. . _tool - { name } :
{ name }
{ underline }
{ description }
. . include : : idf - tools - notes . inc
: start - after : tool - { name } - notes
: end - before : - - -
License : ` { license } < { license_url } > ` _
More info : { info_url }
. . list - table : :
: widths : 10 10 80
: header - rows : 1
* - Platform
- Required
- Download
""" .rstrip().format(name=tool_name,
underline = args . heading_underline_char * len ( tool_name ) ,
description = tool_obj . description ,
license = tool_obj . options . license ,
license_url = license_url ,
info_url = info_url ) )
for platform_name in sorted ( tool_obj . get_supported_platforms ( ) ) :
platform_tool = tool_obj . copy_for_platform ( platform_name )
install_type = platform_tool . get_install_type ( )
if install_type == IDFTool . INSTALL_NEVER :
continue
elif install_type == IDFTool . INSTALL_ALWAYS :
2021-01-25 21:49:01 -05:00
install_type_str = ' required '
2020-01-17 11:38:50 -05:00
elif install_type == IDFTool . INSTALL_ON_REQUEST :
2021-01-25 21:49:01 -05:00
install_type_str = ' optional '
2020-01-17 11:38:50 -05:00
else :
raise NotImplementedError ( )
version = platform_tool . get_recommended_version ( )
version_obj = platform_tool . versions [ version ]
download_obj = version_obj . get_download_for_platform ( platform_name )
# Note: keep the list entries indented to the same number of columns
# as the list header above.
print_out ( """
* - { }
- { }
- { }
. . rst - class : : tool - sha256
SHA256 : { }
""" .strip( ' \n ' ).format(platform_name, install_type_str, download_obj.url, download_obj.sha256))
print_out ( ' ' )
print_out ( ' ' )
2022-08-31 16:59:15 -04:00
def action_check_tool_supported ( args ) : # type: (Any) -> None
"""
Print " True " / " False " to stdout as a result that tool is supported in IDF
Print erorr message to stderr otherwise and set exit code to 1
"""
try :
tools_info = load_tools_info ( )
for _ , v in tools_info . items ( ) :
if v . name == args . tool_name :
print ( v . check_version ( args . exec_path ) )
break
except ( RuntimeError , ToolNotFound , ToolExecError ) as err :
fatal ( f ' Failed to check tool support: (name: { args . tool_name } , exec: { args . exec_path } ) ' )
fatal ( f ' { err } ' )
raise SystemExit ( 1 )
def action_get_tool_supported_versions ( args ) : # type: (Any) -> None
"""
Print supported versions of a tool to stdout
Print erorr message to stderr otherwise and set exit code to 1
"""
try :
tools_info = load_tools_info ( )
for _ , v in tools_info . items ( ) :
if v . name == args . tool_name :
print ( list ( v . versions . keys ( ) ) )
break
except RuntimeError as err :
fatal ( f ' Failed to get tool supported versions. (tool: { args . tool_name } ) ' )
fatal ( f ' { err } ' )
raise SystemExit ( 1 )
2021-03-04 03:53:47 -05:00
def main ( argv ) : # type: (list[str]) -> None
2019-04-17 03:30:30 -04:00
parser = argparse . ArgumentParser ( )
parser . add_argument ( ' --quiet ' , help = ' Don \' t output diagnostic messages to stdout/stderr ' , action = ' store_true ' )
2019-06-05 03:16:19 -04:00
parser . add_argument ( ' --non-interactive ' , help = ' Don \' t output interactive messages and questions ' , action = ' store_true ' )
2019-04-17 03:30:30 -04:00
parser . add_argument ( ' --tools-json ' , help = ' Path to the tools.json file to use ' )
parser . add_argument ( ' --idf-path ' , help = ' ESP-IDF path to use ' )
subparsers = parser . add_subparsers ( dest = ' action ' )
2023-02-05 10:29:03 -05:00
list_parser = subparsers . add_parser ( ' list ' , help = ' List tools and versions available ' )
list_parser . add_argument ( ' --outdated ' , help = ' Print only outdated installed tools ' , action = ' store_true ' )
2019-04-17 03:30:30 -04:00
subparsers . add_parser ( ' check ' , help = ' Print summary of tools installed or found in PATH ' )
export = subparsers . add_parser ( ' export ' , help = ' Output command for setting tool paths, suitable for shell ' )
export . add_argument ( ' --format ' , choices = [ EXPORT_SHELL , EXPORT_KEY_VALUE ] , default = EXPORT_SHELL ,
help = ' Format of the output: shell (suitable for printing into shell), ' +
' or key-value (suitable for parsing by other tools ' )
export . add_argument ( ' --prefer-system ' , help = ' Normally, if the tool is already present in PATH, ' +
' but has an unsupported version, a version from the tools directory ' +
' will be used instead. If this flag is given, the version in PATH ' +
' will be used. ' , action = ' store_true ' )
2022-08-09 12:13:29 -04:00
export . add_argument ( ' --deactivate ' , help = ' Output command for deactivate different ESP-IDF version, previously set with export ' , action = ' store_true ' )
export . add_argument ( ' --unset ' , help = argparse . SUPPRESS , action = ' store_true ' )
export . add_argument ( ' --add_paths_extras ' , help = ' Add idf-related path extras for deactivate option ' )
2019-04-17 03:30:30 -04:00
install = subparsers . add_parser ( ' install ' , help = ' Download and install tools into the tools directory ' )
2019-11-24 06:11:33 -05:00
install . add_argument ( ' tools ' , metavar = ' TOOL ' , nargs = ' * ' , default = [ ' required ' ] ,
help = ' Tools to install. ' +
' To install a specific version use <tool_name>@<version> syntax. ' +
2023-09-14 01:55:24 -04:00
' To install tools by pattern use wildcards in <tool_name_pattern> . ' +
2019-11-24 06:11:33 -05:00
' Use empty or \' required \' to install required tools, not optional ones. ' +
' Use \' all \' to install all tools, including the optional ones. ' )
2021-04-26 15:34:48 -04:00
install . add_argument ( ' --targets ' , default = ' all ' , help = ' A comma separated list of desired chip targets for installing. ' +
' It defaults to installing all supported targets. ' )
2019-04-17 03:30:30 -04:00
2019-07-18 11:20:40 -04:00
download = subparsers . add_parser ( ' download ' , help = ' Download the tools into the dist directory ' )
2021-06-21 04:22:23 -04:00
download . add_argument ( ' --platform ' , default = CURRENT_PLATFORM , help = ' Platform to download the tools for ' )
2019-11-24 06:11:33 -05:00
download . add_argument ( ' tools ' , metavar = ' TOOL ' , nargs = ' * ' , default = [ ' required ' ] ,
help = ' Tools to download. ' +
' To download a specific version use <tool_name>@<version> syntax. ' +
2023-09-14 01:55:24 -04:00
' To download tools by pattern use wildcards in <tool_name_pattern> . ' +
2019-11-24 06:11:33 -05:00
' Use empty or \' required \' to download required tools, not optional ones. ' +
' Use \' all \' to download all tools, including the optional ones. ' )
2021-04-26 15:34:48 -04:00
download . add_argument ( ' --targets ' , default = ' all ' , help = ' A comma separated list of desired chip targets for installing. ' +
' It defaults to installing all supported targets. ' )
2019-07-18 11:20:40 -04:00
2021-12-13 10:45:11 -05:00
uninstall = subparsers . add_parser ( ' uninstall ' , help = ' Remove installed tools, that are not used by current version of ESP-IDF. ' )
uninstall . add_argument ( ' --dry-run ' , help = ' Print unused tools. ' , action = ' store_true ' )
uninstall . add_argument ( ' --remove-archives ' , help = ' Remove old archive versions and archives from unused tools. ' , action = ' store_true ' )
2022-07-01 08:44:12 -04:00
no_constraints_default = os . environ . get ( ' IDF_PYTHON_CHECK_CONSTRAINTS ' , ' ' ) . lower ( ) in [ ' 0 ' , ' n ' , ' no ' ]
2019-07-18 11:20:40 -04:00
if IDF_MAINTAINER :
for subparser in [ download , install ] :
subparser . add_argument ( ' --mirror-prefix-map ' , nargs = ' * ' ,
help = ' Pattern to rewrite download URLs, with source and replacement separated by comma. ' +
' E.g. http://foo.com,http://test.foo.com ' )
2019-04-17 03:30:30 -04:00
install_python_env = subparsers . add_parser ( ' install-python-env ' ,
help = ' Create Python virtual environment and install the ' +
' required Python packages ' )
install_python_env . add_argument ( ' --reinstall ' , help = ' Discard the previously installed environment ' ,
action = ' store_true ' )
install_python_env . add_argument ( ' --extra-wheels-dir ' , help = ' Additional directories with wheels ' +
' to use during installation ' )
2022-11-10 23:00:56 -05:00
install_python_env . add_argument ( ' --extra-wheels-url ' , help = ' Additional URL with wheels ' , default = IDF_PIP_WHEELS_URL )
2020-12-16 11:40:13 -05:00
install_python_env . add_argument ( ' --no-index ' , help = ' Work offline without retrieving wheels index ' )
2021-09-16 10:48:03 -04:00
install_python_env . add_argument ( ' --features ' , default = ' core ' , help = ' A comma separated list of desired features for installing. '
' It defaults to installing just the core funtionality. ' )
2022-07-01 08:44:12 -04:00
install_python_env . add_argument ( ' --no-constraints ' , action = ' store_true ' , default = no_constraints_default ,
2021-09-16 10:48:03 -04:00
help = ' Disable constraint settings. Use with care and only when you want to manage '
2022-07-01 08:44:12 -04:00
' package versions by yourself. It can be set with the IDF_PYTHON_CHECK_CONSTRAINTS '
' environment variable. ' )
2019-04-17 03:30:30 -04:00
if IDF_MAINTAINER :
add_version = subparsers . add_parser ( ' add-version ' , help = ' Add or update download info for a version ' )
add_version . add_argument ( ' --output ' , help = ' Save new tools.json into this file ' )
add_version . add_argument ( ' --tool ' , help = ' Tool name to set add a version for ' , required = True )
add_version . add_argument ( ' --version ' , help = ' Version identifier ' , required = True )
add_version . add_argument ( ' --url-prefix ' , help = ' String to prepend to file names to obtain download URLs ' )
2022-03-15 15:02:37 -04:00
add_version . add_argument ( ' --override ' , action = ' store_true ' , help = ' Override tool versions with new data ' )
add_version_files_group = add_version . add_mutually_exclusive_group ( required = True )
add_version_files_group . add_argument ( ' --checksum-file ' , help = ' URL or path to local file with checksum/size for artifacts ' )
add_version_files_group . add_argument ( ' --artifact-file ' , help = ' File names of the download artifacts ' , nargs = ' * ' )
2019-04-17 03:30:30 -04:00
rewrite = subparsers . add_parser ( ' rewrite ' , help = ' Load tools.json, validate, and save the result back into JSON ' )
rewrite . add_argument ( ' --output ' , help = ' Save new tools.json into this file ' )
subparsers . add_parser ( ' validate ' , help = ' Validate tools.json against schema file ' )
2020-01-17 11:38:50 -05:00
gen_doc = subparsers . add_parser ( ' gen-doc ' , help = ' Write the list of tools as a documentation page ' )
gen_doc . add_argument ( ' --output ' , type = argparse . FileType ( ' w ' ) , default = sys . stdout ,
help = ' Output file name ' )
gen_doc . add_argument ( ' --heading-underline-char ' , help = ' Character to use when generating RST sections ' , default = ' ~ ' )
2021-09-16 10:48:03 -04:00
check_python_dependencies = subparsers . add_parser ( ' check-python-dependencies ' ,
help = ' Check that all required Python packages are installed. ' )
2022-07-01 08:44:12 -04:00
check_python_dependencies . add_argument ( ' --no-constraints ' , action = ' store_true ' , default = no_constraints_default ,
2021-09-16 10:48:03 -04:00
help = ' Disable constraint settings. Use with care and only when you want '
2022-07-01 08:44:12 -04:00
' to manage package versions by yourself. It can be set with the IDF_PYTHON_CHECK_CONSTRAINTS '
' environment variable. ' )
2021-09-16 10:48:03 -04:00
2022-08-31 16:59:15 -04:00
if os . environ . get ( ' IDF_TOOLS_VERSION_HELPER ' ) :
check_tool_supported = subparsers . add_parser ( ' check-tool-supported ' ,
help = ' Check that selected tool is compatible with IDF. Writes " True " / " False " to stdout in success. ' )
check_tool_supported . add_argument ( ' --tool-name ' , required = True , help = ' Tool name (from tools.json) ' )
check_tool_supported . add_argument ( ' --exec-path ' , required = True , help = ' Full path to executable under the test ' )
get_tool_supported_versions = subparsers . add_parser ( ' get-tool-supported-versions ' , help = ' Prints a list of tool \' s supported versions ' )
get_tool_supported_versions . add_argument ( ' --tool-name ' , required = True , help = ' Tool name (from tools.json) ' )
2019-04-17 03:30:30 -04:00
args = parser . parse_args ( argv )
if args . action is None :
parser . print_help ( )
parser . exit ( 1 )
if args . quiet :
global global_quiet
global_quiet = True
2019-06-05 03:16:19 -04:00
if args . non_interactive :
global global_non_interactive
global_non_interactive = True
2022-08-09 12:13:29 -04:00
if ' unset ' in args and args . unset :
args . deactivate = True
2019-04-17 03:30:30 -04:00
global global_idf_path
global_idf_path = os . environ . get ( ' IDF_PATH ' )
if args . idf_path :
global_idf_path = args . idf_path
if not global_idf_path :
2021-01-25 21:49:01 -05:00
global_idf_path = os . path . realpath ( os . path . join ( os . path . dirname ( __file__ ) , ' .. ' ) )
2019-11-18 10:17:30 -05:00
os . environ [ ' IDF_PATH ' ] = global_idf_path
2019-04-17 03:30:30 -04:00
global global_idf_tools_path
global_idf_tools_path = os . environ . get ( ' IDF_TOOLS_PATH ' ) or os . path . expanduser ( IDF_TOOLS_PATH_DEFAULT )
2019-10-30 11:19:17 -04:00
# On macOS, unset __PYVENV_LAUNCHER__ variable if it is set.
# Otherwise sys.executable keeps pointing to the system Python, even when a python binary from a virtualenv is invoked.
# See https://bugs.python.org/issue22490#msg283859.
2020-01-21 04:48:24 -05:00
os . environ . pop ( ' __PYVENV_LAUNCHER__ ' , None )
2019-10-30 11:19:17 -04:00
2019-04-17 03:30:30 -04:00
if sys . version_info . major == 2 :
try :
2021-03-04 03:53:47 -05:00
global_idf_tools_path . decode ( ' ascii ' ) # type: ignore
2019-04-17 03:30:30 -04:00
except UnicodeDecodeError :
fatal ( ' IDF_TOOLS_PATH contains non-ASCII characters: {} ' . format ( global_idf_tools_path ) +
' \n This is not supported yet with Python 2. ' +
' Please set IDF_TOOLS_PATH to a directory with an ASCII name, or switch to Python 3. ' )
raise SystemExit ( 1 )
2022-05-16 16:11:01 -04:00
if CURRENT_PLATFORM is None :
2019-04-17 03:30:30 -04:00
fatal ( ' Platform {} appears to be unsupported ' . format ( PYTHON_PLATFORM ) )
raise SystemExit ( 1 )
global global_tools_json
if args . tools_json :
global_tools_json = args . tools_json
else :
global_tools_json = os . path . join ( global_idf_path , TOOLS_FILE )
action_func_name = ' action_ ' + args . action . replace ( ' - ' , ' _ ' )
action_func = globals ( ) [ action_func_name ]
action_func ( args )
if __name__ == ' __main__ ' :
2021-11-09 07:10:56 -05:00
if ' MSYSTEM ' in os . environ :
fatal ( ' MSys/Mingw is not supported. Please follow the getting started guide of the documentation to set up '
' a supported environment ' )
raise SystemExit ( 1 )
2019-04-17 03:30:30 -04:00
main ( sys . argv [ 1 : ] )