2018-01-23 01:08:28 -05:00
|
|
|
#!/usr/bin/env python
|
|
|
|
#
|
|
|
|
# 'idf.py' is a top-level config/build command line tool for ESP-IDF
|
|
|
|
#
|
|
|
|
# You don't have to use idf.py, you can use cmake directly
|
|
|
|
# (or use cmake in an IDE)
|
|
|
|
#
|
|
|
|
#
|
|
|
|
#
|
|
|
|
# Copyright 2018 Espressif Systems (Shanghai) PTE LTD
|
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
#
|
2018-09-03 06:24:17 -04:00
|
|
|
|
2018-10-25 02:16:30 -04:00
|
|
|
# WARNING: we don't check for Python build-time dependencies until
|
2018-09-03 06:24:17 -04:00
|
|
|
# check_environment() function below. If possible, avoid importing
|
|
|
|
# any external libraries here - put in external script, or import in
|
|
|
|
# their specific function instead.
|
2018-01-23 01:08:28 -05:00
|
|
|
import sys
|
|
|
|
import argparse
|
|
|
|
import os
|
|
|
|
import os.path
|
|
|
|
import subprocess
|
|
|
|
import multiprocessing
|
|
|
|
import re
|
|
|
|
import shutil
|
|
|
|
import json
|
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
|
2018-02-15 23:32:08 -05:00
|
|
|
class FatalError(RuntimeError):
|
|
|
|
"""
|
|
|
|
Wrapper class for runtime errors that aren't caused by bugs in idf.py or the build proces.s
|
|
|
|
"""
|
|
|
|
pass
|
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
|
2018-01-23 01:08:28 -05:00
|
|
|
# Use this Python interpreter for any subprocesses we launch
|
2018-12-04 07:46:48 -05:00
|
|
|
PYTHON = sys.executable
|
2018-01-23 01:08:28 -05:00
|
|
|
|
|
|
|
# note: os.environ changes don't automatically propagate to child processes,
|
2018-06-15 00:59:45 -04:00
|
|
|
# you have to pass env=os.environ explicitly anywhere that we create a process
|
2018-12-04 07:46:48 -05:00
|
|
|
os.environ["PYTHON"] = sys.executable
|
2018-01-23 01:08:28 -05:00
|
|
|
|
|
|
|
# Make flavors, across the various kinds of Windows environments & POSIX...
|
|
|
|
if "MSYSTEM" in os.environ: # MSYS
|
|
|
|
MAKE_CMD = "make"
|
|
|
|
MAKE_GENERATOR = "MSYS Makefiles"
|
|
|
|
elif os.name == 'nt': # other Windows
|
|
|
|
MAKE_CMD = "mingw32-make"
|
|
|
|
MAKE_GENERATOR = "MinGW Makefiles"
|
|
|
|
else:
|
|
|
|
MAKE_CMD = "make"
|
|
|
|
MAKE_GENERATOR = "Unix Makefiles"
|
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
GENERATORS = \
|
|
|
|
[
|
|
|
|
# ('generator name', 'build command line', 'version command line', 'verbose flag')
|
|
|
|
("Ninja", ["ninja"], ["ninja", "--version"], "-v"),
|
2019-05-16 09:20:10 -04:00
|
|
|
(MAKE_GENERATOR, [MAKE_CMD, "-j", str(multiprocessing.cpu_count() + 2)], [MAKE_CMD, "--version"], "VERBOSE=1"),
|
2018-01-23 01:08:28 -05:00
|
|
|
]
|
2018-12-04 07:46:48 -05:00
|
|
|
GENERATOR_CMDS = dict((a[0], a[1]) for a in GENERATORS)
|
|
|
|
GENERATOR_VERBOSE = dict((a[0], a[3]) for a in GENERATORS)
|
|
|
|
|
2018-01-23 01:08:28 -05:00
|
|
|
|
2018-02-15 23:32:08 -05:00
|
|
|
def _run_tool(tool_name, args, cwd):
|
|
|
|
def quote_arg(arg):
|
|
|
|
" Quote 'arg' if necessary "
|
|
|
|
if " " in arg and not (arg.startswith('"') or arg.startswith("'")):
|
|
|
|
return "'" + arg + "'"
|
|
|
|
return arg
|
|
|
|
display_args = " ".join(quote_arg(arg) for arg in args)
|
|
|
|
print("Running %s in directory %s" % (tool_name, quote_arg(cwd)))
|
2018-11-29 09:52:09 -05:00
|
|
|
print('Executing "%s"...' % str(display_args))
|
2018-02-15 23:32:08 -05:00
|
|
|
try:
|
|
|
|
# Note: we explicitly pass in os.environ here, as we may have set IDF_PATH there during startup
|
|
|
|
subprocess.check_call(args, env=os.environ, cwd=cwd)
|
|
|
|
except subprocess.CalledProcessError as e:
|
|
|
|
raise FatalError("%s failed with exit code %d" % (tool_name, e.returncode))
|
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
|
2018-01-23 01:08:28 -05:00
|
|
|
def check_environment():
|
|
|
|
"""
|
|
|
|
Verify the environment contains the top-level tools we need to operate
|
|
|
|
|
|
|
|
(cmake will check a lot of other things)
|
|
|
|
"""
|
|
|
|
if not executable_exists(["cmake", "--version"]):
|
2018-02-15 23:32:08 -05:00
|
|
|
raise FatalError("'cmake' must be available on the PATH to use idf.py")
|
2018-01-23 01:08:28 -05:00
|
|
|
# find the directory idf.py is in, then the parent directory of this, and assume this is IDF_PATH
|
|
|
|
detected_idf_path = os.path.realpath(os.path.join(os.path.dirname(__file__), ".."))
|
|
|
|
if "IDF_PATH" in os.environ:
|
|
|
|
set_idf_path = os.path.realpath(os.environ["IDF_PATH"])
|
|
|
|
if set_idf_path != detected_idf_path:
|
2018-12-04 07:46:48 -05:00
|
|
|
print("WARNING: IDF_PATH environment variable is set to %s but idf.py path indicates IDF directory %s. "
|
|
|
|
"Using the environment variable directory, but results may be unexpected..."
|
2018-01-23 01:08:28 -05:00
|
|
|
% (set_idf_path, detected_idf_path))
|
|
|
|
else:
|
2018-06-15 00:59:45 -04:00
|
|
|
print("Setting IDF_PATH environment variable: %s" % detected_idf_path)
|
2018-01-23 01:08:28 -05:00
|
|
|
os.environ["IDF_PATH"] = detected_idf_path
|
|
|
|
|
2018-09-03 06:24:17 -04:00
|
|
|
# check Python dependencies
|
|
|
|
print("Checking Python dependencies...")
|
|
|
|
try:
|
2018-12-04 07:46:48 -05:00
|
|
|
subprocess.check_call([os.environ["PYTHON"],
|
|
|
|
os.path.join(os.environ["IDF_PATH"], "tools", "check_python_dependencies.py")],
|
2018-09-03 06:24:17 -04:00
|
|
|
env=os.environ)
|
|
|
|
except subprocess.CalledProcessError:
|
|
|
|
raise SystemExit(1)
|
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
|
2018-01-23 01:08:28 -05:00
|
|
|
def executable_exists(args):
|
|
|
|
try:
|
|
|
|
subprocess.check_output(args)
|
|
|
|
return True
|
2018-12-04 07:46:48 -05:00
|
|
|
except Exception:
|
2018-01-23 01:08:28 -05:00
|
|
|
return False
|
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
|
2018-01-23 01:08:28 -05:00
|
|
|
def detect_cmake_generator():
|
|
|
|
"""
|
|
|
|
Find the default cmake generator, if none was specified. Raises an exception if no valid generator is found.
|
|
|
|
"""
|
2018-03-15 20:58:52 -04:00
|
|
|
for (generator, _, version_check, _) in GENERATORS:
|
2018-01-23 01:08:28 -05:00
|
|
|
if executable_exists(version_check):
|
|
|
|
return generator
|
2018-02-15 23:32:08 -05:00
|
|
|
raise FatalError("To use idf.py, either the 'ninja' or 'GNU make' build tool must be available in the PATH")
|
2018-01-23 01:08:28 -05:00
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
|
2019-06-03 06:37:52 -04:00
|
|
|
def _strip_quotes(value, regexp=re.compile(r"^\"(.*)\"$|^'(.*)'$|^(.*)$")):
|
|
|
|
"""
|
|
|
|
Strip quotes like CMake does during parsing cache entries
|
|
|
|
"""
|
|
|
|
|
|
|
|
return [x for x in regexp.match(value).groups() if x is not None][0].rstrip()
|
|
|
|
|
|
|
|
|
|
|
|
def _new_cmakecache_entries(cache_path, new_cache_entries):
|
|
|
|
if not os.path.exists(cache_path):
|
|
|
|
return True
|
|
|
|
|
|
|
|
current_cache = parse_cmakecache(cache_path)
|
|
|
|
|
|
|
|
if new_cache_entries:
|
|
|
|
current_cache = parse_cmakecache(cache_path)
|
|
|
|
|
|
|
|
for entry in new_cache_entries:
|
|
|
|
key, value = entry.split("=", 1)
|
|
|
|
current_value = current_cache.get(key, None)
|
|
|
|
if current_value is None or _strip_quotes(value) != current_value:
|
|
|
|
return True
|
|
|
|
|
|
|
|
return False
|
|
|
|
|
|
|
|
|
2018-02-26 21:21:51 -05:00
|
|
|
def _ensure_build_directory(args, always_run_cmake=False):
|
2018-01-23 01:08:28 -05:00
|
|
|
"""Check the build directory exists and that cmake has been run there.
|
|
|
|
|
|
|
|
If this isn't the case, create the build directory (if necessary) and
|
|
|
|
do an initial cmake run to configure it.
|
|
|
|
|
|
|
|
This function will also check args.generator parameter. If the parameter is incompatible with
|
|
|
|
the build directory, an error is raised. If the parameter is None, this function will set it to
|
|
|
|
an auto-detected default generator or to the value already configured in the build directory.
|
|
|
|
"""
|
|
|
|
project_dir = args.project_dir
|
|
|
|
# Verify the project directory
|
|
|
|
if not os.path.isdir(project_dir):
|
|
|
|
if not os.path.exists(project_dir):
|
2018-02-15 23:32:08 -05:00
|
|
|
raise FatalError("Project directory %s does not exist")
|
2018-01-23 01:08:28 -05:00
|
|
|
else:
|
2018-02-15 23:32:08 -05:00
|
|
|
raise FatalError("%s must be a project directory")
|
2018-01-23 01:08:28 -05:00
|
|
|
if not os.path.exists(os.path.join(project_dir, "CMakeLists.txt")):
|
2018-02-15 23:32:08 -05:00
|
|
|
raise FatalError("CMakeLists.txt not found in project directory %s" % project_dir)
|
2018-01-23 01:08:28 -05:00
|
|
|
|
|
|
|
# Verify/create the build directory
|
|
|
|
build_dir = args.build_dir
|
|
|
|
if not os.path.isdir(build_dir):
|
2018-08-26 22:48:16 -04:00
|
|
|
os.makedirs(build_dir)
|
2018-01-23 01:08:28 -05:00
|
|
|
cache_path = os.path.join(build_dir, "CMakeCache.txt")
|
2019-06-03 06:37:52 -04:00
|
|
|
|
|
|
|
if always_run_cmake or _new_cmakecache_entries(cache_path, args.define_cache_entry):
|
2018-01-23 01:08:28 -05:00
|
|
|
if args.generator is None:
|
|
|
|
args.generator = detect_cmake_generator()
|
|
|
|
try:
|
2018-10-21 09:34:53 -04:00
|
|
|
cmake_args = ["cmake", "-G", args.generator, "-DPYTHON_DEPS_CHECKED=1", "-DESP_PLATFORM=1"]
|
2018-02-21 23:05:05 -05:00
|
|
|
if not args.no_warnings:
|
2018-12-04 07:46:48 -05:00
|
|
|
cmake_args += ["--warn-uninitialized"]
|
2018-03-15 20:58:52 -04:00
|
|
|
if args.no_ccache:
|
2018-12-04 07:46:48 -05:00
|
|
|
cmake_args += ["-DCCACHE_DISABLE=1"]
|
2018-08-26 22:48:16 -04:00
|
|
|
if args.define_cache_entry:
|
2018-12-04 07:46:48 -05:00
|
|
|
cmake_args += ["-D" + d for d in args.define_cache_entry]
|
|
|
|
cmake_args += [project_dir]
|
2018-08-26 22:48:16 -04:00
|
|
|
|
2018-02-21 23:05:05 -05:00
|
|
|
_run_tool("cmake", cmake_args, cwd=args.build_dir)
|
2018-12-04 07:46:48 -05:00
|
|
|
except Exception:
|
2018-02-15 23:32:08 -05:00
|
|
|
# don't allow partially valid CMakeCache.txt files,
|
|
|
|
# to keep the "should I run cmake?" logic simple
|
|
|
|
if os.path.exists(cache_path):
|
2018-01-23 01:08:28 -05:00
|
|
|
os.remove(cache_path)
|
|
|
|
raise
|
|
|
|
|
|
|
|
# Learn some things from the CMakeCache.txt file in the build directory
|
|
|
|
cache = parse_cmakecache(cache_path)
|
|
|
|
try:
|
|
|
|
generator = cache["CMAKE_GENERATOR"]
|
|
|
|
except KeyError:
|
|
|
|
generator = detect_cmake_generator()
|
|
|
|
if args.generator is None:
|
|
|
|
args.generator = generator # reuse the previously configured generator, if none was given
|
|
|
|
if generator != args.generator:
|
2018-02-15 23:32:08 -05:00
|
|
|
raise FatalError("Build is configured for generator '%s' not '%s'. Run 'idf.py fullclean' to start again."
|
2018-12-04 07:46:48 -05:00
|
|
|
% (generator, args.generator))
|
2018-01-23 01:08:28 -05:00
|
|
|
|
|
|
|
try:
|
|
|
|
home_dir = cache["CMAKE_HOME_DIRECTORY"]
|
2018-03-15 20:58:52 -04:00
|
|
|
if os.path.normcase(os.path.realpath(home_dir)) != os.path.normcase(os.path.realpath(project_dir)):
|
2018-02-15 23:32:08 -05:00
|
|
|
raise FatalError("Build directory '%s' configured for project '%s' not '%s'. Run 'idf.py fullclean' to start again."
|
2018-12-04 07:46:48 -05:00
|
|
|
% (build_dir, os.path.realpath(home_dir), os.path.realpath(project_dir)))
|
2018-01-23 01:08:28 -05:00
|
|
|
except KeyError:
|
|
|
|
pass # if cmake failed part way, CMAKE_HOME_DIRECTORY may not be set yet
|
|
|
|
|
|
|
|
|
|
|
|
def parse_cmakecache(path):
|
|
|
|
"""
|
|
|
|
Parse the CMakeCache file at 'path'.
|
|
|
|
|
|
|
|
Returns a dict of name:value.
|
|
|
|
|
|
|
|
CMakeCache entries also each have a "type", but this is currently ignored.
|
|
|
|
"""
|
|
|
|
result = {}
|
|
|
|
with open(path) as f:
|
|
|
|
for line in f:
|
|
|
|
# cmake cache lines look like: CMAKE_CXX_FLAGS_DEBUG:STRING=-g
|
|
|
|
# groups are name, type, value
|
2019-06-03 06:37:52 -04:00
|
|
|
m = re.match(r"^([^#/:=]+):([^:=]+)=(.*)\n$", line)
|
2018-01-23 01:08:28 -05:00
|
|
|
if m:
|
2018-12-04 07:46:48 -05:00
|
|
|
result[m.group(1)] = m.group(3)
|
2018-01-23 01:08:28 -05:00
|
|
|
return result
|
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
|
2018-01-23 01:08:28 -05:00
|
|
|
def build_target(target_name, args):
|
|
|
|
"""
|
|
|
|
Execute the target build system to build target 'target_name'
|
|
|
|
|
|
|
|
Calls _ensure_build_directory() which will run cmake to generate a build
|
|
|
|
directory (with the specified generator) as needed.
|
|
|
|
"""
|
|
|
|
_ensure_build_directory(args)
|
|
|
|
generator_cmd = GENERATOR_CMDS[args.generator]
|
2018-08-26 22:48:16 -04:00
|
|
|
|
2018-03-15 20:58:52 -04:00
|
|
|
if not args.no_ccache:
|
|
|
|
# Setting CCACHE_BASEDIR & CCACHE_NO_HASHDIR ensures that project paths aren't stored in the ccache entries
|
|
|
|
# (this means ccache hits can be shared between different projects. It may mean that some debug information
|
|
|
|
# will point to files in another project, if these files are perfect duplicates of each other.)
|
|
|
|
#
|
|
|
|
# It would be nicer to set these from cmake, but there's no cross-platform way to set build-time environment
|
2018-12-04 07:46:48 -05:00
|
|
|
# os.environ["CCACHE_BASEDIR"] = args.build_dir
|
|
|
|
# os.environ["CCACHE_NO_HASHDIR"] = "1"
|
2018-03-15 20:58:52 -04:00
|
|
|
pass
|
|
|
|
if args.verbose:
|
2018-12-04 07:46:48 -05:00
|
|
|
generator_cmd += [GENERATOR_VERBOSE[args.generator]]
|
2018-03-15 20:58:52 -04:00
|
|
|
|
2018-02-15 23:32:08 -05:00
|
|
|
_run_tool(generator_cmd[0], generator_cmd + [target_name], args.build_dir)
|
|
|
|
|
2018-01-23 01:08:28 -05:00
|
|
|
|
|
|
|
def _get_esptool_args(args):
|
|
|
|
esptool_path = os.path.join(os.environ["IDF_PATH"], "components/esptool_py/esptool/esptool.py")
|
2018-09-06 08:32:49 -04:00
|
|
|
if args.port is None:
|
|
|
|
args.port = get_default_serial_port()
|
2018-12-04 07:46:48 -05:00
|
|
|
result = [PYTHON, esptool_path]
|
|
|
|
result += ["-p", args.port]
|
|
|
|
result += ["-b", str(args.baud)]
|
2018-10-19 15:02:55 -04:00
|
|
|
|
|
|
|
with open(os.path.join(args.build_dir, "flasher_args.json")) as f:
|
|
|
|
flasher_args = json.load(f)
|
|
|
|
|
|
|
|
extra_esptool_args = flasher_args["extra_esptool_args"]
|
2018-12-04 07:46:48 -05:00
|
|
|
result += ["--after", extra_esptool_args["after"]]
|
2018-01-23 01:08:28 -05:00
|
|
|
return result
|
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
|
2018-01-23 01:08:28 -05:00
|
|
|
def flash(action, args):
|
|
|
|
"""
|
|
|
|
Run esptool to flash the entire project, from an argfile generated by the build system
|
|
|
|
"""
|
2018-02-15 23:32:08 -05:00
|
|
|
flasher_args_path = { # action -> name of flasher args file generated by build system
|
2018-01-23 01:08:28 -05:00
|
|
|
"bootloader-flash": "flash_bootloader_args",
|
|
|
|
"partition_table-flash": "flash_partition_table_args",
|
|
|
|
"app-flash": "flash_app_args",
|
|
|
|
"flash": "flash_project_args",
|
|
|
|
}[action]
|
|
|
|
esptool_args = _get_esptool_args(args)
|
2018-12-04 07:46:48 -05:00
|
|
|
esptool_args += ["write_flash", "@" + flasher_args_path]
|
2018-02-15 23:32:08 -05:00
|
|
|
_run_tool("esptool.py", esptool_args, args.build_dir)
|
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
|
2018-01-23 01:08:28 -05:00
|
|
|
def erase_flash(action, args):
|
|
|
|
esptool_args = _get_esptool_args(args)
|
2018-12-04 07:46:48 -05:00
|
|
|
esptool_args += ["erase_flash"]
|
2018-02-15 23:32:08 -05:00
|
|
|
_run_tool("esptool.py", esptool_args, args.build_dir)
|
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
|
2018-01-23 01:08:28 -05:00
|
|
|
def monitor(action, args):
|
|
|
|
"""
|
|
|
|
Run idf_monitor.py to watch build output
|
|
|
|
"""
|
2018-09-06 08:32:49 -04:00
|
|
|
if args.port is None:
|
|
|
|
args.port = get_default_serial_port()
|
2018-01-23 01:08:28 -05:00
|
|
|
desc_path = os.path.join(args.build_dir, "project_description.json")
|
|
|
|
if not os.path.exists(desc_path):
|
|
|
|
_ensure_build_directory(args)
|
|
|
|
with open(desc_path, "r") as f:
|
|
|
|
project_desc = json.load(f)
|
|
|
|
|
|
|
|
elf_file = os.path.join(args.build_dir, project_desc["app_elf"])
|
|
|
|
if not os.path.exists(elf_file):
|
2018-12-04 07:46:48 -05:00
|
|
|
raise FatalError("ELF file '%s' not found. You need to build & flash the project before running 'monitor', "
|
|
|
|
"and the binary on the device must match the one in the build directory exactly. "
|
|
|
|
"Try 'idf.py flash monitor'." % elf_file)
|
2018-01-23 01:08:28 -05:00
|
|
|
idf_monitor = os.path.join(os.environ["IDF_PATH"], "tools/idf_monitor.py")
|
2018-12-04 07:46:48 -05:00
|
|
|
monitor_args = [PYTHON, idf_monitor]
|
2018-01-23 01:08:28 -05:00
|
|
|
if args.port is not None:
|
2018-12-04 07:46:48 -05:00
|
|
|
monitor_args += ["-p", args.port]
|
|
|
|
monitor_args += ["-b", project_desc["monitor_baud"]]
|
|
|
|
monitor_args += [elf_file]
|
2018-05-04 00:06:15 -04:00
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
idf_py = [PYTHON] + get_commandline_options() # commands to re-run idf.py
|
|
|
|
monitor_args += ["-m", " ".join("'%s'" % a for a in idf_py)]
|
2018-05-04 00:06:15 -04:00
|
|
|
|
2018-11-29 09:52:09 -05:00
|
|
|
if "MSYSTEM" in os.environ:
|
2018-12-04 07:46:48 -05:00
|
|
|
monitor_args = ["winpty"] + monitor_args
|
2018-05-04 00:06:15 -04:00
|
|
|
_run_tool("idf_monitor", monitor_args, args.project_dir)
|
2018-02-15 23:32:08 -05:00
|
|
|
|
2018-01-23 01:08:28 -05:00
|
|
|
|
|
|
|
def clean(action, args):
|
|
|
|
if not os.path.isdir(args.build_dir):
|
|
|
|
print("Build directory '%s' not found. Nothing to clean." % args.build_dir)
|
|
|
|
return
|
|
|
|
build_target("clean", args)
|
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
|
2018-02-26 21:21:51 -05:00
|
|
|
def reconfigure(action, args):
|
|
|
|
_ensure_build_directory(args, True)
|
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
|
2019-04-30 03:45:25 -04:00
|
|
|
def _delete_windows_symlinks(directory):
|
|
|
|
"""
|
|
|
|
It deletes symlinks recursively on Windows. It is useful for Python 2 which doesn't detect symlinks on Windows.
|
|
|
|
"""
|
|
|
|
deleted_paths = []
|
|
|
|
if os.name == 'nt':
|
|
|
|
import ctypes
|
|
|
|
for root, dirnames, filenames in os.walk(directory):
|
|
|
|
for d in dirnames:
|
|
|
|
full_path = os.path.join(root, d)
|
|
|
|
try:
|
|
|
|
full_path = full_path.decode('utf-8')
|
|
|
|
except Exception:
|
|
|
|
pass
|
|
|
|
if ctypes.windll.kernel32.GetFileAttributesW(full_path) & 0x0400:
|
|
|
|
os.rmdir(full_path)
|
|
|
|
deleted_paths.append(full_path)
|
|
|
|
return deleted_paths
|
|
|
|
|
|
|
|
|
2018-01-23 01:08:28 -05:00
|
|
|
def fullclean(action, args):
|
|
|
|
build_dir = args.build_dir
|
|
|
|
if not os.path.isdir(build_dir):
|
|
|
|
print("Build directory '%s' not found. Nothing to clean." % build_dir)
|
|
|
|
return
|
|
|
|
if len(os.listdir(build_dir)) == 0:
|
|
|
|
print("Build directory '%s' is empty. Nothing to clean." % build_dir)
|
|
|
|
return
|
|
|
|
|
|
|
|
if not os.path.exists(os.path.join(build_dir, "CMakeCache.txt")):
|
2018-12-04 07:46:48 -05:00
|
|
|
raise FatalError("Directory '%s' doesn't seem to be a CMake build directory. Refusing to automatically "
|
|
|
|
"delete files in this directory. Delete the directory manually to 'clean' it." % build_dir)
|
|
|
|
red_flags = ["CMakeLists.txt", ".git", ".svn"]
|
2018-01-23 01:08:28 -05:00
|
|
|
for red in red_flags:
|
|
|
|
red = os.path.join(build_dir, red)
|
|
|
|
if os.path.exists(red):
|
2018-02-15 23:32:08 -05:00
|
|
|
raise FatalError("Refusing to automatically delete files in directory containing '%s'. Delete files manually if you're sure." % red)
|
2018-01-23 01:08:28 -05:00
|
|
|
# OK, delete everything in the build directory...
|
2019-04-30 03:45:25 -04:00
|
|
|
# Note: Python 2.7 doesn't detect symlinks on Windows (it is supported form 3.2). Tools promising to not
|
|
|
|
# follow symlinks will actually follow them. Deleting the build directory with symlinks deletes also items
|
|
|
|
# outside of this directory.
|
|
|
|
deleted_symlinks = _delete_windows_symlinks(build_dir)
|
|
|
|
if args.verbose and len(deleted_symlinks) > 1:
|
|
|
|
print('The following symlinks were identified and removed:\n%s' % "\n".join(deleted_symlinks))
|
2018-01-23 01:08:28 -05:00
|
|
|
for f in os.listdir(build_dir): # TODO: once we are Python 3 only, this can be os.scandir()
|
|
|
|
f = os.path.join(build_dir, f)
|
2019-04-30 03:45:25 -04:00
|
|
|
if args.verbose:
|
|
|
|
print('Removing: %s' % f)
|
2018-01-23 01:08:28 -05:00
|
|
|
if os.path.isdir(f):
|
|
|
|
shutil.rmtree(f)
|
|
|
|
else:
|
|
|
|
os.remove(f)
|
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
|
2019-03-01 00:12:03 -05:00
|
|
|
def _safe_relpath(path, start=None):
|
|
|
|
""" Return a relative path, same as os.path.relpath, but only if this is possible.
|
|
|
|
|
|
|
|
It is not possible on Windows, if the start directory and the path are on different drives.
|
|
|
|
"""
|
|
|
|
try:
|
|
|
|
return os.path.relpath(path, os.curdir if start is None else start)
|
|
|
|
except ValueError:
|
|
|
|
return os.path.abspath(path)
|
|
|
|
|
|
|
|
|
2018-05-24 02:05:31 -04:00
|
|
|
def print_closing_message(args):
|
|
|
|
# print a closing message of some kind
|
|
|
|
#
|
|
|
|
if "flash" in str(args.actions):
|
|
|
|
print("Done")
|
|
|
|
return
|
|
|
|
|
|
|
|
# Otherwise, if we built any binaries print a message about
|
|
|
|
# how to flash them
|
|
|
|
def print_flashing_message(title, key):
|
|
|
|
print("\n%s build complete. To flash, run this command:" % title)
|
|
|
|
|
|
|
|
with open(os.path.join(args.build_dir, "flasher_args.json")) as f:
|
|
|
|
flasher_args = json.load(f)
|
|
|
|
|
|
|
|
def flasher_path(f):
|
2019-03-01 00:12:03 -05:00
|
|
|
return _safe_relpath(os.path.join(args.build_dir, f))
|
2018-05-24 02:05:31 -04:00
|
|
|
|
2018-08-20 20:13:02 -04:00
|
|
|
if key != "project": # flashing a single item
|
2018-05-24 02:05:31 -04:00
|
|
|
cmd = ""
|
2018-08-20 20:13:02 -04:00
|
|
|
if key == "bootloader": # bootloader needs --flash-mode, etc to be passed in
|
2018-05-24 02:05:31 -04:00
|
|
|
cmd = " ".join(flasher_args["write_flash_args"]) + " "
|
|
|
|
|
|
|
|
cmd += flasher_args[key]["offset"] + " "
|
|
|
|
cmd += flasher_path(flasher_args[key]["file"])
|
2018-08-20 20:13:02 -04:00
|
|
|
else: # flashing the whole project
|
2018-05-24 02:05:31 -04:00
|
|
|
cmd = " ".join(flasher_args["write_flash_args"]) + " "
|
2018-08-31 00:29:38 -04:00
|
|
|
flash_items = sorted(((o,f) for (o,f) in flasher_args["flash_files"].items() if len(o) > 0),
|
2018-12-04 07:46:48 -05:00
|
|
|
key=lambda x: int(x[0], 0))
|
2018-08-20 20:13:02 -04:00
|
|
|
for o,f in flash_items:
|
2018-05-24 02:05:31 -04:00
|
|
|
cmd += o + " " + flasher_path(f) + " "
|
|
|
|
|
2018-10-19 15:02:55 -04:00
|
|
|
print("%s -p %s -b %s --after %s write_flash %s" % (
|
2019-03-01 00:12:03 -05:00
|
|
|
_safe_relpath("%s/components/esptool_py/esptool/esptool.py" % os.environ["IDF_PATH"]),
|
2018-05-24 02:05:31 -04:00
|
|
|
args.port or "(PORT)",
|
|
|
|
args.baud,
|
2018-10-19 15:02:55 -04:00
|
|
|
flasher_args["extra_esptool_args"]["after"],
|
2018-05-24 02:05:31 -04:00
|
|
|
cmd.strip()))
|
2018-08-20 20:13:02 -04:00
|
|
|
print("or run 'idf.py -p %s %s'" % (args.port or "(PORT)", key + "-flash" if key != "project" else "flash",))
|
2018-05-24 02:05:31 -04:00
|
|
|
|
|
|
|
if "all" in args.actions or "build" in args.actions:
|
|
|
|
print_flashing_message("Project", "project")
|
|
|
|
else:
|
|
|
|
if "app" in args.actions:
|
|
|
|
print_flashing_message("App", "app")
|
|
|
|
if "partition_table" in args.actions:
|
|
|
|
print_flashing_message("Partition Table", "partition_table")
|
|
|
|
if "bootloader" in args.actions:
|
|
|
|
print_flashing_message("Bootloader", "bootloader")
|
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
|
2018-01-23 01:08:28 -05:00
|
|
|
ACTIONS = {
|
|
|
|
# action name : ( function (or alias), dependencies, order-only dependencies )
|
2018-12-04 07:46:48 -05:00
|
|
|
"all": (build_target, [], ["reconfigure", "menuconfig", "clean", "fullclean"]),
|
|
|
|
"build": ("all", [], []), # build is same as 'all' target
|
|
|
|
"clean": (clean, [], ["fullclean"]),
|
|
|
|
"fullclean": (fullclean, [], []),
|
|
|
|
"reconfigure": (reconfigure, [], ["menuconfig"]),
|
|
|
|
"menuconfig": (build_target, [], []),
|
|
|
|
"defconfig": (build_target, [], []),
|
|
|
|
"confserver": (build_target, [], []),
|
|
|
|
"size": (build_target, ["app"], []),
|
|
|
|
"size-components": (build_target, ["app"], []),
|
|
|
|
"size-files": (build_target, ["app"], []),
|
|
|
|
"bootloader": (build_target, [], []),
|
|
|
|
"bootloader-clean": (build_target, [], []),
|
|
|
|
"bootloader-flash": (flash, ["bootloader"], ["erase_flash"]),
|
|
|
|
"app": (build_target, [], ["clean", "fullclean", "reconfigure"]),
|
|
|
|
"app-flash": (flash, ["app"], ["erase_flash"]),
|
2018-11-18 22:45:21 -05:00
|
|
|
"efuse_common_table": (build_target, [], ["reconfigure"]),
|
|
|
|
"efuse_custom_table": (build_target, [], ["reconfigure"]),
|
2018-12-06 04:49:02 -05:00
|
|
|
"show_efuse_table": (build_target, [], ["reconfigure"]),
|
2018-12-04 07:46:48 -05:00
|
|
|
"partition_table": (build_target, [], ["reconfigure"]),
|
|
|
|
"partition_table-flash": (flash, ["partition_table"], ["erase_flash"]),
|
|
|
|
"flash": (flash, ["all"], ["erase_flash"]),
|
|
|
|
"erase_flash": (erase_flash, [], []),
|
|
|
|
"monitor": (monitor, [], ["flash", "partition_table-flash", "bootloader-flash", "app-flash"]),
|
|
|
|
"erase_otadata": (build_target, [], []),
|
|
|
|
"read_otadata": (build_target, [], []),
|
2018-01-23 01:08:28 -05:00
|
|
|
}
|
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
|
2018-05-04 00:06:15 -04:00
|
|
|
def get_commandline_options():
|
|
|
|
""" Return all the command line options up to but not including the action """
|
|
|
|
result = []
|
|
|
|
for a in sys.argv:
|
|
|
|
if a in ACTIONS.keys():
|
|
|
|
break
|
|
|
|
else:
|
|
|
|
result.append(a)
|
|
|
|
return result
|
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
|
2018-09-06 08:32:49 -04:00
|
|
|
def get_default_serial_port():
|
|
|
|
""" Return a default serial port. esptool can do this (smarter), but it can create
|
|
|
|
inconsistencies where esptool.py uses one port and idf_monitor uses another.
|
|
|
|
|
|
|
|
Same logic as esptool.py search order, reverse sort by name and choose the first port.
|
|
|
|
"""
|
2018-10-25 02:16:30 -04:00
|
|
|
# Import is done here in order to move it after the check_environment() ensured that pyserial has been installed
|
|
|
|
import serial.tools.list_ports
|
|
|
|
|
2018-09-06 08:32:49 -04:00
|
|
|
ports = list(reversed(sorted(
|
2018-12-04 07:46:48 -05:00
|
|
|
p.device for p in serial.tools.list_ports.comports())))
|
2018-09-06 08:32:49 -04:00
|
|
|
try:
|
2019-05-17 02:25:53 -04:00
|
|
|
print("Choosing default port %s (use '-p PORT' option to set a specific serial port)" % ports[0].encode('ascii', 'ignore'))
|
2018-09-06 08:32:49 -04:00
|
|
|
return ports[0]
|
|
|
|
except IndexError:
|
|
|
|
raise RuntimeError("No serial ports found. Connect a device, or use '-p PORT' option to set a specific port.")
|
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
|
2018-08-26 22:48:16 -04:00
|
|
|
# Import the actions, arguments extension file
|
|
|
|
if os.path.exists(os.path.join(os.getcwd(), "idf_ext.py")):
|
|
|
|
sys.path.append(os.getcwd())
|
|
|
|
try:
|
|
|
|
from idf_ext import add_action_extensions, add_argument_extensions
|
2018-12-04 07:46:48 -05:00
|
|
|
except ImportError:
|
2018-08-26 22:48:16 -04:00
|
|
|
print("Error importing extension file idf_ext.py. Skipping.")
|
|
|
|
print("Please make sure that it contains implementations (even if they're empty implementations) of")
|
|
|
|
print("add_action_extensions and add_argument_extensions.")
|
2018-09-06 08:32:49 -04:00
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
|
2018-01-23 01:08:28 -05:00
|
|
|
def main():
|
2018-05-11 05:20:27 -04:00
|
|
|
if sys.version_info[0] != 2 or sys.version_info[1] != 7:
|
2018-09-18 08:43:23 -04:00
|
|
|
print("Note: You are using Python %d.%d.%d. Python 3 support is new, please report any problems "
|
|
|
|
"you encounter. Search for 'Setting the Python Interpreter' in the ESP-IDF docs if you want to use "
|
|
|
|
"Python 2.7." % sys.version_info[:3])
|
2018-05-11 05:20:27 -04:00
|
|
|
|
2018-08-26 22:48:16 -04:00
|
|
|
# Add actions extensions
|
|
|
|
try:
|
|
|
|
add_action_extensions({
|
2018-12-04 07:46:48 -05:00
|
|
|
"build_target": build_target,
|
|
|
|
"reconfigure": reconfigure,
|
|
|
|
"flash": flash,
|
|
|
|
"monitor": monitor,
|
|
|
|
"clean": clean,
|
|
|
|
"fullclean": fullclean
|
|
|
|
}, ACTIONS)
|
2018-08-26 22:48:16 -04:00
|
|
|
except NameError:
|
|
|
|
pass
|
|
|
|
|
2018-01-23 01:08:28 -05:00
|
|
|
parser = argparse.ArgumentParser(description='ESP-IDF build management tool')
|
2018-05-06 21:31:37 -04:00
|
|
|
parser.add_argument('-p', '--port', help="Serial port",
|
|
|
|
default=os.environ.get('ESPPORT', None))
|
|
|
|
parser.add_argument('-b', '--baud', help="Baud rate",
|
|
|
|
default=os.environ.get('ESPBAUD', 460800))
|
2018-01-23 01:08:28 -05:00
|
|
|
parser.add_argument('-C', '--project-dir', help="Project directory", default=os.getcwd())
|
|
|
|
parser.add_argument('-B', '--build-dir', help="Build directory", default=None)
|
|
|
|
parser.add_argument('-G', '--generator', help="Cmake generator", choices=GENERATOR_CMDS.keys())
|
2018-03-15 20:58:52 -04:00
|
|
|
parser.add_argument('-n', '--no-warnings', help="Disable Cmake warnings", action="store_true")
|
|
|
|
parser.add_argument('-v', '--verbose', help="Verbose build output", action="store_true")
|
2018-08-26 22:48:16 -04:00
|
|
|
parser.add_argument('-D', '--define-cache-entry', help="Create a cmake cache entry", nargs='+')
|
2018-12-04 07:46:48 -05:00
|
|
|
parser.add_argument('--no-ccache', help="Disable ccache. Otherwise, if ccache is available on the PATH then it will be used for faster builds.",
|
|
|
|
action="store_true")
|
2018-01-23 01:08:28 -05:00
|
|
|
parser.add_argument('actions', help="Actions (build targets or other operations)", nargs='+',
|
|
|
|
choices=ACTIONS.keys())
|
|
|
|
|
2018-08-26 22:48:16 -04:00
|
|
|
# Add arguments extensions
|
|
|
|
try:
|
|
|
|
add_argument_extensions(parser)
|
|
|
|
except NameError:
|
|
|
|
pass
|
|
|
|
|
2018-01-23 01:08:28 -05:00
|
|
|
args = parser.parse_args()
|
|
|
|
|
|
|
|
check_environment()
|
|
|
|
|
|
|
|
# Advanced parameter checks
|
|
|
|
if args.build_dir is not None and os.path.realpath(args.project_dir) == os.path.realpath(args.build_dir):
|
2018-12-04 07:46:48 -05:00
|
|
|
raise FatalError("Setting the build directory to the project directory is not supported. Suggest dropping "
|
|
|
|
"--build-dir option, the default is a 'build' subdirectory inside the project directory.")
|
2018-01-23 01:08:28 -05:00
|
|
|
if args.build_dir is None:
|
|
|
|
args.build_dir = os.path.join(args.project_dir, "build")
|
|
|
|
args.build_dir = os.path.realpath(args.build_dir)
|
|
|
|
|
|
|
|
completed_actions = set()
|
2018-12-04 07:46:48 -05:00
|
|
|
|
2018-01-23 01:08:28 -05:00
|
|
|
def execute_action(action, remaining_actions):
|
2018-12-04 07:46:48 -05:00
|
|
|
(function, dependencies, order_dependencies) = ACTIONS[action]
|
2018-01-23 01:08:28 -05:00
|
|
|
# very simple dependency management, build a set of completed actions and make sure
|
|
|
|
# all dependencies are in it
|
|
|
|
for dep in dependencies:
|
2018-12-04 07:46:48 -05:00
|
|
|
if dep not in completed_actions:
|
2018-01-23 01:08:28 -05:00
|
|
|
execute_action(dep, remaining_actions)
|
|
|
|
for dep in order_dependencies:
|
2018-12-04 07:46:48 -05:00
|
|
|
if dep in remaining_actions and dep not in completed_actions:
|
2018-01-23 01:08:28 -05:00
|
|
|
execute_action(dep, remaining_actions)
|
|
|
|
|
|
|
|
if action in completed_actions:
|
|
|
|
pass # we've already done this, don't do it twice...
|
|
|
|
elif function in ACTIONS: # alias of another action
|
|
|
|
execute_action(function, remaining_actions)
|
|
|
|
else:
|
|
|
|
function(action, args)
|
|
|
|
|
|
|
|
completed_actions.add(action)
|
|
|
|
|
2018-05-24 02:05:31 -04:00
|
|
|
actions = list(args.actions)
|
|
|
|
while len(actions) > 0:
|
|
|
|
execute_action(actions[0], actions[1:])
|
|
|
|
actions.pop(0)
|
2018-01-23 01:08:28 -05:00
|
|
|
|
2018-05-24 02:05:31 -04:00
|
|
|
print_closing_message(args)
|
2018-01-23 01:08:28 -05:00
|
|
|
|
2018-12-04 07:46:48 -05:00
|
|
|
|
2018-01-23 01:08:28 -05:00
|
|
|
if __name__ == "__main__":
|
2018-02-15 23:32:08 -05:00
|
|
|
try:
|
2019-02-22 10:07:26 -05:00
|
|
|
# On MSYS2 we need to run idf.py with "winpty" in order to be able to cancel the subprocesses properly on
|
|
|
|
# keyboard interrupt (CTRL+C).
|
|
|
|
# Using an own global variable for indicating that we are running with "winpty" seems to be the most suitable
|
|
|
|
# option as os.environment['_'] contains "winpty" only when it is run manually from console.
|
|
|
|
WINPTY_VAR = 'WINPTY'
|
|
|
|
WINPTY_EXE = 'winpty'
|
|
|
|
if ('MSYSTEM' in os.environ) and (not os.environ['_'].endswith(WINPTY_EXE) and WINPTY_VAR not in os.environ):
|
|
|
|
os.environ[WINPTY_VAR] = '1' # the value is of no interest to us
|
|
|
|
# idf.py calls itself with "winpty" and WINPTY global variable set
|
|
|
|
ret = subprocess.call([WINPTY_EXE, sys.executable] + sys.argv, env=os.environ)
|
|
|
|
if ret:
|
|
|
|
raise SystemExit(ret)
|
|
|
|
else:
|
|
|
|
main()
|
2018-02-15 23:32:08 -05:00
|
|
|
except FatalError as e:
|
|
|
|
print(e)
|
|
|
|
sys.exit(2)
|