Nicolas Borla
/
BBR_1Ebene
BBR 1 Ebene
Diff: mbed-os/tools/build_api.py
- Revision:
- 0:fbdae7e6d805
--- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/mbed-os/tools/build_api.py Mon May 14 11:29:06 2018 +0000 @@ -0,0 +1,1409 @@ +""" +mbed SDK +Copyright (c) 2011-2016 ARM Limited + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +""" +from __future__ import print_function, division, absolute_import + +import re +import tempfile +import datetime +import uuid +import struct +import zlib +import hashlib +from shutil import rmtree +from os.path import join, exists, dirname, basename, abspath, normpath, splitext +from os.path import relpath +from os import linesep, remove, makedirs +from time import time +from intelhex import IntelHex +from json import load, dump +from jinja2 import FileSystemLoader +from jinja2.environment import Environment + +from .arm_pack_manager import Cache +from .utils import (mkdir, run_cmd, run_cmd_ext, NotSupportedException, + ToolException, InvalidReleaseTargetException, + intelhex_offset, integer) +from .paths import (MBED_CMSIS_PATH, MBED_TARGETS_PATH, MBED_LIBRARIES, + MBED_HEADER, MBED_DRIVERS, MBED_PLATFORM, MBED_HAL, + MBED_CONFIG_FILE, MBED_LIBRARIES_DRIVERS, + MBED_LIBRARIES_PLATFORM, MBED_LIBRARIES_HAL, + BUILD_DIR) +from .targets import TARGET_NAMES, TARGET_MAP +from .libraries import Library +from .toolchains import TOOLCHAIN_CLASSES +from .config import Config + +RELEASE_VERSIONS = ['2', '5'] + +def prep_report(report, target_name, toolchain_name, id_name): + """Setup report keys + + Positional arguments: + report - the report to fill + target_name - the target being used + toolchain_name - the toolchain being used + id_name - the name of the executable or library being built + """ + if not target_name in report: + report[target_name] = {} + + if not toolchain_name in report[target_name]: + report[target_name][toolchain_name] = {} + + if not id_name in report[target_name][toolchain_name]: + report[target_name][toolchain_name][id_name] = [] + +def prep_properties(properties, target_name, toolchain_name, vendor_label): + """Setup test properties + + Positional arguments: + properties - the dict to fill + target_name - the target the test is targeting + toolchain_name - the toolchain that will compile the test + vendor_label - the vendor + """ + if not target_name in properties: + properties[target_name] = {} + + if not toolchain_name in properties[target_name]: + properties[target_name][toolchain_name] = {} + + properties[target_name][toolchain_name]["target"] = target_name + properties[target_name][toolchain_name]["vendor"] = vendor_label + properties[target_name][toolchain_name]["toolchain"] = toolchain_name + +def create_result(target_name, toolchain_name, id_name, description): + """Create a result dictionary + + Positional arguments: + target_name - the target being built for + toolchain_name - the toolchain doing the building + id_name - the name of the executable or library being built + description - a human readable description of what's going on + """ + cur_result = {} + cur_result["target_name"] = target_name + cur_result["toolchain_name"] = toolchain_name + cur_result["id"] = id_name + cur_result["description"] = description + cur_result["elapsed_time"] = 0 + cur_result["output"] = "" + + return cur_result + +def add_result_to_report(report, result): + """Add a single result to a report dictionary + + Positional arguments: + report - the report to append to + result - the result to append + """ + result["date"] = datetime.datetime.utcnow().isoformat() + result["uuid"] = str(uuid.uuid1()) + target = result["target_name"] + toolchain = result["toolchain_name"] + id_name = result['id'] + result_wrap = {0: result} + report[target][toolchain][id_name].append(result_wrap) + +def get_config(src_paths, target, toolchain_name, app_config=None): + """Get the configuration object for a target-toolchain combination + + Positional arguments: + src_paths - paths to scan for the configuration files + target - the device we are building for + toolchain_name - the string that identifies the build tools + """ + # Convert src_paths to a list if needed + if not isinstance(src_paths, list): + src_paths = [src_paths] + + # Pass all params to the unified prepare_resources() + toolchain = prepare_toolchain(src_paths, None, target, toolchain_name, + app_config=app_config) + + # Scan src_path for config files + resources = toolchain.scan_resources(src_paths[0]) + for path in src_paths[1:]: + resources.add(toolchain.scan_resources(path)) + + # Update configuration files until added features creates no changes + prev_features = set() + while True: + # Update the configuration with any .json files found while scanning + toolchain.config.add_config_files(resources.json_files) + + # Add features while we find new ones + features = set(toolchain.config.get_features()) + if features == prev_features: + break + + for feature in features: + if feature in resources.features: + resources += resources.features[feature] + + prev_features = features + toolchain.config.validate_config() + if toolchain.config.has_regions: + _ = list(toolchain.config.regions) + + cfg, macros = toolchain.config.get_config_data() + features = toolchain.config.get_features() + return cfg, macros, features + +def is_official_target(target_name, version): + """ Returns True, None if a target is part of the official release for the + given version. Return False, 'reason' if a target is not part of the + official release for the given version. + + Positional arguments: + target_name - Name if the target (ex. 'K64F') + version - The release version string. Should be a string contained within + RELEASE_VERSIONS + """ + + result = True + reason = None + target = TARGET_MAP[target_name] + + if hasattr(target, 'release_versions') \ + and version in target.release_versions: + if version == '2': + # For version 2, either ARM or uARM toolchain support is required + required_toolchains = set(['ARM', 'uARM']) + + if not len(required_toolchains.intersection( + set(target.supported_toolchains))) > 0: + result = False + reason = ("Target '%s' must support " % target.name) + \ + ("one of the folowing toolchains to be included in the") + \ + ((" mbed 2.0 official release: %s" + linesep) % + ", ".join(required_toolchains)) + \ + ("Currently it is only configured to support the ") + \ + ("following toolchains: %s" % + ", ".join(target.supported_toolchains)) + + elif version == '5': + # For version 5, ARM, GCC_ARM, and IAR toolchain support is required + required_toolchains = set(['ARM', 'GCC_ARM', 'IAR']) + required_toolchains_sorted = list(required_toolchains) + required_toolchains_sorted.sort() + supported_toolchains = set(target.supported_toolchains) + supported_toolchains_sorted = list(supported_toolchains) + supported_toolchains_sorted.sort() + + if not required_toolchains.issubset(supported_toolchains): + result = False + reason = ("Target '%s' must support " % target.name) + \ + ("ALL of the folowing toolchains to be included in the") + \ + ((" mbed OS 5.0 official release: %s" + linesep) % + ", ".join(required_toolchains_sorted)) + \ + ("Currently it is only configured to support the ") + \ + ("following toolchains: %s" % + ", ".join(supported_toolchains_sorted)) + + elif not target.default_lib == 'std': + result = False + reason = ("Target '%s' must set the " % target.name) + \ + ("'default_lib' to 'std' to be included in the ") + \ + ("mbed OS 5.0 official release." + linesep) + \ + ("Currently it is set to '%s'" % target.default_lib) + + else: + result = False + reason = ("Target '%s' has set an invalid release version of '%s'" % + version) + \ + ("Please choose from the following release versions: %s" % + ', '.join(RELEASE_VERSIONS)) + + else: + result = False + if not hasattr(target, 'release_versions'): + reason = "Target '%s' " % target.name + reason += "does not have the 'release_versions' key set" + elif not version in target.release_versions: + reason = "Target '%s' does not contain the version '%s' " % \ + (target.name, version) + reason += "in its 'release_versions' key" + + return result, reason + +def transform_release_toolchains(toolchains, version): + """ Given a list of toolchains and a release version, return a list of + only the supported toolchains for that release + + Positional arguments: + toolchains - The list of toolchains + version - The release version string. Should be a string contained within + RELEASE_VERSIONS + """ + if version == '5': + return ['ARM', 'GCC_ARM', 'IAR'] + else: + return toolchains + + +def get_mbed_official_release(version): + """ Given a release version string, return a tuple that contains a target + and the supported toolchains for that release. + Ex. Given '2', return (('LPC1768', ('ARM', 'GCC_ARM')), + ('K64F', ('ARM', 'GCC_ARM')), ...) + + Positional arguments: + version - The version string. Should be a string contained within + RELEASE_VERSIONS + """ + + mbed_official_release = ( + tuple( + tuple( + [ + TARGET_MAP[target].name, + tuple(transform_release_toolchains( + TARGET_MAP[target].supported_toolchains, version)) + ] + ) for target in TARGET_NAMES \ + if (hasattr(TARGET_MAP[target], 'release_versions') + and version in TARGET_MAP[target].release_versions) + ) + ) + + for target in mbed_official_release: + is_official, reason = is_official_target(target[0], version) + + if not is_official: + raise InvalidReleaseTargetException(reason) + + return mbed_official_release + +ARM_COMPILERS = ("ARM", "ARMC6", "uARM") +def target_supports_toolchain(target, toolchain_name): + if toolchain_name in ARM_COMPILERS: + return any(tc in target.supported_toolchains for tc in ARM_COMPILERS) + else: + return toolchain_name in target.supported_toolchains + + +def prepare_toolchain(src_paths, build_dir, target, toolchain_name, + macros=None, clean=False, jobs=1, + notify=None, config=None, app_config=None, + build_profile=None): + """ Prepares resource related objects - toolchain, target, config + + Positional arguments: + src_paths - the paths to source directories + target - ['LPC1768', 'LPC11U24', etc.] + toolchain_name - ['ARM', 'uARM', 'GCC_ARM', 'GCC_CR'] + + Keyword arguments: + macros - additional macros + clean - Rebuild everything if True + jobs - how many compilers we can run at once + notify - Notify function for logs + config - a Config object to use instead of creating one + app_config - location of a chosen mbed_app.json file + build_profile - a list of mergeable build profiles + """ + + # We need to remove all paths which are repeated to avoid + # multiple compilations and linking with the same objects + src_paths = [src_paths[0]] + list(set(src_paths[1:])) + + # If the configuration object was not yet created, create it now + config = config or Config(target, src_paths, app_config=app_config) + target = config.target + if not target_supports_toolchain(target, toolchain_name): + raise NotSupportedException( + "Target {} is not supported by toolchain {}".format( + target.name, toolchain_name)) + + try: + cur_tc = TOOLCHAIN_CLASSES[toolchain_name] + except KeyError: + raise KeyError("Toolchain %s not supported" % toolchain_name) + + profile = {'c': [], 'cxx': [], 'common': [], 'asm': [], 'ld': []} + for contents in build_profile or []: + for key in profile: + profile[key].extend(contents[toolchain_name].get(key, [])) + + toolchain = cur_tc( + target, notify, macros, build_dir=build_dir, build_profile=profile) + + toolchain.config = config + toolchain.jobs = jobs + toolchain.build_all = clean + + return toolchain + +def _printihex(ihex): + import pprint + pprint.PrettyPrinter().pprint(ihex.todict()) + +def _real_region_size(region): + try: + part = intelhex_offset(region.filename, offset=region.start) + return (part.maxaddr() - part.minaddr()) + 1 + except AttributeError: + return region.size + + +def _fill_header(region_list, current_region): + """Fill an application header region + + This is done it three steps: + * Fill the whole region with zeros + * Fill const, timestamp and size entries with their data + * Fill the digests using this header as the header region + """ + region_dict = {r.name: r for r in region_list} + header = IntelHex() + header.puts(current_region.start, b'\x00' * current_region.size) + start = current_region.start + for member in current_region.filename: + _, type, subtype, data = member + member_size = Config.header_member_size(member) + if type == "const": + fmt = { + "8le": ">B", "16le": "<H", "32le": "<L", "64le": "<Q", + "8be": "<B", "16be": ">H", "32be": ">L", "64be": ">Q" + }[subtype] + header.puts(start, struct.pack(fmt, integer(data, 0))) + elif type == "timestamp": + fmt = {"32le": "<L", "64le": "<Q", + "32be": ">L", "64be": ">Q"}[subtype] + header.puts(start, struct.pack(fmt, time())) + elif type == "size": + fmt = {"32le": "<L", "64le": "<Q", + "32be": ">L", "64be": ">Q"}[subtype] + size = sum(_real_region_size(region_dict[r]) for r in data) + header.puts(start, struct.pack(fmt, size)) + elif type == "digest": + if data == "header": + ih = header[:start] + else: + ih = intelhex_offset(region_dict[data].filename, offset=region_dict[data].start) + if subtype.startswith("CRCITT32"): + fmt = {"CRCITT32be": ">l", "CRCITT32le": "<l"}[subtype] + header.puts(start, struct.pack(fmt, zlib.crc32(ih.tobinarray()))) + elif subtype.startswith("SHA"): + if subtype == "SHA256": + hash = hashlib.sha256() + elif subtype == "SHA512": + hash = hashlib.sha512() + hash.update(ih.tobinarray()) + header.puts(start, hash.digest()) + start += Config.header_member_size(member) + return header + +def merge_region_list(region_list, destination, notify, padding=b'\xFF'): + """Merge the region_list into a single image + + Positional Arguments: + region_list - list of regions, which should contain filenames + destination - file name to write all regions to + padding - bytes to fill gapps with + """ + merged = IntelHex() + _, format = splitext(destination) + + notify.info("Merging Regions") + + for region in region_list: + if region.active and not region.filename: + raise ToolException("Active region has no contents: No file found.") + if isinstance(region.filename, list): + header_basename, _ = splitext(destination) + header_filename = header_basename + "_header.hex" + _fill_header(region_list, region).tofile(header_filename, format='hex') + region = region._replace(filename=header_filename) + if region.filename: + notify.info(" Filling region %s with %s" % (region.name, region.filename)) + part = intelhex_offset(region.filename, offset=region.start) + part_size = (part.maxaddr() - part.minaddr()) + 1 + if part_size > region.size: + raise ToolException("Contents of region %s does not fit" + % region.name) + merged.merge(part) + pad_size = region.size - part_size + if pad_size > 0 and region != region_list[-1]: + notify.info(" Padding region %s with 0x%x bytes" % + (region.name, pad_size)) + if format is ".hex": + """The offset will be in the hex file generated when we're done, + so we can skip padding here""" + else: + merged.puts(merged.maxaddr() + 1, padding * pad_size) + + if not exists(dirname(destination)): + makedirs(dirname(destination)) + notify.info("Space used after regions merged: 0x%x" % + (merged.maxaddr() - merged.minaddr() + 1)) + with open(destination, "wb+") as output: + merged.tofile(output, format=format.strip(".")) + +def scan_resources(src_paths, toolchain, dependencies_paths=None, + inc_dirs=None, base_path=None, collect_ignores=False): + """ Scan resources using initialized toolcain + + Positional arguments + src_paths - the paths to source directories + toolchain - valid toolchain object + dependencies_paths - dependency paths that we should scan for include dirs + inc_dirs - additional include directories which should be added to + the scanner resources + """ + + # Scan src_path + resources = toolchain.scan_resources(src_paths[0], base_path=base_path, + collect_ignores=collect_ignores) + for path in src_paths[1:]: + resources.add(toolchain.scan_resources(path, base_path=base_path, + collect_ignores=collect_ignores)) + + # Scan dependency paths for include dirs + if dependencies_paths is not None: + for path in dependencies_paths: + lib_resources = toolchain.scan_resources(path) + resources.inc_dirs.extend(lib_resources.inc_dirs) + + # Add additional include directories if passed + if inc_dirs: + if isinstance(inc_dirs, list): + resources.inc_dirs.extend(inc_dirs) + else: + resources.inc_dirs.append(inc_dirs) + + # Load resources into the config system which might expand/modify resources + # based on config data + resources = toolchain.config.load_resources(resources) + + # Set the toolchain's configuration data + toolchain.set_config_data(toolchain.config.get_config_data()) + + return resources + +def build_project(src_paths, build_path, target, toolchain_name, + libraries_paths=None, linker_script=None, clean=False, + notify=None, name=None, macros=None, inc_dirs=None, jobs=1, + report=None, properties=None, project_id=None, + project_description=None, config=None, + app_config=None, build_profile=None, stats_depth=None): + """ Build a project. A project may be a test or a user program. + + Positional arguments: + src_paths - a path or list of paths that contain all files needed to build + the project + build_path - the directory where all of the object files will be placed + target - the MCU or board that the project will compile for + toolchain_name - the name of the build tools + + Keyword arguments: + libraries_paths - The location of libraries to include when linking + linker_script - the file that drives the linker to do it's job + clean - Rebuild everything if True + notify - Notify function for logs + name - the name of the project + macros - additional macros + inc_dirs - additional directories where include files may be found + jobs - how many compilers we can run at once + report - a dict where a result may be appended + properties - UUUUHHHHH beats me + project_id - the name put in the report + project_description - the human-readable version of what this thing does + config - a Config object to use instead of creating one + app_config - location of a chosen mbed_app.json file + build_profile - a dict of flags that will be passed to the compiler + stats_depth - depth level for memap to display file/dirs + """ + + # Convert src_path to a list if needed + if not isinstance(src_paths, list): + src_paths = [src_paths] + # Extend src_paths wiht libraries_paths + if libraries_paths is not None: + src_paths.extend(libraries_paths) + inc_dirs.extend(map(dirname, libraries_paths)) + + if clean and exists(build_path): + rmtree(build_path) + mkdir(build_path) + + toolchain = prepare_toolchain( + src_paths, build_path, target, toolchain_name, macros=macros, + clean=clean, jobs=jobs, notify=notify, config=config, + app_config=app_config, build_profile=build_profile) + + # The first path will give the name to the library + name = (name or toolchain.config.name or + basename(normpath(abspath(src_paths[0])))) + notify.info("Building project %s (%s, %s)" % + (name, toolchain.target.name, toolchain_name)) + + # Initialize reporting + if report != None: + start = time() + # If project_id is specified, use that over the default name + id_name = project_id.upper() if project_id else name.upper() + description = project_description if project_description else name + vendor_label = toolchain.target.extra_labels[0] + prep_report(report, toolchain.target.name, toolchain_name, id_name) + cur_result = create_result(toolchain.target.name, toolchain_name, + id_name, description) + if properties != None: + prep_properties(properties, toolchain.target.name, toolchain_name, + vendor_label) + + try: + # Call unified scan_resources + resources = scan_resources(src_paths, toolchain, inc_dirs=inc_dirs) + + # Change linker script if specified + if linker_script is not None: + resources.linker_script = linker_script + + # Compile Sources + objects = toolchain.compile_sources(resources, resources.inc_dirs) + resources.objects.extend(objects) + + # Link Program + if toolchain.config.has_regions: + res, _ = toolchain.link_program(resources, build_path, name + "_application") + region_list = list(toolchain.config.regions) + region_list = [r._replace(filename=res) if r.active else r + for r in region_list] + res = "%s.%s" % (join(build_path, name), + getattr(toolchain.target, "OUTPUT_EXT", "bin")) + merge_region_list(region_list, res, notify) + else: + res, _ = toolchain.link_program(resources, build_path, name) + + memap_instance = getattr(toolchain, 'memap_instance', None) + memap_table = '' + if memap_instance: + # Write output to stdout in text (pretty table) format + memap_table = memap_instance.generate_output('table', stats_depth) + notify.info(memap_table) + + # Write output to file in JSON format + map_out = join(build_path, name + "_map.json") + memap_instance.generate_output('json', stats_depth, map_out) + + # Write output to file in CSV format for the CI + map_csv = join(build_path, name + "_map.csv") + memap_instance.generate_output('csv-ci', stats_depth, map_csv) + + resources.detect_duplicates(toolchain) + + if report != None: + end = time() + cur_result["elapsed_time"] = end - start + cur_result["result"] = "OK" + cur_result["memory_usage"] = (memap_instance.mem_report + if memap_instance is not None else None) + cur_result["bin"] = res + cur_result["elf"] = splitext(res)[0] + ".elf" + cur_result.update(toolchain.report) + + add_result_to_report(report, cur_result) + + return res + + except Exception as exc: + if report != None: + end = time() + + if isinstance(exc, NotSupportedException): + cur_result["result"] = "NOT_SUPPORTED" + else: + cur_result["result"] = "FAIL" + + cur_result["elapsed_time"] = end - start + + add_result_to_report(report, cur_result) + # Let Exception propagate + raise + +def build_library(src_paths, build_path, target, toolchain_name, + dependencies_paths=None, name=None, clean=False, + archive=True, notify=None, macros=None, inc_dirs=None, jobs=1, + report=None, properties=None, project_id=None, + remove_config_header_file=False, app_config=None, + build_profile=None): + """ Build a library + + Positional arguments: + src_paths - a path or list of paths that contain all files needed to build + the library + build_path - the directory where all of the object files will be placed + target - the MCU or board that the project will compile for + toolchain_name - the name of the build tools + + Keyword arguments: + dependencies_paths - The location of libraries to include when linking + name - the name of the library + clean - Rebuild everything if True + archive - whether the library will create an archive file + notify - Notify function for logs + macros - additional macros + inc_dirs - additional directories where include files may be found + jobs - how many compilers we can run at once + report - a dict where a result may be appended + properties - UUUUHHHHH beats me + project_id - the name that goes in the report + remove_config_header_file - delete config header file when done building + app_config - location of a chosen mbed_app.json file + build_profile - a dict of flags that will be passed to the compiler + """ + + # Convert src_path to a list if needed + if not isinstance(src_paths, list): + src_paths = [src_paths] + + # Build path + if archive: + # Use temp path when building archive + tmp_path = join(build_path, '.temp') + mkdir(tmp_path) + else: + tmp_path = build_path + + # Clean the build directory + if clean and exists(tmp_path): + rmtree(tmp_path) + mkdir(tmp_path) + + # Pass all params to the unified prepare_toolchain() + toolchain = prepare_toolchain( + src_paths, build_path, target, toolchain_name, macros=macros, + clean=clean, jobs=jobs, notify=notify, app_config=app_config, + build_profile=build_profile) + + # The first path will give the name to the library + if name is None: + name = basename(normpath(abspath(src_paths[0]))) + notify.info("Building library %s (%s, %s)" % + (name, toolchain.target.name, toolchain_name)) + + # Initialize reporting + if report != None: + start = time() + # If project_id is specified, use that over the default name + id_name = project_id.upper() if project_id else name.upper() + description = name + vendor_label = toolchain.target.extra_labels[0] + prep_report(report, toolchain.target.name, toolchain_name, id_name) + cur_result = create_result(toolchain.target.name, toolchain_name, + id_name, description) + cur_result['type'] = 'library' + if properties != None: + prep_properties(properties, toolchain.target.name, toolchain_name, + vendor_label) + + for src_path in src_paths: + if not exists(src_path): + error_msg = "The library source folder does not exist: %s", src_path + if report != None: + cur_result["output"] = error_msg + cur_result["result"] = "FAIL" + add_result_to_report(report, cur_result) + raise Exception(error_msg) + + try: + # Call unified scan_resources + resources = scan_resources(src_paths, toolchain, + dependencies_paths=dependencies_paths, + inc_dirs=inc_dirs) + + + # Copy headers, objects and static libraries - all files needed for + # static lib + toolchain.copy_files(resources.headers, build_path, resources=resources) + toolchain.copy_files(resources.objects, build_path, resources=resources) + toolchain.copy_files(resources.libraries, build_path, + resources=resources) + toolchain.copy_files(resources.json_files, build_path, + resources=resources) + if resources.linker_script: + toolchain.copy_files(resources.linker_script, build_path, + resources=resources) + + if resources.hex_files: + toolchain.copy_files(resources.hex_files, build_path, + resources=resources) + + # Compile Sources + objects = toolchain.compile_sources(resources, resources.inc_dirs) + resources.objects.extend(objects) + + if archive: + toolchain.build_library(objects, build_path, name) + + if remove_config_header_file: + config_header_path = toolchain.get_config_header() + if config_header_path: + remove(config_header_path) + + if report != None: + end = time() + cur_result["elapsed_time"] = end - start + cur_result["result"] = "OK" + + + add_result_to_report(report, cur_result) + return True + + except Exception as exc: + if report != None: + end = time() + + if isinstance(exc, ToolException): + cur_result["result"] = "FAIL" + elif isinstance(exc, NotSupportedException): + cur_result["result"] = "NOT_SUPPORTED" + + cur_result["elapsed_time"] = end - start + + add_result_to_report(report, cur_result) + + # Let Exception propagate + raise + +###################### +### Legacy methods ### +###################### + +def mbed2_obj_path(target_name, toolchain_name): + real_tc_name = TOOLCHAIN_CLASSES[toolchain_name].__name__ + return join("TARGET_" + target_name, "TOOLCHAIN_" + real_tc_name) + +def build_lib(lib_id, target, toolchain_name, clean=False, macros=None, + notify=None, jobs=1, report=None, properties=None, + build_profile=None): + """ Legacy method for building mbed libraries + + Positional arguments: + lib_id - the library's unique identifier + target - the MCU or board that the project will compile for + toolchain_name - the name of the build tools + + Keyword arguments: + clean - Rebuild everything if True + macros - additional macros + notify - Notify function for logs + jobs - how many compilers we can run at once + report - a dict where a result may be appended + properties - UUUUHHHHH beats me + build_profile - a dict of flags that will be passed to the compiler + """ + lib = Library(lib_id) + if not lib.is_supported(target, toolchain_name): + print('Library "%s" is not yet supported on target %s with toolchain %s' + % (lib_id, target.name, toolchain_name)) + return False + + # We need to combine macros from parameter list with macros from library + # definition + lib_macros = lib.macros if lib.macros else [] + if macros: + macros.extend(lib_macros) + else: + macros = lib_macros + + src_paths = lib.source_dir + build_path = lib.build_dir + dependencies_paths = lib.dependencies + inc_dirs = lib.inc_dirs + inc_dirs_ext = lib.inc_dirs_ext + + if not isinstance(src_paths, list): + src_paths = [src_paths] + + # The first path will give the name to the library + name = basename(src_paths[0]) + + if report != None: + start = time() + id_name = name.upper() + description = name + vendor_label = target.extra_labels[0] + cur_result = None + prep_report(report, target.name, toolchain_name, id_name) + cur_result = create_result(target.name, toolchain_name, id_name, + description) + + if properties != None: + prep_properties(properties, target.name, toolchain_name, + vendor_label) + + for src_path in src_paths: + if not exists(src_path): + error_msg = "The library source folder does not exist: %s", src_path + + if report != None: + cur_result["output"] = error_msg + cur_result["result"] = "FAIL" + add_result_to_report(report, cur_result) + + raise Exception(error_msg) + + try: + # Toolchain instance + # Create the desired build directory structure + bin_path = join(build_path, mbed2_obj_path(target.name, toolchain_name)) + mkdir(bin_path) + tmp_path = join(build_path, '.temp', mbed2_obj_path(target.name, + toolchain_name)) + mkdir(tmp_path) + + toolchain = prepare_toolchain( + src_paths, tmp_path, target, toolchain_name, macros=macros, + notify=notify, build_profile=build_profile, jobs=jobs, clean=clean) + + notify.info("Building library %s (%s, %s)" % + (name.upper(), target.name, toolchain_name)) + + # Take into account the library configuration (MBED_CONFIG_FILE) + config = toolchain.config + config.add_config_files([MBED_CONFIG_FILE]) + + # Scan Resources + resources = [] + for src_path in src_paths: + resources.append(toolchain.scan_resources(src_path)) + + # Add extra include directories / files which are required by library + # This files usually are not in the same directory as source files so + # previous scan will not include them + if inc_dirs_ext is not None: + for inc_ext in inc_dirs_ext: + resources.append(toolchain.scan_resources(inc_ext)) + + # Dependencies Include Paths + dependencies_include_dir = [] + if dependencies_paths is not None: + for path in dependencies_paths: + lib_resources = toolchain.scan_resources(path) + dependencies_include_dir.extend(lib_resources.inc_dirs) + dependencies_include_dir.extend(map(dirname, lib_resources.inc_dirs)) + + if inc_dirs: + dependencies_include_dir.extend(inc_dirs) + + # Add other discovered configuration data to the configuration object + for res in resources: + config.load_resources(res) + toolchain.set_config_data(toolchain.config.get_config_data()) + + + # Copy Headers + for resource in resources: + toolchain.copy_files(resource.headers, build_path, + resources=resource) + + dependencies_include_dir.extend( + toolchain.scan_resources(build_path).inc_dirs) + + # Compile Sources + objects = [] + for resource in resources: + objects.extend(toolchain.compile_sources(resource, dependencies_include_dir)) + + needed_update = toolchain.build_library(objects, bin_path, name) + + if report != None and needed_update: + end = time() + cur_result["elapsed_time"] = end - start + cur_result["result"] = "OK" + + add_result_to_report(report, cur_result) + return True + + except Exception: + if report != None: + end = time() + cur_result["result"] = "FAIL" + cur_result["elapsed_time"] = end - start + + add_result_to_report(report, cur_result) + + # Let Exception propagate + raise + +# We do have unique legacy conventions about how we build and package the mbed +# library +def build_mbed_libs(target, toolchain_name, clean=False, macros=None, + notify=None, jobs=1, report=None, properties=None, + build_profile=None): + """ Function returns True is library was built and false if building was + skipped + + Positional arguments: + target - the MCU or board that the project will compile for + toolchain_name - the name of the build tools + + Keyword arguments: + clean - Rebuild everything if True + macros - additional macros + notify - Notify function for logs + jobs - how many compilers we can run at once + report - a dict where a result may be appended + properties - UUUUHHHHH beats me + build_profile - a dict of flags that will be passed to the compiler + """ + + if report != None: + start = time() + id_name = "MBED" + description = "mbed SDK" + vendor_label = target.extra_labels[0] + cur_result = None + prep_report(report, target.name, toolchain_name, id_name) + cur_result = create_result(target.name, toolchain_name, id_name, + description) + + if properties != None: + prep_properties(properties, target.name, toolchain_name, + vendor_label) + + # Check toolchain support + if toolchain_name not in target.supported_toolchains: + supported_toolchains_text = ", ".join(target.supported_toolchains) + print('%s target is not yet supported by toolchain %s' % + (target.name, toolchain_name)) + print('%s target supports %s toolchain%s' % + (target.name, supported_toolchains_text, 's' + if len(target.supported_toolchains) > 1 else '')) + + if report != None: + cur_result["result"] = "SKIP" + add_result_to_report(report, cur_result) + + return False + + try: + # Source and Build Paths + build_target = join(MBED_LIBRARIES, "TARGET_" + target.name) + build_toolchain = join(MBED_LIBRARIES, mbed2_obj_path(target.name, toolchain_name)) + mkdir(build_toolchain) + + # Toolchain + tmp_path = join(MBED_LIBRARIES, '.temp', mbed2_obj_path(target.name, toolchain_name)) + mkdir(tmp_path) + + toolchain = prepare_toolchain( + [""], tmp_path, target, toolchain_name, macros=macros, notify=notify, + build_profile=build_profile, jobs=jobs, clean=clean) + + # Take into account the library configuration (MBED_CONFIG_FILE) + config = toolchain.config + config.add_config_files([MBED_CONFIG_FILE]) + toolchain.set_config_data(toolchain.config.get_config_data()) + + # mbed + notify.info("Building library %s (%s, %s)" % + ('MBED', target.name, toolchain_name)) + + # Common Headers + toolchain.copy_files([MBED_HEADER], MBED_LIBRARIES) + library_incdirs = [dirname(MBED_LIBRARIES), MBED_LIBRARIES] + + for dir, dest in [(MBED_DRIVERS, MBED_LIBRARIES_DRIVERS), + (MBED_PLATFORM, MBED_LIBRARIES_PLATFORM), + (MBED_HAL, MBED_LIBRARIES_HAL)]: + resources = toolchain.scan_resources(dir) + toolchain.copy_files(resources.headers, dest) + library_incdirs.append(dest) + + cmsis_implementation = toolchain.scan_resources(MBED_CMSIS_PATH) + toolchain.copy_files(cmsis_implementation.headers, build_target) + toolchain.copy_files(cmsis_implementation.linker_script, build_toolchain) + toolchain.copy_files(cmsis_implementation.bin_files, build_toolchain) + + hal_implementation = toolchain.scan_resources(MBED_TARGETS_PATH) + toolchain.copy_files(hal_implementation.headers + + hal_implementation.hex_files + + hal_implementation.libraries + + [MBED_CONFIG_FILE], + build_target, resources=hal_implementation) + toolchain.copy_files(hal_implementation.linker_script, build_toolchain) + toolchain.copy_files(hal_implementation.bin_files, build_toolchain) + incdirs = toolchain.scan_resources(build_target).inc_dirs + objects = toolchain.compile_sources(cmsis_implementation + hal_implementation, + library_incdirs + incdirs + [tmp_path]) + toolchain.copy_files(objects, build_toolchain) + + # Common Sources + mbed_resources = None + for dir in [MBED_DRIVERS, MBED_PLATFORM, MBED_HAL]: + mbed_resources += toolchain.scan_resources(dir) + + objects = toolchain.compile_sources(mbed_resources, + library_incdirs + incdirs) + + # A number of compiled files need to be copied as objects as opposed to + # way the linker search for symbols in archives. These are: + # - mbed_retarget.o: to make sure that the C standard lib symbols get + # overridden + # - mbed_board.o: mbed_die is weak + # - mbed_overrides.o: this contains platform overrides of various + # weak SDK functions + # - mbed_main.o: this contains main redirection + separate_names, separate_objects = ['mbed_retarget.o', 'mbed_board.o', + 'mbed_overrides.o', 'mbed_main.o', 'mbed_sdk_boot.o'], [] + + for obj in objects: + for name in separate_names: + if obj.endswith(name): + separate_objects.append(obj) + + for obj in separate_objects: + objects.remove(obj) + + toolchain.build_library(objects, build_toolchain, "mbed") + + for obj in separate_objects: + toolchain.copy_files(obj, build_toolchain) + + if report != None: + end = time() + cur_result["elapsed_time"] = end - start + cur_result["result"] = "OK" + + add_result_to_report(report, cur_result) + + return True + + except Exception as exc: + if report != None: + end = time() + cur_result["result"] = "FAIL" + cur_result["elapsed_time"] = end - start + + cur_result["output"] += str(exc) + + add_result_to_report(report, cur_result) + + # Let Exception propagate + raise + + +def get_unique_supported_toolchains(release_targets=None): + """ Get list of all unique toolchains supported by targets + + Keyword arguments: + release_targets - tuple structure returned from get_mbed_official_release(). + If release_targets is not specified, then it queries all + known targets + """ + unique_supported_toolchains = [] + + if not release_targets: + for target in TARGET_NAMES: + for toolchain in TARGET_MAP[target].supported_toolchains: + if toolchain not in unique_supported_toolchains: + unique_supported_toolchains.append(toolchain) + else: + for target in release_targets: + for toolchain in target[1]: + if toolchain not in unique_supported_toolchains: + unique_supported_toolchains.append(toolchain) + + if "ARM" in unique_supported_toolchains: + unique_supported_toolchains.append("ARMC6") + + return unique_supported_toolchains + +def mcu_toolchain_list(release_version='5'): + """ Shows list of toolchains + + """ + + if isinstance(release_version, basestring): + # Force release_version to lowercase if it is a string + release_version = release_version.lower() + else: + # Otherwise default to printing all known targets and toolchains + release_version = 'all' + + + version_release_targets = {} + version_release_target_names = {} + + for version in RELEASE_VERSIONS: + version_release_targets[version] = get_mbed_official_release(version) + version_release_target_names[version] = [x[0] for x in + version_release_targets[ + version]] + + if release_version in RELEASE_VERSIONS: + release_targets = version_release_targets[release_version] + else: + release_targets = None + + unique_supported_toolchains = get_unique_supported_toolchains( + release_targets) + columns = ["mbed OS %s" % x for x in RELEASE_VERSIONS] + unique_supported_toolchains + return "\n".join(columns) + + +def mcu_target_list(release_version='5'): + """ Shows target list + + """ + + if isinstance(release_version, basestring): + # Force release_version to lowercase if it is a string + release_version = release_version.lower() + else: + # Otherwise default to printing all known targets and toolchains + release_version = 'all' + + + version_release_targets = {} + version_release_target_names = {} + + for version in RELEASE_VERSIONS: + version_release_targets[version] = get_mbed_official_release(version) + version_release_target_names[version] = [x[0] for x in + version_release_targets[ + version]] + + if release_version in RELEASE_VERSIONS: + release_targets = version_release_targets[release_version] + else: + release_targets = None + + target_names = [] + + if release_targets: + target_names = [x[0] for x in release_targets] + else: + target_names = TARGET_NAMES + + return "\n".join(target_names) + + +def mcu_toolchain_matrix(verbose_html=False, platform_filter=None, + release_version='5'): + """ Shows target map using prettytable + + Keyword arguments: + verbose_html - emit html instead of a simple table + platform_filter - remove results that match the string + release_version - get the matrix for this major version number + """ + # Only use it in this function so building works without extra modules + from prettytable import PrettyTable + + if isinstance(release_version, basestring): + # Force release_version to lowercase if it is a string + release_version = release_version.lower() + else: + # Otherwise default to printing all known targets and toolchains + release_version = 'all' + + + version_release_targets = {} + version_release_target_names = {} + + for version in RELEASE_VERSIONS: + version_release_targets[version] = get_mbed_official_release(version) + version_release_target_names[version] = [x[0] for x in + version_release_targets[ + version]] + + if release_version in RELEASE_VERSIONS: + release_targets = version_release_targets[release_version] + else: + release_targets = None + + unique_supported_toolchains = get_unique_supported_toolchains( + release_targets) + prepend_columns = ["Target"] + ["mbed OS %s" % x for x in RELEASE_VERSIONS] + + # All tests status table print + columns = prepend_columns + unique_supported_toolchains + table_printer = PrettyTable(columns) + # Align table + for col in columns: + table_printer.align[col] = "c" + table_printer.align["Target"] = "l" + + perm_counter = 0 + target_counter = 0 + + target_names = [] + + if release_targets: + target_names = [x[0] for x in release_targets] + else: + target_names = TARGET_NAMES + + for target in sorted(target_names): + if platform_filter is not None: + # FIlter out platforms using regex + if re.search(platform_filter, target) is None: + continue + target_counter += 1 + + row = [target] # First column is platform name + + for version in RELEASE_VERSIONS: + if target in version_release_target_names[version]: + text = "Supported" + else: + text = "-" + row.append(text) + + for unique_toolchain in unique_supported_toolchains: + if (unique_toolchain in TARGET_MAP[target].supported_toolchains or + (unique_toolchain == "ARMC6" and + "ARM" in TARGET_MAP[target].supported_toolchains)): + text = "Supported" + perm_counter += 1 + else: + text = "-" + + row.append(text) + table_printer.add_row(row) + + result = table_printer.get_html_string() if verbose_html \ + else table_printer.get_string() + result += "\n" + result += "Supported targets: %d\n"% (target_counter) + if target_counter == 1: + result += "Supported toolchains: %d"% (perm_counter) + return result + + +def get_target_supported_toolchains(target): + """ Returns target supported toolchains list + + Positional arguments: + target - the target to get the supported toolchains of + """ + return TARGET_MAP[target].supported_toolchains if target in TARGET_MAP \ + else None + + +def print_build_results(result_list, build_name): + """ Generate result string for build results + + Positional arguments: + result_list - the list of results to print + build_name - the name of the build we are printing result for + """ + result = "" + if len(result_list) > 0: + result += build_name + "\n" + result += "\n".join([" * %s" % f for f in result_list]) + result += "\n" + return result + +def print_build_memory_usage(report): + """ Generate result table with memory usage values for build results + Aggregates (puts together) reports obtained from self.get_memory_summary() + + Positional arguments: + report - Report generated during build procedure. + """ + from prettytable import PrettyTable + columns_text = ['name', 'target', 'toolchain'] + columns_int = ['static_ram', 'total_flash'] + table = PrettyTable(columns_text + columns_int) + + for col in columns_text: + table.align[col] = 'l' + + for col in columns_int: + table.align[col] = 'r' + + for target in report: + for toolchain in report[target]: + for name in report[target][toolchain]: + for dlist in report[target][toolchain][name]: + for dlistelem in dlist: + # Get 'memory_usage' record and build table with + # statistics + record = dlist[dlistelem] + if 'memory_usage' in record and record['memory_usage']: + # Note that summary should be in the last record of + # 'memory_usage' section. This is why we are + # grabbing last "[-1]" record. + row = [ + record['description'], + record['target_name'], + record['toolchain_name'], + record['memory_usage'][-1]['summary'][ + 'static_ram'], + record['memory_usage'][-1]['summary'][ + 'total_flash'], + ] + table.add_row(row) + + result = "Memory map breakdown for built projects (values in Bytes):\n" + result += table.get_string(sortby='name') + return result + +def write_build_report(build_report, template_filename, filename): + """Write a build report to disk using a template file + + Positional arguments: + build_report - a report generated by the build system + template_filename - a file that contains the template for the style of build + report + filename - the location on disk to write the file to + """ + build_report_failing = [] + build_report_passing = [] + + for report in build_report: + if len(report["failing"]) > 0: + build_report_failing.append(report) + else: + build_report_passing.append(report) + + env = Environment(extensions=['jinja2.ext.with_']) + env.loader = FileSystemLoader('ci_templates') + template = env.get_template(template_filename) + + with open(filename, 'w+') as placeholder: + placeholder.write(template.render( + failing_builds=build_report_failing, + passing_builds=build_report_passing)) + + +def merge_build_data(filename, toolchain_report, app_type): + path_to_file = dirname(abspath(filename)) + try: + build_data = load(open(filename)) + except (IOError, ValueError): + build_data = {'builds': []} + for tgt in toolchain_report.values(): + for tc in tgt.values(): + for project in tc.values(): + for build in project: + try: + build[0]['elf'] = relpath(build[0]['elf'], path_to_file) + build[0]['bin'] = relpath(build[0]['bin'], path_to_file) + except KeyError: + pass + if 'type' not in build[0]: + build[0]['type'] = app_type + build_data['builds'].append(build[0]) + dump(build_data, open(filename, "wb"), indent=4, separators=(',', ': '))