Clone of official tools

Files at this revision

API Documentation at this revision

Comitter:
screamer
Date:
Mon Aug 29 10:55:42 2016 +0100
Parent:
27:5461402c33f8
Child:
29:1210849dba19
Commit message:
Remove obsolete file

Changed in this revision

toolchains/__init__.py.orig Show diff for this revision Revisions of this file
--- a/toolchains/__init__.py.orig	Mon Aug 29 10:55:24 2016 +0100
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,975 +0,0 @@
-"""
-mbed SDK
-Copyright (c) 2011-2013 ARM Limited
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-"""
-
-import re
-import sys
-from os import stat, walk, getcwd, sep
-from copy import copy
-from time import time, sleep
-from types import ListType
-from shutil import copyfile
-from os.path import join, splitext, exists, relpath, dirname, basename, split, abspath, isfile, isdir
-from inspect import getmro
-from copy import deepcopy
-from tools.config import Config
-
-from multiprocessing import Pool, cpu_count
-from tools.utils import run_cmd, mkdir, rel_path, ToolException, NotSupportedException, split_path
-from tools.settings import BUILD_OPTIONS, MBED_ORG_USER
-import tools.hooks as hooks
-from tools.memap import MemapParser
-from hashlib import md5
-import fnmatch
-
-
-#Disables multiprocessing if set to higher number than the host machine CPUs
-CPU_COUNT_MIN = 1
-
-def compile_worker(job):
-    results = []
-    for command in job['commands']:
-        try:
-            _, _stderr, _rc = run_cmd(command, job['work_dir'])
-        except KeyboardInterrupt as e:
-            raise ToolException
-
-        results.append({
-            'code': _rc,
-            'output': _stderr,
-            'command': command
-        })
-
-    return {
-        'source': job['source'],
-        'object': job['object'],
-        'commands': job['commands'],
-        'results': results
-    }
-
-
-class Resources:
-    def __init__(self, base_path=None):
-        self.base_path = base_path
-
-        self.file_basepath = {}
-
-        self.inc_dirs = []
-        self.headers = []
-
-        self.s_sources = []
-        self.c_sources = []
-        self.cpp_sources = []
-
-        self.lib_dirs = set([])
-        self.objects = []
-        self.libraries = []
-
-        # mbed special files
-        self.lib_builds = []
-        self.lib_refs = []
-
-        self.repo_dirs = []
-        self.repo_files = []
-
-        self.linker_script = None
-
-        # Other files
-        self.hex_files = []
-        self.bin_files = []
-        self.json_files = []
-
-        # Features
-        self.features = {}
-
-    def __add__(self, resources):
-        if resources is None:
-            return self
-        else:
-            return self.add(resources)
-
-    def __radd__(self, resources):
-        if resources is None:
-            return self
-        else:
-            return self.add(resources)
-
-    def add(self, resources):
-        for f,p in resources.file_basepath.items():
-            self.file_basepath[f] = p
-
-        self.inc_dirs += resources.inc_dirs
-        self.headers += resources.headers
-
-        self.s_sources += resources.s_sources
-        self.c_sources += resources.c_sources
-        self.cpp_sources += resources.cpp_sources
-
-        self.lib_dirs |= resources.lib_dirs
-        self.objects += resources.objects
-        self.libraries += resources.libraries
-
-        self.lib_builds += resources.lib_builds
-        self.lib_refs += resources.lib_refs
-
-        self.repo_dirs += resources.repo_dirs
-        self.repo_files += resources.repo_files
-
-        if resources.linker_script is not None:
-            self.linker_script = resources.linker_script
-
-        self.hex_files += resources.hex_files
-        self.bin_files += resources.bin_files
-        self.json_files += resources.json_files
-
-        self.features.update(resources.features)
-
-        return self
-
-    def relative_to(self, base, dot=False):
-        for field in ['inc_dirs', 'headers', 's_sources', 'c_sources',
-                      'cpp_sources', 'lib_dirs', 'objects', 'libraries',
-                      'lib_builds', 'lib_refs', 'repo_dirs', 'repo_files',
-                      'hex_files', 'bin_files', 'json_files']:
-            v = [rel_path(f, base, dot) for f in getattr(self, field)]
-            setattr(self, field, v)
-
-        self.features = {k: f.relative_to(base, dot) for k, f in self.features.iteritems() if f}
-
-        if self.linker_script is not None:
-            self.linker_script = rel_path(self.linker_script, base, dot)
-
-    def win_to_unix(self):
-        for field in ['inc_dirs', 'headers', 's_sources', 'c_sources',
-                      'cpp_sources', 'lib_dirs', 'objects', 'libraries',
-                      'lib_builds', 'lib_refs', 'repo_dirs', 'repo_files',
-                      'hex_files', 'bin_files', 'json_files']:
-            v = [f.replace('\\', '/') for f in getattr(self, field)]
-            setattr(self, field, v)
-
-        self.features = {k: f.win_to_unix() for k, f in self.features.iteritems() if f}
-
-        if self.linker_script is not None:
-            self.linker_script = self.linker_script.replace('\\', '/')
-
-    def __str__(self):
-        s = []
-
-        for (label, resources) in (
-                ('Include Directories', self.inc_dirs),
-                ('Headers', self.headers),
-
-                ('Assembly sources', self.s_sources),
-                ('C sources', self.c_sources),
-                ('C++ sources', self.cpp_sources),
-
-                ('Library directories', self.lib_dirs),
-                ('Objects', self.objects),
-                ('Libraries', self.libraries),
-
-                ('Hex files', self.hex_files),
-                ('Bin files', self.bin_files),
-
-                ('Features', self.features),
-            ):
-            if resources:
-                s.append('%s:\n  ' % label + '\n  '.join(resources))
-
-        if self.linker_script:
-            s.append('Linker Script: ' + self.linker_script)
-
-        return '\n'.join(s)
-
-# Support legacy build conventions: the original mbed build system did not have
-# standard labels for the "TARGET_" and "TOOLCHAIN_" specific directories, but
-# had the knowledge of a list of these directories to be ignored.
-LEGACY_IGNORE_DIRS = set([
-    'LPC11U24', 'LPC1768', 'LPC2368', 'LPC4088', 'LPC812', 'KL25Z',
-    'ARM', 'GCC_ARM', 'GCC_CR', 'IAR', 'uARM'
-])
-LEGACY_TOOLCHAIN_NAMES = {
-    'ARM_STD':'ARM', 'ARM_MICRO': 'uARM',
-    'GCC_ARM': 'GCC_ARM', 'GCC_CR': 'GCC_CR',
-    'IAR': 'IAR',
-}
-
-
-class mbedToolchain:
-    VERBOSE = True
-
-    CORTEX_SYMBOLS = {
-        "Cortex-M0" : ["__CORTEX_M0", "ARM_MATH_CM0", "__CMSIS_RTOS", "__MBED_CMSIS_RTOS_CM"],
-        "Cortex-M0+": ["__CORTEX_M0PLUS", "ARM_MATH_CM0PLUS", "__CMSIS_RTOS", "__MBED_CMSIS_RTOS_CM"],
-        "Cortex-M1" : ["__CORTEX_M3", "ARM_MATH_CM1", "__CMSIS_RTOS", "__MBED_CMSIS_RTOS_CM"],
-        "Cortex-M3" : ["__CORTEX_M3", "ARM_MATH_CM3", "__CMSIS_RTOS", "__MBED_CMSIS_RTOS_CM"],
-        "Cortex-M4" : ["__CORTEX_M4", "ARM_MATH_CM4", "__CMSIS_RTOS", "__MBED_CMSIS_RTOS_CM"],
-<<<<<<< HEAD
-        "Cortex-M4F" : ["__CORTEX_M4", "ARM_MATH_CM4", "__FPU_PRESENT=1", "__CMSIS_RTOS", "__MBED_CMSIS_RTOS_CM"],
-        "Cortex-M7" : ["__CORTEX_M7", "ARM_MATH_CM7", "__CMSIS_RTOS", "__MBED_CMSIS_RTOS_CM"],
-        "Cortex-M7F" : ["__CORTEX_M7", "ARM_MATH_CM7", "__FPU_PRESENT=1", "__CMSIS_RTOS", "__MBED_CMSIS_RTOS_CM"],
-        "Cortex-M7FD" : ["__CORTEX_M7", "ARM_MATH_CM7", "__FPU_PRESENT=1", "__CMSIS_RTOS", "__MBED_CMSIS_RTOS_CM"],
-=======
-        "Cortex-M4F": ["__CORTEX_M4", "__FPU_PRESENT=1", "ARM_MATH_CM4", "__CMSIS_RTOS", "__MBED_CMSIS_RTOS_CM"],
-        "Cortex-M7" : ["__CORTEX_M7", "ARM_MATH_CM7", "__CMSIS_RTOS", "__MBED_CMSIS_RTOS_CM"],
-        "Cortex-M7F" : ["__CORTEX_M7", "__FPU_PRESENT=1", "ARM_MATH_CM7", "__CMSIS_RTOS", "__MBED_CMSIS_RTOS_CM"],
-        "Cortex-M7FD" : ["__CORTEX_M7", "__FPU_PRESENT=1", "ARM_MATH_CM7", "__CMSIS_RTOS", "__MBED_CMSIS_RTOS_CM"],
->>>>>>> 359d33cc16e2d8a35c19e95f3fa8dcbb4be1b163
-        "Cortex-A9" : ["__CORTEX_A9", "ARM_MATH_CA9", "__FPU_PRESENT", "__CMSIS_RTOS", "__EVAL", "__MBED_CMSIS_RTOS_CA9"],
-    }
-    
-
-    GOANNA_FORMAT = "[Goanna] warning [%FILENAME%:%LINENO%] - [%CHECKNAME%(%SEVERITY%)] %MESSAGE%"
-    GOANNA_DIAGNOSTIC_PATTERN = re.compile(r'"\[Goanna\] (?P<severity>warning) \[(?P<file>[^:]+):(?P<line>\d+)\] \- (?P<message>.*)"')
-
-    MBED_CONFIG_FILE_NAME="mbed_config.h"
-
-    def __init__(self, target, options=None, notify=None, macros=None, silent=False, extra_verbose=False):
-        self.target = target
-        self.name = self.__class__.__name__
-        
-        # compile/assemble/link/binary hooks
-        self.hook = hooks.Hook(target, self)
-
-        # Toolchain flags
-        self.flags = deepcopy(self.DEFAULT_FLAGS)
-        
-        # User-defined macros
-        self.macros = macros or []
-        
-        # Macros generated from toolchain and target rules/features
-        self.symbols = None
-        
-        # Labels generated from toolchain and target rules/features (used for selective build)
-        self.labels = None
-        
-        # This will hold the configuration data (as returned by Config.get_config_data())
-        self.config_data = None
-
-        # Non-incremental compile
-        self.build_all = False
-        
-        # Build output dir
-        self.build_dir = None
-        self.timestamp = time()
-
-        # Output build naming based on target+toolchain combo (mbed 2.0 builds)
-        self.obj_path = join("TARGET_"+target.name, "TOOLCHAIN_"+self.name)
-
-        # Number of concurrent build jobs. 0 means auto (based on host system cores)
-        self.jobs = 0
-
-        self.CHROOT = None            
-
-        # Ignore patterns from .mbedignore files
-        self.ignore_patterns = []
-
-        # Pre-mbed 2.0 ignore dirs
-        self.legacy_ignore_dirs = LEGACY_IGNORE_DIRS - set([target.name, LEGACY_TOOLCHAIN_NAMES[self.name]])
-
-        # Output notify function
-        if notify:
-            self.notify_fun = notify
-        elif extra_verbose:
-            self.notify_fun = self.print_notify_verbose
-        else:
-            self.notify_fun = self.print_notify
-        
-        # Silent builds (no output)
-        self.silent = silent
-        
-        # Print output buffer
-        self.output = ""
-
-        # Build options passed by -o flag
-        self.options = options if options is not None else []
-
-        # Build options passed by settings.py or mbed_settings.py
-        self.options.extend(BUILD_OPTIONS)
-
-        if self.options:
-            self.info("Build Options: %s" % (', '.join(self.options)))
-        
-        # uVisor spepcific rules
-        if 'UVISOR' in self.target.features and 'UVISOR_SUPPORTED' in self.target.extra_labels:
-            self.target.core = re.sub(r"F$", '', self.target.core)
-
-    def get_output(self):
-        return self.output
-
-    def print_notify(self, event, silent=False):
-        """ Default command line notification
-        """
-        msg = None
-
-        if not self.VERBOSE and event['type'] == 'tool_error':
-            msg = event['message']
-        
-        elif event['type'] in ['info', 'debug']:
-            msg = event['message']
-            
-        elif event['type'] == 'cc':
-            event['severity'] = event['severity'].title()
-            event['file'] = basename(event['file'])
-            msg = '[%(severity)s] %(file)s@%(line)s: %(message)s' % event
-
-        elif event['type'] == 'progress':
-            if not silent:
-                msg = '%s: %s' % (event['action'].title(), basename(event['file']))
-
-        if msg:
-            print msg
-            self.output += msg + "\n"
-
-    def print_notify_verbose(self, event, silent=False):
-        """ Default command line notification with more verbose mode
-        """
-        if event['type'] in ['info', 'debug']:
-            self.print_notify(event) # standard handle
-
-        elif event['type'] == 'cc':
-            event['severity'] = event['severity'].title()
-            event['file'] = basename(event['file'])
-            event['mcu_name'] = "None"
-            event['toolchain'] = "None"
-            event['target_name'] = event['target_name'].upper() if event['target_name'] else "Unknown"
-            event['toolchain_name'] = event['toolchain_name'].upper() if event['toolchain_name'] else "Unknown"
-            msg = '[%(severity)s] %(target_name)s::%(toolchain_name)s::%(file)s@%(line)s: %(message)s' % event
-            print msg
-            self.output += msg + "\n"
-
-        elif event['type'] == 'progress':
-            self.print_notify(event) # standard handle
-
-    def notify(self, event):
-        """ Little closure for notify functions
-        """
-        return self.notify_fun(event, self.silent)
-
-    def goanna_parse_line(self, line):
-        if "analyze" in self.options:
-            return self.GOANNA_DIAGNOSTIC_PATTERN.match(line)
-        else:
-            return None
-
-    def get_symbols(self):
-        if self.symbols is None:
-            # Target and Toolchain symbols
-            labels = self.get_labels()
-            self.symbols = ["TARGET_%s" % t for t in labels['TARGET']]
-            self.symbols.extend(["TOOLCHAIN_%s" % t for t in labels['TOOLCHAIN']])
-
-            # Cortex CPU symbols
-            if self.target.core in mbedToolchain.CORTEX_SYMBOLS:
-                self.symbols.extend(mbedToolchain.CORTEX_SYMBOLS[self.target.core])
-
-            # Symbols defined by the on-line build.system
-            self.symbols.extend(['MBED_BUILD_TIMESTAMP=%s' % self.timestamp, 'TARGET_LIKE_MBED', '__MBED__=1'])
-            if MBED_ORG_USER:
-                self.symbols.append('MBED_USERNAME=' + MBED_ORG_USER)
-
-            # Add target's symbols
-            self.symbols += self.target.macros
-            # Add target's hardware
-            self.symbols += ["DEVICE_" + data + "=1" for data in self.target.device_has]
-            # Add target's features
-            self.symbols += ["FEATURE_" + data + "=1" for data in self.target.features]
-            # Add extra symbols passed via 'macros' parameter
-            self.symbols += self.macros
-
-            # Form factor variables
-            if hasattr(self.target, 'supported_form_factors'):
-                self.symbols.extend(["TARGET_FF_%s" % t for t in self.target.supported_form_factors])
-
-        return list(set(self.symbols))  # Return only unique symbols
-
-    # Extend the internal list of macros
-    def add_macros(self, new_macros):
-        self.macros.extend(new_macros)
-
-    def get_labels(self):
-        if self.labels is None:
-            toolchain_labels = [c.__name__ for c in getmro(self.__class__)]
-            toolchain_labels.remove('mbedToolchain')
-            self.labels = {
-                'TARGET': self.target.get_labels() + ["DEBUG" if "debug-info" in self.options else "RELEASE"],
-                'FEATURE': self.target.features,
-                'TOOLCHAIN': toolchain_labels
-            }
-        return self.labels
-
-    def need_update(self, target, dependencies):
-        if self.build_all:
-            return True
-
-        if not exists(target):
-            return True
-
-        target_mod_time = stat(target).st_mtime
-
-        for d in dependencies:
-
-            # Some objects are not provided with full path and here we do not have
-            # information about the library paths. Safe option: assume an update
-            if not d or not exists(d):
-                return True
-
-            if stat(d).st_mtime >= target_mod_time:
-                return True
-
-        return False
-
-    def is_ignored(self, file_path):
-        for pattern in self.ignore_patterns:
-            if fnmatch.fnmatch(file_path, pattern):
-                return True
-        return False
-
-    # Create a Resources object from the path pointed to by *path* by either traversing a
-    # a directory structure, when *path* is a directory, or adding *path* to the resources,
-    # when *path* is a file.
-    # The parameter *base_path* is used to set the base_path attribute of the Resources
-    # object and the parameter *exclude_paths* is used by the directory traversal to
-    # exclude certain paths from the traversal.
-    def scan_resources(self, path, exclude_paths=None, base_path=None):
-        resources = Resources(path)
-        if not base_path:
-            if isfile(path):
-                base_path = dirname(path)
-            else:
-                base_path = path
-        resources.base_path = base_path
-
-        if isfile(path):
-            self._add_file(path, resources, base_path, exclude_paths=exclude_paths)
-        else:
-            self._add_dir(path, resources, base_path, exclude_paths=exclude_paths)
-        return resources
-
-    # A helper function for scan_resources. _add_dir traverses *path* (assumed to be a
-    # directory) and heeds the ".mbedignore" files along the way. _add_dir calls _add_file
-    # on every file it considers adding to the resources object.
-    def _add_dir(self, path, resources, base_path, exclude_paths=None):
-        """ os.walk(top[, topdown=True[, onerror=None[, followlinks=False]]])
-        When topdown is True, the caller can modify the dirnames list in-place
-        (perhaps using del or slice assignment), and walk() will only recurse into
-        the subdirectories whose names remain in dirnames; this can be used to prune
-        the search, impose a specific order of visiting, or even to inform walk()
-        about directories the caller creates or renames before it resumes walk()
-        again. Modifying dirnames when topdown is False is ineffective, because in
-        bottom-up mode the directories in dirnames are generated before dirpath
-        itself is generated.
-        """
-        labels = self.get_labels()
-        for root, dirs, files in walk(path, followlinks=True):
-            # Check if folder contains .mbedignore
-            if ".mbedignore" in files:
-                with open (join(root,".mbedignore"), "r") as f:
-                    lines=f.readlines()
-                    lines = [l.strip() for l in lines] # Strip whitespaces
-                    lines = [l for l in lines if l != ""] # Strip empty lines
-                    lines = [l for l in lines if not re.match("^#",l)] # Strip comment lines
-                    # Append root path to glob patterns and append patterns to ignore_patterns
-                    self.ignore_patterns.extend([join(root,line.strip()) for line in lines])
-
-            # Skip the whole folder if ignored, e.g. .mbedignore containing '*'
-            if self.is_ignored(join(root,"")):
-                continue
-
-            for d in copy(dirs):
-                dir_path = join(root, d)
-                # Add internal repo folders/files. This is needed for exporters
-                if d == '.hg':
-                    resources.repo_dirs.append(dir_path)
-                    resources.repo_files.extend(self.scan_repository(dir_path))
-
-                if ((d.startswith('.') or d in self.legacy_ignore_dirs) or
-                    # Ignore targets that do not match the TARGET in extra_labels list
-                    (d.startswith('TARGET_') and d[7:] not in labels['TARGET']) or
-                    # Ignore toolchain that do not match the current TOOLCHAIN
-                    (d.startswith('TOOLCHAIN_') and d[10:] not in labels['TOOLCHAIN']) or
-                    # Ignore .mbedignore files
-                    self.is_ignored(join(dir_path,"")) or
-                    # Ignore TESTS dir
-                    (d == 'TESTS')):
-                        dirs.remove(d)
-                elif d.startswith('FEATURE_'):
-                    # Recursively scan features but ignore them in the current scan.
-                    # These are dynamically added by the config system if the conditions are matched
-                    resources.features[d[8:]] = self.scan_resources(dir_path, base_path=base_path)
-                    dirs.remove(d)
-                elif exclude_paths:
-                    for exclude_path in exclude_paths:
-                        rel_path = relpath(dir_path, exclude_path)
-                        if not (rel_path.startswith('..')):
-                            dirs.remove(d)
-                            break
-
-            # Add root to include paths
-            resources.inc_dirs.append(root)
-
-            for file in files:
-                file_path = join(root, file)
-                self._add_file(file_path, resources, base_path)
-
-    # A helper function for both scan_resources and _add_dir. _add_file adds one file
-    # (*file_path*) to the resources object based on the file type.
-    def _add_file(self, file_path, resources, base_path, exclude_paths=None):
-        resources.file_basepath[file_path] = base_path
-
-        if self.is_ignored(file_path):
-            return
-
-        _, ext = splitext(file_path)
-        ext = ext.lower()
-
-        if   ext == '.s':
-            resources.s_sources.append(file_path)
-
-        elif ext == '.c':
-            resources.c_sources.append(file_path)
-
-        elif ext == '.cpp':
-            resources.cpp_sources.append(file_path)
-
-        elif ext == '.h' or ext == '.hpp':
-            resources.headers.append(file_path)
-
-        elif ext == '.o':
-            resources.objects.append(file_path)
-
-        elif ext == self.LIBRARY_EXT:
-            resources.libraries.append(file_path)
-            resources.lib_dirs.add(dirname(file_path))
-
-        elif ext == self.LINKER_EXT:
-            if resources.linker_script is not None:
-                self.info("Warning: Multiple linker scripts detected: %s -> %s" % (resources.linker_script, file_path))
-            resources.linker_script = file_path
-
-        elif ext == '.lib':
-            resources.lib_refs.append(file_path)
-
-        elif ext == '.bld':
-            resources.lib_builds.append(file_path)
-
-        elif file == '.hgignore':
-            resources.repo_files.append(file_path)
-
-        elif ext == '.hex':
-            resources.hex_files.append(file_path)
-
-        elif ext == '.bin':
-            resources.bin_files.append(file_path)
-
-        elif ext == '.json':
-            resources.json_files.append(file_path)
-
-
-    def scan_repository(self, path):
-        resources = []
-
-        for root, dirs, files in walk(path):
-            # Remove ignored directories
-            for d in copy(dirs):
-                if d == '.' or d == '..':
-                    dirs.remove(d)
-
-            for file in files:
-                file_path = join(root, file)
-                resources.append(file_path)
-
-        return resources
-
-    def copy_files(self, files_paths, trg_path, resources=None, rel_path=None):
-
-        # Handle a single file
-        if type(files_paths) != ListType: files_paths = [files_paths]
-
-        for source in files_paths:
-            if source is None:
-                files_paths.remove(source)
-
-        for source in files_paths:
-            if resources is not None:
-                relative_path = relpath(source, resources.file_basepath[source])
-            elif rel_path is not None:
-                relative_path = relpath(source, rel_path)
-            else:
-                _, relative_path = split(source)
-
-            target = join(trg_path, relative_path)
-
-            if (target != source) and (self.need_update(target, [source])):
-                self.progress("copy", relative_path)
-                mkdir(dirname(target))
-                copyfile(source, target)
-
-    def relative_object_path(self, build_path, base_dir, source):
-        source_dir, name, _ = split_path(source)
-        
-        obj_dir = join(build_path, relpath(source_dir, base_dir))
-        mkdir(obj_dir)
-        return join(obj_dir, name + '.o')
-
-    def get_inc_file(self, includes):
-        include_file = join(self.build_dir, ".includes_%s.txt" % self.inc_md5)
-        if not exists(include_file):
-            with open(include_file, "wb") as f:
-                cmd_list = []
-                for c in includes:
-                    if c:
-                        cmd_list.append(('-I%s' % c).replace("\\", "/"))                    
-                string = " ".join(cmd_list)
-                f.write(string)
-        return include_file
-
-    def compile_sources(self, resources, build_path, inc_dirs=None):
-        # Web IDE progress bar for project build
-        files_to_compile = resources.s_sources + resources.c_sources + resources.cpp_sources
-        self.to_be_compiled = len(files_to_compile)
-        self.compiled = 0
-
-        inc_paths = resources.inc_dirs
-        if inc_dirs is not None:
-            inc_paths.extend(inc_dirs)
-        # De-duplicate include paths
-        inc_paths = set(inc_paths)
-        # Sort include paths for consistency
-        inc_paths = sorted(set(inc_paths))
-        # Unique id of all include paths
-        self.inc_md5 = md5(' '.join(inc_paths)).hexdigest()
-        # Where to store response files
-        self.build_dir = build_path
-
-        objects = []
-        queue = []
-        prev_dir = None
-
-        # Sort compile queue for consistency
-        files_to_compile.sort()
-        work_dir = getcwd()
-
-        for source in files_to_compile:
-            _, name, _ = split_path(source)
-            object = self.relative_object_path(build_path, resources.file_basepath[source], source)
-
-            # Queue mode (multiprocessing)
-            commands = self.compile_command(source, object, inc_paths)
-            if commands is not None:
-                queue.append({
-                    'source': source,
-                    'object': object,
-                    'commands': commands,
-                    'work_dir': work_dir,
-                    'chroot': self.CHROOT
-                })
-            else:
-                objects.append(object)
-
-        # Use queues/multiprocessing if cpu count is higher than setting
-        jobs = self.jobs if self.jobs else cpu_count()
-        if jobs > CPU_COUNT_MIN and len(queue) > jobs:
-            return self.compile_queue(queue, objects)
-        else:
-            return self.compile_seq(queue, objects)
-
-    def compile_seq(self, queue, objects):
-        for item in queue:
-            result = compile_worker(item)
-
-            self.compiled += 1
-            self.progress("compile", item['source'], build_update=True)
-            for res in result['results']:
-                self.debug("Command: %s" % ' '.join(res['command']))
-                self.compile_output([
-                    res['code'],
-                    res['output'],
-                    res['command']
-                ])
-            objects.append(result['object'])
-        return objects
-
-    def compile_queue(self, queue, objects):
-        jobs_count = int(self.jobs if self.jobs else cpu_count())
-        p = Pool(processes=jobs_count)
-
-        results = []
-        for i in range(len(queue)):
-            results.append(p.apply_async(compile_worker, [queue[i]]))
-
-        itr = 0
-        while True:
-            itr += 1
-            if itr > 180000:
-                p.terminate()
-                p.join()
-                raise ToolException("Compile did not finish in 5 minutes")
-
-            pending = 0
-            for r in results:
-                if r._ready is True:
-                    try:
-                        result = r.get()
-                        results.remove(r)
-
-                        self.compiled += 1
-                        self.progress("compile", result['source'], build_update=True)
-                        for res in result['results']:
-                            self.debug("Command: %s" % ' '.join(res['command']))
-                            self.compile_output([
-                                res['code'],
-                                res['output'],
-                                res['command']
-                            ])
-                        objects.append(result['object'])
-                    except ToolException, err:
-                        p.terminate()
-                        p.join()
-                        raise ToolException(err)
-                else:
-                    pending += 1
-                    if pending > jobs_count:
-                        break
-
-
-            if len(results) == 0:
-                break
-
-            sleep(0.01)
-
-        results = None
-        p.terminate()
-        p.join()
-
-        return objects
-
-    def compile_command(self, source, object, includes):
-        # Check dependencies
-        _, ext = splitext(source)
-        ext = ext.lower()
-
-        if ext == '.c' or  ext == '.cpp':
-            base, _ = splitext(object)
-            dep_path = base + '.d'
-            deps = self.parse_dependencies(dep_path) if (exists(dep_path)) else []
-            if len(deps) == 0 or self.need_update(object, deps):
-                if ext == '.c':
-                    return self.compile_c(source, object, includes)
-                else:
-                    return self.compile_cpp(source, object, includes)
-        elif ext == '.s':
-            deps = [source]
-            if self.need_update(object, deps):
-                return self.assemble(source, object, includes)
-        else:
-            return False
-
-        return None
-
-    def is_not_supported_error(self, output):
-        return "#error directive: [NOT_SUPPORTED]" in output
-
-    def compile_output(self, output=[]):
-        _rc = output[0]
-        _stderr = output[1]
-        command = output[2]
-
-        # Parse output for Warnings and Errors
-        self.parse_output(_stderr)
-        self.debug("Return: %s"% _rc)
-        for error_line in _stderr.splitlines():
-            self.debug("Output: %s"% error_line)
-
-
-        # Check return code
-        if _rc != 0:
-            for line in _stderr.splitlines():
-                self.tool_error(line)
-
-            if self.is_not_supported_error(_stderr):
-                raise NotSupportedException(_stderr)
-            else:
-                raise ToolException(_stderr)
-
-    def build_library(self, objects, dir, name):
-        needed_update = False
-        lib = self.STD_LIB_NAME % name
-        fout = join(dir, lib)
-        if self.need_update(fout, objects):
-            self.info("Library: %s" % lib)
-            self.archive(objects, fout)
-            needed_update = True
-
-        return needed_update
-
-    def link_program(self, r, tmp_path, name):
-        needed_update = False
-        ext = 'bin'
-        if hasattr(self.target, 'OUTPUT_EXT'):
-            ext = self.target.OUTPUT_EXT
-
-        if hasattr(self.target, 'OUTPUT_NAMING'):
-            self.var("binary_naming", self.target.OUTPUT_NAMING)
-            if self.target.OUTPUT_NAMING == "8.3":
-                name = name[0:8]
-                ext = ext[0:3]
-        
-        # Create destination directory
-        head, tail =  split(name)
-        new_path = join(tmp_path, head)
-        mkdir(new_path)
-        
-        filename = name+'.'+ext
-        elf = join(tmp_path, name + '.elf')
-        bin = join(tmp_path, filename)
-        map = join(tmp_path, name + '.map')
-
-        if self.need_update(elf, r.objects + r.libraries + [r.linker_script]):
-            needed_update = True
-            self.progress("link", name)
-            self.link(elf, r.objects, r.libraries, r.lib_dirs, r.linker_script)
-
-        if self.need_update(bin, [elf]):
-            needed_update = True
-            self.progress("elf2bin", name)
-
-            self.binary(r, elf, bin)
-
-        self.mem_stats(map)
-
-        self.var("compile_succeded", True)
-        self.var("binary", filename)
-
-        return bin, needed_update
-
-    def default_cmd(self, command):
-        self.debug("Command: %s"% ' '.join(command))
-        _stdout, _stderr, _rc = run_cmd(command)
-
-        self.debug("Return: %s"% _rc)
-
-        for output_line in _stdout.splitlines():
-            self.debug("Output: %s"% output_line)
-        for error_line in _stderr.splitlines():
-            self.debug("Errors: %s"% error_line)
-
-        if _rc != 0:
-            for line in _stderr.splitlines():
-                self.tool_error(line)
-            raise ToolException(_stderr)
-
-    ### NOTIFICATIONS ###
-    def info(self, message):
-        self.notify({'type': 'info', 'message': message})
-
-    def debug(self, message):
-        if self.VERBOSE:
-            if type(message) is ListType:
-                message = ' '.join(message)
-            message = "[DEBUG] " + message
-            self.notify({'type': 'debug', 'message': message})
-
-    def cc_info(self, severity, file, line, message, target_name=None, toolchain_name=None):
-        self.notify({'type': 'cc',
-                     'severity': severity,
-                     'file': file,
-                     'line': line,
-                     'message': message,
-                     'target_name': target_name,
-                     'toolchain_name': toolchain_name})
-
-    def progress(self, action, file, build_update=False):
-        msg = {'type': 'progress', 'action': action, 'file': file}
-        if build_update:
-            msg['percent'] = 100. * float(self.compiled) / float(self.to_be_compiled)
-        self.notify(msg)
-
-    def tool_error(self, message):
-        self.notify({'type': 'tool_error', 'message': message})
-
-    def var(self, key, value):
-        self.notify({'type': 'var', 'key': key, 'val': value})
-
-    def mem_stats(self, map):
-        # Creates parser object
-        toolchain = self.__class__.__name__
-
-        # Create memap object
-        memap = MemapParser()
-
-        # Parse and decode a map file
-        if memap.parse(abspath(map), toolchain) is False:
-            self.info("Unknown toolchain for memory statistics %s" % toolchain)
-            return
-
-        # Write output to stdout in text (pretty table) format
-        memap.generate_output('table')
-
-        # Write output to file in JSON format
-        map_out = splitext(map)[0] + "_map.json"
-        memap.generate_output('json', map_out)
- 
-        # Write output to file in CSV format for the CI
-        map_csv = splitext(map)[0] + "_map.csv"
-        memap.generate_output('csv-ci', map_csv)
-
-    # Set the configuration data
-    def set_config_data(self, config_data):
-        self.config_data = config_data
-
-    # Return the location of the config header. This function will create the config
-    # header first if needed. The header will be written in a file called "mbed_conf.h"
-    # located in the project's build directory.
-    # If config headers are not used (self.config_header_content is None), the function
-    # returns None
-    def get_config_header(self):
-        if self.config_data is None:
-            return None
-        config_file = join(self.build_dir, self.MBED_CONFIG_FILE_NAME)
-        if not exists(config_file):
-            with open(config_file, "wt") as f:
-                f.write(Config.config_to_header(self.config_data))
-        return config_file
-
-    # Return the list of macros geenrated by the build system
-    def get_config_macros(self):
-        return Config.config_to_macros(self.config_data) if self.config_data else []
-
-from tools.settings import ARM_BIN
-from tools.settings import GCC_ARM_PATH, GCC_CR_PATH
-from tools.settings import IAR_PATH
-
-TOOLCHAIN_BIN_PATH = {
-    'ARM': ARM_BIN,
-    'uARM': ARM_BIN,
-    'GCC_ARM': GCC_ARM_PATH,
-    'GCC_CR': GCC_CR_PATH,
-    'IAR': IAR_PATH
-}
-
-from tools.toolchains.arm import ARM_STD, ARM_MICRO
-from tools.toolchains.gcc import GCC_ARM, GCC_CR
-from tools.toolchains.iar import IAR
-
-TOOLCHAIN_CLASSES = {
-    'ARM': ARM_STD,
-    'uARM': ARM_MICRO,
-    'GCC_ARM': GCC_ARM,
-    'GCC_CR': GCC_CR,
-    'IAR': IAR
-}
-
-TOOLCHAINS = set(TOOLCHAIN_CLASSES.keys())