X-Git-Url: https://codewiz.org/gitweb?a=blobdiff_plain;f=wizard%2Fbertos_utils.py;h=ce526f013b5ed1f6f4ae6517ea88590f9301391b;hb=30edcd04a72705844739d3bee5dd16dd7c44cf20;hp=b478dc5c864b933280a6ba8c05c3a119ed3d83f6;hpb=55055e1f35dceb091729dd19d03aff49491e16b9;p=bertos.git diff --git a/wizard/bertos_utils.py b/wizard/bertos_utils.py index b478dc5c..ce526f01 100644 --- a/wizard/bertos_utils.py +++ b/wizard/bertos_utils.py @@ -1,10 +1,34 @@ #!/usr/bin/env python # encoding: utf-8 # +# This file is part of BeRTOS. +# +# Bertos is free software; you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation; either version 2 of the License, or +# (at your option) any later version. +# +# This program is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# along with this program; if not, write to the Free Software +# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA +# +# As a special exception, you may use this file as part of a free software +# library without restriction. Specifically, if other files instantiate +# templates or use macros or inline functions from this file, or you compile +# this file and link it with other files to produce an executable, this +# file does not by itself cause the resulting executable to be covered by +# the GNU General Public License. This exception does not however +# invalidate any other reasons why the executable file might be covered by +# the GNU General Public License. +# # Copyright 2008 Develer S.r.l. (http://www.develer.com/) -# All rights reserved. # -# $Id:$ +# $Id$ # # Author: Lorenzo Berni # @@ -14,10 +38,16 @@ import fnmatch import glob import re import shutil +# Use custom copytree function +import copytree +import pickle import const -import codelite_project +import plugins import DefineException +import BProject + +from LoadException import VersionException, ToolchainException def isBertosDir(directory): return os.path.exists(directory + "/VERSION") @@ -25,178 +55,336 @@ def isBertosDir(directory): def bertosVersion(directory): return open(directory + "/VERSION").readline().strip() -def createBertosProject(projectInfo): - directory = projectInfo.info("PROJECT_PATH") - sourcesDir = projectInfo.info("SOURCES_PATH") - if not os.path.isdir(directory): - os.mkdir(directory) +def loadBertosProject(project_file, info_dict): + project_data = pickle.loads(open(project_file, "r").read()) + project_info = BProject.BProject() + project_info.setInfo("PROJECT_PATH", os.path.dirname(project_file)) + if "SOURCES_PATH" in info_dict: + project_data["SOURCES_PATH"] = info_dict["SOURCES_PATH"] + if os.path.exists(project_data["SOURCES_PATH"]): + project_info.setInfo("SOURCES_PATH", project_data["SOURCES_PATH"]) + else: + raise VersionException(project_info) + loadSourceTree(project_info) + cpu_name = project_data["CPU_NAME"] + project_info.setInfo("CPU_NAME", cpu_name) + cpu_info = loadCpuInfos(project_info) + for cpu in cpu_info: + if cpu["CPU_NAME"] == cpu_name: + project_info.setInfo("CPU_INFOS", cpu) + break + tag_list = getTagSet(cpu_info) + # Create, fill and store the dict with the tags + tag_dict = {} + for element in tag_list: + tag_dict[element] = False + infos = project_info.info("CPU_INFOS") + for tag in tag_dict: + if tag in infos["CPU_TAGS"] + [infos["CPU_NAME"], infos["TOOLCHAIN"]]: + tag_dict[tag] = True + else: + tag_dict[tag] = False + project_info.setInfo("ALL_CPU_TAGS", tag_dict) + if "TOOLCHAIN" in info_dict: + project_data["TOOLCHAIN"] = info_dict["TOOLCHAIN"] + if os.path.exists(project_data["TOOLCHAIN"]["path"]): + project_info.setInfo("TOOLCHAIN", project_data["TOOLCHAIN"]) + else: + raise ToolchainException(project_info) + project_info.setInfo("SELECTED_FREQ", project_data["SELECTED_FREQ"]) + project_info.setInfo("OUTPUT", project_data["OUTPUT"]) + loadModuleData(project_info, True) + setEnabledModules(project_info, project_data["ENABLED_MODULES"]) + return project_info + +def setEnabledModules(project_info, enabled_modules): + modules = project_info.info("MODULES") + files = {} + for module, information in modules.items(): + information["enabled"] = module in enabled_modules + for dependency in information["depends"]: + if not dependency in modules: + if dependency in files: + files[dependency] += 1 + else: + files[dependency] = 1 + project_info.setInfo("MODULES", modules) + project_info.setInfo("FILES", files) + +def enabledModules(project_info): + enabled_modules = [] + for name, module in project_info.info("MODULES").items(): + if module["enabled"]: + enabled_modules.append(name) + return enabled_modules + +def mergeSources(srcdir, new_sources, old_sources): + # The current mergeSources function provide only a raw copy of the sources in the + # created project. + # + # TODO: implement the three way merge algorithm + # + shutil.rmtree(srcdir, True) + copytree.copytree(os.path.join(new_sources, "bertos"), srcdir, ignore_list=const.IGNORE_LIST) + +def projectFileGenerator(project_info): + directory = project_info.info("PROJECT_PATH") + project_data = {} + enabled_modules = [] + for module, information in project_info.info("MODULES").items(): + if information["enabled"]: + enabled_modules.append(module) + project_data["ENABLED_MODULES"] = enabled_modules + project_data["SOURCES_PATH"] = project_info.info("SOURCES_PATH") + project_data["TOOLCHAIN"] = project_info.info("TOOLCHAIN") + project_data["CPU_NAME"] = project_info.info("CPU_NAME") + project_data["SELECTED_FREQ"] = project_info.info("SELECTED_FREQ") + project_data["OUTPUT"] = project_info.info("OUTPUT") + return pickle.dumps(project_data) + +def createBertosProject(project_info, edit=False): + directory = project_info.info("PROJECT_PATH") + sources_dir = project_info.info("SOURCES_PATH") + old_sources_dir = project_info.info("OLD_SOURCES_PATH") + if not edit: + if os.path.isdir(directory): + shutil.rmtree(directory, True) + os.makedirs(directory) + # Write the project file f = open(directory + "/project.bertos", "w") - f.write(repr(projectInfo)) + f.write(projectFileGenerator(project_info)) f.close() - ## Destination source dir + # Destination source dir srcdir = directory + "/bertos" - shutil.rmtree(srcdir, True) - shutil.copytree(sourcesDir + "/bertos", srcdir) - ## Destination makefile + if not edit: + # If not in editing mode it copies all the bertos sources in the /bertos subdirectory of the project + shutil.rmtree(srcdir, True) + copytree.copytree(sources_dir + "/bertos", srcdir, ignore_list=const.IGNORE_LIST) + elif old_sources_dir: + # If in editing mode it merges the current bertos sources with the selected ones + # TODO: implement the three way merge algotihm + # + mergeSources(srcdir, sources_dir, old_sources_dir) + # Destination makefile makefile = directory + "/Makefile" - if os.path.exists(makefile): - os.remove(makefile) makefile = open("mktemplates/Makefile").read() - makefile = makefileGenerator(projectInfo, makefile) + makefile = makefileGenerator(project_info, makefile) open(directory + "/Makefile", "w").write(makefile) - ## Destination project dir + # Destination project dir prjdir = directory + "/" + os.path.basename(directory) - shutil.rmtree(prjdir, True) - os.mkdir(prjdir) - ## Destination configurations files + if not edit: + shutil.rmtree(prjdir, True) + os.mkdir(prjdir) + # Destination hw files + hwdir = prjdir + "/hw" + if not edit: + shutil.rmtree(hwdir, True) + os.mkdir(hwdir) + # Copy all the hw files + for module, information in project_info.info("MODULES").items(): + for hwfile in information["hw"]: + string = open(sources_dir + "/" + hwfile, "r").read() + hwfile_path = hwdir + "/" + os.path.basename(hwfile) + if not edit or not os.path.exists(hwfile_path): + # If not in editing mode it copies all the hw files. If in + # editing mode it copies only the files that don't exist yet + open(hwdir + "/" + os.path.basename(hwfile), "w").write(string) + # Destination configurations files cfgdir = prjdir + "/cfg" - shutil.rmtree(cfgdir, True) - os.mkdir(cfgdir) - for key, value in projectInfo.info("CONFIGURATIONS").items(): - string = open(sourcesDir + "/" + key, "r").read() - for parameter, infos in value.items(): + if not edit: + shutil.rmtree(cfgdir, True) + os.mkdir(cfgdir) + # Set properly the autoenabled parameters + for module, information in project_info.info("MODULES").items(): + if "configuration" in information and information["configuration"] != "": + configurations = project_info.info("CONFIGURATIONS") + configuration = configurations[information["configuration"]] + for start, parameter in configuration["paramlist"]: + if "type" in configuration[parameter]["informations"] and configuration[parameter]["informations"]["type"] == "autoenabled": + configuration[parameter]["value"] = "1" if information["enabled"] else "0" + project_info.setInfo("CONFIGURATIONS", configurations) + # Copy all the configuration files + for configuration, information in project_info.info("CONFIGURATIONS").items(): + string = open(sources_dir + "/" + configuration, "r").read() + for start, parameter in information["paramlist"]: + infos = information[parameter] value = infos["value"] - if "type" in infos["informations"] and infos["informations"]["type"] == "autoenabled": - value = "1" - if "unsigned" in infos["informations"].keys() and infos["informations"]["unsigned"]: + if "unsigned" in infos["informations"] and infos["informations"]["unsigned"]: value += "U" - if "long" in infos["informations"].keys() and infos["informations"]["long"]: + if "long" in infos["informations"] and infos["informations"]["long"]: value += "L" string = sub(string, parameter, value) - f = open(cfgdir + "/" + os.path.basename(key), "w") + f = open(cfgdir + "/" + os.path.basename(configuration), "w") f.write(string) f.close() - ## Destinatio mk file - makefile = open("mktemplates/template.mk", "r").read() - makefile = mkGenerator(projectInfo, makefile) - open(prjdir + "/" + os.path.basename(prjdir) + ".mk", "w").write(makefile) - ## Destination main.c file - main = open("srctemplates/main.c", "r").read() - open(prjdir + "/main.c", "w").write(main) - if "codelite" in projectInfo.info("OUTPUT"): - workspace = codeliteWorkspaceGenerator(projectInfo) - open(directory + "/" + os.path.basename(prjdir) + ".workspace", "w").write(workspace) - project = codeliteProjectGenerator(projectInfo) - open(directory + "/" + os.path.basename(prjdir) + ".project", "w").write(project) - -def mkGenerator(projectInfo, makefile): + if not edit: + # Destination user mk file (only on project creation) + makefile = open("mktemplates/template.mk", "r").read() + makefile = mkGenerator(project_info, makefile) + open(prjdir + "/" + os.path.basename(prjdir) + ".mk", "w").write(makefile) + # Destination wizard mk file + makefile = open("mktemplates/template_wiz.mk", "r").read() + makefile = mkGenerator(project_info, makefile) + open(prjdir + "/" + os.path.basename(prjdir) + "_wiz.mk", "w").write(makefile) + # Destination main.c file + if not edit: + main = open("srctemplates/main.c", "r").read() + open(prjdir + "/main.c", "w").write(main) + # Files for selected plugins + relevants_files = {} + for plugin in project_info.info("OUTPUT"): + module = loadPlugin(plugin) + relevants_files[plugin] = module.createProject(project_info) + project_info.setInfo("RELEVANT_FILES", relevants_files) + +def loadPlugin(plugin): + """ + Returns the given plugin module. + """ + return getattr(__import__("plugins", {}, {}, [plugin]), plugin) + +def mkGenerator(project_info, makefile): """ Generates the mk file for the current project. """ - mkData = {} - mkData["$pname"] = os.path.basename(projectInfo.info("PROJECT_PATH")) - mkData["$cpuname"] = projectInfo.info("CPU_INFOS")["CORE_CPU"] - mkData["$cflags"] = " ".join(projectInfo.info("CPU_INFOS")["C_FLAGS"]) - mkData["$ldflags"] = " ".join(projectInfo.info("CPU_INFOS")["LD_FLAGS"]) - mkData["$csrc"], mkData["$pcsrc"], mkData["$constants"] = csrcGenerator(projectInfo) - mkData["$prefix"] = projectInfo.info("TOOLCHAIN")["path"].split("gcc")[0] - mkData["$suffix"] = projectInfo.info("TOOLCHAIN")["path"].split("gcc")[1] - mkData["$cross"] = projectInfo.info("TOOLCHAIN")["path"].split("gcc")[0] - mkData["$main"] = os.path.basename(projectInfo.info("PROJECT_PATH")) + "/main.c" - for key in mkData: + mk_data = {} + mk_data["$pname"] = os.path.basename(project_info.info("PROJECT_PATH")) + mk_data["$cpuclockfreq"] = project_info.info("SELECTED_FREQ") + cpu_mk_parameters = [] + for key, value in project_info.info("CPU_INFOS").items(): + if key.startswith(const.MK_PARAM_ID): + cpu_mk_parameters.append("%s = %s" %(key.replace("MK", mk_data["$pname"]), value)) + mk_data["$cpuparameters"] = "\n".join(cpu_mk_parameters) + mk_data["$csrc"], mk_data["$pcsrc"], mk_data["$cppasrc"], mk_data["$cxxsrc"], mk_data["$asrc"], mk_data["$constants"] = csrcGenerator(project_info) + mk_data["$prefix"] = replaceSeparators(project_info.info("TOOLCHAIN")["path"].split("gcc")[0]) + mk_data["$suffix"] = replaceSeparators(project_info.info("TOOLCHAIN")["path"].split("gcc")[1]) + mk_data["$main"] = os.path.basename(project_info.info("PROJECT_PATH")) + "/main.c" + for key in mk_data: while makefile.find(key) != -1: - makefile = makefile.replace(key, mkData[key]) + makefile = makefile.replace(key, mk_data[key]) return makefile -def makefileGenerator(projectInfo, makefile): +def makefileGenerator(project_info, makefile): """ Generate the Makefile for the current project. """ - # TODO: write a general function that works for both the mk file and the Makefile - while makefile.find("project_name") != -1: - makefile = makefile.replace("project_name", os.path.basename(projectInfo.info("PROJECT_PATH"))) + # TODO write a general function that works for both the mk file and the Makefile + while makefile.find("$pname") != -1: + makefile = makefile.replace("$pname", os.path.basename(project_info.info("PROJECT_PATH"))) return makefile -def csrcGenerator(projectInfo): - modules = projectInfo.info("MODULES") - files = projectInfo.info("FILES") - if "harvard" in projectInfo.info("CPU_INFOS")["CPU_TAGS"]: +def csrcGenerator(project_info): + modules = project_info.info("MODULES") + files = project_info.info("FILES") + if "harvard" in project_info.info("CPU_INFOS")["CPU_TAGS"]: harvard = True else: harvard = False + # file to be included in CSRC variable csrc = [] + # file to be included in PCSRC variable pcsrc = [] + # files to be included in CPPASRC variable + cppasrc = [] + # files to be included in CXXSRC variable + cxxsrc = [] + # files to be included in ASRC variable + asrc = [] + # constants to be included at the beginning of the makefile constants = {} - moduleFiles = set([]) - dependencyFiles = set([]) - asmFiles = set([]) for module, information in modules.items(): + module_files = set([]) + dependency_files = set([]) + # assembly sources + asm_files = set([]) + hwdir = os.path.basename(project_info.info("PROJECT_PATH")) + "/hw" if information["enabled"]: if "constants" in information: constants.update(information["constants"]) - cfiles, sfiles = findModuleFiles(module, projectInfo) - moduleFiles |= set(cfiles) - asmFiles |= set(sfiles) - for fileDependency in information["depends"]: - if fileDependency in files: - dependencyCFiles, dependencySFiles = findModuleFiles(fileDependency, projectInfo) - dependencyFiles |= set(dependencyCFiles) - asmFiles |= set(dependencySFiles) - for file in moduleFiles: - if not harvard or "harvard" not in information or information["harvard"] == "both": + cfiles, sfiles = findModuleFiles(module, project_info) + module_files |= set(cfiles) + asm_files |= set(sfiles) + for file in information["hw"]: + if file.endswith(".c"): + module_files |= set([hwdir + "/" + os.path.basename(file)]) + for file_dependency in information["depends"]: + if file_dependency in files: + dependencyCFiles, dependencySFiles = findModuleFiles(file_dependency, project_info) + dependency_files |= set(dependencyCFiles) + asm_files |= set(dependencySFiles) + for file in module_files: + if not harvard or information.get("harvard", "both") == "both": csrc.append(file) if harvard and "harvard" in information: pcsrc.append(file) - for file in dependencyFiles: + for file in dependency_files: + csrc.append(file) + for file in project_info.info("CPU_INFOS")["C_SRC"]: csrc.append(file) + for file in project_info.info("CPU_INFOS")["PC_SRC"]: + pcsrc.append(file) + for file in asm_files: + cppasrc.append(file) + for file in project_info.info("CPU_INFOS")["CPPA_SRC"]: + cppasrc.append(file) + for file in project_info.info("CPU_INFOS")["CXX_SRC"]: + cxxsrc.append(file) + for file in project_info.info("CPU_INFOS")["ASRC"]: + asrc.append(file) + csrc = set(csrc) csrc = " \\\n\t".join(csrc) + " \\" + pcsrc = set(pcsrc) pcsrc = " \\\n\t".join(pcsrc) + " \\" - constants = "\n".join([os.path.basename(projectInfo.info("PROJECT_PATH")) + "_" + key + " = " + str(value) for key, value in constants.items()]) - print asmFiles - return csrc, pcsrc, constants - -def findModuleFiles(module, projectInfo): + cppasrc = set(cppasrc) + cppasrc = " \\\n\t".join(cppasrc) + " \\" + cxxsrc = set(cxxsrc) + cxxsrc = " \\\n\t".join(cxxsrc) + " \\" + asrc = set(asrc) + asrc = " \\\n\t".join(asrc) + " \\" + constants = "\n".join([os.path.basename(project_info.info("PROJECT_PATH")) + "_" + key + " = " + unicode(value) for key, value in constants.items()]) + return csrc, pcsrc, cppasrc, cxxsrc, asrc, constants + +def findModuleFiles(module, project_info): + # Find the files related to the selected module cfiles = [] sfiles = [] - for filename, path in findDefinitions(module + ".c", projectInfo): - path = path.replace(projectInfo.info("SOURCES_PATH") + "/", "") - cfiles.append(path + "/" + filename) - for filename, path in findDefinitions(module + "_" + projectInfo.info("CPU_INFOS")["TOOLCHAIN"] + ".c", projectInfo): - path = path.replace(projectInfo.info("SOURCES_PATH") + "/", "") + # .c files related to the module and the cpu architecture + for filename, path in findDefinitions(module + ".c", project_info) + \ + findDefinitions(module + "_" + project_info.info("CPU_INFOS")["TOOLCHAIN"] + ".c", project_info): + path = path.replace(project_info.info("SOURCES_PATH") + os.sep, "") + path = replaceSeparators(path) cfiles.append(path + "/" + filename) - for tag in projectInfo.info("CPU_INFOS")["CPU_TAGS"]: - for filename, path in findDefinitions(module + "_" + tag + ".c", projectInfo): - path = path.replace(projectInfo.info("SOURCES_PATH") + "/", "") - cfiles.append(path + "/" + filename) - for filename, path in findDefinitions(module + ".s", projectInfo): - path = path.replace(projectInfo.info("SOURCES_PATH") + "/", "") + # .s files related to the module and the cpu architecture + for filename, path in findDefinitions(module + ".s", project_info) + \ + findDefinitions(module + "_" + project_info.info("CPU_INFOS")["TOOLCHAIN"] + ".s", project_info) + \ + findDefinitions(module + ".S", project_info) + \ + findDefinitions(module + "_" + project_info.info("CPU_INFOS")["TOOLCHAIN"] + ".S", project_info): + path = path.replace(project_info.info("SOURCES_PATH") + os.sep, "") + path = replaceSeparators(path) sfiles.append(path + "/" + filename) - for filename, path in findDefinitions(module + "_" + projectInfo.info("CPU_INFOS")["TOOLCHAIN"] + ".s", projectInfo): - path = path.replace(projectInfo.info("SOURCES_PATH") + "/", "") - sfiles.append(path + "/" + filename) - for tag in projectInfo.info("CPU_INFOS")["CPU_TAGS"]: - for filename, path in findDefinitions(module + "_" + tag + ".s", projectInfo): - path = path.replace(projectInfo.info("SOURCES_PATH") + "/", "") - sfiles.append(path + "/" + filename) - for filename, path in findDefinitions(module + ".S", projectInfo): - path = path.replace(projectInfo.info("SOURCES_PATH") + "/", "") - sfiles.append(path + "/" + filename) - for filename, path in findDefinitions(module + "_" + projectInfo.info("CPU_INFOS")["TOOLCHAIN"] + ".S", projectInfo): - path = path.replace(projectInfo.info("SOURCES_PATH") + "/", "") - sfiles.append(path + "/" + filename) - for tag in projectInfo.info("CPU_INFOS")["CPU_TAGS"]: - for filename, path in findDefinitions(module + "_" + tag + ".S", projectInfo): - path = path.replace(projectInfo.info("SOURCES_PATH") + "/", "") + # .c and .s files related to the module and the cpu tags + for tag in project_info.info("CPU_INFOS")["CPU_TAGS"]: + for filename, path in findDefinitions(module + "_" + tag + ".c", project_info): + path = path.replace(project_info.info("SOURCES_PATH") + os.sep, "") + if os.sep != "/": + path = replaceSeparators(path) + cfiles.append(path + "/" + filename) + for filename, path in findDefinitions(module + "_" + tag + ".s", project_info) + \ + findDefinitions(module + "_" + tag + ".S", project_info): + path = path.replace(project_info.info("SOURCES_PATH") + os.sep, "") + path = replaceSeparators(path) sfiles.append(path + "/" + filename) return cfiles, sfiles -def codeliteProjectGenerator(projectInfo): - template = open("cltemplates/bertos.project").read() - filelist = "\n".join(codelite_project.clFiles(codelite_project.findSources(projectInfo.info("PROJECT_PATH")), projectInfo.info("PROJECT_PATH"))) - while template.find("$filelist") != -1: - template = template.replace("$filelist", filelist) - projectName = os.path.basename(projectInfo.info("PROJECT_PATH")) - while template.find("$project") != -1: - template = template.replace("$project", projectName) - return template - -def codeliteWorkspaceGenerator(projectInfo): - template = open("cltemplates/bertos.workspace").read() - projectName = os.path.basename(projectInfo.info("PROJECT_PATH")) - while template.find("$project") != -1: - template = template.replace("$project", projectName) - return template - +def replaceSeparators(path): + """ + Replace the separators in the given path with unix standard separator. + """ + if os.sep != "/": + while path.find(os.sep) != -1: + path = path.replace(os.sep, "/") + return path + def getSystemPath(): path = os.environ["PATH"] if os.name == "nt": @@ -205,9 +393,9 @@ def getSystemPath(): path = path.split(":") return path -def findToolchains(pathList): +def findToolchains(path_list): toolchains = [] - for element in pathList: + for element in path_list: for toolchain in glob.glob(element+ "/" + const.GCC_NAME): toolchains.append(toolchain) return list(set(toolchains)) @@ -238,6 +426,10 @@ def getToolchainInfo(output): info["thread"] = thread[0].split("Thread model: ")[1] return info +def getToolchainName(toolchain_info): + name = "GCC " + toolchain_info["version"] + " - " + toolchain_info["target"].strip() + return name + def loadSourceTree(project): fileList = [f for f in os.walk(project.info("SOURCES_PATH"))] project.setInfo("FILE_LIST", fileList) @@ -257,6 +449,15 @@ def loadCpuInfos(project): cpuInfos.append(getInfos(definition)) return cpuInfos +def getTagSet(cpu_info): + tag_set = set([]) + for cpu in cpu_info: + tag_set |= set([cpu["CPU_NAME"]]) + tag_set |= set(cpu["CPU_TAGS"]) + tag_set |= set([cpu["TOOLCHAIN"]]) + return tag_set + + def getInfos(definition): D = {} D.update(const.CPU_DEF) @@ -270,71 +471,100 @@ def getInfos(definition): return D def getCommentList(string): - commentList = re.findall(r"/\*{2}\s*([^*]*\*(?:[^/*][^*]*\*+)*)/", string) - commentList = [re.findall(r"^\s*\* *(.*?)$", comment, re.MULTILINE) for comment in commentList] - return commentList + comment_list = re.findall(r"/\*{2}\s*([^*]*\*(?:[^/*][^*]*\*+)*)/", string) + comment_list = [re.findall(r"^\s*\* *(.*?)$", comment, re.MULTILINE) for comment in comment_list] + return comment_list def loadModuleDefinition(first_comment): - toBeParsed = False - moduleDefinition = {} + to_be_parsed = False + module_definition = {} for num, line in enumerate(first_comment): index = line.find("$WIZ$") if index != -1: - toBeParsed = True + to_be_parsed = True try: - exec line[index + len("$WIZ$ "):] in {}, moduleDefinition + exec line[index + len("$WIZ$ "):] in {}, module_definition except: raise ParseError(num, line[index:]) elif line.find("\\brief") != -1: - moduleDefinition["module_description"] = line[line.find("\\brief") + len("\\brief "):] - moduleDict = {} - if "module_name" in moduleDefinition.keys(): - moduleName = moduleDefinition[const.MODULE_DEFINITION["module_name"]] - del moduleDefinition[const.MODULE_DEFINITION["module_name"]] - moduleDict[moduleName] = {} - if const.MODULE_DEFINITION["module_depends"] in moduleDefinition.keys(): - if type(moduleDefinition[const.MODULE_DEFINITION["module_depends"]]) == str: - moduleDefinition[const.MODULE_DEFINITION["module_depends"]] = (moduleDefinition[const.MODULE_DEFINITION["module_depends"]],) - moduleDict[moduleName]["depends"] = moduleDefinition[const.MODULE_DEFINITION["module_depends"]] - del moduleDefinition[const.MODULE_DEFINITION["module_depends"]] + module_definition["module_description"] = line[line.find("\\brief") + len("\\brief "):] + module_dict = {} + if "module_name" in module_definition: + module_name = module_definition[const.MODULE_DEFINITION["module_name"]] + del module_definition[const.MODULE_DEFINITION["module_name"]] + module_dict[module_name] = {} + if const.MODULE_DEFINITION["module_depends"] in module_definition: + depends = module_definition[const.MODULE_DEFINITION["module_depends"]] + del module_definition[const.MODULE_DEFINITION["module_depends"]] + if type(depends) == str: + depends = (depends,) + module_dict[module_name]["depends"] = depends + else: + module_dict[module_name]["depends"] = () + if const.MODULE_DEFINITION["module_configuration"] in module_definition: + module_dict[module_name]["configuration"] = module_definition[const.MODULE_DEFINITION["module_configuration"]] + del module_definition[const.MODULE_DEFINITION["module_configuration"]] else: - moduleDict[moduleName]["depends"] = () - if const.MODULE_DEFINITION["module_configuration"] in moduleDefinition.keys(): - moduleDict[moduleName]["configuration"] = moduleDefinition[const.MODULE_DEFINITION["module_configuration"]] - del moduleDefinition[const.MODULE_DEFINITION["module_configuration"]] + module_dict[module_name]["configuration"] = "" + if "module_description" in module_definition: + module_dict[module_name]["description"] = module_definition["module_description"] + del module_definition["module_description"] + if const.MODULE_DEFINITION["module_harvard"] in module_definition: + harvard = module_definition[const.MODULE_DEFINITION["module_harvard"]] + module_dict[module_name]["harvard"] = harvard + del module_definition[const.MODULE_DEFINITION["module_harvard"]] + if const.MODULE_DEFINITION["module_hw"] in module_definition: + hw = module_definition[const.MODULE_DEFINITION["module_hw"]] + del module_definition[const.MODULE_DEFINITION["module_hw"]] + if type(hw) == str: + hw = (hw, ) + module_dict[module_name]["hw"] = hw else: - moduleDict[moduleName]["configuration"] = "" - if "module_description" in moduleDefinition.keys(): - moduleDict[moduleName]["description"] = moduleDefinition["module_description"] - del moduleDefinition["module_description"] - if const.MODULE_DEFINITION["module_harvard"] in moduleDefinition.keys(): - harvard = moduleDefinition[const.MODULE_DEFINITION["module_harvard"]] - if harvard == "both" or harvard == "pgm_memory": - moduleDict[moduleName]["harvard"] = harvard - del moduleDefinition[const.MODULE_DEFINITION["module_harvard"]] - moduleDict[moduleName]["constants"] = moduleDefinition - moduleDict[moduleName]["enabled"] = False - return toBeParsed, moduleDict - -def loadDefineLists(commentList): - defineList = {} - for comment in commentList: + module_dict[module_name]["hw"] = () + if const.MODULE_DEFINITION["module_supports"] in module_definition: + supports = module_definition[const.MODULE_DEFINITION["module_supports"]] + del module_definition[const.MODULE_DEFINITION["module_supports"]] + module_dict[module_name]["supports"] = supports + module_dict[module_name]["constants"] = module_definition + module_dict[module_name]["enabled"] = False + return to_be_parsed, module_dict + +def isSupported(project, module=None, property_id=None): + if not module and property_id: + item = project.info("CONFIGURATIONS")[property_id[0]][property_id[1]]["informations"] + else: + item = project.info("MODULES")[module] + tag_dict = project.info("ALL_CPU_TAGS") + if "supports" in item: + support_string = item["supports"] + supported = {} + try: + exec "supported = " + support_string in tag_dict, supported + except: + raise SupportedException(support_string) + return supported["supported"] + else: + return True + +def loadDefineLists(comment_list): + define_list = {} + for comment in comment_list: for num, line in enumerate(comment): index = line.find("$WIZ$") if index != -1: try: - exec line[index + len("$WIZ$ "):] in {}, defineList + exec line[index + len("$WIZ$ "):] in {}, define_list except: raise ParseError(num, line[index:]) - for key, value in defineList.items(): + for key, value in define_list.items(): if type(value) == str: - defineList[key] = (value,) - return defineList + define_list[key] = (value,) + return define_list -def getDescriptionInformations(comment): - """ - Take the doxygen comment and strip the wizard informations, returning the tuple - (comment, wizard_information) +def getDescriptionInformations(comment): + """ + Take the doxygen comment and strip the wizard informations, returning the tuple + (comment, wizard_information) """ brief = "" description = "" @@ -363,58 +593,78 @@ def getDefinitionBlocks(text): Take a text and return a list of tuple (description, name-value). """ block = [] - block_tmp = re.findall(r"/\*{2}\s*([^*]*\*(?:[^/*][^*]*\*+)*)/\s*#define\s+((?:[^/]*?/?)+)\s*?(?:/{2,3}[^<].*?)?$", text, re.MULTILINE) - for comment, define in block_tmp: + block_tmp = re.finditer(r"/\*{2}\s*([^*]*\*(?:[^/*][^*]*\*+)*)/\s*#define\s+((?:[^/]*?/?)+)\s*?(?:/{2,3}[^<].*?)?$", text, re.MULTILINE) + for match in block_tmp: # Only the first element is needed - block.append(([re.findall(r"^\s*\* *(.*?)$", line, re.MULTILINE)[0] for line in comment.splitlines()], define)) - for comment, define in re.findall(r"/{3}\s*([^<].*?)\s*#define\s+((?:[^/]*?/?)+)\s*?(?:/{2,3}[^<].*?)?$", text, re.MULTILINE): - block.append(([comment], define)) - for define, comment in re.findall(r"#define\s*(.*?)\s*/{3}<\s*(.+?)\s*?(?:/{2,3}[^<].*?)?$", text, re.MULTILINE): - block.append(([comment], define)) + comment = match.group(1) + define = match.group(2) + start = match.start() + block.append(([re.findall(r"^\s*\* *(.*?)$", line, re.MULTILINE)[0] for line in comment.splitlines()], define, start)) + for match in re.finditer(r"/{3}\s*([^<].*?)\s*#define\s+((?:[^/]*?/?)+)\s*?(?:/{2,3}[^<].*?)?$", text, re.MULTILINE): + comment = match.group(1) + define = match.group(2) + start = match.start() + block.append(([comment], define, start)) + for match in re.finditer(r"#define\s*(.*?)\s*/{3}<\s*(.+?)\s*?(?:/{2,3}[^<].*?)?$", text, re.MULTILINE): + comment = match.group(2) + define = match.group(1) + start = match.start() + block.append(([comment], define, start)) return block -def loadModuleData(project): - moduleInfoDict = {} - listInfoDict = {} - configurationInfoDict = {} - fileDict = {} - for filename, path in findDefinitions("*.h", project): - commentList = getCommentList(open(path + "/" + filename, "r").read()) - if len(commentList) > 0: - moduleInfo = {} - configurationInfo = {} +def loadModuleData(project, edit=False): + module_info_dict = {} + list_info_dict = {} + configuration_info_dict = {} + file_dict = {} + for filename, path in findDefinitions("*.h", project) + findDefinitions("*.c", project) + findDefinitions("*.s", project) + findDefinitions("*.S", project): + comment_list = getCommentList(open(path + "/" + filename, "r").read()) + if len(comment_list) > 0: + module_info = {} + configuration_info = {} try: - toBeParsed, moduleDict = loadModuleDefinition(commentList[0]) + to_be_parsed, module_dict = loadModuleDefinition(comment_list[0]) except ParseError, err: raise DefineException.ModuleDefineException(path, err.line_number, err.line) - for module, information in moduleDict.items(): + for module, information in module_dict.items(): + if "depends" not in information: + information["depends"] = () + information["depends"] += (filename.split(".")[0],) information["category"] = os.path.basename(path) - if "configuration" in information.keys() and len(information["configuration"]): - configuration = moduleDict[module]["configuration"] + if "configuration" in information and len(information["configuration"]): + configuration = module_dict[module]["configuration"] try: - configurationInfo[configuration] = loadConfigurationInfos(project.info("SOURCES_PATH") + "/" + configuration) + configuration_info[configuration] = loadConfigurationInfos(project.info("SOURCES_PATH") + "/" + configuration) except ParseError, err: raise DefineException.ConfigurationDefineException(project.info("SOURCES_PATH") + "/" + configuration, err.line_number, err.line) - moduleInfoDict.update(moduleDict) - configurationInfoDict.update(configurationInfo) - if toBeParsed: + if edit: + try: + path = os.path.basename(project.info("PROJECT_PATH")) + directory = project.info("PROJECT_PATH") + user_configuration = loadConfigurationInfos(directory + "/" + configuration.replace("bertos", path)) + configuration_info[configuration] = updateConfigurationValues(configuration_info[configuration], user_configuration) + except ParseError, err: + raise DefineException.ConfigurationDefineException(directory + "/" + configuration.replace("bertos", path)) + module_info_dict.update(module_dict) + configuration_info_dict.update(configuration_info) + if to_be_parsed: try: - listDict = loadDefineLists(commentList[1:]) - listInfoDict.update(listDict) + list_dict = loadDefineLists(comment_list[1:]) + list_info_dict.update(list_dict) except ParseError, err: raise DefineException.EnumDefineException(path, err.line_number, err.line) for filename, path in findDefinitions("*_" + project.info("CPU_INFOS")["TOOLCHAIN"] + ".h", project): - commentList = getCommentList(open(path + "/" + filename, "r").read()) - listInfoDict.update(loadDefineLists(commentList)) + comment_list = getCommentList(open(path + "/" + filename, "r").read()) + list_info_dict.update(loadDefineLists(comment_list)) for tag in project.info("CPU_INFOS")["CPU_TAGS"]: for filename, path in findDefinitions("*_" + tag + ".h", project): - commentList = getCommentList(open(path + "/" + filename, "r").read()) - listInfoDict.update(loadDefineLists(commentList)) - project.setInfo("MODULES", moduleInfoDict) - project.setInfo("LISTS", listInfoDict) - project.setInfo("CONFIGURATIONS", configurationInfoDict) - project.setInfo("FILES", fileDict) - + comment_list = getCommentList(open(path + "/" + filename, "r").read()) + list_info_dict.update(loadDefineLists(comment_list)) + project.setInfo("MODULES", module_info_dict) + project.setInfo("LISTS", list_info_dict) + project.setInfo("CONFIGURATIONS", configuration_info_dict) + project.setInfo("FILES", file_dict) + def formatParamNameValue(text): """ Take the given string and return a tuple with the name of the parameter in the first position @@ -434,28 +684,54 @@ def loadConfigurationInfos(path): "min": the minimum value for integer parameters "max": the maximum value for integer parameters "long": boolean indicating if the num is a long + "unsigned": boolean indicating if the num is an unsigned "value_list": the name of the enum for enum parameters + "conditional_deps": the list of conditional dependencies for boolean parameters """ - configurationInfos = {} - for comment, define in getDefinitionBlocks(open(path, "r").read()): + configuration_infos = {} + configuration_infos["paramlist"] = [] + for comment, define, start in getDefinitionBlocks(open(path, "r").read()): name, value = formatParamNameValue(define) brief, description, informations = getDescriptionInformations(comment) - configurationInfos[name] = {} - configurationInfos[name]["value"] = value - configurationInfos[name]["informations"] = informations - if ("type" in configurationInfos[name]["informations"].keys() and - configurationInfos[name]["informations"]["type"] == "int" and - configurationInfos[name]["value"].find("L") != -1): - configurationInfos[name]["informations"]["long"] = True - configurationInfos[name]["value"] = configurationInfos[name]["value"].replace("L", "") - if ("type" in configurationInfos[name]["informations"].keys() and - configurationInfos[name]["informations"]["type"] == "int" and - configurationInfos[name]["value"].find("U") != -1): - configurationInfos[name]["informations"]["unsigned"] = True - configurationInfos[name]["value"] = configurationInfos[name]["value"].replace("U", "") - configurationInfos[name]["description"] = description - configurationInfos[name]["brief"] = brief - return configurationInfos + configuration_infos["paramlist"].append((start, name)) + configuration_infos[name] = {} + configuration_infos[name]["value"] = value + configuration_infos[name]["informations"] = informations + if not "type" in configuration_infos[name]["informations"]: + configuration_infos[name]["informations"]["type"] = findParameterType(configuration_infos[name]) + if ("type" in configuration_infos[name]["informations"] and + configuration_infos[name]["informations"]["type"] == "int" and + configuration_infos[name]["value"].find("L") != -1): + configuration_infos[name]["informations"]["long"] = True + configuration_infos[name]["value"] = configuration_infos[name]["value"].replace("L", "") + if ("type" in configuration_infos[name]["informations"] and + configuration_infos[name]["informations"]["type"] == "int" and + configuration_infos[name]["value"].find("U") != -1): + configuration_infos[name]["informations"]["unsigned"] = True + configuration_infos[name]["value"] = configuration_infos[name]["value"].replace("U", "") + if "conditional_deps" in configuration_infos[name]["informations"]: + if (type(configuration_infos[name]["informations"]["conditional_deps"]) == str or + type(configuration_infos[name]["informations"]["conditional_deps"]) == unicode): + configuration_infos[name]["informations"]["conditional_deps"] = (configuration_infos[name]["informations"]["conditional_deps"], ) + elif type(configuration_infos[name]["informations"]["conditional_deps"]) == tuple: + pass + else: + configuration_infos[name]["informations"]["conditional_deps"] = () + configuration_infos[name]["description"] = description + configuration_infos[name]["brief"] = brief + return configuration_infos + +def updateConfigurationValues(def_conf, user_conf): + for param in def_conf["paramlist"]: + if param[1] in user_conf and "value" in user_conf[param[1]]: + def_conf[param[1]]["value"] = user_conf[param[1]]["value"] + return def_conf + +def findParameterType(parameter): + if "value_list" in parameter["informations"]: + return "enum" + if "min" in parameter["informations"] or "max" in parameter["informations"] or re.match(r"^\d+U?L?$", parameter["value"]) != None: + return "int" def sub(string, parameter, value): """ @@ -467,7 +743,7 @@ def isInt(informations): """ Return True if the value is a simple int. """ - if ("long" not in informatios.keys() or not informations["long"]) and ("unsigned" not in informations.keys() or informations["unsigned"]): + if ("long" not in informatios or not informations["long"]) and ("unsigned" not in informations or informations["unsigned"]): return True else: return False @@ -476,7 +752,7 @@ def isLong(informations): """ Return True if the value is a long. """ - if "long" in informations.keys() and informations["long"] and "unsigned" not in informations.keys(): + if "long" in informations and informations["long"] and "unsigned" not in informations: return True else: return False @@ -485,7 +761,7 @@ def isUnsigned(informations): """ Return True if the value is an unsigned. """ - if "unsigned" in informations.keys() and informations["unsigned"] and "long" not in informations.keys(): + if "unsigned" in informations and informations["unsigned"] and "long" not in informations: return True else: return False @@ -494,7 +770,7 @@ def isUnsignedLong(informations): """ Return True if the value is an unsigned long. """ - if "unsigned" in informations.keys() and "long" in informations.keys() and informations["unsigned"] and informations["long"]: + if "unsigned" in informations and "long" in informations and informations["unsigned"] and informations["long"]: return True else: return False @@ -504,3 +780,8 @@ class ParseError(Exception): Exception.__init__(self) self.line_number = line_number self.line = line + +class SupportedException(Exception): + def __init__(self, support_string): + Exception.__init__(self) + self.support_string = support_string