2017-09-21 21:42:57 +00:00
|
|
|
#!/usr/bin/env python3
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
# Copyright 2015 Ben Vanik. All Rights Reserved.
|
|
|
|
|
|
|
|
"""Main build script and tooling for xenia.
|
|
|
|
|
|
|
|
Run with --help or no arguments for possible commands.
|
|
|
|
"""
|
2017-09-21 21:42:57 +00:00
|
|
|
from __future__ import print_function
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
import argparse
|
2019-04-18 15:14:18 +00:00
|
|
|
import json
|
2015-08-01 06:48:24 +00:00
|
|
|
import os
|
|
|
|
import re
|
|
|
|
import shutil
|
|
|
|
import subprocess
|
|
|
|
import sys
|
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
__author__ = 'ben.vanik@gmail.com (Ben Vanik)'
|
|
|
|
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
self_path = os.path.dirname(os.path.abspath(__file__))
|
|
|
|
|
|
|
|
|
|
|
|
def main():
|
2017-09-21 21:42:57 +00:00
|
|
|
# Add self to the root search path.
|
|
|
|
sys.path.insert(0, self_path)
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
# Augment path to include our fancy things.
|
|
|
|
os.environ['PATH'] += os.pathsep + os.pathsep.join([
|
|
|
|
self_path,
|
|
|
|
os.path.abspath(os.path.join('tools', 'build')),
|
|
|
|
])
|
|
|
|
|
|
|
|
# Check git exists.
|
|
|
|
if not has_bin('git'):
|
|
|
|
print('ERROR: git must be installed and on PATH.')
|
|
|
|
sys.exit(1)
|
|
|
|
|
|
|
|
# Check python version.
|
|
|
|
if not sys.version_info[:2] >= (3, 4):
|
|
|
|
print('ERROR: Python 3.4+ must be installed and on PATH')
|
|
|
|
sys.exit(1)
|
|
|
|
|
|
|
|
# Grab Visual Studio version and execute shell to set up environment.
|
|
|
|
if sys.platform == 'win32':
|
|
|
|
vs_version = import_vs_environment()
|
|
|
|
if vs_version is None:
|
|
|
|
print('ERROR: Visual Studio not found!')
|
|
|
|
print('Please refer to the building guide:')
|
2019-06-25 03:32:12 +00:00
|
|
|
print('https://github.com/xenia-project/xenia/blob/master/docs/building.md')
|
2017-09-21 21:42:57 +00:00
|
|
|
sys.exit(1)
|
|
|
|
|
|
|
|
# Setup main argument parser and common arguments.
|
|
|
|
parser = argparse.ArgumentParser(prog='xenia-build')
|
|
|
|
|
|
|
|
# Grab all commands and populate the argument parser for each.
|
|
|
|
subparsers = parser.add_subparsers(title='subcommands',
|
|
|
|
dest='subcommand')
|
|
|
|
commands = discover_commands(subparsers)
|
|
|
|
|
|
|
|
# If the user passed no args, die nicely.
|
|
|
|
if len(sys.argv) == 1:
|
|
|
|
parser.print_help()
|
|
|
|
sys.exit(1)
|
|
|
|
|
|
|
|
# Gather any arguments that we want to pass to child processes.
|
|
|
|
command_args = sys.argv[1:]
|
|
|
|
pass_args = []
|
|
|
|
try:
|
|
|
|
pass_index = command_args.index('--')
|
|
|
|
pass_args = command_args[pass_index + 1:]
|
|
|
|
command_args = command_args[:pass_index]
|
|
|
|
except Exception:
|
|
|
|
pass
|
|
|
|
|
|
|
|
# Parse command name and dispatch.
|
|
|
|
args = vars(parser.parse_args(command_args))
|
|
|
|
command_name = args['subcommand']
|
|
|
|
try:
|
|
|
|
command = commands[command_name]
|
|
|
|
return_code = command.execute(args, pass_args, os.getcwd())
|
|
|
|
except Exception:
|
|
|
|
raise
|
|
|
|
sys.exit(return_code)
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
|
|
|
|
def import_vs_environment():
|
2017-09-21 21:42:57 +00:00
|
|
|
"""Finds the installed Visual Studio version and imports
|
|
|
|
interesting environment variables into os.environ.
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
Returns:
|
2019-04-18 15:14:18 +00:00
|
|
|
A version such as 2015 or None if no installation is found.
|
2017-09-21 21:42:57 +00:00
|
|
|
"""
|
|
|
|
version = 0
|
2019-04-18 15:14:18 +00:00
|
|
|
install_path = None
|
|
|
|
env_tool_args = None
|
|
|
|
|
2019-06-25 03:32:12 +00:00
|
|
|
vswhere = subprocess.check_output('third_party/vswhere/vswhere.exe -version "[15,)" -latest -prerelease -format json -utf8', shell=False, universal_newlines=True)
|
2019-04-18 15:14:18 +00:00
|
|
|
if vswhere:
|
|
|
|
vswhere = json.loads(vswhere)
|
|
|
|
if vswhere and len(vswhere) > 0:
|
|
|
|
version = int(vswhere[0].get("catalog", {}).get("productLineVersion", 2017))
|
|
|
|
install_path = vswhere[0].get("installationPath", None)
|
|
|
|
|
|
|
|
if version < 2017:
|
|
|
|
if 'VS140COMNTOOLS' in os.environ:
|
|
|
|
version = 2015
|
|
|
|
vcvars_path = os.environ['VS140COMNTOOLS']
|
|
|
|
vcvars_path = os.path.join(tools_path, '..\\..\\vc\\vcvarsall.bat')
|
|
|
|
env_tool_args = [vcvars_path, 'x64', '&&', 'set']
|
|
|
|
else:
|
2019-08-17 13:29:47 +00:00
|
|
|
vsdevcmd_path = os.path.join(install_path, 'Common7\\Tools\\VsDevCmd.bat')
|
|
|
|
if os.path.isfile(vsdevcmd_path) and os.access(vsdevcmd_path, os.X_OK):
|
2019-08-17 13:26:31 +00:00
|
|
|
env_tool_args = [vsdevcmd_path, '-arch=amd64', '-host_arch=amd64', '&&', 'set']
|
2019-08-17 13:29:47 +00:00
|
|
|
else:
|
|
|
|
vcvars_path = os.path.join(install_path, 'VC\\Auxiliary\\Build\\vcvarsall.bat')
|
|
|
|
env_tool_args = [vcvars_path, 'x64', '&&', 'set']
|
2017-09-21 21:42:57 +00:00
|
|
|
|
|
|
|
if version == 0:
|
|
|
|
return None
|
|
|
|
|
2019-04-18 15:14:18 +00:00
|
|
|
import_subprocess_environment(env_tool_args)
|
|
|
|
os.environ['VSVERSION'] = str(version)
|
|
|
|
return version
|
|
|
|
|
|
|
|
|
|
|
|
def import_subprocess_environment(args):
|
2017-09-21 21:42:57 +00:00
|
|
|
popen = subprocess.Popen(
|
|
|
|
args, shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, universal_newlines=True)
|
|
|
|
variables, _ = popen.communicate()
|
|
|
|
envvars_to_save = (
|
|
|
|
'devenvdir',
|
|
|
|
'include',
|
|
|
|
'lib',
|
|
|
|
'libpath',
|
|
|
|
'path',
|
|
|
|
'pathext',
|
|
|
|
'systemroot',
|
|
|
|
'temp',
|
|
|
|
'tmp',
|
|
|
|
'windowssdkdir',
|
|
|
|
)
|
|
|
|
for line in variables.splitlines():
|
|
|
|
for envvar in envvars_to_save:
|
|
|
|
if re.match(envvar + '=', line.lower()):
|
|
|
|
var, setting = line.split('=', 1)
|
|
|
|
if envvar == 'path':
|
|
|
|
setting = os.path.dirname(sys.executable) + os.pathsep + setting
|
|
|
|
os.environ[var.upper()] = setting
|
|
|
|
break
|
|
|
|
|
|
|
|
def has_bin(binary):
|
|
|
|
"""Checks whether the given binary is present.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
binary: binary name (without .exe, etc).
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
True if the binary exists.
|
|
|
|
"""
|
|
|
|
bin_path = get_bin(binary)
|
|
|
|
if not bin_path:
|
|
|
|
return False
|
|
|
|
return True
|
|
|
|
|
|
|
|
|
|
|
|
def get_bin(binary):
|
|
|
|
"""Checks whether the given binary is present and returns the path.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
binary: binary name (without .exe, etc).
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
Full path to the binary or None if not found.
|
|
|
|
"""
|
|
|
|
for path in os.environ['PATH'].split(os.pathsep):
|
|
|
|
path = path.strip('"')
|
|
|
|
exe_file = os.path.join(path, binary)
|
|
|
|
if os.path.isfile(exe_file) and os.access(exe_file, os.X_OK):
|
|
|
|
return exe_file
|
|
|
|
exe_file += '.exe'
|
|
|
|
if os.path.isfile(exe_file) and os.access(exe_file, os.X_OK):
|
|
|
|
return exe_file
|
|
|
|
return None
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
|
|
|
|
def shell_call(command, throw_on_error=True, stdout_path=None):
|
2017-09-21 21:42:57 +00:00
|
|
|
"""Executes a shell command.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
command: Command to execute, as a list of parameters.
|
|
|
|
throw_on_error: Whether to throw an error or return the status code.
|
|
|
|
stdout_path: File path to write stdout output to.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
If throw_on_error is False the status code of the call will be returned.
|
|
|
|
"""
|
|
|
|
stdout_file = None
|
|
|
|
if stdout_path:
|
|
|
|
stdout_file = open(stdout_path, 'w')
|
|
|
|
result = 0
|
|
|
|
try:
|
|
|
|
if throw_on_error:
|
|
|
|
result = 1
|
|
|
|
subprocess.check_call(command, shell=False, stdout=stdout_file)
|
|
|
|
result = 0
|
|
|
|
else:
|
|
|
|
result = subprocess.call(command, shell=False, stdout=stdout_file)
|
|
|
|
finally:
|
|
|
|
if stdout_file:
|
|
|
|
stdout_file.close()
|
|
|
|
return result
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
|
2015-12-27 19:53:37 +00:00
|
|
|
def get_git_head_info():
|
2017-09-21 21:42:57 +00:00
|
|
|
"""Queries the current branch and commit checksum from git.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
(branch_name, commit, commit_short)
|
|
|
|
If the user is not on any branch the name will be 'detached'.
|
|
|
|
"""
|
|
|
|
p = subprocess.Popen([
|
|
|
|
'git',
|
|
|
|
'symbolic-ref',
|
|
|
|
'--short',
|
|
|
|
'-q',
|
|
|
|
'HEAD',
|
|
|
|
], stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
|
|
|
(stdout, stderr) = p.communicate()
|
2017-12-16 05:43:27 +00:00
|
|
|
branch_name = stdout.decode('ascii').strip() or 'detached'
|
2020-01-12 05:30:09 +00:00
|
|
|
p = subprocess.Popen([
|
|
|
|
'git',
|
|
|
|
'rev-list',
|
|
|
|
'HEAD',
|
|
|
|
'--count',
|
|
|
|
], stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
|
|
|
(stdout, stderr) = p.communicate()
|
|
|
|
commit_number = stdout.decode('ascii').strip() or '0000'
|
2017-09-21 21:42:57 +00:00
|
|
|
p = subprocess.Popen([
|
|
|
|
'git',
|
|
|
|
'rev-parse',
|
|
|
|
'HEAD',
|
|
|
|
], stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
|
|
|
(stdout, stderr) = p.communicate()
|
2017-12-16 05:43:27 +00:00
|
|
|
commit = stdout.decode('ascii').strip() or 'unknown'
|
2017-09-21 21:42:57 +00:00
|
|
|
p = subprocess.Popen([
|
|
|
|
'git',
|
|
|
|
'rev-parse',
|
|
|
|
'--short',
|
|
|
|
'HEAD',
|
|
|
|
], stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
|
|
|
(stdout, stderr) = p.communicate()
|
2017-12-16 05:43:27 +00:00
|
|
|
commit_short = stdout.decode('ascii').strip() or 'unknown'
|
2020-01-12 05:30:09 +00:00
|
|
|
return branch_name, commit_number, commit, commit_short
|
2015-12-27 19:53:37 +00:00
|
|
|
|
|
|
|
|
|
|
|
def generate_version_h():
|
2017-09-21 21:42:57 +00:00
|
|
|
"""Generates a build/version.h file that contains current git info.
|
|
|
|
"""
|
2020-01-12 05:30:09 +00:00
|
|
|
(branch_name, commit_number, commit, commit_short) = get_git_head_info()
|
2017-09-21 21:42:57 +00:00
|
|
|
contents = '''// Autogenerated by `xb premake`.
|
|
|
|
#ifndef GENERATED_VERSION_H_
|
|
|
|
#define GENERATED_VERSION_H_
|
|
|
|
#define XE_BUILD_BRANCH "%s"
|
2020-01-12 05:30:09 +00:00
|
|
|
#define XE_BUILD_COMMIT_NUMBER "%s"
|
2017-09-21 21:42:57 +00:00
|
|
|
#define XE_BUILD_COMMIT "%s"
|
|
|
|
#define XE_BUILD_COMMIT_SHORT "%s"
|
|
|
|
#define XE_BUILD_DATE __DATE__
|
|
|
|
#endif // GENERATED_VERSION_H_
|
2020-01-12 05:30:09 +00:00
|
|
|
''' % (branch_name, commit_number, commit, commit_short)
|
2017-09-21 21:42:57 +00:00
|
|
|
with open('build/version.h', 'w') as f:
|
|
|
|
f.write(contents)
|
2015-12-27 19:53:37 +00:00
|
|
|
|
|
|
|
|
2015-08-01 07:05:55 +00:00
|
|
|
def git_submodule_update():
|
2017-09-21 21:42:57 +00:00
|
|
|
"""Runs a full recursive git submodule init and update.
|
2015-08-01 07:05:55 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
Older versions of git do not support 'update --init --recursive'. We could
|
|
|
|
check and run it on versions that do support it and speed things up a bit.
|
|
|
|
"""
|
|
|
|
if True:
|
|
|
|
shell_call([
|
|
|
|
'git',
|
|
|
|
'submodule',
|
|
|
|
'update',
|
|
|
|
'--init',
|
|
|
|
'--recursive',
|
|
|
|
])
|
|
|
|
else:
|
|
|
|
shell_call([
|
|
|
|
'git',
|
|
|
|
'submodule',
|
|
|
|
'init',
|
|
|
|
])
|
|
|
|
shell_call([
|
|
|
|
'git',
|
|
|
|
'submodule',
|
|
|
|
'foreach',
|
|
|
|
'--recursive',
|
|
|
|
'git',
|
|
|
|
'submodule',
|
|
|
|
'init',
|
|
|
|
])
|
|
|
|
shell_call([
|
|
|
|
'git',
|
|
|
|
'submodule',
|
|
|
|
'update',
|
|
|
|
'--recursive',
|
|
|
|
])
|
2015-08-01 07:05:55 +00:00
|
|
|
|
|
|
|
|
2015-08-01 07:41:46 +00:00
|
|
|
def get_clang_format_binary():
|
2017-09-21 21:42:57 +00:00
|
|
|
"""Finds a clang-format binary. Aborts if none is found.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
A path to the clang-format executable.
|
|
|
|
"""
|
|
|
|
attempts = [
|
|
|
|
'C:\\Program Files\\LLVM\\bin\\clang-format.exe',
|
|
|
|
'C:\\Program Files (x86)\\LLVM\\bin\\clang-format.exe',
|
|
|
|
'clang-format-3.8',
|
|
|
|
'clang-format',
|
|
|
|
]
|
|
|
|
for binary in attempts:
|
|
|
|
if has_bin(binary):
|
|
|
|
return binary
|
|
|
|
print('ERROR: clang-format is not on PATH')
|
2018-11-22 15:20:09 +00:00
|
|
|
print('LLVM is available from https://llvm.org/releases/download.html')
|
2017-09-21 21:42:57 +00:00
|
|
|
print('At least version 3.8 is required.')
|
|
|
|
print('See docs/style_guide.md for instructions on how to get it.')
|
|
|
|
sys.exit(1)
|
|
|
|
|
2015-08-01 07:41:46 +00:00
|
|
|
|
2018-04-04 00:02:49 +00:00
|
|
|
def run_premake(target_os, action, cc=None):
|
2017-09-21 21:42:57 +00:00
|
|
|
"""Runs premake on the main project with the given format.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
target_os: target --os to pass to premake.
|
|
|
|
action: action to preform.
|
|
|
|
"""
|
2019-04-18 10:26:20 +00:00
|
|
|
args = [
|
2018-04-29 01:15:16 +00:00
|
|
|
sys.executable,
|
2017-09-21 21:42:57 +00:00
|
|
|
os.path.join('tools', 'build', 'premake'),
|
|
|
|
'--file=premake5.lua',
|
|
|
|
'--os=%s' % target_os,
|
|
|
|
'--test-suite-mode=combined',
|
|
|
|
'--verbose',
|
|
|
|
action,
|
2019-04-18 10:26:20 +00:00
|
|
|
]
|
|
|
|
if cc:
|
|
|
|
args.insert(4, '--cc=%s' % cc)
|
|
|
|
|
|
|
|
ret = subprocess.call(args, shell=False)
|
2017-09-21 21:42:57 +00:00
|
|
|
|
|
|
|
if ret == 0:
|
|
|
|
generate_version_h()
|
|
|
|
|
|
|
|
return ret
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
def run_premake_clean():
|
2017-09-21 21:42:57 +00:00
|
|
|
"""Runs a premake clean operation.
|
|
|
|
"""
|
|
|
|
if sys.platform == 'darwin':
|
|
|
|
return run_premake('macosx', 'clean')
|
|
|
|
elif sys.platform == 'win32':
|
|
|
|
return run_premake('windows', 'clean')
|
|
|
|
else:
|
|
|
|
return run_premake('linux', 'clean')
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
|
2018-04-04 00:02:49 +00:00
|
|
|
def run_platform_premake(cc=None):
|
2017-09-21 21:42:57 +00:00
|
|
|
"""Runs all gyp configurations.
|
|
|
|
"""
|
|
|
|
if sys.platform == 'darwin':
|
|
|
|
return run_premake('macosx', 'xcode4')
|
|
|
|
elif sys.platform == 'win32':
|
|
|
|
vs_version = '2015'
|
|
|
|
if 'VSVERSION' in os.environ:
|
|
|
|
vs_version = os.environ['VSVERSION']
|
|
|
|
|
|
|
|
return run_premake('windows', 'vs' + vs_version)
|
|
|
|
else:
|
2018-04-04 00:02:49 +00:00
|
|
|
ret = run_premake('linux', 'gmake', cc)
|
2017-09-21 21:42:57 +00:00
|
|
|
ret = ret != 0 and run_premake('linux', 'codelite') or ret
|
|
|
|
return ret
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
|
2015-12-31 08:37:13 +00:00
|
|
|
def run_premake_export_commands():
|
2017-09-21 21:42:57 +00:00
|
|
|
"""Runs premake to generate an LLVM compile_commands.json file.
|
|
|
|
"""
|
|
|
|
# TODO(benvanik): only do linux? whatever clang-tidy is ok with.
|
|
|
|
if sys.platform == 'darwin':
|
|
|
|
run_premake('macosx', 'export-compile-commands')
|
|
|
|
elif sys.platform == 'win32':
|
|
|
|
run_premake('windows', 'export-compile-commands')
|
|
|
|
else:
|
|
|
|
run_premake('linux', 'export-compile-commands')
|
2015-12-31 08:37:13 +00:00
|
|
|
|
|
|
|
|
2015-08-01 06:48:24 +00:00
|
|
|
def get_build_bin_path(args):
|
2017-09-21 21:42:57 +00:00
|
|
|
"""Returns the path of the bin/ path with build results based on the
|
|
|
|
configuration specified in the parsed arguments.
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
Args:
|
|
|
|
args: Parsed arguments.
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
Returns:
|
|
|
|
A full path for the bin folder.
|
|
|
|
"""
|
|
|
|
if sys.platform == 'darwin':
|
|
|
|
platform = 'macosx'
|
|
|
|
elif sys.platform == 'win32':
|
|
|
|
platform = 'windows'
|
|
|
|
else:
|
|
|
|
platform = 'linux'
|
2018-02-15 04:57:45 +00:00
|
|
|
return os.path.join(self_path, 'build', 'bin', platform.capitalize(), args['config'].capitalize())
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
|
|
|
|
def discover_commands(subparsers):
|
2017-09-21 21:42:57 +00:00
|
|
|
"""Looks for all commands and returns a dictionary of them.
|
|
|
|
In the future commands could be discovered on disk.
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
Args:
|
|
|
|
subparsers: Argument subparsers parent used to add command parsers.
|
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
Returns:
|
|
|
|
A dictionary containing name-to-Command mappings.
|
|
|
|
"""
|
|
|
|
commands = {
|
|
|
|
'setup': SetupCommand(subparsers),
|
|
|
|
'pull': PullCommand(subparsers),
|
|
|
|
'premake': PremakeCommand(subparsers),
|
|
|
|
'build': BuildCommand(subparsers),
|
|
|
|
'genspirv': GenSpirvCommand(subparsers),
|
|
|
|
'gentests': GenTestsCommand(subparsers),
|
|
|
|
'test': TestCommand(subparsers),
|
|
|
|
'gputest': GpuTestCommand(subparsers),
|
|
|
|
'clean': CleanCommand(subparsers),
|
|
|
|
'nuke': NukeCommand(subparsers),
|
|
|
|
'lint': LintCommand(subparsers),
|
|
|
|
'format': FormatCommand(subparsers),
|
|
|
|
'style': StyleCommand(subparsers),
|
|
|
|
'tidy': TidyCommand(subparsers),
|
|
|
|
}
|
|
|
|
if sys.platform == 'win32':
|
|
|
|
commands['devenv'] = DevenvCommand(subparsers)
|
2018-07-18 10:16:12 +00:00
|
|
|
commands['buildhlsl'] = BuildHlslCommand(subparsers)
|
2017-09-21 21:42:57 +00:00
|
|
|
return commands
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
class Command(object):
|
|
|
|
"""Base type for commands.
|
2015-08-01 06:48:24 +00:00
|
|
|
"""
|
2017-09-21 21:42:57 +00:00
|
|
|
|
|
|
|
def __init__(self, subparsers, name, help_short=None, help_long=None,
|
|
|
|
*args, **kwargs):
|
|
|
|
"""Initializes a command.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
subparsers: Argument subparsers parent used to add command parsers.
|
|
|
|
name: The name of the command exposed to the management script.
|
|
|
|
help_short: Help text printed alongside the command when queried.
|
|
|
|
help_long: Extended help text when viewing command help.
|
|
|
|
"""
|
|
|
|
self.name = name
|
|
|
|
self.help_short = help_short
|
|
|
|
self.help_long = help_long
|
|
|
|
|
|
|
|
self.parser = subparsers.add_parser(name,
|
|
|
|
help=help_short,
|
|
|
|
description=help_long)
|
|
|
|
self.parser.set_defaults(command_handler=self)
|
|
|
|
|
|
|
|
def execute(self, args, pass_args, cwd):
|
|
|
|
"""Executes the command.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
args: Arguments hash for the command.
|
|
|
|
pass_args: Arguments list to pass to child commands.
|
|
|
|
cwd: Current working directory.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
Return code of the command.
|
|
|
|
"""
|
|
|
|
return 1
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
|
|
|
|
class SetupCommand(Command):
|
2017-09-21 21:42:57 +00:00
|
|
|
"""'setup' command."""
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
def __init__(self, subparsers, *args, **kwargs):
|
|
|
|
super(SetupCommand, self).__init__(
|
|
|
|
subparsers,
|
|
|
|
name='setup',
|
|
|
|
help_short='Setup the build environment.',
|
|
|
|
*args, **kwargs)
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
def execute(self, args, pass_args, cwd):
|
|
|
|
print('Setting up the build environment...')
|
|
|
|
print('')
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
# Setup submodules.
|
|
|
|
print('- git submodule init / update...')
|
|
|
|
git_submodule_update()
|
|
|
|
print('')
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
print('- running premake...')
|
|
|
|
if run_platform_premake() == 0:
|
|
|
|
print('')
|
|
|
|
print('Success!')
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
return 0
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
|
|
|
|
class PullCommand(Command):
|
2017-09-21 21:42:57 +00:00
|
|
|
"""'pull' command."""
|
|
|
|
|
|
|
|
def __init__(self, subparsers, *args, **kwargs):
|
|
|
|
super(PullCommand, self).__init__(
|
|
|
|
subparsers,
|
|
|
|
name='pull',
|
|
|
|
help_short='Pulls the repo and all dependencies and rebases changes.',
|
|
|
|
*args, **kwargs)
|
|
|
|
self.parser.add_argument('--merge', action='store_true',
|
|
|
|
help='Merges on master instead of rebasing.')
|
|
|
|
|
|
|
|
def execute(self, args, pass_args, cwd):
|
|
|
|
print('Pulling...')
|
|
|
|
print('')
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
print('- switching to master...')
|
|
|
|
shell_call([
|
|
|
|
'git',
|
|
|
|
'checkout',
|
|
|
|
'master',
|
|
|
|
])
|
|
|
|
print('')
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
print('- pulling self...')
|
|
|
|
if args['merge']:
|
|
|
|
shell_call([
|
|
|
|
'git',
|
|
|
|
'pull',
|
|
|
|
])
|
|
|
|
else:
|
|
|
|
shell_call([
|
|
|
|
'git',
|
|
|
|
'pull',
|
|
|
|
'--rebase',
|
|
|
|
])
|
|
|
|
print('')
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
print('- pulling dependencies...')
|
|
|
|
git_submodule_update()
|
|
|
|
print('')
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
print('- running premake...')
|
|
|
|
if run_platform_premake() == 0:
|
|
|
|
print('')
|
|
|
|
print('Success!')
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
return 0
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
class PremakeCommand(Command):
|
|
|
|
"""'premake' command."""
|
|
|
|
|
|
|
|
def __init__(self, subparsers, *args, **kwargs):
|
|
|
|
super(PremakeCommand, self).__init__(
|
|
|
|
subparsers,
|
|
|
|
name='premake',
|
|
|
|
help_short='Runs premake to update all projects.',
|
|
|
|
*args, **kwargs)
|
2018-04-04 00:02:49 +00:00
|
|
|
self.parser.add_argument(
|
|
|
|
'--cc', default='clang', help='Compiler toolchain passed to premake')
|
2017-09-21 21:42:57 +00:00
|
|
|
|
|
|
|
def execute(self, args, pass_args, cwd):
|
|
|
|
# Update premake. If no binary found, it will be built from source.
|
|
|
|
print('Running premake...')
|
|
|
|
print('')
|
2018-04-04 00:02:49 +00:00
|
|
|
if run_platform_premake(args['cc']) == 0:
|
2017-09-21 21:42:57 +00:00
|
|
|
print('Success!')
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
return 0
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
|
|
|
|
class BaseBuildCommand(Command):
|
2017-09-21 21:42:57 +00:00
|
|
|
"""Base command for things that require building."""
|
|
|
|
|
|
|
|
def __init__(self, subparsers, *args, **kwargs):
|
|
|
|
super(BaseBuildCommand, self).__init__(
|
|
|
|
subparsers,
|
|
|
|
*args, **kwargs)
|
2018-04-04 00:02:49 +00:00
|
|
|
self.parser.add_argument(
|
|
|
|
'--cc', default='clang', help='Compiler toolchain passed to premake')
|
2017-09-21 21:42:57 +00:00
|
|
|
self.parser.add_argument(
|
2019-10-20 20:51:32 +00:00
|
|
|
'--config', choices=['release', 'debug', 'checked'], default='release',
|
2017-09-21 21:42:57 +00:00
|
|
|
type=str.lower, help='Chooses the build configuration.')
|
2019-10-20 20:51:32 +00:00
|
|
|
self.parser.add_argument(
|
2020-01-17 15:16:11 +00:00
|
|
|
'--cpu_arch', choices=['avx', 'avx2', 'avx512'], default='avx',
|
2019-10-20 20:51:32 +00:00
|
|
|
type=str.lower, help='Chooses the CPU architecture.')
|
2017-09-21 21:42:57 +00:00
|
|
|
self.parser.add_argument(
|
|
|
|
'--target', action='append', default=[],
|
|
|
|
help='Builds only the given target(s).')
|
|
|
|
self.parser.add_argument(
|
|
|
|
'--force', action='store_true',
|
|
|
|
help='Forces a full rebuild.')
|
|
|
|
self.parser.add_argument(
|
|
|
|
'--no_premake', action='store_true',
|
|
|
|
help='Skips running premake before building.')
|
2017-12-20 01:25:25 +00:00
|
|
|
self.parser.add_argument(
|
2019-05-08 17:11:09 +00:00
|
|
|
'-j', default=4, type=int, help='Number of parallel threads')
|
2017-09-21 21:42:57 +00:00
|
|
|
|
|
|
|
def execute(self, args, pass_args, cwd):
|
|
|
|
if not args['no_premake']:
|
|
|
|
print('- running premake...')
|
2018-04-04 00:02:49 +00:00
|
|
|
run_platform_premake(args['cc'])
|
2017-09-21 21:42:57 +00:00
|
|
|
print('')
|
|
|
|
|
2017-12-20 01:25:25 +00:00
|
|
|
threads = args['j']
|
|
|
|
if threads < 0:
|
|
|
|
threads = 0
|
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
print('- building (%s):%s...' % (
|
|
|
|
'all' if not len(args['target']) else ', '.join(args['target']),
|
|
|
|
args['config']))
|
|
|
|
if sys.platform == 'win32':
|
|
|
|
targets = None
|
|
|
|
if len(args['target']):
|
|
|
|
targets = '/t:' + ';'.join(target + (':Rebuild' if args['force'] else '')
|
|
|
|
for target in args['target'])
|
|
|
|
else:
|
|
|
|
targets = '/t:Rebuild' if args['force'] else None
|
|
|
|
|
|
|
|
result = subprocess.call([
|
|
|
|
'msbuild',
|
|
|
|
'build/xenia.sln',
|
|
|
|
'/nologo',
|
|
|
|
'/m',
|
|
|
|
'/v:m',
|
|
|
|
'/p:Configuration=' + args['config'],
|
2019-10-20 20:51:32 +00:00
|
|
|
'/p:VCBuildAdditionalOptions=/arch:' + args['cpu_arch'],
|
2017-09-21 21:42:57 +00:00
|
|
|
] + ([targets] if targets is not None else []) + pass_args, shell=False)
|
|
|
|
elif sys.platform == 'darwin':
|
|
|
|
# TODO(benvanik): other platforms.
|
|
|
|
print('ERROR: don\'t know how to build on this platform.')
|
|
|
|
result = 1
|
|
|
|
else:
|
|
|
|
result = subprocess.call([
|
|
|
|
'make',
|
2017-12-20 01:25:25 +00:00
|
|
|
'-j' if threads is 0 else '-j%d' % threads,
|
2017-09-21 21:42:57 +00:00
|
|
|
'-Cbuild/',
|
|
|
|
'config=%s_linux' % (args['config']),
|
2017-12-20 01:25:25 +00:00
|
|
|
] + pass_args + args['target'], shell=False, env=dict(os.environ))
|
2017-09-21 21:42:57 +00:00
|
|
|
print('')
|
|
|
|
if result != 0:
|
|
|
|
print('ERROR: build failed with one or more errors.')
|
|
|
|
return result
|
|
|
|
return 0
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
|
|
|
|
class BuildCommand(BaseBuildCommand):
|
2017-09-21 21:42:57 +00:00
|
|
|
"""'build' command."""
|
|
|
|
|
|
|
|
def __init__(self, subparsers, *args, **kwargs):
|
|
|
|
super(BuildCommand, self).__init__(
|
|
|
|
subparsers,
|
|
|
|
name='build',
|
|
|
|
help_short='Builds the project.',
|
|
|
|
*args, **kwargs)
|
|
|
|
|
|
|
|
def execute(self, args, pass_args, cwd):
|
|
|
|
print('Building %s...' % (args['config']))
|
|
|
|
print('')
|
|
|
|
|
|
|
|
result = super(BuildCommand, self).execute(args, pass_args, cwd)
|
|
|
|
if not result:
|
|
|
|
print('Success!')
|
|
|
|
return result
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
|
2016-02-21 00:24:42 +00:00
|
|
|
class GenSpirvCommand(Command):
|
2017-09-21 21:42:57 +00:00
|
|
|
"""'genspirv' command."""
|
|
|
|
|
|
|
|
def __init__(self, subparsers, *args, **kwargs):
|
|
|
|
super(GenSpirvCommand, self).__init__(
|
|
|
|
subparsers,
|
|
|
|
name='genspirv',
|
|
|
|
help_short='Generates SPIR-V binaries and header files.',
|
|
|
|
help_long='''
|
|
|
|
Generates the .spv/.h binaries under src/xenia/*/vulkan/shaders/bin/).
|
|
|
|
Run after modifying any .vert/.geom/.frag files.
|
|
|
|
''',
|
|
|
|
*args, **kwargs)
|
|
|
|
|
|
|
|
def execute(self, args, pass_args, cwd):
|
|
|
|
print('Generating SPIR-V binaries...')
|
|
|
|
print('')
|
2016-02-21 00:24:42 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
vulkan_sdk_path = os.environ['VULKAN_SDK']
|
|
|
|
vulkan_bin_path = os.path.join(vulkan_sdk_path, 'bin')
|
|
|
|
glslang = os.path.join(vulkan_bin_path, 'glslangValidator')
|
|
|
|
spirv_dis = os.path.join(vulkan_bin_path, 'spirv-dis')
|
|
|
|
spirv_remap = os.path.join(vulkan_bin_path, 'spirv-remap')
|
|
|
|
|
|
|
|
# Ensure we have the tools.
|
|
|
|
if not os.path.exists(vulkan_sdk_path):
|
|
|
|
print('ERROR: could not find the Vulkan SDK')
|
|
|
|
return 1
|
|
|
|
elif not has_bin(glslang):
|
|
|
|
print('ERROR: could not find glslangValidator')
|
|
|
|
return 1
|
|
|
|
elif not has_bin(spirv_dis):
|
|
|
|
print('ERROR: could not find spirv-dis')
|
|
|
|
return 1
|
|
|
|
elif not has_bin(spirv_remap):
|
|
|
|
print('ERROR: could not find spirv-remap')
|
|
|
|
return 1
|
|
|
|
|
|
|
|
src_files = [os.path.join(root, name)
|
|
|
|
for root, dirs, files in os.walk('src')
|
|
|
|
for name in files
|
|
|
|
if (name.endswith('.vert') or name.endswith('.geom') or
|
|
|
|
name.endswith('.frag'))]
|
|
|
|
|
|
|
|
any_errors = False
|
|
|
|
for src_file in src_files:
|
|
|
|
print('- %s' % (src_file))
|
|
|
|
src_name = os.path.splitext(os.path.basename(src_file))[0]
|
|
|
|
identifier = os.path.basename(src_file).replace('.', '_')
|
|
|
|
|
|
|
|
bin_path = os.path.join(os.path.dirname(src_file), 'bin')
|
|
|
|
spv_file = os.path.join(bin_path, identifier) + '.spv'
|
|
|
|
txt_file = os.path.join(bin_path, identifier) + '.txt'
|
|
|
|
h_file = os.path.join(bin_path, identifier) + '.h'
|
|
|
|
|
|
|
|
# GLSL source -> .spv binary
|
|
|
|
shell_call([
|
|
|
|
glslang,
|
2018-11-26 17:46:09 +00:00
|
|
|
'-Os',
|
2017-09-21 21:42:57 +00:00
|
|
|
'-V', src_file,
|
|
|
|
'-o', spv_file,
|
|
|
|
])
|
|
|
|
|
|
|
|
# Disassemble binary into human-readable text.
|
|
|
|
shell_call([
|
|
|
|
spirv_dis,
|
|
|
|
'-o', txt_file,
|
|
|
|
spv_file,
|
|
|
|
])
|
|
|
|
|
|
|
|
# TODO(benvanik): remap?
|
|
|
|
|
|
|
|
# bin2c so we get a header file we can compile in.
|
2017-12-21 22:11:35 +00:00
|
|
|
with open(h_file, 'w') as out_file:
|
2017-09-21 21:42:57 +00:00
|
|
|
out_file.write('// generated from `xb genspirv`\n')
|
|
|
|
out_file.write('// source: %s\n' % os.path.basename(src_file))
|
|
|
|
out_file.write('const uint8_t %s[] = {' % (identifier))
|
|
|
|
with open(spv_file, 'rb') as in_file:
|
|
|
|
index = 0
|
|
|
|
c = in_file.read(1)
|
2017-12-21 22:11:35 +00:00
|
|
|
while len(c) != 0:
|
2017-09-21 21:42:57 +00:00
|
|
|
if index % 12 == 0:
|
|
|
|
out_file.write('\n ')
|
|
|
|
else:
|
|
|
|
out_file.write(' ')
|
|
|
|
index += 1
|
|
|
|
out_file.write('0x%02X,' % ord(c))
|
|
|
|
c = in_file.read(1)
|
|
|
|
out_file.write('\n};\n')
|
|
|
|
|
|
|
|
if any_errors:
|
|
|
|
print('ERROR: failed to build one or more SPIR-V files.')
|
|
|
|
return 1
|
2016-02-21 00:24:42 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
return 0
|
2016-02-21 00:24:42 +00:00
|
|
|
|
|
|
|
|
2018-07-18 10:16:12 +00:00
|
|
|
class BuildHlslCommand(Command):
|
|
|
|
"""'buildhlsl' command."""
|
|
|
|
|
|
|
|
def __init__(self, subparsers, *args, **kwargs):
|
|
|
|
super(BuildHlslCommand, self).__init__(
|
|
|
|
subparsers,
|
|
|
|
name='buildhlsl',
|
|
|
|
help_short='Generates Direct3D shader binaries and header files.',
|
|
|
|
help_long='''
|
2018-09-16 17:35:10 +00:00
|
|
|
Generates the .cso/.h binaries under src/xenia/*/d3d12/shaders/dxbc/.
|
2018-12-12 19:08:20 +00:00
|
|
|
Run after modifying any .hs/vs/ds/gs/ps/cs.hlsl files.
|
2018-07-18 10:16:12 +00:00
|
|
|
''',
|
|
|
|
*args, **kwargs)
|
|
|
|
|
|
|
|
def execute(self, args, pass_args, cwd):
|
|
|
|
print('Building Direct3D shaders...')
|
|
|
|
print('')
|
|
|
|
|
|
|
|
windows_sdk_bin_path = os.path.join(os.environ['ProgramFiles(x86)'],
|
|
|
|
'Windows Kits/10/bin/x64')
|
|
|
|
fxc = os.path.join(windows_sdk_bin_path, 'fxc')
|
|
|
|
|
|
|
|
# Ensure we have the tools.
|
|
|
|
if not os.path.exists(windows_sdk_bin_path):
|
|
|
|
print('ERROR: could not find Windows 10 SDK binaries')
|
|
|
|
return 1
|
|
|
|
elif not has_bin(fxc):
|
|
|
|
print('ERROR: could not find fxc')
|
|
|
|
return 1
|
|
|
|
|
|
|
|
src_files = [os.path.join(root, name)
|
|
|
|
for root, dirs, files in os.walk('src')
|
|
|
|
for name in files
|
2018-12-12 19:08:20 +00:00
|
|
|
if (name.endswith('.hs.hlsl') or
|
|
|
|
name.endswith('.vs.hlsl') or
|
|
|
|
name.endswith('.ds.hlsl') or
|
2018-07-18 10:16:12 +00:00
|
|
|
name.endswith('.gs.hlsl') or
|
2018-12-12 19:08:20 +00:00
|
|
|
name.endswith('.ps.hlsl') or
|
2018-07-18 10:16:12 +00:00
|
|
|
name.endswith('.cs.hlsl'))]
|
|
|
|
|
|
|
|
# TODO(Triang3l): Handle any_errors.
|
|
|
|
any_errors = False
|
|
|
|
for src_file in src_files:
|
|
|
|
print('- %s' % (src_file))
|
|
|
|
src_name = os.path.splitext(os.path.basename(src_file))[0]
|
|
|
|
identifier = os.path.basename(src_file)[:-5].replace('.', '_')
|
|
|
|
|
2018-09-16 17:35:10 +00:00
|
|
|
bin_path = os.path.join(os.path.dirname(src_file), 'dxbc')
|
2018-07-18 10:16:12 +00:00
|
|
|
if not os.path.exists(bin_path):
|
|
|
|
os.mkdir(bin_path)
|
|
|
|
cso_file = os.path.join(bin_path, identifier) + '.cso'
|
|
|
|
txt_file = os.path.join(bin_path, identifier) + '.txt'
|
|
|
|
h_file = os.path.join(bin_path, identifier) + '.h'
|
|
|
|
|
|
|
|
# HLSL source -> .cso binary and DXBC disassembly.
|
|
|
|
shell_call([
|
|
|
|
fxc,
|
|
|
|
'/nologo',
|
2018-07-20 17:46:01 +00:00
|
|
|
'/T', identifier[-2:] + '_5_1',
|
2018-07-18 10:16:12 +00:00
|
|
|
'/Fo', cso_file,
|
|
|
|
'/Fc', txt_file,
|
|
|
|
src_file,
|
|
|
|
])
|
|
|
|
|
|
|
|
# bin2c so we get a header file we can compile in.
|
|
|
|
with open(h_file, 'w') as out_file:
|
|
|
|
out_file.write('// generated from `xb buildhlsl`\n')
|
|
|
|
out_file.write('// source: %s\n' % os.path.basename(src_file))
|
|
|
|
out_file.write('const uint8_t %s[] = {' % (identifier))
|
|
|
|
with open(cso_file, 'rb') as in_file:
|
|
|
|
index = 0
|
|
|
|
c = in_file.read(1)
|
|
|
|
while len(c) != 0:
|
|
|
|
if index % 12 == 0:
|
|
|
|
out_file.write('\n ')
|
|
|
|
else:
|
|
|
|
out_file.write(' ')
|
|
|
|
index += 1
|
|
|
|
out_file.write('0x%02X,' % ord(c))
|
|
|
|
c = in_file.read(1)
|
|
|
|
out_file.write('\n};\n')
|
|
|
|
|
|
|
|
if any_errors:
|
|
|
|
print('ERROR: failed to build one or more shaders.')
|
|
|
|
return 1
|
|
|
|
|
|
|
|
return 0
|
|
|
|
|
|
|
|
|
2015-08-01 06:48:24 +00:00
|
|
|
class TestCommand(BaseBuildCommand):
|
2017-09-21 21:42:57 +00:00
|
|
|
"""'test' command."""
|
|
|
|
|
|
|
|
def __init__(self, subparsers, *args, **kwargs):
|
|
|
|
super(TestCommand, self).__init__(
|
|
|
|
subparsers,
|
|
|
|
name='test',
|
|
|
|
help_short='Runs automated tests that have been built with `xb build`.',
|
|
|
|
help_long='''
|
|
|
|
To pass arguments to the test executables separate them with `--`.
|
|
|
|
For example, you can run only the instr_foo.s tests with:
|
|
|
|
$ xb test -- instr_foo
|
|
|
|
''',
|
|
|
|
*args, **kwargs)
|
|
|
|
self.parser.add_argument(
|
|
|
|
'--no_build', action='store_true',
|
|
|
|
help='Don\'t build before running tests.')
|
|
|
|
self.parser.add_argument(
|
|
|
|
'--continue', action='store_true',
|
|
|
|
help='Don\'t stop when a test errors, but continue running all.')
|
|
|
|
|
|
|
|
def execute(self, args, pass_args, cwd):
|
|
|
|
print('Testing...')
|
|
|
|
print('')
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
# The test executables that will be built and run.
|
|
|
|
test_targets = args['target'] or [
|
|
|
|
'xenia-cpu-ppc-tests',
|
|
|
|
]
|
|
|
|
args['target'] = test_targets
|
|
|
|
|
|
|
|
# Build all targets (if desired).
|
|
|
|
if not args['no_build']:
|
|
|
|
result = super(TestCommand, self).execute(args, [], cwd)
|
|
|
|
if result:
|
|
|
|
print('Failed to build, aborting test run.')
|
|
|
|
return result
|
|
|
|
|
|
|
|
# Ensure all targets exist before we run.
|
|
|
|
test_executables = [
|
|
|
|
get_bin(os.path.join(get_build_bin_path(args), test_target))
|
|
|
|
for test_target in test_targets]
|
|
|
|
for test_executable in test_executables:
|
|
|
|
if not has_bin(test_executable):
|
|
|
|
print('ERROR: Unable to find %s - build it.' % (test_executable))
|
|
|
|
return 1
|
|
|
|
|
|
|
|
# Run tests.
|
|
|
|
any_failed = False
|
|
|
|
for test_executable in test_executables:
|
|
|
|
print('- %s' % test_executable)
|
|
|
|
result = shell_call([test_executable] + pass_args,
|
|
|
|
throw_on_error=False)
|
|
|
|
if result:
|
|
|
|
any_failed = True
|
|
|
|
if args['continue']:
|
|
|
|
print('ERROR: test failed but continuing due to --continue.')
|
|
|
|
else:
|
|
|
|
print('ERROR: test failed, aborting, use --continue to keep going.')
|
|
|
|
return result
|
|
|
|
|
|
|
|
if any_failed:
|
|
|
|
print('ERROR: one or more tests failed.')
|
|
|
|
result = 1
|
2015-08-01 06:48:24 +00:00
|
|
|
return result
|
|
|
|
|
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
class GenTestsCommand(Command):
|
|
|
|
"""'gentests' command."""
|
|
|
|
|
|
|
|
def __init__(self, subparsers, *args, **kwargs):
|
|
|
|
super(GenTestsCommand, self).__init__(
|
|
|
|
subparsers,
|
|
|
|
name='gentests',
|
|
|
|
help_short='Generates test binaries.',
|
|
|
|
help_long='''
|
|
|
|
Generates test binaries (under src/xenia/cpu/ppc/testing/bin/).
|
|
|
|
Run after modifying test .s files.
|
|
|
|
''',
|
|
|
|
*args, **kwargs)
|
|
|
|
|
|
|
|
def execute(self, args, pass_args, cwd):
|
|
|
|
print('Generating test binaries...')
|
|
|
|
print('')
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
binutils_path = os.path.join('third_party', 'binutils-ppc-cygwin')
|
|
|
|
ppc_as = os.path.join(binutils_path, 'powerpc-none-elf-as')
|
|
|
|
ppc_ld = os.path.join(binutils_path, 'powerpc-none-elf-ld')
|
|
|
|
ppc_objdump = os.path.join(binutils_path, 'powerpc-none-elf-objdump')
|
|
|
|
ppc_nm = os.path.join(binutils_path, 'powerpc-none-elf-nm')
|
|
|
|
|
|
|
|
test_src = os.path.join('src', 'xenia', 'cpu', 'ppc', 'testing')
|
|
|
|
test_bin = os.path.join(test_src, 'bin')
|
|
|
|
|
|
|
|
# Ensure the test output path exists.
|
|
|
|
if not os.path.exists(test_bin):
|
|
|
|
os.mkdir(test_bin)
|
|
|
|
|
|
|
|
src_files = [os.path.join(root, name)
|
|
|
|
for root, dirs, files in os.walk('src')
|
|
|
|
for name in files
|
|
|
|
if (name.startswith('instr_') or name.startswith('seq_'))
|
|
|
|
and name.endswith(('.s'))]
|
|
|
|
|
|
|
|
def make_unix_path(p):
|
|
|
|
"""Forces a unix path separator style, as required by binutils.
|
|
|
|
"""
|
|
|
|
return p.replace(os.sep, '/')
|
|
|
|
|
|
|
|
any_errors = False
|
|
|
|
for src_file in src_files:
|
|
|
|
print('- %s' % src_file)
|
|
|
|
src_name = os.path.splitext(os.path.basename(src_file))[0]
|
|
|
|
obj_file = os.path.join(test_bin, src_name) + '.o'
|
|
|
|
shell_call([
|
|
|
|
ppc_as,
|
|
|
|
'-a32',
|
|
|
|
'-be',
|
|
|
|
'-mregnames',
|
|
|
|
'-mpower7',
|
|
|
|
'-maltivec',
|
|
|
|
'-mvsx',
|
|
|
|
'-mvmx128',
|
|
|
|
'-R',
|
|
|
|
'-o%s' % (make_unix_path(obj_file)),
|
|
|
|
make_unix_path(src_file),
|
|
|
|
])
|
|
|
|
dis_file = os.path.join(test_bin, src_name) + '.dis'
|
|
|
|
shell_call([
|
|
|
|
ppc_objdump,
|
|
|
|
'--adjust-vma=0x100000',
|
|
|
|
'-Mpower7',
|
|
|
|
'-Mvmx128',
|
|
|
|
'-D',
|
|
|
|
'-EB',
|
|
|
|
make_unix_path(obj_file),
|
|
|
|
], stdout_path=dis_file)
|
|
|
|
# Eat the first 4 lines to kill the file path that'll differ across machines.
|
|
|
|
with open(dis_file) as f:
|
|
|
|
dis_file_lines = f.readlines()
|
|
|
|
with open(dis_file, 'w') as f:
|
|
|
|
f.writelines(dis_file_lines[4:])
|
|
|
|
shell_call([
|
|
|
|
ppc_ld,
|
|
|
|
'-A powerpc:common32',
|
|
|
|
'-melf32ppc',
|
|
|
|
'-EB',
|
|
|
|
'-nostdlib',
|
|
|
|
'--oformat=binary',
|
|
|
|
'-Ttext=0x80000000',
|
|
|
|
'-e0x80000000',
|
|
|
|
'-o%s' % (make_unix_path(os.path.join(test_bin, src_name) + '.bin')),
|
|
|
|
make_unix_path(obj_file),
|
|
|
|
])
|
|
|
|
shell_call([
|
|
|
|
ppc_nm,
|
|
|
|
'--numeric-sort',
|
|
|
|
make_unix_path(obj_file),
|
|
|
|
], stdout_path=os.path.join(test_bin, src_name) + '.map')
|
|
|
|
|
|
|
|
if any_errors:
|
|
|
|
print('ERROR: failed to build one or more tests.')
|
|
|
|
return 1
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
return 0
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
|
2015-12-31 20:27:26 +00:00
|
|
|
class GpuTestCommand(BaseBuildCommand):
|
2017-09-21 21:42:57 +00:00
|
|
|
"""'gputest' command."""
|
|
|
|
|
|
|
|
def __init__(self, subparsers, *args, **kwargs):
|
|
|
|
super(GpuTestCommand, self).__init__(
|
|
|
|
subparsers,
|
|
|
|
name='gputest',
|
|
|
|
help_short='Runs automated GPU diff tests against reference imagery.',
|
|
|
|
help_long='''
|
|
|
|
To pass arguments to the test executables separate them with `--`.
|
|
|
|
''',
|
|
|
|
*args, **kwargs)
|
|
|
|
self.parser.add_argument(
|
|
|
|
'--no_build', action='store_true',
|
|
|
|
help='Don\'t build before running tests.')
|
|
|
|
self.parser.add_argument(
|
|
|
|
'--update_reference_files', action='store_true',
|
|
|
|
help='Update all reference imagery.')
|
|
|
|
self.parser.add_argument(
|
|
|
|
'--generate_missing_reference_files', action='store_true',
|
|
|
|
help='Create reference files for new traces.')
|
|
|
|
|
|
|
|
def execute(self, args, pass_args, cwd):
|
|
|
|
print('Testinging...')
|
|
|
|
print('')
|
2015-12-31 20:27:26 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
# The test executables that will be built and run.
|
|
|
|
test_targets = args['target'] or [
|
2019-05-08 15:31:04 +00:00
|
|
|
'xenia-gpu-vulkan-trace-dump',
|
2017-09-21 21:42:57 +00:00
|
|
|
]
|
|
|
|
args['target'] = test_targets
|
|
|
|
|
|
|
|
# Build all targets (if desired).
|
|
|
|
if not args['no_build']:
|
|
|
|
result = super(GpuTestCommand, self).execute(args, [], cwd)
|
|
|
|
if result:
|
|
|
|
print('Failed to build, aborting test run.')
|
|
|
|
return result
|
|
|
|
|
|
|
|
# Ensure all targets exist before we run.
|
|
|
|
test_executables = [
|
|
|
|
get_bin(os.path.join(get_build_bin_path(args), test_target))
|
|
|
|
for test_target in test_targets]
|
|
|
|
for test_executable in test_executables:
|
|
|
|
if not has_bin(test_executable):
|
|
|
|
print('ERROR: Unable to find %s - build it.' % (test_executable))
|
|
|
|
return 1
|
|
|
|
|
|
|
|
output_path = os.path.join(self_path, 'build', 'gputest')
|
|
|
|
if os.path.isdir(output_path):
|
|
|
|
shutil.rmtree(output_path)
|
|
|
|
os.makedirs(output_path)
|
|
|
|
print('Running tests and outputting to %s...' % (output_path))
|
|
|
|
|
|
|
|
reference_trace_root = os.path.join(self_path, 'testdata',
|
|
|
|
'reference-gpu-traces')
|
|
|
|
|
|
|
|
# Run tests.
|
|
|
|
any_failed = False
|
|
|
|
result = shell_call([
|
2018-04-29 01:15:16 +00:00
|
|
|
sys.executable,
|
2017-09-21 21:42:57 +00:00
|
|
|
os.path.join(self_path, 'tools', 'gpu-trace-diff'),
|
|
|
|
'--executable=' + test_executables[0],
|
|
|
|
'--trace_path=' + os.path.join(reference_trace_root, 'traces'),
|
|
|
|
'--output_path=' + output_path,
|
|
|
|
'--reference_path=' + os.path.join(reference_trace_root, 'references'),
|
|
|
|
] + (['--generate_missing_reference_files'] if args['generate_missing_reference_files'] else []) +
|
|
|
|
(['--update_reference_files'] if args['update_reference_files'] else []) +
|
|
|
|
pass_args,
|
|
|
|
throw_on_error=False)
|
|
|
|
if result:
|
|
|
|
any_failed = True
|
|
|
|
|
|
|
|
if any_failed:
|
|
|
|
print('ERROR: one or more tests failed.')
|
|
|
|
result = 1
|
|
|
|
print('Check %s/results.html for more details.' % (output_path))
|
2015-12-31 20:27:26 +00:00
|
|
|
return result
|
|
|
|
|
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
class CleanCommand(Command):
|
|
|
|
"""'clean' command."""
|
2015-12-31 20:27:26 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
def __init__(self, subparsers, *args, **kwargs):
|
|
|
|
super(CleanCommand, self).__init__(
|
|
|
|
subparsers,
|
|
|
|
name='clean',
|
|
|
|
help_short='Removes intermediate files and build outputs.',
|
|
|
|
*args, **kwargs)
|
2015-12-31 20:27:26 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
def execute(self, args, pass_args, cwd):
|
|
|
|
print('Cleaning build artifacts...')
|
|
|
|
print('')
|
2015-12-31 20:27:26 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
print('- premake clean...')
|
|
|
|
run_premake_clean()
|
|
|
|
print('')
|
2015-12-31 20:27:26 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
print('Success!')
|
|
|
|
return 0
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
class NukeCommand(Command):
|
|
|
|
"""'nuke' command."""
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
def __init__(self, subparsers, *args, **kwargs):
|
|
|
|
super(NukeCommand, self).__init__(
|
|
|
|
subparsers,
|
|
|
|
name='nuke',
|
|
|
|
help_short='Removes all build/ output.',
|
|
|
|
*args, **kwargs)
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
def execute(self, args, pass_args, cwd):
|
|
|
|
print('Cleaning build artifacts...')
|
|
|
|
print('')
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
print('- removing build/...')
|
|
|
|
if os.path.isdir('build/'):
|
|
|
|
shutil.rmtree('build/')
|
|
|
|
print('')
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
print('- git reset to master...')
|
|
|
|
shell_call([
|
|
|
|
'git',
|
|
|
|
'reset',
|
|
|
|
'--hard',
|
|
|
|
'master',
|
|
|
|
])
|
|
|
|
print('')
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
print('- running premake...')
|
|
|
|
run_platform_premake()
|
|
|
|
print('')
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
print('Success!')
|
|
|
|
return 0
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
|
2015-08-01 17:49:58 +00:00
|
|
|
def find_xenia_source_files():
|
2017-09-21 21:42:57 +00:00
|
|
|
"""Gets all xenia source files in the project.
|
2015-08-01 09:51:49 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
Returns:
|
|
|
|
A list of file paths.
|
|
|
|
"""
|
|
|
|
return [os.path.join(root, name)
|
|
|
|
for root, dirs, files in os.walk('src')
|
|
|
|
for name in files
|
|
|
|
if name.endswith(('.cc', '.c', '.h', '.inl'))]
|
2015-08-01 17:49:58 +00:00
|
|
|
|
|
|
|
|
|
|
|
def find_all_source_files():
|
2017-09-21 21:42:57 +00:00
|
|
|
"""Gets all interesting source files in the project.
|
2015-08-01 17:49:58 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
Returns:
|
|
|
|
A list of file paths.
|
|
|
|
"""
|
|
|
|
return find_xenia_source_files()
|
2015-08-01 09:51:49 +00:00
|
|
|
|
|
|
|
|
2015-08-01 06:48:24 +00:00
|
|
|
class LintCommand(Command):
|
2017-09-21 21:42:57 +00:00
|
|
|
"""'lint' command."""
|
|
|
|
|
|
|
|
def __init__(self, subparsers, *args, **kwargs):
|
|
|
|
super(LintCommand, self).__init__(
|
|
|
|
subparsers,
|
|
|
|
name='lint',
|
|
|
|
help_short='Checks for lint errors with clang-format.',
|
|
|
|
*args, **kwargs)
|
|
|
|
self.parser.add_argument(
|
|
|
|
'--all', action='store_true',
|
|
|
|
help='Lint all files, not just those changed.')
|
|
|
|
self.parser.add_argument(
|
|
|
|
'--origin', action='store_true',
|
|
|
|
help='Lints all files changed relative to origin/master.')
|
|
|
|
|
|
|
|
def execute(self, args, pass_args, cwd):
|
|
|
|
clang_format_binary = get_clang_format_binary()
|
|
|
|
|
|
|
|
difftemp = '.difftemp.txt'
|
|
|
|
|
|
|
|
if args['all']:
|
|
|
|
all_files = find_all_source_files()
|
|
|
|
print('- linting %d files' % (len(all_files)))
|
|
|
|
any_errors = False
|
|
|
|
for file_path in all_files:
|
|
|
|
if os.path.exists(difftemp): os.remove(difftemp)
|
|
|
|
ret = shell_call([
|
|
|
|
clang_format_binary,
|
|
|
|
'-output-replacements-xml',
|
|
|
|
'-style=file',
|
|
|
|
file_path,
|
|
|
|
], throw_on_error=False, stdout_path=difftemp)
|
|
|
|
with open(difftemp) as f:
|
|
|
|
had_errors = '<replacement ' in f.read()
|
|
|
|
if os.path.exists(difftemp): os.remove(difftemp)
|
|
|
|
if had_errors:
|
|
|
|
any_errors = True
|
|
|
|
print('')
|
|
|
|
print(file_path)
|
|
|
|
shell_call([
|
|
|
|
clang_format_binary,
|
|
|
|
'-style=file',
|
|
|
|
file_path,
|
|
|
|
], throw_on_error=False, stdout_path=difftemp)
|
|
|
|
shell_call([
|
2018-04-29 01:15:16 +00:00
|
|
|
sys.executable,
|
2017-09-21 21:42:57 +00:00
|
|
|
'tools/diff.py',
|
|
|
|
file_path,
|
|
|
|
difftemp,
|
|
|
|
difftemp,
|
|
|
|
])
|
|
|
|
shell_call([
|
|
|
|
'type' if sys.platform == 'win32' else 'cat',
|
|
|
|
difftemp,
|
|
|
|
])
|
|
|
|
if os.path.exists(difftemp): os.remove(difftemp)
|
|
|
|
print('')
|
|
|
|
print('')
|
|
|
|
if any_errors:
|
|
|
|
print('ERROR: 1+ diffs. Stage changes and run \'xb format\' to fix.')
|
|
|
|
return 1
|
|
|
|
else:
|
|
|
|
print('Linting completed successfully.')
|
|
|
|
return 0
|
|
|
|
else:
|
|
|
|
print('- git-clang-format --diff')
|
|
|
|
if os.path.exists(difftemp): os.remove(difftemp)
|
|
|
|
ret = shell_call([
|
2018-04-29 01:15:16 +00:00
|
|
|
sys.executable,
|
2017-09-21 21:42:57 +00:00
|
|
|
'third_party/clang-format/git-clang-format',
|
|
|
|
'--binary=%s' % (clang_format_binary),
|
|
|
|
'--commit=%s' % ('origin/master' if args['origin'] else 'HEAD'),
|
|
|
|
'--diff',
|
|
|
|
], throw_on_error=False, stdout_path=difftemp)
|
|
|
|
with open(difftemp) as f:
|
|
|
|
contents = f.read()
|
|
|
|
not_modified = 'no modified files' in contents
|
|
|
|
not_modified = not_modified or 'did not modify' in contents
|
|
|
|
f.close()
|
|
|
|
if os.path.exists(difftemp): os.remove(difftemp)
|
|
|
|
if not not_modified:
|
|
|
|
any_errors = True
|
|
|
|
print('')
|
|
|
|
shell_call([
|
2018-04-29 01:15:16 +00:00
|
|
|
sys.executable,
|
2017-09-21 21:42:57 +00:00
|
|
|
'third_party/clang-format/git-clang-format',
|
|
|
|
'--binary=%s' % (clang_format_binary),
|
|
|
|
'--commit=%s' % ('origin/master' if args['origin'] else 'HEAD'),
|
|
|
|
'--diff',
|
|
|
|
])
|
|
|
|
print('ERROR: 1+ diffs. Stage changes and run \'xb format\' to fix.')
|
|
|
|
return 1
|
|
|
|
else:
|
|
|
|
print('Linting completed successfully.')
|
|
|
|
return 0
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
|
|
|
|
class FormatCommand(Command):
|
2017-09-21 21:42:57 +00:00
|
|
|
"""'format' command."""
|
|
|
|
|
|
|
|
def __init__(self, subparsers, *args, **kwargs):
|
|
|
|
super(FormatCommand, self).__init__(
|
|
|
|
subparsers,
|
|
|
|
name='format',
|
|
|
|
help_short='Reformats staged code with clang-format.',
|
|
|
|
*args, **kwargs)
|
|
|
|
self.parser.add_argument(
|
|
|
|
'--all', action='store_true',
|
|
|
|
help='Format all files, not just those changed.')
|
|
|
|
self.parser.add_argument(
|
|
|
|
'--origin', action='store_true',
|
|
|
|
help='Formats all files changed relative to origin/master.')
|
|
|
|
|
|
|
|
def execute(self, args, pass_args, cwd):
|
|
|
|
clang_format_binary = get_clang_format_binary()
|
|
|
|
|
|
|
|
if args['all']:
|
|
|
|
all_files = find_all_source_files()
|
|
|
|
print('- clang-format [%d files]' % (len(all_files)))
|
|
|
|
any_errors = False
|
|
|
|
for file_path in all_files:
|
|
|
|
ret = shell_call([
|
|
|
|
clang_format_binary,
|
|
|
|
'-i',
|
|
|
|
'-style=file',
|
|
|
|
file_path,
|
|
|
|
], throw_on_error=False)
|
|
|
|
if ret:
|
|
|
|
any_errors = True
|
|
|
|
print('')
|
|
|
|
if any_errors:
|
|
|
|
print('ERROR: 1+ clang-format calls failed.')
|
|
|
|
print('Ensure all files are staged.')
|
|
|
|
return 1
|
|
|
|
else:
|
|
|
|
print('Formatting completed successfully.')
|
|
|
|
return 0
|
|
|
|
else:
|
|
|
|
print('- git-clang-format')
|
|
|
|
shell_call([
|
2018-04-29 01:15:16 +00:00
|
|
|
sys.executable,
|
2017-09-21 21:42:57 +00:00
|
|
|
'third_party/clang-format/git-clang-format',
|
|
|
|
'--binary=%s' % (clang_format_binary),
|
|
|
|
'--commit=%s' % ('origin/master' if args['origin'] else 'HEAD'),
|
|
|
|
])
|
|
|
|
print('')
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
return 0
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
|
2015-08-01 17:49:58 +00:00
|
|
|
# TODO(benvanik): merge into linter, or as lint --anal?
|
|
|
|
class StyleCommand(Command):
|
2017-09-21 21:42:57 +00:00
|
|
|
"""'style' command."""
|
|
|
|
|
|
|
|
def __init__(self, subparsers, *args, **kwargs):
|
|
|
|
super(StyleCommand, self).__init__(
|
|
|
|
subparsers,
|
|
|
|
name='style',
|
|
|
|
help_short='Runs the style checker on all code.',
|
|
|
|
*args, **kwargs)
|
|
|
|
|
|
|
|
def execute(self, args, pass_args, cwd):
|
|
|
|
all_files = [file_path for file_path in find_all_source_files()
|
|
|
|
if not file_path.endswith('_test.cc')]
|
|
|
|
print('- cpplint [%d files]' % (len(all_files)))
|
|
|
|
ret = shell_call([
|
2018-04-29 01:15:16 +00:00
|
|
|
sys.executable,
|
2017-09-21 21:42:57 +00:00
|
|
|
'third_party/google-styleguide/cpplint/cpplint.py',
|
|
|
|
'--output=vs7',
|
|
|
|
'--linelength=80',
|
|
|
|
'--filter=-build/c++11,+build/include_alpha',
|
|
|
|
'--root=src',
|
|
|
|
] + all_files, throw_on_error=False)
|
|
|
|
print('')
|
|
|
|
if ret:
|
|
|
|
print('ERROR: 1+ cpplint calls failed.')
|
|
|
|
return 1
|
|
|
|
else:
|
|
|
|
print('Style linting completed successfully.')
|
|
|
|
return 0
|
2015-08-01 17:49:58 +00:00
|
|
|
|
|
|
|
|
2016-01-01 19:15:07 +00:00
|
|
|
# TODO(benvanik): merge into linter, or as lint --anal?
|
|
|
|
class TidyCommand(Command):
|
2017-09-21 21:42:57 +00:00
|
|
|
"""'tidy' command."""
|
|
|
|
|
|
|
|
def __init__(self, subparsers, *args, **kwargs):
|
|
|
|
super(TidyCommand, self).__init__(
|
|
|
|
subparsers,
|
|
|
|
name='tidy',
|
2019-07-18 00:52:01 +00:00
|
|
|
help_short='Runs the clang-tidy checker on all code.',
|
2017-09-21 21:42:57 +00:00
|
|
|
*args, **kwargs)
|
|
|
|
self.parser.add_argument(
|
|
|
|
'--fix', action='store_true',
|
|
|
|
help='Applies suggested fixes, where possible.')
|
|
|
|
|
|
|
|
def execute(self, args, pass_args, cwd):
|
|
|
|
# Run premake to generate our compile_commands.json file for clang to use.
|
|
|
|
run_premake_export_commands()
|
|
|
|
|
|
|
|
platform_name = ''
|
|
|
|
if sys.platform == 'darwin':
|
|
|
|
platform_name = 'darwin'
|
|
|
|
elif sys.platform == 'win32':
|
|
|
|
platform_name = 'windows'
|
|
|
|
else:
|
|
|
|
platform_name = 'linux'
|
|
|
|
tool_root = 'build/llvm_tools/debug_%s' % platform_name
|
|
|
|
|
|
|
|
all_files = [file_path for file_path in find_all_source_files()
|
|
|
|
if not file_path.endswith('_test.cc')]
|
|
|
|
# Tidy only likes .cc files.
|
|
|
|
all_files = [file_path for file_path in all_files
|
|
|
|
if file_path.endswith('.cc')]
|
|
|
|
|
|
|
|
any_errors = False
|
|
|
|
for file in all_files:
|
|
|
|
print('- clang-tidy %s' % (file))
|
|
|
|
ret = shell_call([
|
|
|
|
'clang-tidy',
|
|
|
|
'-p', tool_root,
|
|
|
|
'-checks=' + ','.join([
|
|
|
|
'clang-analyzer-*',
|
|
|
|
'google-*',
|
|
|
|
'misc-*',
|
|
|
|
'modernize-*'
|
|
|
|
# TODO(benvanik): pick the ones we want - some are silly.
|
|
|
|
# 'readability-*',
|
|
|
|
]),
|
|
|
|
] + (['-fix'] if args['fix'] else []) + [
|
|
|
|
file,
|
|
|
|
], throw_on_error=False)
|
|
|
|
if ret:
|
|
|
|
any_errors = True
|
2016-01-01 19:15:07 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
print('')
|
|
|
|
if any_errors:
|
|
|
|
print('ERROR: 1+ clang-tidy calls failed.')
|
|
|
|
return 1
|
|
|
|
else:
|
|
|
|
print('Tidy completed successfully.')
|
|
|
|
return 0
|
2016-01-01 19:15:07 +00:00
|
|
|
|
|
|
|
|
2015-08-01 06:48:24 +00:00
|
|
|
class DevenvCommand(Command):
|
2017-09-21 21:42:57 +00:00
|
|
|
"""'devenv' command."""
|
|
|
|
|
|
|
|
def __init__(self, subparsers, *args, **kwargs):
|
|
|
|
super(DevenvCommand, self).__init__(
|
|
|
|
subparsers,
|
|
|
|
name='devenv',
|
|
|
|
help_short='Launches Visual Studio with the sln.',
|
|
|
|
*args, **kwargs)
|
|
|
|
|
|
|
|
def execute(self, args, pass_args, cwd):
|
|
|
|
print('Launching Visual Studio...')
|
|
|
|
print('')
|
|
|
|
|
|
|
|
print('- running premake...')
|
|
|
|
run_platform_premake()
|
|
|
|
print('')
|
2015-08-01 06:48:24 +00:00
|
|
|
|
2017-09-21 21:42:57 +00:00
|
|
|
print('- launching devenv...')
|
|
|
|
shell_call([
|
|
|
|
'devenv',
|
|
|
|
'build\\xenia.sln',
|
|
|
|
])
|
|
|
|
print('')
|
|
|
|
|
|
|
|
return 0
|
2015-08-01 06:48:24 +00:00
|
|
|
|
|
|
|
|
|
|
|
if __name__ == '__main__':
|
2017-09-21 21:42:57 +00:00
|
|
|
main()
|
|
|
|
|