Files
python/bin/generate_tests.py

267 lines
8.2 KiB
Python
Raw Normal View History

#!/usr/bin/env python3.7
"""
Generates exercise test suites using an exercise's canonical-data.json
(found in problem-specifications) and $exercise/.meta/template.j2.
If either does not exist, generation will not be attempted.
Usage:
generate_tests.py Generates tests for all exercises
generate_tests.py two-fer Generates tests for two-fer exercise
generate_tests.py t* Generates tests for all exercises matching t*
generate_tests.py --check Checks if test files are out of sync with templates
generate_tests.py --check two-fer Checks if two-fer test file is out of sync with template
"""
import argparse
import filecmp
import importlib.util
import json
import logging
import os
import posixpath
import re
import shutil
import sys
from glob import glob
from itertools import repeat
from string import punctuation, whitespace
from subprocess import check_call
from tempfile import NamedTemporaryFile
from jinja2 import Environment, FileSystemLoader, TemplateNotFound, UndefinedError
VERSION = "0.2.0"
DEFAULT_SPEC_LOCATION = os.path.join("..", "problem-specifications")
RGX_WORDS = re.compile(r"[-_\s]|(?=[A-Z])")
logging.basicConfig()
logger = logging.getLogger("generator")
logger.setLevel(logging.WARN)
def replace_all(string, chars, rep):
"""
Replace any char in chars with rep, reduce runs and strip terminal ends.
"""
trans = str.maketrans(dict(zip(chars, repeat(rep))))
return re.sub("{0}+".format(re.escape(rep)), rep, string.translate(trans)).strip(
rep
)
def to_snake(string):
"""
Convert pretty much anything to to_snake.
"""
clean = re.sub("(.)([A-Z][a-z]+)", r"\1_\2", string)
clean = re.sub("([a-z0-9])([A-Z])", r"\1_\2", clean).lower()
return replace_all(clean, whitespace + punctuation, "_")
def camel_case(string):
"""
Convert pretty much anything to CamelCase.
"""
return "".join(w.title() for w in to_snake(string).split("_"))
def get_tested_properties(spec):
"""
Get set of tested properties from spec. Include nested cases.
"""
props = set()
for case in spec["cases"]:
if "property" in case:
props.add(case["property"])
if "cases" in case:
props.update(get_tested_properties(case))
return sorted(props)
def error_case(case):
return (
"expected" in case
and isinstance(case["expected"], dict)
and "error" in case["expected"]
)
def has_error_case(cases):
cases = cases[:]
while cases:
case = cases.pop(0)
if error_case(case):
return True
cases.extend(case.get("cases", []))
return False
def regex_replace(s, find, repl):
return re.sub(find, repl, s)
def load_canonical(exercise, spec_path):
"""
Loads the canonical data for an exercise as a nested dictionary
"""
full_path = os.path.join(spec_path, "exercises", exercise, "canonical-data.json")
with open(full_path) as f:
spec = json.load(f)
spec["properties"] = get_tested_properties(spec)
return spec
def load_additional_tests(exercise):
"""
Loads additional tests from .meta/additional_tests.json
"""
full_path = os.path.join("exercises", exercise, ".meta", "additional_tests.json")
try:
with open(full_path) as f:
data = json.load(f)
return data.get("cases", [])
except FileNotFoundError:
return []
def format_file(path):
"""
Runs black auto-formatter on file at path
"""
check_call(["black", "-q", path])
def compare_existing(rendered, tests_path):
"""
Returns true if contents of file at tests_path match rendered
"""
if not os.path.isfile(tests_path):
return False
with open(tests_path) as f:
current = f.read()
return rendered == current
def generate_exercise(env, spec_path, exercise, check=False):
"""
Renders test suite for exercise and if check is:
True: verifies that current tests file matches rendered
False: saves rendered to tests file
"""
slug = os.path.basename(exercise)
meta_dir = os.path.join(exercise, ".meta")
plugins_module = None
plugins_name = "plugins"
plugins_source = os.path.join(meta_dir, f"{plugins_name}.py")
try:
if os.path.isfile(plugins_source):
plugins_spec = importlib.util.spec_from_file_location(
plugins_name, plugins_source
)
plugins_module = importlib.util.module_from_spec(plugins_spec)
sys.modules[plugins_name] = plugins_module
plugins_spec.loader.exec_module(plugins_module)
spec = load_canonical(slug, spec_path)
additional_tests = load_additional_tests(slug)
2019-08-06 14:45:35 -04:00
spec["additional_cases"] = additional_tests
template_path = posixpath.join(slug, ".meta", "template.j2")
template = env.get_template(template_path)
tests_path = os.path.join(exercise, f"{to_snake(slug)}_test.py")
spec["has_error_case"] = has_error_case(spec["cases"])
if plugins_module is not None:
spec[plugins_name] = plugins_module
logger.info(f"{slug}: attempting render")
rendered = template.render(**spec)
with NamedTemporaryFile("w", delete=False) as tmp:
tmp.write(rendered)
try:
logger.debug(f"{slug}: formatting tmp file")
format_file(tmp.name)
except FileNotFoundError as e:
logger.error(f"{slug}: the black utility must be installed")
return False
if check:
try:
if not filecmp.cmp(tmp.name, tests_path):
logger.error(
f"{slug}: check failed; tests must be regenerated with bin/generate_tests.py"
)
return False
finally:
os.remove(tmp.name)
else:
shutil.move(tmp.name, tests_path)
print(f"{slug} generated at {tests_path}")
except (TypeError, UndefinedError, SyntaxError) as e:
logger.debug(str(e))
logger.error(f"{slug}: generation failed")
return False
except TemplateNotFound as e:
logger.debug(str(e))
logger.info(f"{slug}: no template found; skipping")
except FileNotFoundError as e:
logger.debug(str(e))
logger.info(f"{slug}: no canonical data found; skipping")
return True
def generate(
exercise_glob,
spec_path=DEFAULT_SPEC_LOCATION,
stop_on_failure=False,
check=False,
**kwargs,
):
"""
Primary entry point. Generates test files for all exercises matching exercise_glob
"""
# black must be installed or all test files will error
if not shutil.which("black"):
logger.error("the black utility must be installed")
sys.exit(1)
loader = FileSystemLoader(["config", "exercises"])
env = Environment(loader=loader, keep_trailing_newline=True)
env.filters["to_snake"] = to_snake
env.filters["camel_case"] = camel_case
env.filters["regex_replace"] = regex_replace
env.tests["error_case"] = error_case
result = True
for exercise in glob(os.path.join("exercises", exercise_glob)):
if not generate_exercise(env, spec_path, exercise, check):
result = False
if stop_on_failure:
break
if not result:
sys.exit(1)
if __name__ == "__main__":
parser = argparse.ArgumentParser()
parser.add_argument("exercise_glob", nargs="?", default="*", metavar="EXERCISE")
parser.add_argument(
"--version",
action="version",
version="%(prog)s {} for Python {}".format(VERSION, sys.version.split("\n")[0]),
)
parser.add_argument("-v", "--verbose", action="store_true")
parser.add_argument(
"-p",
"--spec-path",
default=DEFAULT_SPEC_LOCATION,
help=(
"path to clone of exercism/problem-specifications " "(default: %(default)s)"
),
)
parser.add_argument("--stop-on-failure", action="store_true")
parser.add_argument(
"--check",
action="store_true",
help="check if tests are up-to-date, but do not modify test files",
)
opts = parser.parse_args()
if opts.verbose:
logger.setLevel(logging.DEBUG)
generate(**opts.__dict__)