123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873 |
- import os
- import typing
- from enum import auto, StrEnum
- import ymake
- import _dart_fields as df
- import ytest
- from _common import (
- rootrel_arc_src,
- sort_uniq,
- strip_roots,
- to_yesno,
- )
- from _dart_fields import create_dart_record
- # 1 is 60 files per chunk for TIMEOUT(60) - default timeout for SIZE(SMALL)
- # 0.5 is 120 files per chunk for TIMEOUT(60) - default timeout for SIZE(SMALL)
- # 0.2 is 300 files per chunk for TIMEOUT(60) - default timeout for SIZE(SMALL)
- ESLINT_FILE_PROCESSING_TIME_DEFAULT = 0.2 # seconds per file
- class TsTestType(StrEnum):
- JEST = auto()
- HERMIONE = auto()
- PLAYWRIGHT = auto()
- ESLINT = auto()
- TSC_TYPECHECK = auto()
- TS_STYLELINT = auto()
- TS_TEST_FIELDS_BASE = (
- df.BinaryPath.normalized,
- df.BuildFolderPath.normalized,
- df.ForkMode.test_fork_mode,
- df.NodejsRootVarName.value,
- df.ScriptRelPath.first_flat,
- df.SourceFolderPath.normalized,
- df.SplitFactor.from_unit,
- df.TestData.from_unit,
- df.TestedProjectName.filename_without_ext,
- df.TestEnv.value,
- df.TestName.value,
- df.TestRecipes.value,
- df.TestTimeout.from_unit,
- )
- TS_TEST_SPECIFIC_FIELDS = {
- TsTestType.JEST: (
- df.Size.from_unit,
- df.Tag.from_unit,
- df.Requirements.from_unit,
- df.ConfigPath.value,
- df.TsTestDataDirs.value,
- df.TsTestDataDirsRename.value,
- df.TsResources.value,
- df.TsTestForPath.value,
- ),
- TsTestType.HERMIONE: (
- df.Tag.from_unit_fat_external_no_retries,
- df.Requirements.from_unit_with_full_network,
- df.ConfigPath.value,
- df.TsTestDataDirs.value,
- df.TsTestDataDirsRename.value,
- df.TsResources.value,
- df.TsTestForPath.value,
- ),
- TsTestType.PLAYWRIGHT: (
- df.Size.from_unit,
- df.Tag.from_unit,
- df.Requirements.from_unit,
- df.ConfigPath.value,
- df.TsTestDataDirs.value,
- df.TsTestDataDirsRename.value,
- df.TsResources.value,
- df.TsTestForPath.value,
- ),
- TsTestType.ESLINT: (
- df.Size.from_unit,
- df.TestCwd.moddir,
- df.Tag.from_unit,
- df.Requirements.from_unit,
- df.EslintConfigPath.value,
- ),
- TsTestType.TSC_TYPECHECK: (
- df.Size.from_unit,
- df.TestCwd.moddir,
- df.Tag.from_unit,
- df.Requirements.from_unit,
- ),
- TsTestType.TS_STYLELINT: (
- df.TsStylelintConfig.value,
- df.TestFiles.stylesheets,
- df.NodeModulesBundleFilename.value,
- ),
- }
- class PluginLogger(object):
- def __init__(self):
- self.unit = None
- self.prefix = ""
- def reset(self, unit, prefix=""):
- self.unit = unit
- self.prefix = prefix
- def get_state(self):
- return (self.unit, self.prefix)
- def _stringify_messages(self, messages):
- parts = []
- for m in messages:
- if m is None:
- parts.append("None")
- else:
- parts.append(m if isinstance(m, str) else repr(m))
- # cyan color (code 36) for messages
- return "\033[0;32m{}\033[0;49m\n\033[0;36m{}\033[0;49m".format(self.prefix, " ".join(parts))
- def info(self, *messages):
- if self.unit:
- self.unit.message(["INFO", self._stringify_messages(messages)])
- def warn(self, *messages):
- if self.unit:
- self.unit.message(["WARN", self._stringify_messages(messages)])
- def error(self, *messages):
- if self.unit:
- self.unit.message(["ERROR", self._stringify_messages(messages)])
- def print_vars(self, *variables):
- if self.unit:
- values = ["{}={}".format(v, self.unit.get(v)) for v in variables]
- self.info("\n".join(values))
- logger = PluginLogger()
- def _with_report_configure_error(fn):
- def _wrapper(*args, **kwargs):
- last_state = logger.get_state()
- unit = args[0]
- logger.reset(unit if unit.get("TS_LOG") == "yes" else None, fn.__name__)
- try:
- fn(*args, **kwargs)
- except Exception as exc:
- ymake.report_configure_error(str(exc))
- if unit.get("TS_RAISE") == "yes":
- raise
- else:
- unit.message(["WARN", "Configure error is reported. Add -DTS_RAISE to see actual exception"])
- finally:
- logger.reset(*last_state)
- return _wrapper
- def _build_directives(name, flags, paths):
- # type: (str, list[str]|tuple[str], list[str]) -> str
- parts = [p for p in [name] + (flags or []) if p]
- parts_str = ";".join(parts)
- expressions = ['${{{parts}:"{path}"}}'.format(parts=parts_str, path=path) for path in paths]
- return " ".join(expressions)
- def _build_cmd_input_paths(paths, hide=False, disable_include_processor=False):
- # type: (list[str]|tuple[str], bool, bool) -> str
- hide_part = "hide" if hide else ""
- disable_ip_part = "context=TEXT" if disable_include_processor else ""
- return _build_directives("input", [hide_part, disable_ip_part], paths)
- def _create_erm_json(unit):
- from lib.nots.erm_json_lite import ErmJsonLite
- erm_packages_path = unit.get("ERM_PACKAGES_PATH")
- path = unit.resolve(unit.resolve_arc_path(erm_packages_path))
- return ErmJsonLite.load(path)
- def _get_pm_type(unit) -> typing.Literal["pnpm", "npm"]:
- resolved = unit.get("PM_TYPE")
- if not resolved:
- raise Exception("PM_TYPE is not set yet. Macro _SET_PACKAGE_MANAGER() should be called before.")
- return resolved
- def _get_source_path(unit):
- sources_path = unit.get("TS_TEST_FOR_DIR") if unit.get("TS_TEST_FOR") else unit.path()
- return sources_path
- def _create_pm(unit):
- from lib.nots.package_manager import get_package_manager_type
- sources_path = _get_source_path(unit)
- module_path = unit.get("TS_TEST_FOR_PATH") if unit.get("TS_TEST_FOR") else unit.get("MODDIR")
- # noinspection PyPep8Naming
- PackageManager = get_package_manager_type(_get_pm_type(unit))
- return PackageManager(
- sources_path=unit.resolve(sources_path),
- build_root="$B",
- build_path=unit.path().replace("$S", "$B", 1),
- contribs_path=unit.get("NPM_CONTRIBS_PATH"),
- nodejs_bin_path=None,
- script_path=None,
- module_path=module_path,
- )
- @_with_report_configure_error
- def on_set_package_manager(unit):
- pm_type = "pnpm" # projects without any lockfile are processed by pnpm
- source_path = _get_source_path(unit)
- for pm_key, lockfile_name in [("pnpm", "pnpm-lock.yaml"), ("npm", "package-lock.json")]:
- lf_path = os.path.join(source_path, lockfile_name)
- lf_path_resolved = unit.resolve_arc_path(strip_roots(lf_path))
- if lf_path_resolved:
- pm_type = pm_key
- break
- if pm_type == 'npm' and "devtools/dummy_arcadia/typescript/npm" not in source_path:
- ymake.report_configure_error(
- "\n"
- "Project is configured to use npm as a package manager. \n"
- "Only pnpm is supported at the moment.\n"
- "Please follow the instruction to migrate your project:\n"
- "https://docs.yandex-team.ru/frontend-in-arcadia/tutorials/migrate#migrate-to-pnpm"
- )
- unit.on_peerdir_ts_resource(pm_type)
- unit.set(["PM_TYPE", pm_type])
- unit.set(["PM_SCRIPT", f"${pm_type.upper()}_SCRIPT"])
- @_with_report_configure_error
- def on_set_append_with_directive(unit, var_name, dir, *values):
- wrapped = ['${{{dir}:"{v}"}}'.format(dir=dir, v=v) for v in values]
- __set_append(unit, var_name, " ".join(wrapped))
- @_with_report_configure_error
- def on_from_npm_lockfiles(unit, *args):
- from lib.nots.package_manager.base import PackageManagerError
- # This is contrib with pnpm-lock.yaml files only
- # Force set to pnpm
- unit.set(["PM_TYPE", "pnpm"])
- pm = _create_pm(unit)
- lf_paths = []
- for lf_path in args:
- abs_lf_path = unit.resolve(unit.resolve_arc_path(lf_path))
- if abs_lf_path:
- lf_paths.append(abs_lf_path)
- elif unit.get("TS_STRICT_FROM_NPM_LOCKFILES") == "yes":
- ymake.report_configure_error("lockfile not found: {}".format(lf_path))
- try:
- for pkg in pm.extract_packages_meta_from_lockfiles(lf_paths):
- unit.on_from_npm([pkg.tarball_url, pkg.sky_id, pkg.integrity, pkg.integrity_algorithm, pkg.tarball_path])
- except PackageManagerError as e:
- logger.warn(str(e))
- pass
- def _check_nodejs_version(unit, major):
- if major < 14:
- raise Exception(
- "Node.js {} is unsupported. Update Node.js please. See https://nda.ya.ru/t/joB9Mivm6h4znu".format(major)
- )
- if major < 18:
- unit.message(
- [
- "WARN",
- "Node.js {} is deprecated. Update Node.js please. See https://nda.ya.ru/t/joB9Mivm6h4znu".format(major),
- ]
- )
- @_with_report_configure_error
- def on_peerdir_ts_resource(unit, *resources):
- from lib.nots.package_manager import BasePackageManager
- pj = BasePackageManager.load_package_json_from_dir(unit.resolve(_get_source_path(unit)))
- erm_json = _create_erm_json(unit)
- dirs = []
- nodejs_version = _select_matching_version(erm_json, "nodejs", pj.get_nodejs_version())
- _check_nodejs_version(unit, nodejs_version.major)
- for tool in resources:
- dir_name = erm_json.canonize_name(tool)
- if erm_json.use_resource_directly(tool):
- # raises the configuration error when the version is unsupported
- _select_matching_version(erm_json, tool, pj.get_dep_specifier(tool), dep_is_required=True)
- elif tool == "nodejs":
- dirs.append(os.path.join("build", "platform", dir_name, str(nodejs_version)))
- _set_resource_vars(unit, erm_json, tool, nodejs_version)
- elif erm_json.is_resource_multiplatform(tool):
- v = _select_matching_version(erm_json, tool, pj.get_dep_specifier(tool))
- sb_resources = [
- sbr for sbr in erm_json.get_sb_resources(tool, v) if sbr.get("nodejs") == nodejs_version.major
- ]
- nodejs_dir = "NODEJS_{}".format(nodejs_version.major)
- if len(sb_resources) > 0:
- dirs.append(os.path.join("build", "external_resources", dir_name, str(v), nodejs_dir))
- _set_resource_vars(unit, erm_json, tool, v, nodejs_version.major)
- else:
- unit.message(["WARN", "Missing {}@{} for {}".format(tool, str(v), nodejs_dir)])
- else:
- v = _select_matching_version(erm_json, tool, pj.get_dep_specifier(tool))
- dirs.append(os.path.join("build", "external_resources", dir_name, str(v)))
- _set_resource_vars(unit, erm_json, tool, v, nodejs_version.major)
- if dirs:
- unit.onpeerdir(dirs)
- @_with_report_configure_error
- def on_ts_configure(unit):
- # type: (Unit) -> None
- from lib.nots.package_manager.base import PackageJson
- from lib.nots.package_manager.base.utils import build_pj_path
- from lib.nots.typescript import TsConfig
- tsconfig_paths = unit.get("TS_CONFIG_PATH").split()
- # for use in CMD as inputs
- __set_append(
- unit, "TS_CONFIG_FILES", _build_cmd_input_paths(tsconfig_paths, hide=True, disable_include_processor=True)
- )
- mod_dir = unit.get("MODDIR")
- cur_dir = unit.get("TS_TEST_FOR_PATH") if unit.get("TS_TEST_FOR") else mod_dir
- pj_path = build_pj_path(unit.resolve(unit.resolve_arc_path(cur_dir)))
- dep_paths = PackageJson.load(pj_path).get_dep_paths_by_names()
- # reversed for using the first tsconfig as the config for include processor (legacy)
- for tsconfig_path in reversed(tsconfig_paths):
- abs_tsconfig_path = unit.resolve(unit.resolve_arc_path(tsconfig_path))
- if not abs_tsconfig_path:
- raise Exception("tsconfig not found: {}".format(tsconfig_path))
- tsconfig = TsConfig.load(abs_tsconfig_path)
- config_files = tsconfig.inline_extend(dep_paths)
- config_files = _resolve_module_files(unit, mod_dir, config_files)
- use_tsconfig_outdir = unit.get("TS_CONFIG_USE_OUTDIR") == "yes"
- tsconfig.validate(use_tsconfig_outdir)
- # add tsconfig files from which root tsconfig files were extended
- __set_append(
- unit, "TS_CONFIG_FILES", _build_cmd_input_paths(config_files, hide=True, disable_include_processor=True)
- )
- # region include processor
- unit.set(["TS_CONFIG_ROOT_DIR", tsconfig.compiler_option("rootDir")]) # also for hermione
- if use_tsconfig_outdir:
- unit.set(["TS_CONFIG_OUT_DIR", tsconfig.compiler_option("outDir")]) # also for hermione
- unit.set(["TS_CONFIG_SOURCE_MAP", to_yesno(tsconfig.compiler_option("sourceMap"))])
- unit.set(["TS_CONFIG_DECLARATION", to_yesno(tsconfig.compiler_option("declaration"))])
- unit.set(["TS_CONFIG_DECLARATION_MAP", to_yesno(tsconfig.compiler_option("declarationMap"))])
- unit.set(["TS_CONFIG_PRESERVE_JSX", to_yesno(tsconfig.compiler_option("jsx") == "preserve")])
- # endregion
- _filter_inputs_by_rules_from_tsconfig(unit, tsconfig)
- # Code navigation
- if unit.get("TS_YNDEXING") == "yes":
- unit.on_do_ts_yndexing()
- # Style tests
- _setup_eslint(unit)
- _setup_tsc_typecheck(unit)
- _setup_stylelint(unit)
- @_with_report_configure_error
- def on_setup_build_env(unit): # type: (Unit) -> None
- build_env_var = unit.get("TS_BUILD_ENV") # type: str
- if not build_env_var:
- return
- options = []
- for name in build_env_var.split(","):
- options.append("--env")
- value = unit.get(f"TS_ENV_{name}")
- if value is None:
- ymake.report_configure_error(f"Env var '{name}' is provided in a list, but var value is not provided")
- continue
- double_quote_escaped_value = value.replace('"', '\\"')
- options.append(f'"{name}={double_quote_escaped_value}"')
- unit.set(["NOTS_TOOL_BUILD_ENV", " ".join(options)])
- def __set_append(unit, var_name, value):
- # type: (Unit, str, str|list[str]|tuple[str]) -> None
- """
- SET_APPEND() python naive implementation - append value/values to the list of values
- """
- previous_value = unit.get(var_name) or ""
- value_in_str = " ".join(value) if isinstance(value, list) or isinstance(value, tuple) else value
- new_value = previous_value + " " + value_in_str
- unit.set([var_name, new_value])
- def __strip_prefix(prefix, line):
- # type: (str, str) -> str
- if line.startswith(prefix):
- prefix_len = len(prefix)
- return line[prefix_len:]
- return line
- def _filter_inputs_by_rules_from_tsconfig(unit, tsconfig):
- """
- Reduce file list from the TS_GLOB_FILES variable following tsconfig.json rules
- """
- mod_dir = unit.get("MODDIR")
- target_path = os.path.join("${ARCADIA_ROOT}", mod_dir, "") # To have "/" in the end
- all_files = [__strip_prefix(target_path, f) for f in unit.get("TS_GLOB_FILES").split(" ")]
- filtered_files = tsconfig.filter_files(all_files)
- __set_append(unit, "TS_INPUT_FILES", [os.path.join(target_path, f) for f in filtered_files])
- def _is_tests_enabled(unit):
- if unit.get("TIDY") == "yes":
- return False
- return True
- def _setup_eslint(unit):
- if not _is_tests_enabled(unit):
- return
- if unit.get("_NO_LINT_VALUE") == "none":
- return
- test_files = df.TestFiles.ts_lint_srcs(unit, (), {})[df.TestFiles.KEY]
- if not test_files:
- return
- unit.on_peerdir_ts_resource("eslint")
- user_recipes = unit.get("TEST_RECIPES_VALUE")
- unit.on_setup_install_node_modules_recipe()
- test_type = TsTestType.ESLINT
- from lib.nots.package_manager import constants
- peers = _create_pm(unit).get_peers_from_package_json()
- deps = df.CustomDependencies.nots_with_recipies(unit, (peers,), {})[df.CustomDependencies.KEY].split()
- if deps:
- joined_deps = "\n".join(deps)
- logger.info(f"{test_type} deps: \n{joined_deps}")
- unit.ondepends(deps)
- flat_args = (test_type, "MODDIR")
- dart_record = create_dart_record(
- TS_TEST_FIELDS_BASE + TS_TEST_SPECIFIC_FIELDS[test_type],
- unit,
- flat_args,
- {},
- )
- dart_record[df.TestFiles.KEY] = test_files
- dart_record[df.NodeModulesBundleFilename.KEY] = constants.NODE_MODULES_WORKSPACE_BUNDLE_FILENAME
- extra_deps = df.CustomDependencies.test_depends_only(unit, (), {})[df.CustomDependencies.KEY].split()
- dart_record[df.CustomDependencies.KEY] = " ".join(sort_uniq(deps + extra_deps))
- dart_record[df.LintFileProcessingTime.KEY] = str(ESLINT_FILE_PROCESSING_TIME_DEFAULT)
- data = ytest.dump_test(unit, dart_record)
- if data:
- unit.set_property(["DART_DATA", data])
- unit.set(["TEST_RECIPES_VALUE", user_recipes])
- @_with_report_configure_error
- def _setup_tsc_typecheck(unit):
- if not _is_tests_enabled(unit):
- return
- if unit.get("_TS_TYPECHECK_VALUE") == "none":
- return
- test_files = df.TestFiles.ts_input_files(unit, (), {})[df.TestFiles.KEY]
- if not test_files:
- return
- tsconfig_paths = unit.get("TS_CONFIG_PATH").split()
- tsconfig_path = tsconfig_paths[0]
- if len(tsconfig_paths) > 1:
- tsconfig_path = unit.get("_TS_TYPECHECK_TSCONFIG")
- if not tsconfig_path:
- macros = " or ".join([f"TS_TYPECHECK({p})" for p in tsconfig_paths])
- raise Exception(f"Module uses several tsconfig files, specify which one to use for typecheck: {macros}")
- abs_tsconfig_path = unit.resolve(unit.resolve_arc_path(tsconfig_path))
- if not abs_tsconfig_path:
- raise Exception(f"tsconfig for typecheck not found: {tsconfig_path}")
- unit.on_peerdir_ts_resource("typescript")
- user_recipes = unit.get("TEST_RECIPES_VALUE")
- unit.on_setup_install_node_modules_recipe()
- unit.on_setup_extract_output_tars_recipe([unit.get("MODDIR")])
- test_type = TsTestType.TSC_TYPECHECK
- from lib.nots.package_manager import constants
- peers = _create_pm(unit).get_peers_from_package_json()
- deps = df.CustomDependencies.nots_with_recipies(unit, (peers,), {})[df.CustomDependencies.KEY].split()
- if deps:
- joined_deps = "\n".join(deps)
- logger.info(f"{test_type} deps: \n{joined_deps}")
- unit.ondepends(deps)
- flat_args = (test_type,)
- dart_record = create_dart_record(
- TS_TEST_FIELDS_BASE + TS_TEST_SPECIFIC_FIELDS[test_type],
- unit,
- flat_args,
- {},
- )
- dart_record[df.TestFiles.KEY] = test_files
- dart_record[df.NodeModulesBundleFilename.KEY] = constants.NODE_MODULES_WORKSPACE_BUNDLE_FILENAME
- extra_deps = df.CustomDependencies.test_depends_only(unit, (), {})[df.CustomDependencies.KEY].split()
- dart_record[df.CustomDependencies.KEY] = " ".join(sort_uniq(deps + extra_deps))
- dart_record[df.TsConfigPath.KEY] = tsconfig_path
- data = ytest.dump_test(unit, dart_record)
- if data:
- unit.set_property(["DART_DATA", data])
- unit.set(["TEST_RECIPES_VALUE", user_recipes])
- @_with_report_configure_error
- def _setup_stylelint(unit):
- if not _is_tests_enabled(unit):
- return
- if unit.get("_TS_STYLELINT_VALUE") == "no":
- return
- test_files = df.TestFiles.stylesheets(unit, (), {})[df.TestFiles.KEY]
- if not test_files:
- return
- from lib.nots.package_manager import constants
- recipes_value = unit.get("TEST_RECIPES_VALUE")
- unit.on_setup_install_node_modules_recipe()
- unit.on_setup_extract_output_tars_recipe([unit.get("MODDIR")])
- test_type = TsTestType.TS_STYLELINT
- peers = _create_pm(unit).get_peers_from_package_json()
- deps = df.CustomDependencies.nots_with_recipies(unit, (peers,), {})[df.CustomDependencies.KEY].split()
- if deps:
- joined_deps = "\n".join(deps)
- logger.info(f"{test_type} deps: \n{joined_deps}")
- unit.ondepends(deps)
- flat_args = (test_type,)
- spec_args = dict(nm_bundle=constants.NODE_MODULES_WORKSPACE_BUNDLE_FILENAME)
- dart_record = create_dart_record(
- TS_TEST_FIELDS_BASE + TS_TEST_SPECIFIC_FIELDS[test_type], unit, flat_args, spec_args
- )
- extra_deps = df.CustomDependencies.test_depends_only(unit, (), {})[df.CustomDependencies.KEY].split()
- dart_record[df.CustomDependencies.KEY] = " ".join(sort_uniq(deps + extra_deps))
- data = ytest.dump_test(unit, dart_record)
- if data:
- unit.set_property(["DART_DATA", data])
- unit.set(["TEST_RECIPES_VALUE", recipes_value])
- def _resolve_module_files(unit, mod_dir, file_paths):
- mod_dir_with_sep_len = len(mod_dir) + 1
- resolved_files = []
- for path in file_paths:
- resolved = rootrel_arc_src(path, unit)
- if resolved.startswith(mod_dir):
- resolved = resolved[mod_dir_with_sep_len:]
- resolved_files.append(resolved)
- return resolved_files
- def _set_resource_vars(unit, erm_json, tool, version, nodejs_major=None):
- # type: (any, ErmJsonLite, Version, str|None, int|None) -> None
- resource_name = erm_json.canonize_name(tool).upper()
- # example: NODEJS_12_18_4 | HERMIONE_7_0_4_NODEJS_18
- version_str = str(version).replace(".", "_")
- yamake_resource_name = "{}_{}".format(resource_name, version_str)
- if erm_json.is_resource_multiplatform(tool):
- yamake_resource_name += "_NODEJS_{}".format(nodejs_major)
- yamake_resource_var = "{}_RESOURCE_GLOBAL".format(yamake_resource_name)
- unit.set(["{}_ROOT".format(resource_name), "${}".format(yamake_resource_var)])
- unit.set(["{}-ROOT-VAR-NAME".format(resource_name), yamake_resource_var])
- def _select_matching_version(erm_json, resource_name, range_str, dep_is_required=False):
- # type: (ErmJsonLite, str, str, bool) -> Version
- if dep_is_required and range_str is None:
- raise Exception(
- "Please install the '{tool}' package to the project. Run the command:\n"
- " ya tool nots add -D {tool}".format(tool=resource_name)
- )
- try:
- version = erm_json.select_version_of(resource_name, range_str)
- if version:
- return version
- raise ValueError("There is no allowed version to satisfy this range: '{}'".format(range_str))
- except Exception as error:
- toolchain_versions = erm_json.get_versions_of(erm_json.get_resource(resource_name))
- raise Exception(
- "Requested {} version range '{}' could not be satisfied. \n"
- "Please use a range that would include one of the following: {}. \n"
- "For further details please visit the link: {} \nOriginal error: {} \n".format(
- resource_name,
- range_str,
- ", ".join(map(str, toolchain_versions)),
- "https://docs.yandex-team.ru/frontend-in-arcadia/_generated/toolchain",
- str(error),
- )
- )
- @_with_report_configure_error
- def on_prepare_deps_configure(unit):
- contrib_path = unit.get("NPM_CONTRIBS_PATH")
- if contrib_path == '-':
- unit.on_prepare_deps_configure_no_contrib()
- return
- unit.onpeerdir(contrib_path)
- pm = _create_pm(unit)
- pj = pm.load_package_json_from_dir(pm.sources_path)
- has_deps = pj.has_dependencies()
- ins, outs = pm.calc_prepare_deps_inouts(unit.get("_TARBALLS_STORE"), has_deps)
- if has_deps:
- unit.onpeerdir(pm.get_local_peers_from_package_json())
- __set_append(unit, "_PREPARE_DEPS_INOUTS", _build_directives("input", ["hide"], sorted(ins)))
- __set_append(unit, "_PREPARE_DEPS_INOUTS", _build_directives("output", ["hide"], sorted(outs)))
- else:
- __set_append(unit, "_PREPARE_DEPS_INOUTS", _build_directives("output", [], sorted(outs)))
- unit.set(["_PREPARE_DEPS_CMD", "$_PREPARE_NO_DEPS_CMD"])
- @_with_report_configure_error
- def on_prepare_deps_configure_no_contrib(unit):
- pm = _create_pm(unit)
- pj = pm.load_package_json_from_dir(pm.sources_path)
- has_deps = pj.has_dependencies()
- ins, outs, resources = pm.calc_prepare_deps_inouts_and_resources(unit.get("_TARBALLS_STORE"), has_deps)
- if has_deps:
- unit.onpeerdir(pm.get_local_peers_from_package_json())
- __set_append(unit, "_PREPARE_DEPS_INOUTS", _build_directives("input", ["hide"], sorted(ins)))
- __set_append(unit, "_PREPARE_DEPS_INOUTS", _build_directives("output", ["hide"], sorted(outs)))
- unit.set(["_PREPARE_DEPS_RESOURCES", " ".join([f'${{resource:"{uri}"}}' for uri in sorted(resources)])])
- unit.set(["_PREPARE_DEPS_USE_RESOURCES_FLAG", "--resource-root $(RESOURCE_ROOT)"])
- else:
- __set_append(unit, "_PREPARE_DEPS_INOUTS", _build_directives("output", [], sorted(outs)))
- unit.set(["_PREPARE_DEPS_CMD", "$_PREPARE_NO_DEPS_CMD"])
- @_with_report_configure_error
- def on_node_modules_configure(unit):
- pm = _create_pm(unit)
- pj = pm.load_package_json_from_dir(pm.sources_path)
- if pj.has_dependencies():
- unit.onpeerdir(pm.get_local_peers_from_package_json())
- local_cli = unit.get("TS_LOCAL_CLI") == "yes"
- ins, outs = pm.calc_node_modules_inouts(local_cli)
- __set_append(unit, "_NODE_MODULES_INOUTS", _build_directives("input", ["hide"], sorted(ins)))
- if not unit.get("TS_TEST_FOR"):
- __set_append(unit, "_NODE_MODULES_INOUTS", _build_directives("output", ["hide"], sorted(outs)))
- if pj.get_use_prebuilder():
- unit.on_peerdir_ts_resource("@yatool/prebuilder")
- unit.set(
- [
- "_YATOOL_PREBUILDER_ARG",
- "--yatool-prebuilder-path $YATOOL_PREBUILDER_ROOT/node_modules/@yatool/prebuilder",
- ]
- )
- # YATOOL_PREBUILDER_0_7_0_RESOURCE_GLOBAL
- prebuilder_major = unit.get("YATOOL_PREBUILDER-ROOT-VAR-NAME").split("_")[2]
- logger.info(f"Detected prebuilder \033[0;32mv{prebuilder_major}.x.x\033[0;49m")
- if prebuilder_major == "0":
- # TODO: FBP-1408
- lf = pm.load_lockfile_from_dir(pm.sources_path)
- is_valid, invalid_keys = lf.validate_has_addons_flags()
- if not is_valid:
- ymake.report_configure_error(
- "Project is configured to use @yatool/prebuilder. \n"
- + "Some packages in the pnpm-lock.yaml are misconfigured.\n"
- + "Run \033[0;32m`ya tool nots update-lockfile`\033[0;49m to fix lockfile.\n"
- + "All packages with `requiresBuild:true` have to be marked with `hasAddons:true/false`.\n"
- + "Misconfigured keys: \n"
- + " - "
- + "\n - ".join(invalid_keys)
- )
- else:
- lf = pm.load_lockfile_from_dir(pm.sources_path)
- requires_build_packages = lf.get_requires_build_packages()
- is_valid, validation_messages = pj.validate_prebuilds(requires_build_packages)
- if not is_valid:
- ymake.report_configure_error(
- "Project is configured to use @yatool/prebuilder. \n"
- + "Some packages are misconfigured.\n"
- + "Run \033[0;32m`ya tool nots update-lockfile`\033[0;49m to fix pnpm-lock.yaml and package.json.\n"
- + "Validation details: \n"
- + "\n".join(validation_messages)
- )
- @_with_report_configure_error
- def on_ts_test_for_configure(unit, test_runner, default_config, node_modules_filename):
- if not _is_tests_enabled(unit):
- return
- if unit.enabled('TS_COVERAGE'):
- unit.on_peerdir_ts_resource("nyc")
- for_mod_path = df.TsTestForPath.value(unit, (), {})[df.TsTestForPath.KEY]
- unit.onpeerdir([for_mod_path])
- unit.on_setup_extract_node_modules_recipe([for_mod_path])
- unit.on_setup_extract_output_tars_recipe([for_mod_path])
- build_root = "$B" if test_runner == TsTestType.HERMIONE else "$(BUILD_ROOT)"
- unit.set(["TS_TEST_NM", os.path.join(build_root, for_mod_path, node_modules_filename)])
- config_path = unit.get("TS_TEST_CONFIG_PATH")
- if not config_path:
- config_path = os.path.join(for_mod_path, default_config)
- unit.set(["TS_TEST_CONFIG_PATH", config_path])
- test_files = df.TestFiles.ts_test_srcs(unit, (), {})[df.TestFiles.KEY]
- if not test_files:
- ymake.report_configure_error("No tests found")
- return
- from lib.nots.package_manager import constants
- peers = _create_pm(unit).get_peers_from_package_json()
- deps = df.CustomDependencies.nots_with_recipies(unit, (peers,), {})[df.CustomDependencies.KEY].split()
- if deps:
- joined_deps = "\n".join(deps)
- logger.info(f"{test_runner} deps: \n{joined_deps}")
- unit.ondepends(deps)
- flat_args = (test_runner, "TS_TEST_FOR_PATH")
- spec_args = {'erm_json': _create_erm_json(unit)}
- dart_record = create_dart_record(
- TS_TEST_FIELDS_BASE + TS_TEST_SPECIFIC_FIELDS[test_runner],
- unit,
- flat_args,
- spec_args,
- )
- dart_record[df.TestFiles.KEY] = test_files
- dart_record[df.NodeModulesBundleFilename.KEY] = constants.NODE_MODULES_WORKSPACE_BUNDLE_FILENAME
- extra_deps = df.CustomDependencies.test_depends_only(unit, (), {})[df.CustomDependencies.KEY].split()
- dart_record[df.CustomDependencies.KEY] = " ".join(sort_uniq(deps + extra_deps))
- if test_runner == TsTestType.HERMIONE:
- dart_record[df.Size.KEY] = "LARGE"
- data = ytest.dump_test(unit, dart_record)
- if data:
- unit.set_property(["DART_DATA", data])
- @_with_report_configure_error
- def on_validate_ts_test_for_args(unit, for_mod, root):
- # FBP-1085
- is_arc_root = root == "${ARCADIA_ROOT}"
- is_rel_for_mod = for_mod.startswith(".")
- if is_arc_root and is_rel_for_mod:
- ymake.report_configure_error(
- "You are using a relative path for a module. "
- + "You have to add RELATIVE key, like (RELATIVE {})".format(for_mod)
- )
- @_with_report_configure_error
- def on_set_ts_test_for_vars(unit, for_mod):
- unit.set(["TS_TEST_FOR", "yes"])
- unit.set(["TS_TEST_FOR_DIR", unit.resolve_arc_path(for_mod)])
- unit.set(["TS_TEST_FOR_PATH", rootrel_arc_src(for_mod, unit)])
- @_with_report_configure_error
- def on_ts_files(unit, *files):
- new_cmds = ['$COPY_CMD ${{input;context=TEXT:"{0}"}} ${{output;noauto:"{0}"}}'.format(f) for f in files]
- all_cmds = unit.get("_TS_FILES_COPY_CMD")
- if all_cmds:
- new_cmds.insert(0, all_cmds)
- unit.set(["_TS_FILES_COPY_CMD", " && ".join(new_cmds)])
- @_with_report_configure_error
- def on_ts_package_check_files(unit):
- ts_files = unit.get("_TS_FILES_COPY_CMD")
- if ts_files == "":
- ymake.report_configure_error(
- "\n"
- "In the TS_PACKAGE module, you should define at least one file using the TS_FILES() macro.\n"
- "If you use the TS_FILES_GLOB, check the expression. For example, use `src/**/*` instead of `src/*`.\n"
- "Docs: https://docs.yandex-team.ru/frontend-in-arcadia/references/TS_PACKAGE#ts-files."
- )
- @_with_report_configure_error
- def on_depends_on_mod(unit):
- if unit.get("_TS_TEST_DEPENDS_ON_BUILD"):
- for_mod_path = unit.get("TS_TEST_FOR_PATH")
- unit.ondepends([for_mod_path])
|