| {"repo": "common-workflow-language/cwltool", "pull_number": 95, "instance_id": "common-workflow-language__cwltool-95", "issue_numbers": "", "base_commit": "efc29f3610e7062287037fa9cd834c56397bff62", "patch": "diff --git a/cwltool/builder.py b/cwltool/builder.py\n--- a/cwltool/builder.py\n+++ b/cwltool/builder.py\n@@ -34,9 +34,9 @@ class Builder(object):\n def __init__(self): # type: () -> None\n self.names = None # type: avro.schema.Names\n self.schemaDefs = None # type: Dict[str,Dict[unicode, Any]]\n- self.files = None # type: List[Dict[str, str]]\n+ self.files = None # type: List[Dict[unicode, unicode]]\n self.fs_access = None # type: StdFsAccess\n- self.job = None # type: Dict[str, Any]\n+ self.job = None # type: Dict[unicode, Union[Dict[unicode, Any], List, unicode]]\n self.requirements = None # type: List[Dict[str,Any]]\n self.outdir = None # type: str\n self.tmpdir = None # type: str\ndiff --git a/cwltool/cwlrdf.py b/cwltool/cwlrdf.py\n--- a/cwltool/cwlrdf.py\n+++ b/cwltool/cwlrdf.py\n@@ -6,7 +6,7 @@\n from typing import Any, Union, Dict, IO\n \n def makerdf(workflow, wf, ctx):\n- # type: (Union[str, unicode], Dict[str,Any], Loader.ContextType) -> Graph\n+ # type: (Union[str, unicode], Union[List[Dict[unicode, Any]], Dict[unicode, Any]], Loader.ContextType) -> Graph\n prefixes = {}\n for k,v in ctx.iteritems():\n if isinstance(v, dict):\n@@ -18,7 +18,11 @@ def makerdf(workflow, wf, ctx):\n p, _ = frg.split(\"/\")\n prefixes[p] = u\"%s#%s/\" % (doc_url, p)\n \n- wf[\"@context\"] = ctx\n+ if isinstance(wf, list):\n+ for entry in wf:\n+ entry[\"@context\"] = ctx\n+ else:\n+ wf[\"@context\"] = ctx\n g = Graph().parse(data=json.dumps(wf), format='json-ld', location=workflow)\n \n # Bug in json-ld loader causes @id fields to be added to the graph\n@@ -31,7 +35,7 @@ def makerdf(workflow, wf, ctx):\n return g\n \n def printrdf(workflow, wf, ctx, sr, stdout):\n- # type: (Union[str, unicode], Dict[str, Any], Loader.ContextType, str, IO[Any]) -> None\n+ # type: (Union[str, unicode], Union[List[Dict[unicode, Any]], Dict[unicode, Any]], Loader.ContextType, str, IO[Any]) -> None\n stdout.write(makerdf(workflow, wf, ctx).serialize(format=sr))\n \n def lastpart(uri): # type: (Any) -> str\n@@ -172,7 +176,7 @@ def dot_without_parameters(g, stdout): # type: (Graph, IO[Any]) -> None\n \n \n def printdot(workflow, wf, ctx, stdout, include_parameters=False):\n- # type: (Union[str, unicode], Dict[str, Any], Loader.ContextType, Any, bool) -> None\n+ # type: (Union[str, unicode], Union[List[Dict[unicode, Any]], Dict[unicode, Any]], Loader.ContextType, Any, bool) -> None\n g = makerdf(workflow, wf, ctx)\n \n stdout.write(\"digraph {\")\ndiff --git a/cwltool/draft2tool.py b/cwltool/draft2tool.py\n--- a/cwltool/draft2tool.py\n+++ b/cwltool/draft2tool.py\n@@ -30,7 +30,7 @@\n \n class ExpressionTool(Process):\n def __init__(self, toolpath_object, **kwargs):\n- # type: (Dict[str,List[None]], **Any) -> None\n+ # type: (Dict[unicode, Any], **Any) -> None\n super(ExpressionTool, self).__init__(toolpath_object, **kwargs)\n \n class ExpressionJob(object):\n@@ -53,7 +53,7 @@ def run(self, **kwargs): # type: (**Any) -> None\n self.output_callback({}, \"permanentFail\")\n \n def job(self, joborder, output_callback, **kwargs):\n- # type: (Dict[str,str], str, Callable[[Any, Any], Any], **Any) -> Generator[ExpressionTool.ExpressionJob, None, None]\n+ # type: (Dict[unicode, unicode], Callable[[Any, Any], Any], **Any) -> Generator[ExpressionTool.ExpressionJob, None, None]\n builder = self._init_job(joborder, **kwargs)\n \n j = ExpressionTool.ExpressionJob()\n@@ -113,14 +113,14 @@ def run(self, **kwargs):\n \n class CommandLineTool(Process):\n def __init__(self, toolpath_object, **kwargs):\n- # type: (Dict[str,Any], **Any) -> None\n+ # type: (Dict[unicode, Any], **Any) -> None\n super(CommandLineTool, self).__init__(toolpath_object, **kwargs)\n \n def makeJobRunner(self): # type: () -> CommandLineJob\n return CommandLineJob()\n \n def makePathMapper(self, reffiles, **kwargs):\n- # type: (Set[str], str, **Any) -> PathMapper\n+ # type: (Set[unicode], **Any) -> PathMapper\n dockerReq, _ = self.get_requirement(\"DockerRequirement\")\n try:\n if dockerReq and kwargs.get(\"use_container\"):\n@@ -132,7 +132,7 @@ def makePathMapper(self, reffiles, **kwargs):\n raise WorkflowException(u\"Missing input file %s\" % e)\n \n def job(self, joborder, output_callback, **kwargs):\n- # type: (Dict[str,str], str, Callable[..., Any], **Any) -> Generator[Union[CommandLineJob, CallbackJob], None, None]\n+ # type: (Dict[unicode, unicode], Callable[..., Any], **Any) -> Generator[Union[CommandLineJob, CallbackJob], None, None]\n \n jobname = uniquename(kwargs.get(\"name\", shortname(self.tool.get(\"id\", \"job\"))))\n \n@@ -149,7 +149,7 @@ def job(self, joborder, output_callback, **kwargs):\n if docker_req and kwargs.get(\"use_container\") is not False:\n dockerimg = docker_req.get(\"dockerImageId\") or docker_req.get(\"dockerPull\")\n cmdline = [\"docker\", \"run\", dockerimg] + cmdline\n- keydict = {\"cmdline\": cmdline}\n+ keydict = {u\"cmdline\": cmdline}\n \n for _,f in cachebuilder.pathmapper.items():\n st = os.stat(f[0])\n@@ -200,7 +200,7 @@ def rm_pending_output_callback(output_callback, jobcachepending,\n \n builder = self._init_job(joborder, **kwargs)\n \n- reffiles = set((f[\"path\"] for f in builder.files))\n+ reffiles = set((f[u\"path\"] for f in builder.files))\n \n j = self.makeJobRunner()\n j.builder = builder\n@@ -292,9 +292,9 @@ def _check_adjust(f): # type: (Dict[str,Any]) -> Dict[str,Any]\n yield j\n \n def collect_output_ports(self, ports, builder, outdir):\n- # type: (Set[Dict[str,Any]], Builder, str) -> Dict[str,Union[str,List[Any],Dict[str,Any]]]\n+ # type: (Set[Dict[str,Any]], Builder, str) -> Dict[unicode, Union[unicode, List[Any], Dict[unicode, Any]]]\n try:\n- ret = {} # type: Dict[str,Union[str,List[Any],Dict[str,Any]]]\n+ ret = {} # type: Dict[unicode, Union[unicode, List[Any], Dict[unicode, Any]]]\n custom_output = os.path.join(outdir, \"cwl.output.json\")\n if builder.fs_access.exists(custom_output):\n with builder.fs_access.open(custom_output, \"r\") as f:\n@@ -323,7 +323,7 @@ def collect_output_ports(self, ports, builder, outdir):\n raise WorkflowException(\"Error validating output record, \" + str(e) + \"\\n in \" + json.dumps(ret, indent=4))\n \n def collect_output(self, schema, builder, outdir):\n- # type: (Dict[str,Any], Builder, str) -> Union[Dict[str, Any], List[Union[Dict[str, Any], str]]]\n+ # type: (Dict[str,Any], Builder, str) -> Union[Dict[unicode, Any], List[Union[Dict[unicode, Any], unicode]]]\n r = [] # type: List[Any]\n if \"outputBinding\" in schema:\n binding = schema[\"outputBinding\"]\ndiff --git a/cwltool/expression.py b/cwltool/expression.py\n--- a/cwltool/expression.py\n+++ b/cwltool/expression.py\n@@ -19,7 +19,7 @@ def jshead(engineConfig, rootvars):\n return u\"\\n\".join(engineConfig + [u\"var %s = %s;\" % (k, json.dumps(v, indent=4)) for k, v in rootvars.items()])\n \n def exeval(ex, jobinput, requirements, outdir, tmpdir, context, pull_image):\n- # type: (Dict[str,Any], Dict[str,str], List[Dict[str, Any]], str, str, Any, bool) -> sandboxjs.JSON\n+ # type: (Dict[str, Any], Dict[unicode, Union[Dict, List, unicode]], List[Dict[str, Any]], str, str, Any, bool) -> sandboxjs.JSON\n \n if ex[\"engine\"] == \"https://w3id.org/cwl/cwl#JavascriptEngine\":\n engineConfig = [] # type: List[unicode]\n@@ -126,7 +126,7 @@ def param_interpolate(ex, obj, strip=True):\n \n def do_eval(ex, jobinput, requirements, outdir, tmpdir, resources,\n context=None, pull_image=True, timeout=None):\n- # type: (Any, Dict[str,str], List[Dict[str,Any]], str, str, Dict[str, Union[int, str]], Any, bool, int) -> Any\n+ # type: (Union[dict, unicode], Dict[unicode, Union[Dict, List, unicode]], List[Dict[str, Any]], str, str, Dict[str, Union[int, str]], Any, bool, int) -> Any\n \n runtime = resources.copy()\n runtime[\"tmpdir\"] = tmpdir\n@@ -140,7 +140,7 @@ def do_eval(ex, jobinput, requirements, outdir, tmpdir, resources,\n \n if isinstance(ex, dict) and \"engine\" in ex and \"script\" in ex:\n return exeval(ex, jobinput, requirements, outdir, tmpdir, context, pull_image)\n- if isinstance(ex, basestring):\n+ if isinstance(ex, (str, unicode)):\n for r in requirements:\n if r[\"class\"] == \"InlineJavascriptRequirement\":\n return sandboxjs.interpolate(str(ex), jshead(r.get(\"expressionLib\", []), rootvars),\ndiff --git a/cwltool/job.py b/cwltool/job.py\n--- a/cwltool/job.py\n+++ b/cwltool/job.py\n@@ -42,7 +42,7 @@ class CommandLineJob(object):\n \n def __init__(self): # type: () -> None\n self.builder = None # type: Builder\n- self.joborder = None # type: Dict[str,str]\n+ self.joborder = None # type: Dict[unicode, Union[Dict[unicode, Any], List, unicode]]\n self.stdin = None # type: str\n self.stdout = None # type: str\n self.successCodes = None # type: Iterable[int]\ndiff --git a/cwltool/load_tool.py b/cwltool/load_tool.py\n--- a/cwltool/load_tool.py\n+++ b/cwltool/load_tool.py\n@@ -44,7 +44,7 @@ def fetch_document(argsworkflow):\n \n def validate_document(document_loader, workflowobj, uri,\n enable_dev=False, strict=True, preprocess_only=False):\n- # type: (Loader, Dict[unicode, Any], unicode, bool, bool, bool) -> Tuple[Loader, Names, Any, Dict[str, str], unicode]\n+ # type: (Loader, Dict[unicode, Any], unicode, bool, bool, bool) -> Tuple[Loader, Names, Union[Dict[unicode, Any], List[Dict[unicode, Any]]], Dict[unicode, Any], unicode]\n \"\"\"Validate a CWL document.\"\"\"\n jobobj = None\n if \"cwl:tool\" in workflowobj:\n@@ -83,8 +83,13 @@ def validate_document(document_loader, workflowobj, uri,\n \n workflowobj[\"id\"] = fileuri\n processobj, metadata = document_loader.resolve_all(workflowobj, fileuri)\n+ if not isinstance(processobj, (dict, list)):\n+ raise validate.ValidationException(\"Workflow must be a dict or list.\")\n \n if not metadata:\n+ if not isinstance(processobj, dict):\n+ raise validate.ValidationException(\n+ \"Draft-2 workflows must be a dict.\")\n metadata = {\"$namespaces\": processobj.get(\"$namespaces\", {}),\n \"$schemas\": processobj.get(\"$schemas\", []),\n \"cwlVersion\": processobj[\"cwlVersion\"]}\n@@ -92,7 +97,6 @@ def validate_document(document_loader, workflowobj, uri,\n if preprocess_only:\n return document_loader, avsc_names, processobj, metadata, uri\n \n- document_loader.validate_links(processobj)\n schema.validate_doc(avsc_names, processobj, document_loader, strict)\n \n if metadata.get(\"cwlVersion\") != update.LATEST:\n@@ -100,14 +104,13 @@ def validate_document(document_loader, workflowobj, uri,\n processobj, document_loader, fileuri, enable_dev, metadata)\n \n if jobobj:\n- metadata[\"cwl:defaults\"] = jobobj\n+ metadata[u\"cwl:defaults\"] = jobobj\n \n return document_loader, avsc_names, processobj, metadata, uri\n \n \n-def make_tool(document_loader, avsc_names, processobj, metadata, uri, makeTool,\n- kwargs):\n- # type: (Loader, Names, Dict[str, Any], Dict[str, Any], unicode, Callable[..., Process], Dict[str, Any]) -> Process\n+def make_tool(document_loader, avsc_names, metadata, uri, makeTool, kwargs):\n+ # type: (Loader, Names, Dict[unicode, Any], unicode, Callable[..., Process], Dict[str, Any]) -> Process\n \"\"\"Make a Python CWL object.\"\"\"\n resolveduri = document_loader.resolve_ref(uri)[0]\n \n@@ -121,7 +124,7 @@ def make_tool(document_loader, avsc_names, processobj, metadata, uri, makeTool,\n urlparse.urldefrag(i[\"id\"])[1] for i in resolveduri\n if \"id\" in i))\n else:\n- processobj = cast(Dict[str, Any], resolveduri)\n+ processobj = resolveduri\n \n kwargs = kwargs.copy()\n kwargs.update({\n@@ -149,5 +152,5 @@ def load_tool(argsworkflow, makeTool, kwargs=None,\n document_loader, avsc_names, processobj, metadata, uri = validate_document(\n document_loader, workflowobj, uri, enable_dev=enable_dev,\n strict=strict)\n- return make_tool(document_loader, avsc_names, processobj, metadata, uri,\n+ return make_tool(document_loader, avsc_names, metadata, uri,\n makeTool, kwargs if kwargs else {})\ndiff --git a/cwltool/main.py b/cwltool/main.py\n--- a/cwltool/main.py\n+++ b/cwltool/main.py\n@@ -152,7 +152,7 @@ def arg_parser(): # type: () -> argparse.ArgumentParser\n \n \n def single_job_executor(t, job_order_object, **kwargs):\n- # type: (Process, Dict[str,Any], str, argparse.Namespace,**Any) -> Union[str,Dict[str,str]]\n+ # type: (Process, Dict[unicode, Any], **Any) -> Union[str,Dict[str,str]]\n final_output = []\n final_status = []\n \n@@ -238,7 +238,7 @@ def __call__(self, parser, namespace, values, option_string=None):\n \n \n def generate_parser(toolparser, tool, namemap):\n- # type: (argparse.ArgumentParser, Process,Dict[str,str]) -> argparse.ArgumentParser\n+ # type: (argparse.ArgumentParser, Process, Dict[unicode, unicode]) -> argparse.ArgumentParser\n toolparser.add_argument(\"job_order\", nargs=\"?\", help=\"Job input json file\")\n namemap[\"job_order\"] = \"job_order\"\n \n@@ -298,14 +298,15 @@ def generate_parser(toolparser, tool, namemap):\n else:\n typekw = {}\n \n- toolparser.add_argument(flag + name, required=required,\n- help=ahelp, action=action, default=default, **typekw)\n+ toolparser.add_argument( # type: ignore\n+ args=flag + name, required=required, help=ahelp, action=action,\n+ default=default, **typekw)\n \n return toolparser\n \n \n def load_job_order(args, t, stdin, print_input_deps=False, relative_deps=False, stdout=sys.stdout):\n- # type: (argparse.Namespace, Process, argparse.ArgumentParser, IO[Any], bool, bool, IO[Any]) -> Union[int,Tuple[Dict[str,Any],str]]\n+ # type: (argparse.Namespace, Process, IO[Any], bool, bool, IO[Any]) -> Union[int,Tuple[Dict[str,Any],str]]\n \n job_order_object = None\n \n@@ -313,9 +314,9 @@ def load_job_order(args, t, stdin, print_input_deps=False, relative_deps=False,\n loader = Loader({})\n else:\n jobloaderctx = {\n- \"path\": {\"@type\": \"@id\"},\n- \"format\": {\"@type\": \"@id\"},\n- \"id\": \"@id\"}\n+ u\"path\": {u\"@type\": u\"@id\"},\n+ u\"format\": {u\"@type\": u\"@id\"},\n+ u\"id\": u\"@id\"}\n jobloaderctx.update(t.metadata.get(\"$namespaces\", {}))\n loader = Loader(jobloaderctx)\n \n@@ -332,14 +333,14 @@ def load_job_order(args, t, stdin, print_input_deps=False, relative_deps=False,\n elif job_order_file:\n input_basedir = args.basedir if args.basedir else os.path.abspath(os.path.dirname(job_order_file))\n try:\n- job_order_object, _ = loader.resolve_ref(job_order_file)\n+ job_order_object, _ = loader.resolve_ref(job_order_file, checklinks=False)\n except Exception as e:\n _logger.error(str(e), exc_info=(e if args.debug else False))\n return 1\n toolparser = None\n else:\n input_basedir = args.basedir if args.basedir else os.getcwd()\n- namemap = {} # type: Dict[str,str]\n+ namemap = {} # type: Dict[unicode, unicode]\n toolparser = generate_parser(argparse.ArgumentParser(prog=args.workflow), t, namemap)\n if toolparser:\n if args.tool_help:\n@@ -419,7 +420,7 @@ def makeRelative(u):\n \n stdout.write(json.dumps(deps, indent=4))\n \n-def flatten_deps(d, files):\n+def flatten_deps(d, files): # type: (Any, Set[unicode]) -> None\n if isinstance(d, list):\n for s in d:\n flatten_deps(s, files)\n@@ -428,33 +429,36 @@ def flatten_deps(d, files):\n if \"secondaryFiles\" in d:\n flatten_deps(d[\"secondaryFiles\"], files)\n \n-def find_run(d, runs):\n+def find_run(d, runs): # type: (Any, Set[unicode]) -> None\n if isinstance(d, list):\n for s in d:\n find_run(s, runs)\n elif isinstance(d, dict):\n- if \"run\" in d and isinstance(d[\"run\"], basestring):\n+ if \"run\" in d and isinstance(d[\"run\"], (str, unicode)):\n runs.add(d[\"run\"])\n for s in d.values():\n find_run(s, runs)\n \n def replace_refs(d, rewrite, stem, newstem):\n+ # type: (Any, Dict[unicode, unicode], unicode, unicode) -> None\n if isinstance(d, list):\n for s,v in enumerate(d):\n- if isinstance(v, basestring) and v.startswith(stem):\n+ if isinstance(v, (str, unicode)) and v.startswith(stem):\n d[s] = newstem + v[len(stem):]\n else:\n replace_refs(v, rewrite, stem, newstem)\n elif isinstance(d, dict):\n- if \"run\" in d and isinstance(d[\"run\"], basestring):\n+ if \"run\" in d and isinstance(d[\"run\"], (str, unicode)):\n d[\"run\"] = rewrite[d[\"run\"]]\n for s,v in d.items():\n- if isinstance(v, basestring) and v.startswith(stem):\n+ if isinstance(v, (str, unicode)) and v.startswith(stem):\n d[s] = newstem + v[len(stem):]\n replace_refs(v, rewrite, stem, newstem)\n \n def print_pack(document_loader, processobj, uri, metadata):\n+ # type: (Loader, Any, unicode, Dict[unicode, unicode]) -> str\n def loadref(b, u):\n+ # type: (unicode, unicode) -> Union[Dict, List, unicode]\n return document_loader.resolve_ref(u, base_url=b)[0]\n deps = process.scandeps(uri, processobj,\n set((\"run\",)), set(), loadref)\n@@ -462,7 +466,7 @@ def loadref(b, u):\n fdeps = set((uri,))\n flatten_deps(deps, fdeps)\n \n- runs = set()\n+ runs = set() # type: Set[unicode]\n for f in fdeps:\n find_run(document_loader.idx[f], runs)\n \n@@ -476,9 +480,10 @@ def loadref(b, u):\n for r in runs:\n rewrite[r] = \"#\" + shortname(r)\n \n- packed = {\"$graph\": [], \"cwlVersion\": metadata[\"cwlVersion\"]}\n+ packed = {\"$graph\": [], \"cwlVersion\": metadata[\"cwlVersion\"]\n+ } # type: Dict[unicode, Any]\n for r,v in rewrite.items():\n- dc = copy.deepcopy(document_loader.idx[r])\n+ dc = cast(Dict[unicode, Any], copy.deepcopy(document_loader.idx[r]))\n dc[\"id\"] = v\n dc[\"name\"] = v\n replace_refs(dc, rewrite, r+\"/\" if \"#\" in r else r+\"#\", v+\"/\")\n@@ -508,7 +513,7 @@ def main(argsl=None,\n stderr=sys.stderr,\n versionfunc=versionstring,\n job_order_object=None):\n- # type: (List[str],Callable[...,Union[str,Dict[str,str]]],Callable[...,Process],Callable[[Dict[str,int]],Dict[str,int]],argparse.ArgumentParser,IO[Any],IO[Any],IO[Any],Callable[[],unicode]) -> int\n+ # type: (List[str], argparse.Namespace, Callable[..., Union[str, Dict[str, str]]], Callable[..., Process], Callable[[Dict[str, int]], Dict[str, int]], IO[Any], IO[Any], IO[Any], Callable[[], unicode], Union[int, Tuple[Dict[str, Any], str]]) -> int\n \n _logger.removeHandler(defaultStreamHandler)\n stderr_handler = logging.StreamHandler(stderr)\n@@ -588,8 +593,8 @@ def main(argsl=None,\n printdot(uri, processobj, document_loader.ctx, stdout)\n return 0\n \n- tool = make_tool(document_loader, avsc_names, processobj, metadata,\n- uri, makeTool, {})\n+ tool = make_tool(document_loader, avsc_names, metadata, uri,\n+ makeTool, {})\n except (validate.ValidationException) as exc:\n _logger.error(u\"Tool definition failed validation:\\n%s\", exc,\n exc_info=(exc if args.debug else False))\n@@ -611,14 +616,15 @@ def main(argsl=None,\n \n if args.tmp_outdir_prefix != 'tmp':\n # Use user defined temp directory (if it exists)\n- args.tmp_outdir_prefix = os.path.abspath(args.tmp_outdir_prefix)\n+ setattr(args, 'tmp_outdir_prefix',\n+ os.path.abspath(args.tmp_outdir_prefix))\n if not os.path.exists(args.tmp_outdir_prefix):\n _logger.error(\"Intermediate output directory prefix doesn't exist.\")\n return 1\n \n if args.tmpdir_prefix != 'tmp':\n # Use user defined prefix (if the folder exists)\n- args.tmpdir_prefix = os.path.abspath(args.tmpdir_prefix)\n+ setattr(args, 'tmpdir_prefix', os.path.abspath(args.tmpdir_prefix))\n if not os.path.exists(args.tmpdir_prefix):\n _logger.error(\"Temporary directory prefix doesn't exist.\")\n return 1\n@@ -633,12 +639,13 @@ def main(argsl=None,\n return job_order_object\n \n if args.cachedir:\n- args.cachedir = os.path.abspath(args.cachedir)\n- args.move_outputs = False\n+ setattr(args, 'cachedir', os.path.abspath(args.cachedir))\n+ setattr(args, 'move_outputs', False)\n \n try:\n- args.tmp_outdir_prefix = args.cachedir if args.cachedir else args.tmp_outdir_prefix\n- args.basedir = job_order_object[1]\n+ setattr(args, 'tmp_outdir_prefix',\n+ args.cachedir if args.cachedir else args.tmp_outdir_prefix)\n+ setattr(args, 'basedir', job_order_object[1])\n del args.workflow\n del args.job_order\n out = executor(tool, job_order_object[0],\ndiff --git a/cwltool/pathmapper.py b/cwltool/pathmapper.py\n--- a/cwltool/pathmapper.py\n+++ b/cwltool/pathmapper.py\n@@ -7,8 +7,8 @@\n _logger = logging.getLogger(\"cwltool\")\n \n \n-def abspath(src, basedir): # type: (str,str) -> str\n- if src.startswith(\"file://\"):\n+def abspath(src, basedir): # type: (unicode, unicode) -> unicode\n+ if src.startswith(u\"file://\"):\n ab = src[7:]\n else:\n ab = src if os.path.isabs(src) else os.path.join(basedir, src)\n@@ -21,31 +21,31 @@ class PathMapper(object):\n (absolute local path, absolute container path)\"\"\"\n \n def __init__(self, referenced_files, basedir):\n- # type: (Set[str], str) -> None\n- self._pathmap = {} # type: Dict[str, Tuple[str, str]]\n+ # type: (Set[unicode], unicode) -> None\n+ self._pathmap = {} # type: Dict[unicode, Tuple[unicode, unicode]]\n self.setup(referenced_files, basedir)\n \n def setup(self, referenced_files, basedir):\n- # type: (Set[str], str) -> None\n+ # type: (Set[unicode], unicode) -> None\n for src in referenced_files:\n ab = abspath(src, basedir)\n self._pathmap[src] = (ab, ab)\n \n- def mapper(self, src): # type: (str) -> Tuple[str,str]\n- if \"#\" in src:\n- i = src.index(\"#\")\n+ def mapper(self, src): # type: (unicode) -> Tuple[unicode, unicode]\n+ if u\"#\" in src:\n+ i = src.index(u\"#\")\n p = self._pathmap[src[:i]]\n return (p[0], p[1] + src[i:])\n else:\n return self._pathmap[src]\n \n- def files(self): # type: () -> List[str]\n+ def files(self): # type: () -> List[unicode]\n return self._pathmap.keys()\n \n- def items(self): # type: () -> List[Tuple[str,Tuple[str,str]]]\n+ def items(self): # type: () -> List[Tuple[unicode, Tuple[unicode, unicode]]]\n return self._pathmap.items()\n \n- def reversemap(self, target): # type: (str) -> Tuple[str, str]\n+ def reversemap(self, target): # type: (unicode) -> Tuple[unicode, unicode]\n for k, v in self._pathmap.items():\n if v[1] == target:\n return (k, v[0])\n@@ -55,8 +55,8 @@ def reversemap(self, target): # type: (str) -> Tuple[str, str]\n class DockerPathMapper(PathMapper):\n \n def __init__(self, referenced_files, basedir):\n- # type: (Set[str], str) -> None\n- self.dirs = {} # type: Dict[str, Union[bool, str]]\n+ # type: (Set[unicode], unicode) -> None\n+ self.dirs = {} # type: Dict[unicode, Union[bool, unicode]]\n super(DockerPathMapper, self).__init__(referenced_files, basedir)\n \n def setup(self, referenced_files, basedir):\n@@ -77,16 +77,16 @@ def setup(self, referenced_files, basedir):\n del self.dirs[d]\n self.dirs[dirn] = True\n \n- prefix = \"job\" + str(random.randint(1, 1000000000)) + \"_\"\n+ prefix = u\"job\" + str(random.randint(1, 1000000000)) + u\"_\"\n \n- names = set() # type: Set[str]\n+ names = set() # type: Set[unicode]\n for d in self.dirs:\n- name = os.path.join(\"/var/lib/cwl\", prefix + os.path.basename(d))\n+ name = os.path.join(u\"/var/lib/cwl\", prefix + os.path.basename(d))\n i = 1\n while name in names:\n i += 1\n- name = os.path.join(\"/var/lib/cwl\",\n- prefix + os.path.basename(d) + str(i))\n+ name = os.path.join(u\"/var/lib/cwl\",\n+ prefix + os.path.basename(d) + unicode(i))\n names.add(name)\n self.dirs[d] = name\n \ndiff --git a/cwltool/process.py b/cwltool/process.py\n--- a/cwltool/process.py\n+++ b/cwltool/process.py\n@@ -20,7 +20,8 @@\n import glob\n from .errors import WorkflowException\n from .pathmapper import abspath\n-from typing import Any, Callable, Generator, Union, IO, AnyStr, Tuple\n+from typing import (Any, AnyStr, Callable, cast, Dict, List, Generator, IO,\n+ Tuple, Union)\n from collections import Iterable\n from rdflib import URIRef\n from rdflib.namespace import RDFS, OWL\n@@ -52,6 +53,7 @@\n \"invocation.md\")\n \n salad_files = ('metaschema.yml',\n+ 'metaschema_base.yml',\n 'salad.md',\n 'field_name.yml',\n 'import_include.md',\n@@ -72,9 +74,9 @@\n 'vocab_res_src.yml',\n 'vocab_res_proc.yml')\n \n-SCHEMA_CACHE = {} # type: Dict[str, Tuple[Loader, Union[avro.schema.Names, avro.schema.SchemaParseException], Dict[unicode,Any], Loader]]\n-SCHEMA_FILE = None\n-SCHEMA_ANY = None\n+SCHEMA_CACHE = {} # type: Dict[str, Tuple[Loader, Union[avro.schema.Names, avro.schema.SchemaParseException], Dict[unicode, Any], Loader]]\n+SCHEMA_FILE = None # type: Dict[unicode, Any]\n+SCHEMA_ANY = None # type: Dict[unicode, Any]\n \n def get_schema(version):\n # type: (str) -> Tuple[Loader, Union[avro.schema.Names, avro.schema.SchemaParseException], Dict[unicode,Any], Loader]\n@@ -107,18 +109,20 @@ def get_schema(version):\n \"https://w3id.org/cwl/CommonWorkflowLanguage.yml\", cache=cache)\n \n global SCHEMA_FILE, SCHEMA_ANY # pylint: disable=global-statement\n- SCHEMA_FILE = SCHEMA_CACHE[version][3].idx[\"https://w3id.org/cwl/cwl#File\"]\n- SCHEMA_ANY = SCHEMA_CACHE[version][3].idx[\"https://w3id.org/cwl/salad#Any\"]\n+ SCHEMA_FILE = cast(Dict[unicode, Any],\n+ SCHEMA_CACHE[version][3].idx[\"https://w3id.org/cwl/cwl#File\"])\n+ SCHEMA_ANY = cast(Dict[unicode, Any],\n+ SCHEMA_CACHE[version][3].idx[\"https://w3id.org/cwl/salad#Any\"])\n \n return SCHEMA_CACHE[version]\n \n def shortname(inputid):\n- # type: (Union[str, unicode]) -> str\n+ # type: (unicode) -> unicode\n d = urlparse.urlparse(inputid)\n if d.fragment:\n- return d.fragment.split(\"/\")[-1].split(\".\")[-1]\n+ return d.fragment.split(u\"/\")[-1].split(u\".\")[-1]\n else:\n- return d.path.split(\"/\")[-1]\n+ return d.path.split(u\"/\")[-1]\n \n \n class UnsupportedRequirement(Exception):\n@@ -202,7 +206,7 @@ def formatSubclassOf(fmt, cls, ontology, visited):\n \n \n def checkFormat(actualFile, inputFormats, ontology):\n- # type: (Union[Dict[str, Any], List[Dict[str, Any]]], Any, Graph) -> None\n+ # type: (Union[Dict[unicode, Any], List, unicode], Union[List[unicode], unicode], Graph) -> None\n for af in aslist(actualFile):\n if \"format\" not in af:\n raise validate.ValidationException(u\"Missing required 'format' for File %s\" % af)\n@@ -212,13 +216,13 @@ def checkFormat(actualFile, inputFormats, ontology):\n raise validate.ValidationException(u\"Incompatible file format %s required format(s) %s\" % (af[\"format\"], inputFormats))\n \n def fillInDefaults(inputs, job):\n- # type: (List[Dict[str, str]], Dict[str, str]) -> None\n+ # type: (List[Dict[unicode, unicode]], Dict[unicode, Union[Dict[unicode, Any], List, unicode]]) -> None\n for inp in inputs:\n- if shortname(inp[\"id\"]) in job:\n+ if shortname(inp[u\"id\"]) in job:\n pass\n- elif shortname(inp[\"id\"]) not in job and \"default\" in inp:\n- job[shortname(inp[\"id\"])] = copy.copy(inp[\"default\"])\n- elif shortname(inp[\"id\"]) not in job and inp[\"type\"][0] == \"null\":\n+ elif shortname(inp[u\"id\"]) not in job and u\"default\" in inp:\n+ job[shortname(inp[u\"id\"])] = copy.copy(inp[u\"default\"])\n+ elif shortname(inp[u\"id\"]) not in job and inp[u\"type\"][0] == u\"null\":\n pass\n else:\n raise validate.ValidationException(\"Missing input parameter `%s`\" % shortname(inp[\"id\"]))\n@@ -227,7 +231,7 @@ class Process(object):\n __metaclass__ = abc.ABCMeta\n \n def __init__(self, toolpath_object, **kwargs):\n- # type: (Dict[str, Any], **Any) -> None\n+ # type: (Dict[unicode, Any], **Any) -> None\n self.metadata = kwargs.get(\"metadata\", {}) # type: Dict[str,Any]\n self.names = None # type: avro.schema.Names\n names = schema_salad.schema.make_avro_schema(\n@@ -298,11 +302,12 @@ def __init__(self, toolpath_object, **kwargs):\n \n \n def _init_job(self, joborder, **kwargs):\n- # type: (Dict[str, str], str, **Any) -> Builder\n+ # type: (Dict[unicode, unicode], **Any) -> Builder\n builder = Builder()\n- builder.job = copy.deepcopy(joborder)\n+ builder.job = cast(Dict[unicode, Union[Dict[unicode, Any], List,\n+ unicode]], copy.deepcopy(joborder))\n \n- fillInDefaults(self.tool[\"inputs\"], builder.job)\n+ fillInDefaults(self.tool[u\"inputs\"], builder.job)\n \n # Validate job order\n try:\n@@ -427,7 +432,7 @@ def visit(self, op):\n \n @abc.abstractmethod\n def job(self, job_order, output_callbacks, **kwargs):\n- # type: (Dict[str, str], str, Callable[[Any, Any], Any], **Any) -> Generator[Any, None, None]\n+ # type: (Dict[unicode, unicode], Callable[[Any, Any], Any], **Any) -> Generator[Any, None, None]\n return None\n \n def empty_subtree(dirpath): # type: (AnyStr) -> bool\n@@ -461,7 +466,7 @@ def uniquename(stem): # type: (unicode) -> unicode\n return u\n \n def scandeps(base, doc, reffields, urlfields, loadref):\n- # type: (str, Any, Set[str], Set[str], Callable[[str, str], Any]) -> List[Dict[str, str]]\n+ # type: (unicode, Any, Set[str], Set[str], Callable[[unicode, str], Any]) -> List[Dict[str, str]]\n r = []\n if isinstance(doc, dict):\n if \"id\" in doc:\ndiff --git a/cwltool/schemas/draft-4/salad/schema_salad/__init__.py b/cwltool/schemas/draft-4/salad/schema_salad/__init__.py\n--- a/cwltool/schemas/draft-4/salad/schema_salad/__init__.py\n+++ b/cwltool/schemas/draft-4/salad/schema_salad/__init__.py\n@@ -1,7 +1,6 @@\n import logging\n import sys\n-if sys.version_info >= (2,7):\n- import typing\n+import typing\n \n __author__ = 'peter.amstutz@curoverse.com'\n \ndiff --git a/cwltool/schemas/draft-4/salad/schema_salad/__main__.py b/cwltool/schemas/draft-4/salad/schema_salad/__main__.py\n--- a/cwltool/schemas/draft-4/salad/schema_salad/__main__.py\n+++ b/cwltool/schemas/draft-4/salad/schema_salad/__main__.py\n@@ -1,6 +1,5 @@\n from . import main\n import sys\n-if sys.version_info >= (2,7):\n- import typing\n+import typing\n \n sys.exit(main.main())\ndiff --git a/cwltool/schemas/draft-4/salad/schema_salad/add_dictlist.py b/cwltool/schemas/draft-4/salad/schema_salad/add_dictlist.py\n--- a/cwltool/schemas/draft-4/salad/schema_salad/add_dictlist.py\n+++ b/cwltool/schemas/draft-4/salad/schema_salad/add_dictlist.py\n@@ -1,8 +1,7 @@\n import sys\n-if sys.version_info >= (2,7):\n- import typing\n+from typing import Any, Dict\n \n-def add_dictlist(di, key, val):\n+def add_dictlist(di, key, val): # type: (Dict, Any, Any) -> None\n if key not in di:\n di[key] = []\n di[key].append(val)\ndiff --git a/cwltool/schemas/draft-4/salad/schema_salad/aslist.py b/cwltool/schemas/draft-4/salad/schema_salad/aslist.py\n--- a/cwltool/schemas/draft-4/salad/schema_salad/aslist.py\n+++ b/cwltool/schemas/draft-4/salad/schema_salad/aslist.py\n@@ -1,8 +1,7 @@\n import sys\n-if sys.version_info >= (2,7):\n- import typing\n+from typing import Any, List\n \n-def aslist(l):\n+def aslist(l): # type: (Any) -> List\n \"\"\"Convenience function to wrap single items and lists, and return lists unchanged.\"\"\"\n \n if isinstance(l, list):\ndiff --git a/cwltool/schemas/draft-4/salad/schema_salad/flatten.py b/cwltool/schemas/draft-4/salad/schema_salad/flatten.py\n--- a/cwltool/schemas/draft-4/salad/schema_salad/flatten.py\n+++ b/cwltool/schemas/draft-4/salad/schema_salad/flatten.py\n@@ -1,9 +1,9 @@\n import sys\n-if sys.version_info >= (2,7):\n- import typing\n+from typing import Any, Tuple\n \n # http://rightfootin.blogspot.com/2006/09/more-on-python-flatten.html\n def flatten(l, ltypes=(list, tuple)):\n+ # type: (Any, Any) -> Any\n if l is None:\n return []\n if not isinstance(l, ltypes):\n@@ -21,4 +21,4 @@ def flatten(l, ltypes=(list, tuple)):\n else:\n l[i:i + 1] = l[i]\n i += 1\n- return ltype(l)\n+ return ltype(l) # type: ignore\ndiff --git a/cwltool/schemas/draft-4/salad/schema_salad/jsonld_context.py b/cwltool/schemas/draft-4/salad/schema_salad/jsonld_context.py\n--- a/cwltool/schemas/draft-4/salad/schema_salad/jsonld_context.py\n+++ b/cwltool/schemas/draft-4/salad/schema_salad/jsonld_context.py\n@@ -1,6 +1,11 @@\n+import collections\n import shutil\n import json\n-import yaml\n+import ruamel.yaml as yaml\n+try:\n+ from ruamel.yaml import CSafeLoader as SafeLoader\n+except ImportError:\n+ from ruamel.yaml import SafeLoader # type: ignore\n import os\n import subprocess\n import copy\n@@ -11,62 +16,74 @@\n from rdflib import Graph, URIRef\n import rdflib.namespace\n from rdflib.namespace import RDF, RDFS\n-try:\n- import urlparse\n-except ImportError:\n- import urllib.parse as urlparse\n- unicode=str\n- basestring=str\n+import urlparse\n import logging\n from .aslist import aslist\n-if sys.version_info >= (2,7):\n- import typing\n+from typing import Any, cast, Dict, Iterable, Tuple, Union\n+from .ref_resolver import Loader\n \n _logger = logging.getLogger(\"salad\")\n \n+\n def pred(datatype, field, name, context, defaultBase, namespaces):\n+ # type: (Dict[str, Union[Dict, str]], Dict, str, Loader.ContextType, str, Dict[str, rdflib.namespace.Namespace]) -> Union[Dict, str]\n split = urlparse.urlsplit(name)\n \n- v = None\n+ vee = None # type: Union[str, unicode]\n \n if split.scheme:\n- v = name\n- (ns, ln) = rdflib.namespace.split_uri(unicode(v))\n+ vee = name\n+ (ns, ln) = rdflib.namespace.split_uri(unicode(vee))\n name = ln\n if ns[0:-1] in namespaces:\n- v = unicode(namespaces[ns[0:-1]][ln])\n- _logger.debug(\"name, v %s %s\", name, v)\n+ vee = unicode(namespaces[ns[0:-1]][ln])\n+ _logger.debug(\"name, v %s %s\", name, vee)\n+\n+ v = None # type: Any\n \n if field and \"jsonldPredicate\" in field:\n if isinstance(field[\"jsonldPredicate\"], dict):\n v = {}\n for k, val in field[\"jsonldPredicate\"].items():\n- v[(\"@\"+k[1:] if k.startswith(\"_\") else k)] = val\n+ v[(\"@\" + k[1:] if k.startswith(\"_\") else k)] = val\n else:\n v = field[\"jsonldPredicate\"]\n elif \"jsonldPredicate\" in datatype:\n- for d in datatype[\"jsonldPredicate\"]:\n- if d[\"symbol\"] == name:\n- v = d[\"predicate\"]\n+ if isinstance(datatype[\"jsonldPredicate\"], collections.Iterable):\n+ for d in datatype[\"jsonldPredicate\"]:\n+ if isinstance(d, dict):\n+ if d[\"symbol\"] == name:\n+ v = d[\"predicate\"]\n+ else:\n+ raise Exception(\n+ \"entries in the jsonldPredicate List must be \"\n+ \"Dictionaries\")\n+ else:\n+ raise Exception(\"jsonldPredicate must be a List of Dictionaries.\")\n # if not v:\n # if field and \"jsonldPrefix\" in field:\n # defaultBase = field[\"jsonldPrefix\"]\n # elif \"jsonldPrefix\" in datatype:\n # defaultBase = datatype[\"jsonldPrefix\"]\n \n- if not v:\n- v = defaultBase + name\n+ ret = v or vee\n+\n+ if not ret:\n+ ret = defaultBase + name\n \n if name in context:\n- if context[name] != v:\n- raise Exception(\"Predicate collision on %s, '%s' != '%s'\" % (name, context[name], v))\n+ if context[name] != ret:\n+ raise Exception(\"Predicate collision on %s, '%s' != '%s'\" %\n+ (name, context[name], ret))\n else:\n- _logger.debug(\"Adding to context '%s' %s (%s)\", name, v, type(v))\n- context[name] = v\n+ _logger.debug(\"Adding to context '%s' %s (%s)\", name, ret, type(ret))\n+ context[name] = ret\n+\n+ return ret\n \n- return v\n \n def process_type(t, g, context, defaultBase, namespaces, defaultPrefix):\n+ # type: (Dict[str, Any], Graph, Loader.ContextType, str, Dict[str, rdflib.namespace.Namespace], str) -> None\n if t[\"type\"] == \"record\":\n recordname = t[\"name\"]\n \n@@ -86,12 +103,14 @@ def process_type(t, g, context, defaultBase, namespaces, defaultPrefix):\n predicate = \"%s:%s\" % (defaultPrefix, recordname)\n \n if context.get(recordname, predicate) != predicate:\n- raise Exception(\"Predicate collision on '%s', '%s' != '%s'\" % (recordname, context[recordname], predicate))\n+ raise Exception(\"Predicate collision on '%s', '%s' != '%s'\" % (\n+ recordname, context[recordname], predicate))\n \n if not recordname:\n raise Exception()\n \n- _logger.debug(\"Adding to context '%s' %s (%s)\", recordname, predicate, type(predicate))\n+ _logger.debug(\"Adding to context '%s' %s (%s)\",\n+ recordname, predicate, type(predicate))\n context[recordname] = predicate\n \n for i in t.get(\"fields\", []):\n@@ -119,7 +138,8 @@ def process_type(t, g, context, defaultBase, namespaces, defaultPrefix):\n # TODO generate range from datatype.\n \n if isinstance(i[\"type\"], dict) and \"name\" in i[\"type\"]:\n- process_type(i[\"type\"], g, context, defaultBase, namespaces, defaultPrefix)\n+ process_type(i[\"type\"], g, context, defaultBase,\n+ namespaces, defaultPrefix)\n \n if \"extends\" in t:\n for e in aslist(t[\"extends\"]):\n@@ -132,31 +152,26 @@ def process_type(t, g, context, defaultBase, namespaces, defaultPrefix):\n \n \n def salad_to_jsonld_context(j, schema_ctx):\n- context = {}\n+ # type: (Iterable, Dict[str, Any]) -> Tuple[Loader.ContextType, Graph]\n+ context = {} # type: Loader.ContextType\n namespaces = {}\n g = Graph()\n defaultPrefix = \"\"\n \n- for k,v in schema_ctx.items():\n+ for k, v in schema_ctx.items():\n context[k] = v\n namespaces[k] = rdflib.namespace.Namespace(v)\n \n if \"@base\" in context:\n- defaultBase = context[\"@base\"]\n+ defaultBase = cast(str, context[\"@base\"])\n del context[\"@base\"]\n else:\n defaultBase = \"\"\n \n- for k,v in namespaces.items():\n+ for k, v in namespaces.items():\n g.bind(k, v)\n \n for t in j:\n process_type(t, g, context, defaultBase, namespaces, defaultPrefix)\n \n return (context, g)\n-\n-if __name__ == \"__main__\":\n- with open(sys.argv[1]) as f:\n- j = yaml.load(f)\n- (ctx, g) = salad_to_jsonld_context(j)\n- print(json.dumps(ctx, indent=4, sort_keys=True))\ndiff --git a/cwltool/schemas/draft-4/salad/schema_salad/main.py b/cwltool/schemas/draft-4/salad/schema_salad/main.py\n--- a/cwltool/schemas/draft-4/salad/schema_salad/main.py\n+++ b/cwltool/schemas/draft-4/salad/schema_salad/main.py\n@@ -9,31 +9,29 @@\n import json\n from rdflib import Graph, plugin\n from rdflib.serializer import Serializer\n-import yaml\n import os\n-try:\n- import urlparse\n-except ImportError:\n- import urllib.parse as urlparse\n+import urlparse\n \n from .ref_resolver import Loader\n from . import validate\n-if sys.version_info >= (2,7):\n- import typing\n+from typing import Any, Dict, List, Union\n \n _logger = logging.getLogger(\"salad\")\n \n from rdflib.plugin import register, Parser\n-import rdflib_jsonld.parser\n register('json-ld', Parser, 'rdflib_jsonld.parser', 'JsonLDParser')\n \n+\n def printrdf(workflow, wf, ctx, sr):\n- g = Graph().parse(data=json.dumps(wf), format='json-ld', location=workflow, context=ctx)\n+ # type: (str, Union[List, Dict[Any, Any], str, unicode], Dict[unicode, Any], str) -> None\n+ g = Graph().parse(data=json.dumps(wf), format='json-ld',\n+ location=workflow, context=ctx)\n print(g.serialize(format=sr))\n \n-def main(args=None):\n- if args is None:\n- args = sys.argv[1:]\n+\n+def main(argsl=None): # type: (List[str]) -> int\n+ if argsl is None:\n+ argsl = sys.argv[1:]\n \n parser = argparse.ArgumentParser()\n parser.add_argument(\"--rdf-serializer\",\n@@ -41,15 +39,23 @@ def main(args=None):\n default=\"turtle\")\n \n exgroup = parser.add_mutually_exclusive_group()\n- exgroup.add_argument(\"--print-jsonld-context\", action=\"store_true\", help=\"Print JSON-LD context for schema\")\n- exgroup.add_argument(\"--print-rdfs\", action=\"store_true\", help=\"Print RDF schema\")\n- exgroup.add_argument(\"--print-avro\", action=\"store_true\", help=\"Print Avro schema\")\n-\n- exgroup.add_argument(\"--print-rdf\", action=\"store_true\", help=\"Print corresponding RDF graph for document\")\n- exgroup.add_argument(\"--print-pre\", action=\"store_true\", help=\"Print document after preprocessing\")\n- exgroup.add_argument(\"--print-index\", action=\"store_true\", help=\"Print node index\")\n- exgroup.add_argument(\"--print-metadata\", action=\"store_true\", help=\"Print document metadata\")\n- exgroup.add_argument(\"--version\", action=\"store_true\", help=\"Print version\")\n+ exgroup.add_argument(\"--print-jsonld-context\", action=\"store_true\",\n+ help=\"Print JSON-LD context for schema\")\n+ exgroup.add_argument(\n+ \"--print-rdfs\", action=\"store_true\", help=\"Print RDF schema\")\n+ exgroup.add_argument(\"--print-avro\", action=\"store_true\",\n+ help=\"Print Avro schema\")\n+\n+ exgroup.add_argument(\"--print-rdf\", action=\"store_true\",\n+ help=\"Print corresponding RDF graph for document\")\n+ exgroup.add_argument(\"--print-pre\", action=\"store_true\",\n+ help=\"Print document after preprocessing\")\n+ exgroup.add_argument(\n+ \"--print-index\", action=\"store_true\", help=\"Print node index\")\n+ exgroup.add_argument(\"--print-metadata\",\n+ action=\"store_true\", help=\"Print document metadata\")\n+ exgroup.add_argument(\"--version\", action=\"store_true\",\n+ help=\"Print version\")\n \n exgroup = parser.add_mutually_exclusive_group()\n exgroup.add_argument(\"--strict\", action=\"store_true\", help=\"Strict validation (unrecognized or out of place fields are error)\",\n@@ -58,14 +64,17 @@ def main(args=None):\n default=True, dest=\"strict\")\n \n exgroup = parser.add_mutually_exclusive_group()\n- exgroup.add_argument(\"--verbose\", action=\"store_true\", help=\"Default logging\")\n- exgroup.add_argument(\"--quiet\", action=\"store_true\", help=\"Only print warnings and errors.\")\n- exgroup.add_argument(\"--debug\", action=\"store_true\", help=\"Print even more logging\")\n+ exgroup.add_argument(\"--verbose\", action=\"store_true\",\n+ help=\"Default logging\")\n+ exgroup.add_argument(\"--quiet\", action=\"store_true\",\n+ help=\"Only print warnings and errors.\")\n+ exgroup.add_argument(\"--debug\", action=\"store_true\",\n+ help=\"Print even more logging\")\n \n parser.add_argument(\"schema\", type=str)\n parser.add_argument(\"document\", type=str, nargs=\"?\", default=None)\n \n- args = parser.parse_args(args)\n+ args = parser.parse_args(argsl)\n \n if args.quiet:\n _logger.setLevel(logging.WARN)\n@@ -89,7 +98,16 @@ def main(args=None):\n if not urlparse.urlparse(schema_uri)[0]:\n schema_uri = \"file://\" + os.path.abspath(schema_uri)\n schema_raw_doc = metaschema_loader.fetch(schema_uri)\n- schema_doc, schema_metadata = metaschema_loader.resolve_all(schema_raw_doc, schema_uri)\n+\n+ try:\n+ schema_doc, schema_metadata = metaschema_loader.resolve_all(\n+ schema_raw_doc, schema_uri)\n+ except (validate.ValidationException) as e:\n+ _logger.error(\"Schema `%s` failed link checking:\\n%s\",\n+ args.schema, e, exc_info=(e if args.debug else False))\n+ _logger.debug(\"Index is %s\", metaschema_loader.idx.keys())\n+ _logger.debug(\"Vocabulary is %s\", metaschema_loader.vocab.keys())\n+ return 1\n \n # Optionally print the schema after ref resolution\n if not args.document and args.print_pre:\n@@ -100,20 +118,13 @@ def main(args=None):\n print(json.dumps(metaschema_loader.idx.keys(), indent=4))\n return 0\n \n- # Validate links in the schema document\n- try:\n- metaschema_loader.validate_links(schema_doc)\n- except (validate.ValidationException) as e:\n- _logger.error(\"Schema `%s` failed link checking:\\n%s\", args.schema, e, exc_info=(e if args.debug else False))\n- _logger.debug(\"Index is %s\", metaschema_loader.idx.keys())\n- _logger.debug(\"Vocabulary is %s\", metaschema_loader.vocab.keys())\n- return 1\n-\n # Validate the schema document against the metaschema\n try:\n- schema.validate_doc(metaschema_names, schema_doc, metaschema_loader, args.strict)\n+ schema.validate_doc(metaschema_names, schema_doc,\n+ metaschema_loader, args.strict)\n except validate.ValidationException as e:\n- _logger.error(\"While validating schema `%s`:\\n%s\" % (args.schema, str(e)))\n+ _logger.error(\"While validating schema `%s`:\\n%s\" %\n+ (args.schema, str(e)))\n return 1\n \n # Get the json-ld context and RDFS representation from the schema\n@@ -127,11 +138,18 @@ def main(args=None):\n # Create the loader that will be used to load the target document.\n document_loader = Loader(schema_ctx)\n \n- # Make the Avro validation that will be used to validate the target document\n- (avsc_names, avsc_obj) = schema.make_avro_schema(schema_doc, document_loader)\n+ # Make the Avro validation that will be used to validate the target\n+ # document\n+ if isinstance(schema_doc, list):\n+ (avsc_names, avsc_obj) = schema.make_avro_schema(\n+ schema_doc, document_loader)\n+ else:\n+ _logger.error(\"Schema `%s` must be a list.\", args.schema)\n+ return 1\n \n if isinstance(avsc_names, Exception):\n- _logger.error(\"Schema `%s` error:\\n%s\", args.schema, avsc_names, exc_info=(avsc_names if args.debug else False))\n+ _logger.error(\"Schema `%s` error:\\n%s\", args.schema,\n+ avsc_names, exc_info=(avsc_names if args.debug else False))\n if args.print_avro:\n print(json.dumps(avsc_obj, indent=4))\n return 1\n@@ -168,7 +186,8 @@ def main(args=None):\n doc = \"file://\" + os.path.abspath(uri)\n document, doc_metadata = document_loader.resolve_ref(uri)\n except (validate.ValidationException, RuntimeError) as e:\n- _logger.error(\"Document `%s` failed validation:\\n%s\", args.document, e, exc_info=(e if args.debug else False))\n+ _logger.error(\"Document `%s` failed validation:\\n%s\",\n+ args.document, e, exc_info=(e if args.debug else False))\n return 1\n \n # Optionally print the document after ref resolution\n@@ -180,19 +199,13 @@ def main(args=None):\n print(json.dumps(document_loader.idx.keys(), indent=4))\n return 0\n \n- # Validate links in the target document\n- try:\n- document_loader.validate_links(document)\n- except (validate.ValidationException) as e:\n- _logger.error(\"Document `%s` failed link checking:\\n%s\", args.document, e, exc_info=(e if args.debug else False))\n- _logger.debug(\"Index is %s\", json.dumps(document_loader.idx.keys(), indent=4))\n- return 1\n-\n # Validate the schema document against the metaschema\n try:\n- schema.validate_doc(avsc_names, document, document_loader, args.strict)\n+ schema.validate_doc(avsc_names, document,\n+ document_loader, args.strict)\n except validate.ValidationException as e:\n- _logger.error(\"While validating document `%s`:\\n%s\" % (args.document, str(e)))\n+ _logger.error(\"While validating document `%s`:\\n%s\" %\n+ (args.document, str(e)))\n return 1\n \n # Optionally convert the document to RDF\ndiff --git a/cwltool/schemas/draft-4/salad/schema_salad/makedoc.py b/cwltool/schemas/draft-4/salad/schema_salad/makedoc.py\n--- a/cwltool/schemas/draft-4/salad/schema_salad/makedoc.py\n+++ b/cwltool/schemas/draft-4/salad/schema_salad/makedoc.py\n@@ -1,30 +1,23 @@\n import mistune\n from . import schema\n import json\n-import yaml\n import os\n import copy\n import re\n import sys\n-try:\n- import StringIO\n-except ImportError:\n- from io import StringIO\n+import StringIO\n import logging\n-try:\n- import urlparse\n-except ImportError:\n- import urllib.parse as urlparse\n+import urlparse\n from .aslist import aslist\n from .add_dictlist import add_dictlist\n import re\n import argparse\n-if sys.version_info >= (2,7):\n- import typing\n+from typing import Any, IO, Union\n \n _logger = logging.getLogger(\"salad\")\n \n-def has_types(items):\n+\n+def has_types(items): # type: (Any) -> List[basestring]\n r = [] # type: List\n if isinstance(items, dict):\n if items[\"type\"] == \"https://w3id.org/cwl/salad#record\":\n@@ -41,36 +34,46 @@ def has_types(items):\n return [items]\n return []\n \n+\n def linkto(item):\n _, frg = urlparse.urldefrag(item)\n return \"[%s](#%s)\" % (frg, to_id(frg))\n \n+\n class MyRenderer(mistune.Renderer):\n+\n+ def __init__(self): # type: () -> None\n+ super(mistune.Renderer, self).__init__()\n+ self.options = {}\n+\n def header(self, text, level, raw=None):\n return \"\"\"<h%i id=\"%s\">%s</h%i>\"\"\" % (level, to_id(text), text, level)\n \n-def to_id(text):\n+\n+def to_id(text): # type: (Union[str, unicode]) -> Union[str, unicode]\n textid = text\n if text[0] in (\"0\", \"1\", \"2\", \"3\", \"4\", \"5\", \"6\", \"7\", \"8\", \"9\"):\n try:\n- textid = text[text.index(\" \")+1:]\n+ textid = text[text.index(\" \") + 1:]\n except ValueError:\n pass\n textid = textid.replace(\" \", \"_\")\n return textid\n \n+\n class ToC(object):\n- def __init__(self):\n+\n+ def __init__(self): # type: () -> None\n self.first_toc_entry = True\n self.numbering = [0]\n self.toc = \"\"\n self.start_numbering = True\n \n- def add_entry(self, thisdepth, title):\n+ def add_entry(self, thisdepth, title): # type: (int, str) -> str\n depth = len(self.numbering)\n if thisdepth < depth:\n self.toc += \"</ol>\"\n- for n in range(0, depth-thisdepth):\n+ for n in range(0, depth - thisdepth):\n self.numbering.pop()\n self.toc += \"</li></ol>\"\n self.numbering[-1] += 1\n@@ -84,16 +87,17 @@ def add_entry(self, thisdepth, title):\n self.numbering.append(1)\n \n if self.start_numbering:\n- num = \"%i.%s\" % (self.numbering[0], \".\".join([str(n) for n in self.numbering[1:]]))\n+ num = \"%i.%s\" % (self.numbering[0], \".\".join(\n+ [str(n) for n in self.numbering[1:]]))\n else:\n num = \"\"\n- self.toc += \"\"\"<li><a href=\"#%s\">%s %s</a><ol>\\n\"\"\" %(to_id(title),\n- num, title)\n+ self.toc += \"\"\"<li><a href=\"#%s\">%s %s</a><ol>\\n\"\"\" % (to_id(title),\n+ num, title)\n return num\n \n- def contents(self, id):\n+ def contents(self, idn): # type: (str) -> str\n c = \"\"\"<h1 id=\"%s\">Table of contents</h1>\n- <nav class=\"tocnav\"><ol>%s\"\"\" % (id, self.toc)\n+ <nav class=\"tocnav\"><ol>%s\"\"\" % (idn, self.toc)\n c += \"</ol>\"\n for i in range(0, len(self.numbering)):\n c += \"</li></ol>\"\n@@ -111,7 +115,8 @@ def contents(self, id):\n \"https://w3id.org/cwl/salad#enum\",\n \"https://w3id.org/cwl/salad#array\")\n \n-def number_headings(toc, maindoc):\n+\n+def number_headings(toc, maindoc): # type: (ToC, str) -> str\n mdlines = []\n skip = False\n for line in maindoc.splitlines():\n@@ -133,19 +138,27 @@ def number_headings(toc, maindoc):\n maindoc = '\\n'.join(mdlines)\n return maindoc\n \n-def fix_doc(doc):\n+\n+def fix_doc(doc): # type: (Union[List[str], str]) -> str\n if isinstance(doc, list):\n- doc = \"\".join(doc)\n- return \"\\n\".join([re.sub(r\"<([^>@]+@[^>]+)>\", r\"[\\1](mailto:\\1)\", d) for d in doc.splitlines()])\n+ docstr = \"\".join(doc)\n+ else:\n+ docstr = doc\n+ return \"\\n\".join(\n+ [re.sub(r\"<([^>@]+@[^>]+)>\", r\"[\\1](mailto:\\1)\", d)\n+ for d in docstr.splitlines()])\n+\n \n class RenderType(object):\n+\n def __init__(self, toc, j, renderlist, redirects):\n+ # type: (ToC, List[Dict], str, Dict) -> None\n self.typedoc = StringIO.StringIO()\n self.toc = toc\n- self.subs = {} # type: Dict\n- self.docParent = {} # type: Dict\n- self.docAfter = {} # type: Dict\n- self.rendered = set() # type: Set\n+ self.subs = {} # type: Dict[str, str]\n+ self.docParent = {} # type: Dict[str, List]\n+ self.docAfter = {} # type: Dict[str, List]\n+ self.rendered = set() # type: Set[str]\n self.redirects = redirects\n self.title = None # type: str\n \n@@ -153,7 +166,7 @@ def __init__(self, toc, j, renderlist, redirects):\n if \"extends\" in t:\n for e in aslist(t[\"extends\"]):\n add_dictlist(self.subs, e, t[\"name\"])\n- #if \"docParent\" not in t and \"docAfter\" not in t:\n+ # if \"docParent\" not in t and \"docAfter\" not in t:\n # add_dictlist(self.docParent, e, t[\"name\"])\n \n if t.get(\"docParent\"):\n@@ -166,7 +179,7 @@ def __init__(self, toc, j, renderlist, redirects):\n if t.get(\"docAfter\"):\n add_dictlist(self.docAfter, t[\"docAfter\"], t[\"name\"])\n \n- _, _, metaschema_loader = schema.get_metaschema()\n+ metaschema_loader = schema.get_metaschema()[2]\n alltypes = schema.extend_and_specialize(j, metaschema_loader)\n \n self.typemap = {} # type: Dict\n@@ -187,7 +200,7 @@ def __init__(self, toc, j, renderlist, redirects):\n _, frg2 = urlparse.urldefrag(f[\"name\"])\n self.uses[tp].append((frg1, frg2))\n if tp not in basicTypes and tp not in self.record_refs[t[\"name\"]]:\n- self.record_refs[t[\"name\"]].append(tp)\n+ self.record_refs[t[\"name\"]].append(tp)\n except KeyError as e:\n _logger.error(\"Did not find 'type' in %s\", t)\n raise\n@@ -201,6 +214,7 @@ def __init__(self, toc, j, renderlist, redirects):\n self.render_type(f, 1)\n \n def typefmt(self, tp, redirects, nbsp=False):\n+ # type: (Any, Dict[str, str], bool) -> Union[str, unicode]\n global primitiveType\n if isinstance(tp, list):\n if nbsp and len(tp) <= 3:\n@@ -231,8 +245,7 @@ def typefmt(self, tp, redirects, nbsp=False):\n tp = frg\n return \"\"\"<a href=\"#%s\">%s</a>\"\"\" % (to_id(tp), tp)\n \n-\n- def render_type(self, f, depth):\n+ def render_type(self, f, depth): # type: (Dict[str, Any], int) -> None\n if f[\"name\"] in self.rendered or f[\"name\"] in self.redirects:\n return\n self.rendered.add(f[\"name\"])\n@@ -248,6 +261,7 @@ def render_type(self, f, depth):\n f[\"doc\"] = \"\"\n \n def extendsfrom(item, ex):\n+ # type: (Dict[str, Any], List[Dict[str, Any]]) -> None\n if \"extends\" in item:\n for e in aslist(item[\"extends\"]):\n ex.insert(0, self.typemap[e])\n@@ -262,8 +276,9 @@ def extendsfrom(item, ex):\n for i in e[\"doc\"]:\n idx = i.find(\":\")\n if idx > -1:\n- enumDesc[i[:idx]] = i[idx+1:]\n- e[\"doc\"] = [i for i in e[\"doc\"] if i.find(\":\") == -1 or i.find(\" \") < i.find(\":\")]\n+ enumDesc[i[:idx]] = i[idx + 1:]\n+ e[\"doc\"] = [i for i in e[\"doc\"] if i.find(\n+ \":\") == -1 or i.find(\" \") < i.find(\":\")]\n \n f[\"doc\"] = fix_doc(f[\"doc\"])\n \n@@ -294,15 +309,16 @@ def extendsfrom(item, ex):\n if f[\"type\"] == \"documentation\":\n f[\"doc\"] = number_headings(self.toc, f[\"doc\"])\n \n- #if \"extends\" in f:\n+ # if \"extends\" in f:\n # doc += \"\\n\\nExtends \"\n # doc += \", \".join([\" %s\" % linkto(ex) for ex in aslist(f[\"extends\"])])\n- #if f[\"name\"] in self.subs:\n+ # if f[\"name\"] in self.subs:\n # doc += \"\\n\\nExtended by\"\n # doc += \", \".join([\" %s\" % linkto(s) for s in self.subs[f[\"name\"]]])\n- #if f[\"name\"] in self.uses:\n+ # if f[\"name\"] in self.uses:\n # doc += \"\\n\\nReferenced by\"\n- # doc += \", \".join([\" [%s.%s](#%s)\" % (s[0], s[1], to_id(s[0])) for s in self.uses[f[\"name\"]]])\n+ # doc += \", \".join([\" [%s.%s](#%s)\" % (s[0], s[1], to_id(s[0]))\n+ # for s in self.uses[f[\"name\"]]])\n \n doc = doc + \"\\n\\n\" + f[\"doc\"]\n \n@@ -323,16 +339,19 @@ def extendsfrom(item, ex):\n opt = True\n \n desc = i[\"doc\"]\n- #if \"inherited_from\" in i:\n+ # if \"inherited_from\" in i:\n # desc = \"%s _Inherited from %s_\" % (desc, linkto(i[\"inherited_from\"]))\n \n- frg = schema.avro_name(i[\"name\"])\n- tr = \"<td><code>%s</code></td><td>%s</td><td>%s</td><td>%s</td>\" % (frg, self.typefmt(tp, self.redirects), opt, mistune.markdown(desc))\n+ rfrg = schema.avro_name(i[\"name\"])\n+ tr = \"<td><code>%s</code></td><td>%s</td><td>%s</td>\"\\\n+ \"<td>%s</td>\" % (\n+ rfrg, self.typefmt(tp, self.redirects), opt,\n+ mistune.markdown(desc))\n if opt:\n required.append(tr)\n else:\n optional.append(tr)\n- for i in required+optional:\n+ for i in required + optional:\n doc += \"<tr>\" + i + \"</tr>\"\n doc += \"\"\"</table>\"\"\"\n elif f[\"type\"] == \"enum\":\n@@ -342,30 +361,34 @@ def extendsfrom(item, ex):\n for e in ex:\n for i in e.get(\"symbols\", []):\n doc += \"<tr>\"\n- frg = schema.avro_name(i)\n- doc += \"<td><code>%s</code></td><td>%s</td>\" % (frg, enumDesc.get(frg, \"\"))\n+ efrg = schema.avro_name(i)\n+ doc += \"<td><code>%s</code></td><td>%s</td>\" % (\n+ efrg, enumDesc.get(efrg, \"\"))\n doc += \"</tr>\"\n doc += \"\"\"</table>\"\"\"\n f[\"doc\"] = doc\n \n self.typedoc.write(f[\"doc\"])\n \n- subs = self.docParent.get(f[\"name\"], []) + self.record_refs.get(f[\"name\"], [])\n+ subs = self.docParent.get(f[\"name\"], []) + \\\n+ self.record_refs.get(f[\"name\"], [])\n if len(subs) == 1:\n self.render_type(self.typemap[subs[0]], depth)\n else:\n for s in subs:\n- self.render_type(self.typemap[s], depth+1)\n+ self.render_type(self.typemap[s], depth + 1)\n \n for s in self.docAfter.get(f[\"name\"], []):\n self.render_type(self.typemap[s], depth)\n \n+\n def avrold_doc(j, outdoc, renderlist, redirects, brand, brandlink):\n+ # type: (List[Dict[unicode, Any]], IO[Any], str, Dict, str, str) -> None\n toc = ToC()\n toc.start_numbering = False\n \n rt = RenderType(toc, j, renderlist, redirects)\n- content = rt.typedoc.getvalue()\n+ content = rt.typedoc.getvalue() # type: unicode\n \n outdoc.write(\"\"\"\n <!DOCTYPE html>\n@@ -452,22 +475,24 @@ def avrold_doc(j, outdoc, renderlist, redirects, brand, brandlink):\n \n args = parser.parse_args()\n \n- s = []\n+ s = [] # type: List[Dict[unicode, Any]]\n a = args.schema\n with open(a) as f:\n if a.endswith(\"md\"):\n s.append({\"name\": os.path.splitext(os.path.basename(a))[0],\n- \"type\": \"documentation\",\n- \"doc\": f.read().decode(\"utf-8\")\n- })\n+ \"type\": \"documentation\",\n+ \"doc\": f.read().decode(\"utf-8\")\n+ })\n else:\n uri = \"file://\" + os.path.abspath(a)\n- _, _, metaschema_loader = schema.get_metaschema()\n+ metaschema_loader = schema.get_metaschema()[2]\n j, schema_metadata = metaschema_loader.resolve_ref(uri, \"\")\n if isinstance(j, list):\n s.extend(j)\n- else:\n+ elif isinstance(j, dict):\n s.append(j)\n+ else:\n+ raise ValueError(\"Schema must resolve to a list or a dict\")\n \n primitiveType = args.primtype\n redirect = {}\ndiff --git a/cwltool/schemas/draft-4/salad/schema_salad/ref_resolver.py b/cwltool/schemas/draft-4/salad/schema_salad/ref_resolver.py\n--- a/cwltool/schemas/draft-4/salad/schema_salad/ref_resolver.py\n+++ b/cwltool/schemas/draft-4/salad/schema_salad/ref_resolver.py\n@@ -5,31 +5,30 @@\n import logging\n import collections\n import requests\n+import urlparse\n+import re\n+import ruamel.yaml as yaml\n try:\n- import urlparse\n-except:\n- import urllib.parse as urlparse\n-import yaml\n+ from ruamel.yaml import CSafeLoader as SafeLoader\n+except ImportError:\n+ from ruamel.yaml import SafeLoader # type: ignore\n from . import validate\n import pprint\n-try:\n- from StringIO import StringIO\n-except ImportError:\n- from io import StringIO\n- unicode=str\n- basestring=str\n+from StringIO import StringIO\n from .aslist import aslist\n+from .flatten import flatten\n import rdflib\n from rdflib.namespace import RDF, RDFS, OWL\n+from rdflib.plugins.parsers.notation3 import BadSyntax\n import xml.sax\n-if sys.version_info >= (2,7):\n- import typing\n- from typing import Union, Tuple\n+from typing import (Any, AnyStr, Callable, cast, Dict, List, Iterable, Tuple,\n+ TypeVar, Union)\n \n _logger = logging.getLogger(\"salad\")\n \n class NormDict(dict):\n- def __init__(self, normalize=unicode):\n+\n+ def __init__(self, normalize=unicode): # type: (type) -> None\n super(NormDict, self).__init__()\n self.normalize = normalize\n \n@@ -45,6 +44,7 @@ def __delitem__(self, key):\n def __contains__(self, key):\n return super(NormDict, self).__contains__(self.normalize(key))\n \n+\n def merge_properties(a, b):\n c = {}\n for i in a:\n@@ -59,22 +59,32 @@ def merge_properties(a, b):\n \n return c\n \n-def SubLoader(loader):\n- return Loader(loader.ctx, schemagraph=loader.graph, foreign_properties=loader.foreign_properties, idx=loader.idx, cache=loader.cache)\n+\n+def SubLoader(loader): # type: (Loader) -> Loader\n+ return Loader(loader.ctx, schemagraph=loader.graph,\n+ foreign_properties=loader.foreign_properties, idx=loader.idx,\n+ cache=loader.cache)\n+\n \n class Loader(object):\n- def __init__(self, ctx, schemagraph=None, foreign_properties=None, idx=None, cache=None):\n+\n+ ContextType = Dict[unicode, Union[Dict, unicode, Iterable[unicode]]]\n+ DocumentType = TypeVar('DocumentType', List, Dict[unicode, Any])\n+\n+ def __init__(self, ctx, schemagraph=None, foreign_properties=None,\n+ idx=None, cache=None):\n+ # type: (Loader.ContextType, rdflib.Graph, Set[unicode], Dict[unicode, Union[List, Dict[unicode, Any], unicode]], Dict[unicode, Any]) -> None\n normalize = lambda url: urlparse.urlsplit(url).geturl()\n if idx is not None:\n self.idx = idx\n else:\n self.idx = NormDict(normalize)\n \n- self.ctx = {} # type: Dict[str, Union[str, dict, basestring]]\n+ self.ctx = {} # type: Loader.ContextType\n if schemagraph is not None:\n self.graph = schemagraph\n else:\n- self.graph = rdflib.Graph()\n+ self.graph = rdflib.graph.Graph()\n \n if foreign_properties is not None:\n self.foreign_properties = foreign_properties\n@@ -86,41 +96,49 @@ def __init__(self, ctx, schemagraph=None, foreign_properties=None, idx=None, cac\n else:\n self.cache = {}\n \n- self.url_fields = set() # type: Set[str]\n- self.vocab_fields = set() # type: Set\n- self.identifiers = set() # type: Set\n- self.identity_links = set() # type: Set\n- self.standalone = set() # type: Set\n- self.nolinkcheck = set() # type: Set\n- self.vocab = {} # type: Dict\n- self.rvocab = {} # type: Dict\n+ self.url_fields = None # type: Set[unicode]\n+ self.scoped_ref_fields = None # type: Dict[unicode, int]\n+ self.vocab_fields = None # type: Set[unicode]\n+ self.identifiers = None # type: Set[unicode]\n+ self.identity_links = None # type: Set[unicode]\n+ self.standalone = None # type: Set[unicode]\n+ self.nolinkcheck = None # type: Set[unicode]\n+ self.vocab = {} # type: Dict[unicode, unicode]\n+ self.rvocab = {} # type: Dict[unicode, unicode]\n+ self.idmap = None # type: Dict[unicode, Any]\n+ self.mapPredicate = None # type: Dict[unicode, unicode]\n+ self.type_dsl_fields = None # type: Set[unicode]\n \n self.add_context(ctx)\n \n- def expand_url(self, url, base_url, scoped=False, vocab_term=False):\n- if url in (\"@id\", \"@type\"):\n+ def expand_url(self, url, base_url, scoped_id=False, vocab_term=False, scoped_ref=None):\n+ # type: (unicode, unicode, bool, bool, int) -> unicode\n+ if url in (u\"@id\", u\"@type\"):\n return url\n \n if vocab_term and url in self.vocab:\n return url\n \n- if self.vocab and \":\" in url:\n- prefix = url.split(\":\")[0]\n+ if self.vocab and u\":\" in url:\n+ prefix = url.split(u\":\")[0]\n if prefix in self.vocab:\n- url = self.vocab[prefix] + url[len(prefix)+1:]\n+ url = self.vocab[prefix] + url[len(prefix) + 1:]\n \n split = urlparse.urlsplit(url)\n \n- if split.scheme or url.startswith(\"$(\") or url.startswith(\"${\"):\n+ if split.scheme or url.startswith(u\"$(\") or url.startswith(u\"${\"):\n pass\n- elif scoped and not split.fragment:\n+ elif scoped_id and not split.fragment:\n splitbase = urlparse.urlsplit(base_url)\n- frg = \"\"\n+ frg = u\"\"\n if splitbase.fragment:\n- frg = splitbase.fragment + \"/\" + split.path\n+ frg = splitbase.fragment + u\"/\" + split.path\n else:\n frg = split.path\n- url = urlparse.urlunsplit((splitbase.scheme, splitbase.netloc, splitbase.path, splitbase.query, frg))\n+ url = urlparse.urlunsplit(\n+ (splitbase.scheme, splitbase.netloc, splitbase.path, splitbase.query, frg))\n+ elif scoped_ref is not None and not split.fragment:\n+ pass\n else:\n url = urlparse.urljoin(base_url, url)\n \n@@ -129,43 +147,56 @@ def expand_url(self, url, base_url, scoped=False, vocab_term=False):\n else:\n return url\n \n- def _add_properties(self, s):\n- for _, _, rng in self.graph.triples( (s, RDFS.range, None) ):\n- literal = ((str(rng).startswith(\"http://www.w3.org/2001/XMLSchema#\") and not str(rng) == \"http://www.w3.org/2001/XMLSchema#anyURI\") or\n- str(rng) == \"http://www.w3.org/2000/01/rdf-schema#Literal\")\n+ def _add_properties(self, s): # type: (unicode) -> None\n+ for _, _, rng in self.graph.triples((s, RDFS.range, None)):\n+ literal = ((unicode(rng).startswith(\n+ u\"http://www.w3.org/2001/XMLSchema#\") and\n+ not unicode(rng) == u\"http://www.w3.org/2001/XMLSchema#anyURI\")\n+ or unicode(rng) ==\n+ u\"http://www.w3.org/2000/01/rdf-schema#Literal\")\n if not literal:\n- self.url_fields.add(str(s))\n- self.foreign_properties.add(str(s))\n+ self.url_fields.add(unicode(s))\n+ self.foreign_properties.add(s)\n \n- def add_namespaces(self, ns):\n+ def add_namespaces(self, ns): # type: (Dict[unicode, unicode]) -> None\n self.vocab.update(ns)\n \n def add_schemas(self, ns, base_url):\n+ # type: (Union[List[unicode], unicode], unicode) -> None\n for sch in aslist(ns):\n- try:\n- self.graph.parse(urlparse.urljoin(base_url, sch), format=\"xml\")\n- except xml.sax.SAXParseException:\n- self.graph.parse(urlparse.urljoin(base_url, sch), format=\"turtle\")\n-\n- for s, _, _ in self.graph.triples( (None, RDF.type, RDF.Property) ):\n+ for fmt in ['xml', 'turtle', 'rdfa']:\n+ try:\n+ self.graph.parse(urlparse.urljoin(base_url, sch),\n+ format=fmt)\n+ break\n+ except xml.sax.SAXParseException: # type: ignore\n+ pass\n+ except TypeError:\n+ pass\n+ except BadSyntax:\n+ pass\n+\n+ for s, _, _ in self.graph.triples((None, RDF.type, RDF.Property)):\n self._add_properties(s)\n- for s, _, o in self.graph.triples( (None, RDFS.subPropertyOf, None) ):\n+ for s, _, o in self.graph.triples((None, RDFS.subPropertyOf, None)):\n self._add_properties(s)\n self._add_properties(o)\n- for s, _, _ in self.graph.triples( (None, RDFS.range, None) ):\n+ for s, _, _ in self.graph.triples((None, RDFS.range, None)):\n self._add_properties(s)\n- for s, _, _ in self.graph.triples( (None, RDF.type, OWL.ObjectProperty) ):\n+ for s, _, _ in self.graph.triples((None, RDF.type, OWL.ObjectProperty)):\n self._add_properties(s)\n \n- for s, _, _ in self.graph.triples( (None, None, None) ):\n- self.idx[str(s)] = True\n-\n+ for s, _, _ in self.graph.triples((None, None, None)):\n+ self.idx[unicode(s)] = None\n \n def add_context(self, newcontext, baseuri=\"\"):\n+ # type: (Loader.ContextType, unicode) -> None\n if self.vocab:\n- raise validate.ValidationException(\"Refreshing context that already has stuff in it\")\n+ raise validate.ValidationException(\n+ \"Refreshing context that already has stuff in it\")\n \n self.url_fields = set()\n+ self.scoped_ref_fields = {}\n self.vocab_fields = set()\n self.identifiers = set()\n self.identity_links = set()\n@@ -175,39 +206,45 @@ def add_context(self, newcontext, baseuri=\"\"):\n self.mapPredicate = {}\n self.vocab = {}\n self.rvocab = {}\n+ self.type_dsl_fields = set()\n \n- self.ctx.update(_copy_dict_without_key(newcontext, \"@context\"))\n+ self.ctx.update(_copy_dict_without_key(newcontext, u\"@context\"))\n \n _logger.debug(\"ctx is %s\", self.ctx)\n \n- for c in self.ctx:\n- if self.ctx[c] == \"@id\":\n- self.identifiers.add(c)\n- self.identity_links.add(c)\n- elif isinstance(self.ctx[c], dict) and self.ctx[c].get(\"@type\") == \"@id\":\n- self.url_fields.add(c)\n- if self.ctx[c].get(\"identity\", False):\n- self.identity_links.add(c)\n- elif isinstance(self.ctx[c], dict) and self.ctx[c].get(\"@type\") == \"@vocab\":\n- self.url_fields.add(c)\n- self.vocab_fields.add(c)\n-\n- if isinstance(self.ctx[c], dict) and self.ctx[c].get(\"noLinkCheck\"):\n- self.nolinkcheck.add(c)\n-\n- if isinstance(self.ctx[c], dict) and self.ctx[c].get(\"mapSubject\"):\n- self.idmap[c] = self.ctx[c][\"mapSubject\"]\n-\n- if isinstance(self.ctx[c], dict) and self.ctx[c].get(\"mapPredicate\"):\n- self.mapPredicate[c] = self.ctx[c][\"mapPredicate\"]\n-\n- if isinstance(self.ctx[c], dict) and \"@id\" in self.ctx[c]:\n- self.vocab[c] = self.ctx[c][\"@id\"]\n- elif isinstance(self.ctx[c], basestring):\n- self.vocab[c] = self.ctx[c]\n-\n- for k,v in self.vocab.items():\n- self.rvocab[self.expand_url(v, \"\", scoped=False)] = k\n+ for key, value in self.ctx.items():\n+ if value == u\"@id\":\n+ self.identifiers.add(key)\n+ self.identity_links.add(key)\n+ elif isinstance(value, dict) and value.get(u\"@type\") == u\"@id\":\n+ self.url_fields.add(key)\n+ if u\"refScope\" in value:\n+ self.scoped_ref_fields[key] = value[u\"refScope\"]\n+ if value.get(u\"identity\", False):\n+ self.identity_links.add(key)\n+ elif isinstance(value, dict) and value.get(u\"@type\") == u\"@vocab\":\n+ self.url_fields.add(key)\n+ self.vocab_fields.add(key)\n+ if u\"refScope\" in value:\n+ self.scoped_ref_fields[key] = value[u\"refScope\"]\n+ if value.get(u\"typeDSL\"):\n+ self.type_dsl_fields.add(key)\n+ if isinstance(value, dict) and value.get(u\"noLinkCheck\"):\n+ self.nolinkcheck.add(key)\n+\n+ if isinstance(value, dict) and value.get(u\"mapSubject\"):\n+ self.idmap[key] = value[u\"mapSubject\"]\n+\n+ if isinstance(value, dict) and value.get(u\"mapPredicate\"):\n+ self.mapPredicate[key] = value[u\"mapPredicate\"]\n+\n+ if isinstance(value, dict) and u\"@id\" in value:\n+ self.vocab[key] = value[u\"@id\"]\n+ elif isinstance(value, basestring):\n+ self.vocab[key] = value\n+\n+ for k, v in self.vocab.items():\n+ self.rvocab[self.expand_url(v, u\"\", scoped_id=False)] = k\n \n _logger.debug(\"identifiers is %s\", self.identifiers)\n _logger.debug(\"identity_links is %s\", self.identity_links)\n@@ -215,29 +252,31 @@ def add_context(self, newcontext, baseuri=\"\"):\n _logger.debug(\"vocab_fields is %s\", self.vocab_fields)\n _logger.debug(\"vocab is %s\", self.vocab)\n \n+ def resolve_ref(self, ref, base_url=None, checklinks=True):\n+ # type: (Union[Dict[unicode, Any], unicode], unicode, bool) -> Tuple[Union[List, Dict[unicode, Any], unicode], Dict[unicode, Any]]\n+ base_url = base_url or u'file://%s/' % os.path.abspath('.')\n \n- def resolve_ref(self, ref, base_url=None):\n- base_url = base_url or 'file://%s/' % os.path.abspath('.')\n-\n- obj = None\n+ obj = None # type: Dict[unicode, Any]\n inc = False\n \n # If `ref` is a dict, look for special directives.\n if isinstance(ref, dict):\n obj = ref\n- if \"$import\" in ref:\n+ if u\"$import\" in ref:\n if len(obj) == 1:\n- ref = obj[\"$import\"]\n+ ref = obj[u\"$import\"]\n obj = None\n else:\n- raise ValueError(\"'$import' must be the only field in %s\" % (str(obj)))\n- elif \"$include\" in obj:\n+ raise ValueError(\n+ \"'$import' must be the only field in %s\" % (str(obj)))\n+ elif u\"$include\" in obj:\n if len(obj) == 1:\n- ref = obj[\"$include\"]\n+ ref = obj[u\"$include\"]\n inc = True\n obj = None\n else:\n- raise ValueError(\"'$include' must be the only field in %s\" % (str(obj)))\n+ raise ValueError(\n+ \"'$include' must be the only field in %s\" % (str(obj)))\n else:\n ref = None\n for identifier in self.identifiers:\n@@ -245,12 +284,13 @@ def resolve_ref(self, ref, base_url=None):\n ref = obj[identifier]\n break\n if not ref:\n- raise ValueError(\"Object `%s` does not have identifier field in %s\" % (obj, self.identifiers))\n+ raise ValueError(\n+ \"Object `%s` does not have identifier field in %s\" % (obj, self.identifiers))\n \n- if not isinstance(ref, basestring):\n+ if not isinstance(ref, (str, unicode)):\n raise ValueError(\"Must be string: `%s`\" % str(ref))\n \n- url = self.expand_url(ref, base_url, scoped=(obj is not None))\n+ url = self.expand_url(ref, base_url, scoped_id=(obj is not None))\n \n # Has this reference been loaded already?\n if url in self.idx:\n@@ -260,6 +300,7 @@ def resolve_ref(self, ref, base_url=None):\n if inc:\n return self.fetch_text(url), {}\n \n+ doc = None\n if obj:\n for identifier in self.identifiers:\n obj[identifier] = url\n@@ -268,136 +309,242 @@ def resolve_ref(self, ref, base_url=None):\n # Load structured document\n doc_url, frg = urlparse.urldefrag(url)\n if doc_url in self.idx:\n- raise validate.ValidationException(\"Reference `#%s` not found in file `%s`.\" % (frg, doc_url))\n- obj = self.fetch(doc_url)\n+ raise validate.ValidationException(\n+ \"Reference `#%s` not found in file `%s`.\" % (frg, doc_url))\n+ doc = self.fetch(doc_url)\n \n # Recursively expand urls and resolve directives\n- obj, metadata = self.resolve_all(obj, doc_url)\n+ resolved_obj, metadata = self.resolve_all(\n+ doc if doc else obj, doc_url, checklinks=checklinks)\n \n- # Requested reference should be in the index now, otherwise it's a bad reference\n+ # Requested reference should be in the index now, otherwise it's a bad\n+ # reference\n if url is not None:\n if url in self.idx:\n- obj = self.idx[url]\n+ resolved_obj = self.idx[url]\n else:\n- raise RuntimeError(\"Reference `%s` is not in the index. Index contains:\\n %s\" % (url, \"\\n \".join(self.idx)))\n+ raise RuntimeError(\"Reference `%s` is not in the index. \"\n+ \"Index contains:\\n %s\" % (url, \"\\n \".join(self.idx)))\n \n- try:\n- if \"$graph\" in obj:\n- metadata = _copy_dict_without_key(obj, \"$graph\")\n- obj = obj[\"$graph\"]\n- return obj, metadata\n+ if isinstance(resolved_obj, (dict)):\n+ if u\"$graph\" in resolved_obj:\n+ metadata = _copy_dict_without_key(resolved_obj, u\"$graph\")\n+ return resolved_obj[u\"$graph\"], metadata\n else:\n- return obj, metadata\n- except TypeError:\n- return obj, metadata\n+ return resolved_obj, metadata\n+ else:\n+ return resolved_obj, metadata\n+\n+\n+ def _resolve_idmap(self, document, loader):\n+ # type: (Dict[unicode, Union[Dict[unicode, Dict[unicode, unicode]], List[Dict[unicode, Any]]]], Loader) -> None\n+ # Convert fields with mapSubject into lists\n+ # use mapPredicate if the mapped value isn't a dict.\n+ for idmapField in loader.idmap:\n+ if (idmapField in document):\n+ idmapFieldValue = document[idmapField]\n+ if (isinstance(idmapFieldValue, dict)\n+ and \"$import\" not in idmapFieldValue\n+ and \"$include\" not in idmapFieldValue):\n+ ls = []\n+ for k in sorted(idmapFieldValue.keys()):\n+ val = idmapFieldValue[k]\n+ v = None # type: Dict[unicode, Any]\n+ if not isinstance(val, dict):\n+ if idmapField in loader.mapPredicate:\n+ v = {loader.mapPredicate[idmapField]: val}\n+ else:\n+ raise validate.ValidationException(\n+ \"mapSubject '%s' value '%s' is not a dict\"\n+ \"and does not have a mapPredicate\", k, v)\n+ else:\n+ v = val\n+ v[loader.idmap[idmapField]] = k\n+ ls.append(v)\n+ document[idmapField] = ls\n \n- def resolve_all(self, document, base_url, file_base=None):\n+ typeDSLregex = re.compile(ur\"^([^[?]+)(\\[\\])?(\\?)?$\")\n+\n+ def _type_dsl(self, t):\n+ # type: (Union[unicode, Dict, List]) -> Union[unicode, Dict[unicode, unicode], List[Union[unicode, Dict[unicode, unicode]]]]\n+ if not isinstance(t, (str, unicode)):\n+ return t\n+\n+ m = Loader.typeDSLregex.match(t)\n+ if not m:\n+ return t\n+ first = m.group(1)\n+ second = third = None\n+ if m.group(2):\n+ second = {u\"type\": u\"array\",\n+ u\"items\": first}\n+ if m.group(3):\n+ third = [u\"null\", second or first]\n+ return third or second or first\n+\n+ def _resolve_type_dsl(self, document, loader):\n+ # type: (Dict[unicode, Union[unicode, Dict[unicode, unicode], List]], Loader) -> None\n+ for d in loader.type_dsl_fields:\n+ if d in document:\n+ datum = document[d]\n+ if isinstance(datum, (str, unicode)):\n+ document[d] = self._type_dsl(datum)\n+ elif isinstance(datum, list):\n+ document[d] = [self._type_dsl(t) for t in datum]\n+ datum2 = document[d]\n+ if isinstance(datum2, list):\n+ document[d] = flatten(datum2)\n+ seen = [] # type: List[unicode]\n+ uniq = []\n+ for item in document[d]:\n+ if item not in seen:\n+ uniq.append(item)\n+ seen.append(item)\n+ document[d] = uniq\n+\n+ def _resolve_identifier(self, document, loader, base_url):\n+ # type: (Dict[unicode, unicode], Loader, unicode) -> unicode\n+ # Expand identifier field (usually 'id') to resolve scope\n+ for identifer in loader.identifiers:\n+ if identifer in document:\n+ if isinstance(document[identifer], basestring):\n+ document[identifer] = loader.expand_url(\n+ document[identifer], base_url, scoped_id=True)\n+ if (document[identifer] not in loader.idx\n+ or isinstance(\n+ loader.idx[document[identifer]], basestring)):\n+ loader.idx[document[identifer]] = document\n+ base_url = document[identifer]\n+ else:\n+ raise validate.ValidationException(\n+ \"identifier field '%s' must be a string\"\n+ % (document[identifer]))\n+ return base_url\n+\n+ def _resolve_identity(self, document, loader, base_url):\n+ # type: (Dict[unicode, List[unicode]], Loader, unicode) -> None\n+ # Resolve scope for identity fields (fields where the value is the\n+ # identity of a standalone node, such as enum symbols)\n+ for identifer in loader.identity_links:\n+ if identifer in document and isinstance(document[identifer], list):\n+ for n, v in enumerate(document[identifer]):\n+ if isinstance(document[identifer][n], basestring):\n+ document[identifer][n] = loader.expand_url(\n+ document[identifer][n], base_url, scoped_id=True)\n+ if document[identifer][n] not in loader.idx:\n+ loader.idx[document[identifer][\n+ n]] = document[identifer][n]\n+\n+ def _normalize_fields(self, document, loader):\n+ # type: (Dict[unicode, unicode], Loader) -> None\n+ # Normalize fields which are prefixed or full URIn to vocabulary terms\n+ for d in document:\n+ d2 = loader.expand_url(d, u\"\", scoped_id=False, vocab_term=True)\n+ if d != d2:\n+ document[d2] = document[d]\n+ del document[d]\n+\n+ def _resolve_uris(self, document, loader, base_url):\n+ # type: (Dict[unicode, Union[unicode, List[unicode]]], Loader, unicode) -> None\n+ # Resolve remaining URLs based on document base\n+ for d in loader.url_fields:\n+ if d in document:\n+ datum = document[d]\n+ if isinstance(datum, (str, unicode)):\n+ document[d] = loader.expand_url(\n+ datum, base_url, scoped_id=False,\n+ vocab_term=(d in loader.vocab_fields),\n+ scoped_ref=self.scoped_ref_fields.get(d))\n+ elif isinstance(datum, list):\n+ document[d] = [\n+ loader.expand_url(\n+ url, base_url, scoped_id=False,\n+ vocab_term=(d in loader.vocab_fields),\n+ scoped_ref=self.scoped_ref_fields.get(d))\n+ if isinstance(url, (str, unicode))\n+ else url for url in datum]\n+\n+\n+ def resolve_all(self, document, base_url, file_base=None, checklinks=True):\n+ # type: (DocumentType, unicode, unicode, bool) -> Tuple[Union[List, Dict[unicode, Any], unicode], Dict[unicode, Any]]\n loader = self\n- metadata = {} # type: Dict[str, str]\n+ metadata = {} # type: Dict[unicode, Any]\n if file_base is None:\n file_base = base_url\n \n if isinstance(document, dict):\n # Handle $import and $include\n- if ('$import' in document or '$include' in document):\n- return self.resolve_ref(document, file_base)\n+ if (u'$import' in document or u'$include' in document):\n+ return self.resolve_ref(document, base_url=file_base, checklinks=checklinks)\n elif isinstance(document, list):\n pass\n else:\n- return document, metadata\n+ return (document, metadata)\n \n- newctx = None\n+ newctx = None # type: Loader\n if isinstance(document, dict):\n # Handle $base, $profile, $namespaces, $schemas and $graph\n- if \"$base\" in document:\n- base_url = document[\"$base\"]\n+ if u\"$base\" in document:\n+ base_url = document[u\"$base\"]\n \n- if \"$profile\" in document:\n+ if u\"$profile\" in document:\n if not newctx:\n newctx = SubLoader(self)\n- prof = self.fetch(document[\"$profile\"])\n- newctx.add_namespaces(document.get(\"$namespaces\", {}), document[\"$profile\"])\n- newctx.add_schemas(document.get(\"$schemas\", []), document[\"$profile\"])\n+ prof = self.fetch(document[u\"$profile\"])\n+ newctx.add_namespaces(document.get(u\"$namespaces\", {}))\n+ newctx.add_schemas(document.get(\n+ u\"$schemas\", []), document[u\"$profile\"])\n \n- if \"$namespaces\" in document:\n+ if u\"$namespaces\" in document:\n if not newctx:\n newctx = SubLoader(self)\n- newctx.add_namespaces(document[\"$namespaces\"])\n+ newctx.add_namespaces(document[u\"$namespaces\"])\n \n- if \"$schemas\" in document:\n+ if u\"$schemas\" in document:\n if not newctx:\n newctx = SubLoader(self)\n- newctx.add_schemas(document[\"$schemas\"], file_base)\n+ newctx.add_schemas(document[u\"$schemas\"], file_base)\n \n if newctx:\n loader = newctx\n \n- if \"$graph\" in document:\n- metadata = _copy_dict_without_key(document, \"$graph\")\n- document = document[\"$graph\"]\n- metadata, _ = loader.resolve_all(metadata, base_url, file_base)\n+ if u\"$graph\" in document:\n+ metadata = _copy_dict_without_key(document, u\"$graph\")\n+ document = document[u\"$graph\"]\n+ resolved_metadata = loader.resolve_all(metadata, base_url,\n+ file_base=file_base, checklinks=False)[0]\n+ if isinstance(resolved_metadata, dict):\n+ metadata = resolved_metadata\n+ else:\n+ raise validate.ValidationException(\n+ \"Validation error, metadata must be dict: %s\"\n+ % (resolved_metadata))\n \n if isinstance(document, dict):\n- for idmapField in loader.idmap:\n- if idmapField in document and isinstance(document[idmapField], dict):\n- ls = []\n- for k,v in document[idmapField].items():\n- if not isinstance(v, dict):\n- if idmapField in loader.mapPredicate:\n- v = {loader.mapPredicate[idmapField]: v}\n- else:\n- raise validate.ValidationException(\"mapSubject '%s' value '%s' is not a dict and does not have a mapPredicate\", k, v)\n- v[loader.idmap[idmapField]] = k\n- ls.append(v)\n- document[idmapField] = ls\n-\n- for identifer in loader.identifiers:\n- if identifer in document:\n- if isinstance(document[identifer], basestring):\n- document[identifer] = loader.expand_url(document[identifer], base_url, scoped=True)\n- if document[identifer] not in loader.idx or isinstance(loader.idx[document[identifer]], basestring):\n- loader.idx[document[identifer]] = document\n- base_url = document[identifer]\n- else:\n- raise validate.ValidationException(\"identifier field '%s' must be a string\" % (document[identifer]))\n-\n- for identifer in loader.identity_links:\n- if identifer in document and isinstance(document[identifer], list):\n- for n, v in enumerate(document[identifer]):\n- document[identifer][n] = loader.expand_url(document[identifer][n], base_url, scoped=True)\n- if document[identifer][n] not in loader.idx:\n- loader.idx[document[identifer][n]] = document[identifer][n]\n- #else:\n- # raise validate.ValidationException(\"identity field '%s' must be a string\" % (document[identifer]))\n-\n- for d in document:\n- d2 = loader.expand_url(d, \"\", scoped=False, vocab_term=True)\n- if d != d2:\n- document[d2] = document[d]\n- del document[d]\n-\n- for d in loader.url_fields:\n- if d in document:\n- if isinstance(document[d], basestring):\n- document[d] = loader.expand_url(document[d], base_url, scoped=False, vocab_term=(d in loader.vocab_fields))\n- elif isinstance(document[d], list):\n- document[d] = [loader.expand_url(url, base_url, scoped=False, vocab_term=(d in loader.vocab_fields)) if isinstance(url, basestring) else url for url in document[d] ]\n+ self._normalize_fields(document, loader)\n+ self._resolve_idmap(document, loader)\n+ self._resolve_type_dsl(document, loader)\n+ base_url = self._resolve_identifier(document, loader, base_url)\n+ self._resolve_identity(document, loader, base_url)\n+ self._resolve_uris(document, loader, base_url)\n \n try:\n for key, val in document.items():\n- document[key], _ = loader.resolve_all(val, base_url, file_base)\n+ document[key], _ = loader.resolve_all(\n+ val, base_url, file_base=file_base, checklinks=False)\n except validate.ValidationException as v:\n _logger.debug(\"loader is %s\", id(loader))\n- raise validate.ValidationException(\"(%s) (%s) Validation error in field %s:\\n%s\" % (id(loader), file_base, key, validate.indent(str(v))))\n+ raise validate.ValidationException(\"(%s) (%s) Validation error in field %s:\\n%s\" % (\n+ id(loader), file_base, key, validate.indent(str(v))))\n \n elif isinstance(document, list):\n i = 0\n try:\n while i < len(document):\n val = document[i]\n- if isinstance(val, dict) and \"$import\" in val:\n- l, _ = loader.resolve_ref(val, file_base)\n- if isinstance(l, list):\n+ if isinstance(val, dict) and u\"$import\" in val:\n+ l, _ = loader.resolve_ref(val, base_url=file_base, checklinks=False)\n+ if isinstance(l, list): # never true?\n del document[i]\n for item in aslist(l):\n document.insert(i, item)\n@@ -406,27 +553,34 @@ def resolve_all(self, document, base_url, file_base=None):\n document[i] = l\n i += 1\n else:\n- document[i], _ = loader.resolve_all(val, base_url, file_base)\n+ document[i], _ = loader.resolve_all(\n+ val, base_url, file_base=file_base, checklinks=False)\n i += 1\n except validate.ValidationException as v:\n- raise validate.ValidationException(\"(%s) (%s) Validation error in position %i:\\n%s\" % (id(loader), file_base, i, validate.indent(str(v))))\n+ raise validate.ValidationException(\"(%s) (%s) Validation error in position %i:\\n%s\" % (\n+ id(loader), file_base, i, validate.indent(str(v))))\n \n for identifer in loader.identity_links:\n if identifer in metadata:\n- if isinstance(metadata[identifer], basestring):\n- metadata[identifer] = loader.expand_url(metadata[identifer], base_url, scoped=True)\n+ if isinstance(metadata[identifer], (str, unicode)):\n+ metadata[identifer] = loader.expand_url(\n+ metadata[identifer], base_url, scoped_id=True)\n loader.idx[metadata[identifer]] = document\n \n+ if checklinks:\n+ document = self.validate_links(document, u\"\")\n+\n return document, metadata\n \n def fetch_text(self, url):\n+ # type: (unicode) -> unicode\n if url in self.cache:\n return self.cache[url]\n \n split = urlparse.urlsplit(url)\n scheme, path = split.scheme, split.path\n \n- if scheme in ['http', 'https'] and requests:\n+ if scheme in [u'http', u'https'] and requests:\n try:\n resp = requests.get(url)\n resp.raise_for_status()\n@@ -446,18 +600,18 @@ def fetch_text(self, url):\n else:\n raise ValueError('Unsupported scheme in url: %s' % url)\n \n- def fetch(self, url):\n+ def fetch(self, url): # type: (unicode) -> Any\n if url in self.idx:\n return self.idx[url]\n try:\n text = self.fetch_text(url)\n if isinstance(text, bytes):\n- text = StringIO(text.decode('utf-8'))\n+ textIO = StringIO(text.decode('utf-8'))\n else:\n- text = StringIO(text)\n- text.name = url\n- result = yaml.load(text)\n- except yaml.parser.ParserError as e:\n+ textIO = StringIO(text)\n+ textIO.name = url # type: ignore\n+ result = yaml.load(textIO, Loader=SafeLoader)\n+ except yaml.parser.ParserError as e: # type: ignore\n raise validate.ValidationException(\"Syntax error %s\" % (e))\n if isinstance(result, dict) and self.identifiers:\n for identifier in self.identifiers:\n@@ -468,58 +622,97 @@ def fetch(self, url):\n self.idx[url] = result\n return result\n \n- def check_file(self, fn):\n+ def check_file(self, fn): # type: (unicode) -> bool\n if fn.startswith(\"file://\"):\n u = urlparse.urlsplit(fn)\n return os.path.exists(u.path)\n else:\n return False\n \n- def validate_link(self, field, link):\n+ FieldType = TypeVar('FieldType', unicode, List[unicode], Dict[unicode, Any])\n+\n+ def validate_scoped(self, field, link, docid):\n+ # type: (unicode, unicode, unicode) -> unicode\n+ split = urlparse.urlsplit(docid)\n+ sp = split.fragment.split(u\"/\")\n+ n = self.scoped_ref_fields[field]\n+ while n > 0 and len(sp) > 0:\n+ sp.pop()\n+ n -= 1\n+ tried = []\n+ while True:\n+ sp.append(link)\n+ url = urlparse.urlunsplit((\n+ split.scheme, split.netloc, split.path, split.query,\n+ u\"/\".join(sp)))\n+ tried.append(url)\n+ if url in self.idx:\n+ return url\n+ sp.pop()\n+ if len(sp) == 0:\n+ break\n+ sp.pop()\n+ raise validate.ValidationException(\n+ \"Field `%s` contains undefined reference to `%s`, tried %s\" % (field, link, tried))\n+\n+ def validate_link(self, field, link, docid):\n+ # type: (unicode, FieldType, unicode) -> FieldType\n if field in self.nolinkcheck:\n- return True\n- if isinstance(link, basestring):\n+ return link\n+ if isinstance(link, (str, unicode)):\n if field in self.vocab_fields:\n if link not in self.vocab and link not in self.idx and link not in self.rvocab:\n- if not self.check_file(link):\n- raise validate.ValidationException(\"Field `%s` contains undefined reference to `%s`\" % (field, link))\n+ if field in self.scoped_ref_fields:\n+ return self.validate_scoped(field, link, docid)\n+ elif not self.check_file(link):\n+ raise validate.ValidationException(\n+ \"Field `%s` contains undefined reference to `%s`\" % (field, link))\n elif link not in self.idx and link not in self.rvocab:\n- if not self.check_file(link):\n- raise validate.ValidationException(\"Field `%s` contains undefined reference to `%s`\" % (field, link))\n+ if field in self.scoped_ref_fields:\n+ return self.validate_scoped(field, link, docid)\n+ elif not self.check_file(link):\n+ raise validate.ValidationException(\n+ \"Field `%s` contains undefined reference to `%s`\" % (field, link))\n elif isinstance(link, list):\n errors = []\n- for i in link:\n+ for n, i in enumerate(link):\n try:\n- self.validate_link(field, i)\n+ link[n] = self.validate_link(field, i, docid)\n except validate.ValidationException as v:\n errors.append(v)\n if errors:\n- raise validate.ValidationException(\"\\n\".join([str(e) for e in errors]))\n+ raise validate.ValidationException(\n+ \"\\n\".join([str(e) for e in errors]))\n elif isinstance(link, dict):\n- self.validate_links(link)\n- return True\n+ self.validate_links(link, docid)\n+ else:\n+ raise validate.ValidationException(\"Link must be a str, unicode, \"\n+ \"list, or a dict.\")\n+ return link\n \n- def getid(self, d):\n+ def getid(self, d): # type: (Any) -> unicode\n if isinstance(d, dict):\n for i in self.identifiers:\n if i in d:\n- if isinstance(d[i], basestring):\n+ if isinstance(d[i], (str, unicode)):\n return d[i]\n return None\n \n- def validate_links(self, document):\n+ def validate_links(self, document, base_url):\n+ # type: (DocumentType, unicode) -> DocumentType\n docid = self.getid(document)\n- if docid is None:\n- docid = \"\"\n+ if not docid:\n+ docid = base_url\n \n errors = []\n+ iterator = None # type: Any\n if isinstance(document, list):\n iterator = enumerate(document)\n elif isinstance(document, dict):\n try:\n for d in self.url_fields:\n- if d not in self.identity_links and d in document:\n- self.validate_link(d, document[d])\n+ if d in document and d not in self.identity_links:\n+ document[d] = self.validate_link(d, document[d], docid)\n except validate.ValidationException as v:\n errors.append(v)\n if hasattr(document, \"iteritems\"):\n@@ -527,31 +720,36 @@ def validate_links(self, document):\n else:\n iterator = document.items()\n else:\n- return\n+ return document\n \n for key, val in iterator:\n try:\n- self.validate_links(val)\n+ document[key] = self.validate_links(val, docid) # type: ignore\n except validate.ValidationException as v:\n if key not in self.nolinkcheck:\n- docid = self.getid(val)\n- if docid:\n- errors.append(validate.ValidationException(\"While checking object `%s`\\n%s\" % (docid, validate.indent(str(v)))))\n+ docid2 = self.getid(val)\n+ if docid2:\n+ errors.append(validate.ValidationException(\n+ \"While checking object `%s`\\n%s\" % (docid2, validate.indent(str(v)))))\n else:\n if isinstance(key, basestring):\n- errors.append(validate.ValidationException(\"While checking field `%s`\\n%s\" % (key, validate.indent(str(v)))))\n+ errors.append(validate.ValidationException(\n+ \"While checking field `%s`\\n%s\" % (key, validate.indent(str(v)))))\n else:\n- errors.append(validate.ValidationException(\"While checking position %s\\n%s\" % (key, validate.indent(str(v)))))\n+ errors.append(validate.ValidationException(\n+ \"While checking position %s\\n%s\" % (key, validate.indent(str(v)))))\n \n if errors:\n if len(errors) > 1:\n- raise validate.ValidationException(\"\\n\".join([str(e) for e in errors]))\n+ raise validate.ValidationException(\n+ \"\\n\".join([str(e) for e in errors]))\n else:\n raise errors[0]\n- return\n+ return document\n \n \n def _copy_dict_without_key(from_dict, filtered_key):\n+ # type: (Dict, Any) -> Dict\n new_dict = {}\n for key, value in from_dict.items():\n if key != filtered_key:\ndiff --git a/cwltool/schemas/draft-4/salad/schema_salad/schema.py b/cwltool/schemas/draft-4/salad/schema_salad/schema.py\n--- a/cwltool/schemas/draft-4/salad/schema_salad/schema.py\n+++ b/cwltool/schemas/draft-4/salad/schema_salad/schema.py\n@@ -4,49 +4,51 @@\n import sys\n import pprint\n from pkg_resources import resource_stream\n-import yaml\n+import ruamel.yaml as yaml\n+try:\n+ from ruamel.yaml import CSafeLoader as SafeLoader\n+except ImportError:\n+ from ruamel.yaml import SafeLoader # type: ignore\n import avro.schema\n from . import validate\n import json\n-try:\n- import urlparse\n- AvroSchemaFromJSONData=avro.schema.make_avsc_object\n-except ImportError:\n- import urllib.parse as urlparse\n- basestring=str\n- AvroSchemaFromJSONData=avro.schema.SchemaFromJSONData\n+import urlparse\n+AvroSchemaFromJSONData = avro.schema.make_avsc_object\n+# AvroSchemaFromJSONData=avro.schema.SchemaFromJSONData\n from . import ref_resolver\n from .flatten import flatten\n import logging\n from .aslist import aslist\n from . import jsonld_context\n-if sys.version_info >= (2,7):\n- import typing\n+from typing import Any, AnyStr, cast, Dict, List, Tuple, TypeVar, Union\n \n _logger = logging.getLogger(\"salad\")\n \n salad_files = ('metaschema.yml',\n- 'salad.md',\n- 'field_name.yml',\n- 'import_include.md',\n- 'link_res.yml',\n- 'ident_res.yml',\n- 'vocab_res.yml',\n- 'vocab_res.yml',\n- 'field_name_schema.yml',\n- 'field_name_src.yml',\n- 'field_name_proc.yml',\n- 'ident_res_schema.yml',\n- 'ident_res_src.yml',\n- 'ident_res_proc.yml',\n- 'link_res_schema.yml',\n- 'link_res_src.yml',\n- 'link_res_proc.yml',\n- 'vocab_res_schema.yml',\n- 'vocab_res_src.yml',\n- 'vocab_res_proc.yml')\n+ 'metaschema_base.yml',\n+ 'salad.md',\n+ 'field_name.yml',\n+ 'import_include.md',\n+ 'link_res.yml',\n+ 'ident_res.yml',\n+ 'vocab_res.yml',\n+ 'vocab_res.yml',\n+ 'field_name_schema.yml',\n+ 'field_name_src.yml',\n+ 'field_name_proc.yml',\n+ 'ident_res_schema.yml',\n+ 'ident_res_src.yml',\n+ 'ident_res_proc.yml',\n+ 'link_res_schema.yml',\n+ 'link_res_src.yml',\n+ 'link_res_proc.yml',\n+ 'vocab_res_schema.yml',\n+ 'vocab_res_src.yml',\n+ 'vocab_res_proc.yml')\n+\n \n def get_metaschema():\n+ # type: () -> Tuple[avro.schema.Names, List[Dict[unicode, Any]], ref_resolver.Loader]\n loader = ref_resolver.Loader({\n \"Any\": \"https://w3id.org/cwl/salad#Any\",\n \"ArraySchema\": \"https://w3id.org/cwl/salad#ArraySchema\",\n@@ -78,21 +80,22 @@ def get_metaschema():\n \"@id\": \"https://w3id.org/cwl/salad#docAfter\",\n \"@type\": \"@id\"\n },\n- \"docParent\": {\n- \"@id\": \"https://w3id.org/cwl/salad#docParent\",\n- \"@type\": \"@id\"\n- },\n \"docChild\": {\n \"@id\": \"https://w3id.org/cwl/salad#docChild\",\n \"@type\": \"@id\"\n },\n+ \"docParent\": {\n+ \"@id\": \"https://w3id.org/cwl/salad#docParent\",\n+ \"@type\": \"@id\"\n+ },\n \"documentRoot\": \"https://w3id.org/cwl/salad#SchemaDefinedType/documentRoot\",\n \"documentation\": \"https://w3id.org/cwl/salad#documentation\",\n \"double\": \"http://www.w3.org/2001/XMLSchema#double\",\n \"enum\": \"https://w3id.org/cwl/salad#enum\",\n \"extends\": {\n \"@id\": \"https://w3id.org/cwl/salad#extends\",\n- \"@type\": \"@id\"\n+ \"@type\": \"@id\",\n+ \"refScope\": 1\n },\n \"fields\": \"sld:fields\",\n \"float\": \"http://www.w3.org/2001/XMLSchema#float\",\n@@ -100,16 +103,20 @@ def get_metaschema():\n \"int\": \"http://www.w3.org/2001/XMLSchema#int\",\n \"items\": {\n \"@id\": \"https://w3id.org/cwl/salad#items\",\n- \"@type\": \"@vocab\"\n+ \"@type\": \"@vocab\",\n+ \"refScope\": 2\n },\n \"jsonldPredicate\": \"sld:jsonldPredicate\",\n \"long\": \"http://www.w3.org/2001/XMLSchema#long\",\n+ \"mapPredicate\": \"https://w3id.org/cwl/salad#JsonldPredicate/mapPredicate\",\n+ \"mapSubject\": \"https://w3id.org/cwl/salad#JsonldPredicate/mapSubject\",\n \"name\": \"@id\",\n \"noLinkCheck\": \"https://w3id.org/cwl/salad#JsonldPredicate/noLinkCheck\",\n \"null\": \"https://w3id.org/cwl/salad#null\",\n \"rdf\": \"http://www.w3.org/1999/02/22-rdf-syntax-ns#\",\n \"rdfs\": \"http://www.w3.org/2000/01/rdf-schema#\",\n \"record\": \"https://w3id.org/cwl/salad#record\",\n+ \"refScope\": \"https://w3id.org/cwl/salad#JsonldPredicate/refScope\",\n \"sld\": \"https://w3id.org/cwl/salad#\",\n \"specialize\": \"https://w3id.org/cwl/salad#SaladRecordSchema/specialize\",\n \"specializeFrom\": {\n@@ -128,8 +135,11 @@ def get_metaschema():\n },\n \"type\": {\n \"@id\": \"https://w3id.org/cwl/salad#type\",\n- \"@type\": \"@vocab\"\n+ \"@type\": \"@vocab\",\n+ \"refScope\": 2,\n+ \"typeDSL\": True\n },\n+ \"typeDSL\": \"https://w3id.org/cwl/salad#JsonldPredicate/typeDSL\",\n \"xsd\": \"http://www.w3.org/2001/XMLSchema#\"\n })\n \n@@ -142,62 +152,74 @@ def get_metaschema():\n loader.cache[\"https://w3id.org/cwl/salad\"] = rs.read()\n rs.close()\n \n- j = yaml.load(loader.cache[\"https://w3id.org/cwl/salad\"])\n+ j = yaml.load(loader.cache[\"https://w3id.org/cwl/salad\"],\n+ Loader=SafeLoader)\n j, _ = loader.resolve_all(j, \"https://w3id.org/cwl/salad#\")\n \n- #pprint.pprint(j)\n+ # pprint.pprint(j)\n \n (sch_names, sch_obj) = make_avro_schema(j, loader)\n if isinstance(sch_names, Exception):\n- _logger.error(\"Metaschema error, avro was:\\n%s\", json.dumps(sch_obj, indent=4))\n+ _logger.error(\"Metaschema error, avro was:\\n%s\",\n+ json.dumps(sch_obj, indent=4))\n raise sch_names\n validate_doc(sch_names, j, loader, strict=True)\n return (sch_names, j, loader)\n \n+\n def load_schema(schema_ref, cache=None):\n+ # type: (Union[unicode, Dict[unicode, Any]], Dict) -> Tuple[ref_resolver.Loader, Union[avro.schema.Names, avro.schema.SchemaParseException], Dict[unicode, Any], ref_resolver.Loader]\n metaschema_names, metaschema_doc, metaschema_loader = get_metaschema()\n if cache is not None:\n metaschema_loader.cache = cache\n schema_doc, schema_metadata = metaschema_loader.resolve_ref(schema_ref, \"\")\n \n+ if not isinstance(schema_doc, list):\n+ raise ValueError(\"Schema reference must resolve to a list.\")\n+\n validate_doc(metaschema_names, schema_doc, metaschema_loader, True)\n metactx = schema_metadata.get(\"@context\", {})\n metactx.update(schema_metadata.get(\"$namespaces\", {}))\n- (schema_ctx, rdfs) = jsonld_context.salad_to_jsonld_context(schema_doc, metactx)\n+ (schema_ctx, rdfs) = jsonld_context.salad_to_jsonld_context(\n+ schema_doc, metactx)\n \n # Create the loader that will be used to load the target document.\n document_loader = ref_resolver.Loader(schema_ctx, cache=cache)\n \n- # Make the Avro validation that will be used to validate the target document\n+ # Make the Avro validation that will be used to validate the target\n+ # document\n (avsc_names, avsc_obj) = make_avro_schema(schema_doc, document_loader)\n \n- return document_loader, avsc_names, schema_metadata\n+ return document_loader, avsc_names, schema_metadata, metaschema_loader\n \n def load_and_validate(document_loader, avsc_names, document, strict):\n+ # type: (ref_resolver.Loader, avro.schema.Names, Union[Dict[unicode, Any], unicode], bool) -> Tuple[Any, Dict[unicode, Any]]\n if isinstance(document, dict):\n data, metadata = document_loader.resolve_all(document, document[\"id\"])\n else:\n data, metadata = document_loader.resolve_ref(document)\n \n- document_loader.validate_links(data)\n validate_doc(avsc_names, data, document_loader, strict)\n return data, metadata\n \n-def validate_doc(schema_names, validate_doc, loader, strict):\n+\n+def validate_doc(schema_names, doc, loader, strict):\n+ # type: (avro.schema.Names, Union[Dict[unicode, Any], List[Dict[unicode, Any]], unicode], ref_resolver.Loader, bool) -> None\n has_root = False\n for r in schema_names.names.values():\n- if ((hasattr(r, 'get_prop') and r.get_prop(\"documentRoot\")) or (\n- \"documentRoot\" in r.props)):\n+ if ((hasattr(r, 'get_prop') and r.get_prop(u\"documentRoot\")) or (\n+ u\"documentRoot\" in r.props)):\n has_root = True\n break\n \n if not has_root:\n- raise validate.ValidationException(\"No document roots defined in the schema\")\n+ raise validate.ValidationException(\n+ \"No document roots defined in the schema\")\n \n- if isinstance(validate_doc, list):\n- pass\n- elif isinstance(validate_doc, dict):\n- validate_doc = [validate_doc]\n+ if isinstance(doc, list):\n+ validate_doc = doc\n+ elif isinstance(doc, dict):\n+ validate_doc = [doc]\n else:\n raise validate.ValidationException(\"Document must be dict or list\")\n \n@@ -206,30 +228,34 @@ def validate_doc(schema_names, validate_doc, loader, strict):\n errors = []\n success = False\n for r in schema_names.names.values():\n- if ((hasattr(r,\"get_prop\") and r.get_prop(\"documentRoot\")) or (\n- \"documentRoot\" in r.props)):\n+ if ((hasattr(r, \"get_prop\") and r.get_prop(u\"documentRoot\")) or (\n+ u\"documentRoot\" in r.props)):\n try:\n- validate.validate_ex(r, item, loader.identifiers, strict, foreign_properties=loader.foreign_properties)\n+ validate.validate_ex(\n+ r, item, loader.identifiers, strict, foreign_properties=loader.foreign_properties)\n success = True\n break\n except validate.ValidationException as e:\n if hasattr(r, \"get_prop\"):\n- name = r.get_prop(\"name\")\n+ name = r.get_prop(u\"name\")\n elif hasattr(r, \"name\"):\n name = r.name\n- errors.append(\"Could not validate as `%s` because\\n%s\" % (name, validate.indent(str(e), nolead=False)))\n+ errors.append(\"Could not validate as `%s` because\\n%s\" % (\n+ name, validate.indent(str(e), nolead=False)))\n if not success:\n objerr = \"Validation error at position %i\" % pos\n for ident in loader.identifiers:\n if ident in item:\n objerr = \"Validation error in object %s\" % (item[ident])\n break\n- anyerrors.append(\"%s\\n%s\" % (objerr, validate.indent(\"\\n\".join(errors))))\n+ anyerrors.append(\"%s\\n%s\" %\n+ (objerr, validate.indent(\"\\n\".join(errors))))\n if anyerrors:\n raise validate.ValidationException(\"\\n\".join(anyerrors))\n \n \n def replace_type(items, spec, loader, found):\n+ # type: (Any, Dict[unicode, Any], ref_resolver.Loader, Set[unicode]) -> Any\n \"\"\" Go through and replace types in the 'spec' mapping\"\"\"\n \n items = copy.deepcopy(items)\n@@ -252,11 +278,12 @@ def replace_type(items, spec, loader, found):\n elif isinstance(items, list):\n # recursively transform list\n return [replace_type(i, spec, loader, found) for i in items]\n- elif isinstance(items, basestring):\n+ elif isinstance(items, (str, unicode)):\n # found a string which is a symbol corresponding to a type.\n replace_with = None\n if items in loader.vocab:\n- # If it's a vocabulary term, first expand it to its fully qualified URI\n+ # If it's a vocabulary term, first expand it to its fully qualified\n+ # URI\n items = loader.vocab[items]\n \n if items in spec:\n@@ -267,16 +294,20 @@ def replace_type(items, spec, loader, found):\n return replace_type(replace_with, spec, loader, found)\n return items\n \n-def avro_name(url):\n+\n+def avro_name(url): # type: (AnyStr) -> AnyStr\n doc_url, frg = urlparse.urldefrag(url)\n if frg:\n if '/' in frg:\n- return frg[frg.rindex('/')+1:]\n+ return frg[frg.rindex('/') + 1:]\n else:\n return frg\n return url\n \n+Avro = TypeVar('Avro', Dict[unicode, Any], List[Any], unicode)\n+\n def make_valid_avro(items, alltypes, found, union=False):\n+ # type: (Avro, Dict[unicode, Dict[unicode, Any]], Set[unicode], bool) -> Union[Avro, Dict]\n items = copy.deepcopy(items)\n if isinstance(items, dict):\n if items.get(\"name\"):\n@@ -284,10 +315,11 @@ def make_valid_avro(items, alltypes, found, union=False):\n \n if \"type\" in items and items[\"type\"] in (\"https://w3id.org/cwl/salad#record\", \"https://w3id.org/cwl/salad#enum\", \"record\", \"enum\"):\n if (hasattr(items, \"get\") and items.get(\"abstract\")) or (\"abstract\"\n- in items):\n+ in items):\n return items\n if not items.get(\"name\"):\n- raise Exception(\"Named schemas must have a non-empty name: %s\" % items)\n+ raise Exception(\n+ \"Named schemas must have a non-empty name: %s\" % items)\n \n if items[\"name\"] in found:\n return items[\"name\"]\n@@ -295,7 +327,8 @@ def make_valid_avro(items, alltypes, found, union=False):\n found.add(items[\"name\"])\n for n in (\"type\", \"items\", \"values\", \"fields\"):\n if n in items:\n- items[n] = make_valid_avro(items[n], alltypes, found, union=True)\n+ items[n] = make_valid_avro(\n+ items[n], alltypes, found, union=True)\n if \"symbols\" in items:\n items[\"symbols\"] = [avro_name(sym) for sym in items[\"symbols\"]]\n return items\n@@ -304,18 +337,21 @@ def make_valid_avro(items, alltypes, found, union=False):\n for i in items:\n ret.append(make_valid_avro(i, alltypes, found, union=union))\n return ret\n- if union and isinstance(items, basestring):\n+ if union and isinstance(items, (str, unicode)):\n if items in alltypes and avro_name(items) not in found:\n- return make_valid_avro(alltypes[items], alltypes, found, union=union)\n- items = avro_name(items)\n+ return cast(Dict, make_valid_avro(alltypes[items], alltypes, found,\n+ union=union))\n+ items = avro_name(items) # type: ignore\n+ # bug in mypy 0.3.1, fixed in 0.4-dev\n return items\n \n \n def extend_and_specialize(items, loader):\n+ # type: (List[Dict[unicode, Any]], ref_resolver.Loader) -> List[Dict[unicode, Any]]\n \"\"\"Apply 'extend' and 'specialize' to fully materialize derived record\n types.\"\"\"\n \n- types = {}\n+ types = {} # type: Dict[unicode, Any]\n for t in items:\n types[t[\"name\"]] = t\n n = []\n@@ -323,22 +359,24 @@ def extend_and_specialize(items, loader):\n for t in items:\n t = copy.deepcopy(t)\n if \"extends\" in t:\n- spec = {}\n+ spec = {} # type: Dict[unicode, unicode]\n if \"specialize\" in t:\n for sp in aslist(t[\"specialize\"]):\n spec[sp[\"specializeFrom\"]] = sp[\"specializeTo\"]\n \n- exfields = [] # type: List[str]\n- exsym = [] # type: List[str]\n+ exfields = [] # type: List[unicode]\n+ exsym = [] # type: List[unicode]\n for ex in aslist(t[\"extends\"]):\n if ex not in types:\n- raise Exception(\"Extends %s in %s refers to invalid base type\" % (t[\"extends\"], t[\"name\"]))\n+ raise Exception(\"Extends %s in %s refers to invalid base type\" % (\n+ t[\"extends\"], t[\"name\"]))\n \n basetype = copy.deepcopy(types[ex])\n \n if t[\"type\"] == \"record\":\n if spec:\n- basetype[\"fields\"] = replace_type(basetype.get(\"fields\", []), spec, loader, set())\n+ basetype[\"fields\"] = replace_type(\n+ basetype.get(\"fields\", []), spec, loader, set())\n \n for f in basetype.get(\"fields\", []):\n if \"inherited_from\" not in f:\n@@ -352,19 +390,13 @@ def extend_and_specialize(items, loader):\n exfields.extend(t.get(\"fields\", []))\n t[\"fields\"] = exfields\n \n- fieldnames = set() # type: Set[str]\n+ fieldnames = set() # type: Set[unicode]\n for field in t[\"fields\"]:\n if field[\"name\"] in fieldnames:\n- raise validate.ValidationException(\"Field name %s appears twice in %s\" % (field[\"name\"], t[\"name\"]))\n+ raise validate.ValidationException(\n+ \"Field name %s appears twice in %s\" % (field[\"name\"], t[\"name\"]))\n else:\n fieldnames.add(field[\"name\"])\n-\n- for y in [x for x in t[\"fields\"] if x[\"name\"] == \"class\"]:\n- y[\"type\"] = {\"type\": \"enum\",\n- \"symbols\": [r[\"name\"]],\n- \"name\": r[\"name\"]+\"_class\",\n- }\n- y[\"doc\"] = \"Must be `%s` to indicate this is a %s object.\" % (r[\"name\"], r[\"name\"])\n elif t[\"type\"] == \"enum\":\n exsym.extend(t.get(\"symbols\", []))\n t[\"symbol\"] = exsym\n@@ -377,7 +409,7 @@ def extend_and_specialize(items, loader):\n for t in n:\n ex_types[t[\"name\"]] = t\n \n- extended_by = {} # type: Dict[str, str]\n+ extended_by = {} # type: Dict[unicode, unicode]\n for t in n:\n if \"extends\" in t:\n for ex in aslist(t[\"extends\"]):\n@@ -385,29 +417,34 @@ def extend_and_specialize(items, loader):\n add_dictlist(extended_by, ex, ex_types[t[\"name\"]])\n add_dictlist(extended_by, avro_name(ex), ex_types[ex])\n \n+ for t in n:\n+ if t.get(\"abstract\") and t[\"name\"] not in extended_by:\n+ raise validate.ValidationException(\"%s is abstract but missing a concrete subtype\" % t[\"name\"])\n+\n for t in n:\n if \"fields\" in t:\n t[\"fields\"] = replace_type(t[\"fields\"], extended_by, loader, set())\n \n return n\n \n-def make_avro_schema(j, loader):\n- names = avro.schema.Names()\n \n- #pprint.pprint(j)\n+def make_avro_schema(i, loader):\n+ # type: (List[Dict[unicode, Any]], ref_resolver.Loader) -> Tuple[Union[avro.schema.Names,avro.schema.SchemaParseException], List[Dict[unicode, Any]]]\n+ names = avro.schema.Names()\n \n- j = extend_and_specialize(j, loader)\n+ j = extend_and_specialize(i, loader)\n \n- name_dict = {}\n+ name_dict = {} # type: Dict[unicode, Dict[unicode, Any]]\n for t in j:\n name_dict[t[\"name\"]] = t\n j2 = make_valid_avro(j, name_dict, set())\n \n- j3 = [t for t in j2 if isinstance(t, dict) and not t.get(\"abstract\") and t.get(\"type\") != \"documentation\"]\n+ j3 = [t for t in j2 if isinstance(t, dict) and not t.get(\n+ \"abstract\") and t.get(\"type\") != \"documentation\"]\n \n try:\n AvroSchemaFromJSONData(j3, names)\n except avro.schema.SchemaParseException as e:\n- names = e\n+ return (e, j3)\n \n return (names, j3)\ndiff --git a/cwltool/schemas/draft-4/salad/schema_salad/validate.py b/cwltool/schemas/draft-4/salad/schema_salad/validate.py\n--- a/cwltool/schemas/draft-4/salad/schema_salad/validate.py\n+++ b/cwltool/schemas/draft-4/salad/schema_salad/validate.py\n@@ -1,19 +1,14 @@\n import pprint\n import avro.schema\n-import yaml\n import sys\n-try:\n- import urlparse\n-except:\n- import urllib.parse as urlparse\n- basestring=str\n-if sys.version_info >= (2,7):\n- import typing\n+import urlparse\n+from typing import Any\n \n class ValidationException(Exception):\n pass\n \n-def validate(expected_schema, datum, identifiers=[], strict=False, foreign_properties=set()):\n+def validate(expected_schema, datum, identifiers=set(), strict=False, foreign_properties=set()):\n+ # type: (avro.schema.Schema, Any, Set[unicode], bool, Set[unicode]) -> bool\n try:\n return validate_ex(expected_schema, datum, identifiers, strict=strict, foreign_properties=foreign_properties)\n except ValidationException:\n@@ -24,13 +19,13 @@ def validate(expected_schema, datum, identifiers=[], strict=False, foreign_prope\n LONG_MIN_VALUE = -(1 << 63)\n LONG_MAX_VALUE = (1 << 63) - 1\n \n-def indent(v, nolead=False):\n+def indent(v, nolead=False): # type: (str, bool) -> str\n if nolead:\n return v.splitlines()[0] + \"\\n\".join([\" \" + l for l in v.splitlines()[1:]])\n else:\n return \"\\n\".join([\" \" + l for l in v.splitlines()])\n \n-def friendly(v):\n+def friendly(v): # type: (Any) -> Any\n if isinstance(v, avro.schema.NamedSchema):\n return v.name\n if isinstance(v, avro.schema.ArraySchema):\n@@ -42,21 +37,29 @@ def friendly(v):\n else:\n return v\n \n-def multi(v, q=\"\"):\n+def multi(v, q=\"\"): # type: (str, str) -> str\n if '\\n' in v:\n return \"%s%s%s\\n\" % (q, v, q)\n else:\n return \"%s%s%s\" % (q, v, q)\n \n-def vpformat(datum):\n+def vpformat(datum): # type: (Any) -> str\n a = pprint.pformat(datum)\n if len(a) > 160:\n a = a[0:160] + \"[...]\"\n return a\n \n-def validate_ex(expected_schema, datum, identifiers=set(), strict=False, foreign_properties=set()):\n+def validate_ex(expected_schema, datum, identifiers=None, strict=False,\n+ foreign_properties=None):\n+ # type: (avro.schema.Schema, Any, Set[unicode], bool, Set[unicode]) -> bool\n \"\"\"Determine if a python datum is an instance of a schema.\"\"\"\n \n+ if not identifiers:\n+ identifiers = set()\n+\n+ if not foreign_properties:\n+ foreign_properties = set()\n+\n schema_type = expected_schema.type\n \n if schema_type == 'null':\n@@ -100,12 +103,12 @@ def validate_ex(expected_schema, datum, identifiers=set(), strict=False, foreign\n return True\n else:\n raise ValidationException(\"the value `%s` is not float or double\" % vpformat(datum))\n- elif schema_type == 'fixed':\n+ elif isinstance(expected_schema, avro.schema.FixedSchema):\n if isinstance(datum, str) and len(datum) == expected_schema.size:\n return True\n else:\n raise ValidationException(\"the value `%s` is not fixed\" % vpformat(datum))\n- elif schema_type == 'enum':\n+ elif isinstance(expected_schema, avro.schema.EnumSchema):\n if expected_schema.name == \"Any\":\n if datum is not None:\n return True\n@@ -115,7 +118,7 @@ def validate_ex(expected_schema, datum, identifiers=set(), strict=False, foreign\n return True\n else:\n raise ValidationException(\"the value `%s`\\n is not a valid symbol in enum %s, expected one of %s\" % (vpformat(datum), expected_schema.name, \"'\" + \"', '\".join(expected_schema.symbols) + \"'\"))\n- elif schema_type == 'array':\n+ elif isinstance(expected_schema, avro.schema.ArraySchema):\n if isinstance(datum, list):\n for i, d in enumerate(datum):\n try:\n@@ -125,14 +128,14 @@ def validate_ex(expected_schema, datum, identifiers=set(), strict=False, foreign\n return True\n else:\n raise ValidationException(\"the value `%s` is not a list, expected list of %s\" % (vpformat(datum), friendly(expected_schema.items)))\n- elif schema_type == 'map':\n+ elif isinstance(expected_schema, avro.schema.MapSchema):\n if (isinstance(datum, dict) and\n False not in [isinstance(k, basestring) for k in datum.keys()] and\n False not in [validate(expected_schema.values, v, strict=strict) for v in datum.values()]):\n return True\n else:\n raise ValidationException(\"`%s` is not a valid map value, expected\\n %s\" % (vpformat(datum), vpformat(expected_schema.values)))\n- elif schema_type in ['union', 'error_union']:\n+ elif isinstance(expected_schema, avro.schema.UnionSchema):\n if True in [validate(s, datum, identifiers, strict=strict) for s in expected_schema.schemas]:\n return True\n else:\n@@ -144,7 +147,7 @@ def validate_ex(expected_schema, datum, identifiers=set(), strict=False, foreign\n errors.append(str(e))\n raise ValidationException(\"the value %s is not a valid type in the union, expected one of:\\n%s\" % (multi(vpformat(datum), '`'), \"\\n\".join([\"- %s, but\\n %s\" % (friendly(expected_schema.schemas[i]), indent(multi(errors[i]))) for i in range(0, len(expected_schema.schemas))])))\n \n- elif schema_type in ['record', 'error', 'request']:\n+ elif isinstance(expected_schema, avro.schema.RecordSchema):\n if not isinstance(datum, dict):\n raise ValidationException(\"`%s`\\n is not a dict\" % vpformat(datum))\n \ndiff --git a/cwltool/schemas/draft-4/salad/setup.py b/cwltool/schemas/draft-4/salad/setup.py\n--- a/cwltool/schemas/draft-4/salad/setup.py\n+++ b/cwltool/schemas/draft-4/salad/setup.py\n@@ -24,21 +24,24 @@\n # In tox, it will cover them anyway.\n requirements = []\n \n-install_requires=[\n- 'requests',\n- 'PyYAML',\n- 'rdflib >= 4.1.0',\n- 'rdflib-jsonld >= 0.3.0',\n- 'mistune',\n- 'typing']\n+install_requires = [\n+ 'requests',\n+ 'ruamel.yaml',\n+ 'rdflib >= 4.1.0',\n+ 'rdflib-jsonld >= 0.3.0',\n+ 'mistune',\n+ 'typing']\n \n-if sys.version_info.major < 3:\n- install_requires.append(\"avro\")\n-else:\n- install_requires.append(\"avro-python3\")\n+install_requires.append(\"avro\") # TODO: remove me once cwltool is\n+# available in Debian Stable, Ubuntu 12.04 LTS\n+\n+# extras_require={ # TODO: uncomment me, same conditions as above\n+# ':python_version<\"3\"': ['avro'],\n+# ':python_version>=\"3\"': ['avro-python3']}\n+extras_require = {} # TODO: to be removed when the above is added\n \n setup(name='schema-salad',\n- version='1.9',\n+ version='1.12',\n description='Schema Annotations for Linked Avro Data (SALAD)',\n long_description=open(README).read(),\n author='Common workflow language working group',\n@@ -49,6 +52,7 @@\n packages=[\"schema_salad\"],\n package_data={'schema_salad': ['metaschema/*']},\n install_requires=install_requires,\n+ extras_require=extras_require,\n test_suite='tests',\n tests_require=[],\n entry_points={\n@@ -63,7 +67,8 @@\n \"Operating System :: MacOS :: MacOS X\",\n \"Development Status :: 4 - Beta\",\n \"Programming Language :: Python :: 2.7\",\n- \"Programming Language :: Python :: 3.3\",\n- \"Programming Language :: Python :: 3.4\",\n- \"Programming Language :: Python :: 3.5\"]\n+ #\"Programming Language :: Python :: 3.3\", # TODO: uncomment these\n+ #\"Programming Language :: Python :: 3.4\", # lines\n+ #\"Programming Language :: Python :: 3.5\"\n+ ]\n )\ndiff --git a/cwltool/schemas/draft-4/salad/typeshed/2.7/ruamel/__init__.py b/cwltool/schemas/draft-4/salad/typeshed/2.7/ruamel/__init__.py\nnew file mode 100644\ndiff --git a/cwltool/stdfsaccess.py b/cwltool/stdfsaccess.py\n--- a/cwltool/stdfsaccess.py\n+++ b/cwltool/stdfsaccess.py\n@@ -1,4 +1,4 @@\n-from typing import Any, IO\n+from typing import Any, BinaryIO\n from .pathmapper import abspath\n import glob\n import os\n@@ -6,17 +6,17 @@\n \n class StdFsAccess(object):\n \n- def __init__(self, basedir): # type: (str) -> None\n+ def __init__(self, basedir): # type: (unicode) -> None\n self.basedir = basedir\n \n- def _abs(self, p): # type: (str) -> str\n+ def _abs(self, p): # type: (unicode) -> unicode\n return abspath(p, self.basedir)\n \n- def glob(self, pattern): # type: (str) -> List[str]\n+ def glob(self, pattern): # type: (unicode) -> List[unicode]\n return glob.glob(self._abs(pattern))\n \n- def open(self, fn, mode): # type: (str, str) -> IO[Any]\n+ def open(self, fn, mode): # type: (unicode, str) -> BinaryIO\n return open(self._abs(fn), mode)\n \n- def exists(self, fn): # type: (str) -> bool\n+ def exists(self, fn): # type: (unicode) -> bool\n return os.path.exists(self._abs(fn))\ndiff --git a/cwltool/update.py b/cwltool/update.py\n--- a/cwltool/update.py\n+++ b/cwltool/update.py\n@@ -341,6 +341,25 @@ def draft3toDraft4dev1(doc, loader, baseuri):\n \"\"\"Public updater for draft-3 to draft-4.dev1.\"\"\"\n return (_draft3toDraft4dev1(doc, loader, baseuri), \"draft-4.dev1\")\n \n+def _draft4Dev1toDev2(doc, loader, baseuri):\n+ # type: (Any, Loader, str) -> Any\n+ if isinstance(doc, dict):\n+ if \"class\" in doc and doc[\"class\"] == \"Workflow\":\n+ for out in doc[\"outputs\"]:\n+ out[\"outputSource\"] = out[\"source\"]\n+ del out[\"source\"]\n+ for key, value in doc.items():\n+ doc[key] = _draft4Dev1toDev2(value, loader, baseuri)\n+ elif isinstance(doc, list):\n+ doc = [_draft4Dev1toDev2(item, loader, baseuri) for item in doc]\n+\n+ return doc\n+\n+def draft4Dev1toDev2(doc, loader, baseuri):\n+ # type: (Any, Loader, str) -> Tuple[Any, str]\n+ \"\"\"Public updater for draft-4.dev1 to draft-4.dev2.\"\"\"\n+ return (_draft4Dev1toDev2(doc, loader, baseuri), \"draft-4.dev2\")\n+\n UPDATES = {\n \"draft-2\": draft2toDraft3dev1,\n \"draft-3\": draft3toDraft4dev1\n@@ -352,13 +371,14 @@ def draft3toDraft4dev1(doc, loader, baseuri):\n \"draft-3.dev3\": draftDraft3dev3toDev4,\n \"draft-3.dev4\": draftDraft3dev4toDev5,\n \"draft-3.dev5\": draftDraft3dev5toFinal,\n- \"draft-4.dev1\": None\n+ \"draft-4.dev1\": draft4Dev1toDev2,\n+ \"draft-4.dev2\": None\n } # type: Dict[unicode, Callable[[Any, Loader, str], Tuple[Any, str]]]\n \n ALLUPDATES = UPDATES.copy()\n ALLUPDATES.update(DEVUPDATES)\n \n-LATEST = \"draft-4.dev1\"\n+LATEST = \"draft-4.dev2\"\n \n def identity(doc, loader, baseuri): # pylint: disable=unused-argument\n # type: (Any, Loader, str) -> Tuple[Any, Union[str, unicode]]\n@@ -366,19 +386,19 @@ def identity(doc, loader, baseuri): # pylint: disable=unused-argument\n return (doc, doc[\"cwlVersion\"])\n \n def checkversion(doc, metadata, enable_dev):\n- # type: (Union[List, Dict[str, Any]], Dict[str, Any], bool) -> Tuple[Dict[str, Any], Union[str, unicode]] # pylint: disable=line-too-long\n+ # type: (Union[List[Dict[unicode, Any]], Dict[unicode, Any]], Dict[unicode, Any], bool) -> Tuple[Dict[unicode, Any], unicode] # pylint: disable=line-too-long\n \"\"\"Checks the validity of the version of the give CWL document.\n \n Returns the document and the validated version string.\n \"\"\"\n if isinstance(doc, list):\n metadata = metadata.copy()\n- metadata[\"$graph\"] = doc\n+ metadata[u\"$graph\"] = doc\n cdoc = metadata\n else:\n cdoc = doc\n \n- version = cdoc[\"cwlVersion\"]\n+ version = cdoc[u\"cwlVersion\"]\n \n if version not in UPDATES:\n if version in DEVUPDATES:\n@@ -398,16 +418,16 @@ def checkversion(doc, metadata, enable_dev):\n return (cdoc, version)\n \n def update(doc, loader, baseuri, enable_dev, metadata):\n- # type: (Any, Loader, str, bool, Any) -> Any\n+ # type: (Union[List[Dict[unicode, Any]], Dict[unicode, Any]], Loader, str, bool, Any) -> Dict[unicode, Any]\n \n- (doc, version) = checkversion(doc, metadata, enable_dev)\n+ (cdoc, version) = checkversion(doc, metadata, enable_dev)\n \n- nextupdate = identity\n+ nextupdate = identity # type: Callable[[Any, Loader, str], Tuple[Any, str]]\n \n while nextupdate:\n- (doc, version) = nextupdate(doc, loader, baseuri)\n+ (cdoc, version) = nextupdate(cdoc, loader, baseuri)\n nextupdate = ALLUPDATES[version]\n \n- doc[\"cwlVersion\"] = version\n+ cdoc[u\"cwlVersion\"] = version\n \n- return doc\n+ return cdoc\ndiff --git a/cwltool/workflow.py b/cwltool/workflow.py\n--- a/cwltool/workflow.py\n+++ b/cwltool/workflow.py\n@@ -54,7 +54,7 @@ def findfiles(wo, fn=None): # type: (Any, List) -> List[Dict[str, Any]]\n \n \n def match_types(sinktype, src, iid, inputobj, linkMerge, valueFrom):\n- # type: (Union[List[str],str], WorkflowStateItem, str, Dict[str, Any], str, str) -> bool\n+ # type: (Union[List[str],str], WorkflowStateItem, str, Dict[unicode, Any], str, str) -> bool\n if isinstance(sinktype, list):\n # Sink is union type\n for st in sinktype:\n@@ -104,25 +104,37 @@ def are_same_type(src, sink): # type: (Any, Any) -> bool\n return src == sink\n \n \n-def object_from_state(state, parms, frag_only, supportsMultipleInput):\n- # type: (Dict[str,WorkflowStateItem], List[Dict[str, Any]], bool, bool) -> Dict[str, str]\n- inputobj = {} # type: Dict[str, str]\n+def object_from_state(state, parms, frag_only, supportsMultipleInput, sourceField):\n+ # type: (Dict[unicode, WorkflowStateItem], List[Dict[unicode, Any]], bool, bool, unicode) -> Dict[unicode, Any]\n+ inputobj = {} # type: Dict[unicode, Any]\n for inp in parms:\n iid = inp[\"id\"]\n if frag_only:\n iid = shortname(iid)\n- if \"source\" in inp:\n- if isinstance(inp[\"source\"], list) and not supportsMultipleInput:\n- raise WorkflowException(\"Workflow contains multiple inbound links to a single parameter but MultipleInputFeatureRequirement is not declared.\")\n- connections = aslist(inp[\"source\"])\n+ if sourceField in inp:\n+ if (isinstance(inp[sourceField], list) and not\n+ supportsMultipleInput):\n+ raise WorkflowException(\n+ \"Workflow contains multiple inbound links to a single \"\n+ \"parameter but MultipleInputFeatureRequirement is not \"\n+ \"declared.\")\n+ connections = aslist(inp[sourceField])\n for src in connections:\n if src in state and state[src] is not None:\n- if not match_types(inp[\"type\"], state[src], iid, inputobj,\n- inp.get(\"linkMerge\", (\"merge_nested\" if len(connections) > 1 else None)),\n- valueFrom=inp.get(\"valueFrom\")):\n- raise WorkflowException(u\"Type mismatch between source '%s' (%s) and sink '%s' (%s)\" % (src, state[src].parameter[\"type\"], inp[\"id\"], inp[\"type\"]))\n+ if not match_types(\n+ inp[\"type\"], state[src], iid, inputobj,\n+ inp.get(\"linkMerge\", (\"merge_nested\"\n+ if len(connections) > 1 else None)),\n+ valueFrom=inp.get(\"valueFrom\")):\n+ raise WorkflowException(\n+ u\"Type mismatch between source '%s' (%s) and \"\n+ \"sink '%s' (%s)\" % (src,\n+ state[src].parameter[\"type\"], inp[\"id\"],\n+ inp[\"type\"]))\n elif src not in state:\n- raise WorkflowException(u\"Connect source '%s' on parameter '%s' does not exist\" % (src, inp[\"id\"]))\n+ raise WorkflowException(\n+ u\"Connect source '%s' on parameter '%s' does not \"\n+ \"exist\" % (src, inp[\"id\"]))\n else:\n return None\n elif \"default\" in inp:\n@@ -146,7 +158,7 @@ def __init__(self, step): # type: (Any) -> None\n self.name = uniquename(u\"step %s\" % shortname(self.id))\n \n def job(self, joborder, output_callback, **kwargs):\n- # type: (Dict[str,str], str, functools.partial[None], **Any) -> Generator\n+ # type: (Dict[unicode, unicode], functools.partial[None], **Any) -> Generator\n kwargs[\"part_of\"] = self.name\n kwargs[\"name\"] = shortname(self.id)\n for j in self.step.job(joborder, output_callback, **kwargs):\n@@ -161,7 +173,7 @@ def __init__(self, workflow, **kwargs):\n self.tool = workflow.tool\n self.steps = [WorkflowJobStep(s) for s in workflow.steps]\n self.id = workflow.tool[\"id\"]\n- self.state = None # type: Dict[str, WorkflowStateItem]\n+ self.state = None # type: Dict[unicode, WorkflowStateItem]\n self.processStatus = None # type: str\n if \"outdir\" in kwargs:\n self.outdir = kwargs[\"outdir\"]\n@@ -198,14 +210,17 @@ def receive_output(self, step, outputparms, jobout, processStatus):\n step.completed = True\n \n def try_make_job(self, step, **kwargs):\n- # type: (WorkflowJobStep, str, **Any) -> Generator\n+ # type: (WorkflowJobStep, **Any) -> Generator\n inputparms = step.tool[\"inputs\"]\n outputparms = step.tool[\"outputs\"]\n \n- supportsMultipleInput = bool(self.workflow.get_requirement(\"MultipleInputFeatureRequirement\")[0])\n+ supportsMultipleInput = bool(self.workflow.get_requirement(\n+ \"MultipleInputFeatureRequirement\")[0])\n \n try:\n- inputobj = object_from_state(self.state, inputparms, False, supportsMultipleInput)\n+ inputobj = object_from_state(\n+ self.state, inputparms, False, supportsMultipleInput,\n+ \"source\")\n if inputobj is None:\n _logger.debug(u\"[%s] job step %s not ready\", self.name, step.id)\n return\n@@ -217,7 +232,9 @@ def try_make_job(self, step, **kwargs):\n \n callback = functools.partial(self.receive_output, step, outputparms)\n \n- valueFrom = {i[\"id\"]: i[\"valueFrom\"] for i in step.tool[\"inputs\"] if \"valueFrom\" in i}\n+ valueFrom = {\n+ i[\"id\"]: i[\"valueFrom\"] for i in step.tool[\"inputs\"]\n+ if \"valueFrom\" in i}\n \n if len(valueFrom) > 0 and not bool(self.workflow.get_requirement(\"StepInputExpressionRequirement\")[0]):\n raise WorkflowException(\"Workflow step contains valueFrom but StepInputExpressionRequirement not in requirements\")\n@@ -225,8 +242,9 @@ def try_make_job(self, step, **kwargs):\n vfinputs = {shortname(k): v for k,v in inputobj.iteritems()}\n def valueFromFunc(k, v): # type: (Any, Any) -> Any\n if k in valueFrom:\n- return expression.do_eval(valueFrom[k], vfinputs, self.workflow.requirements,\n- None, None, {}, context=v)\n+ return expression.do_eval(\n+ valueFrom[k], vfinputs, self.workflow.requirements,\n+ None, None, {}, context=v)\n else:\n return v\n \n@@ -279,7 +297,7 @@ def run(self, **kwargs):\n _logger.debug(u\"[%s] workflow starting\", self.name)\n \n def job(self, joborder, output_callback, move_outputs=True, **kwargs):\n- # type: (Dict[str,str], str, Callable[[Any, Any], Any], bool, **Any) -> Generator[WorkflowJob, None, None]\n+ # type: (Dict[unicode, Any], Callable[[Any, Any], Any], bool, **Any) -> Generator[WorkflowJob, None, None]\n self.state = {}\n self.processStatus = \"success\"\n \n@@ -326,7 +344,7 @@ def job(self, joborder, output_callback, move_outputs=True, **kwargs):\n \n supportsMultipleInput = bool(self.workflow.get_requirement(\"MultipleInputFeatureRequirement\")[0])\n \n- wo = object_from_state(self.state, self.tool[\"outputs\"], True, supportsMultipleInput)\n+ wo = object_from_state(self.state, self.tool[\"outputs\"], True, supportsMultipleInput, \"outputSource\")\n \n if wo is None:\n raise WorkflowException(\"Output for workflow not available\")\n@@ -375,7 +393,7 @@ def job(self, joborder, output_callback, move_outputs=True, **kwargs):\n \n class Workflow(Process):\n def __init__(self, toolpath_object, **kwargs):\n- # type: (Dict[str, Any], **Any) -> None\n+ # type: (Dict[unicode, Any], **Any) -> None\n super(Workflow, self).__init__(toolpath_object, **kwargs)\n \n kwargs[\"requirements\"] = self.requirements\n@@ -388,7 +406,7 @@ def __init__(self, toolpath_object, **kwargs):\n # TODO: statically validate data links instead of doing it at runtime.\n \n def job(self, joborder, output_callback, **kwargs):\n- # type: (Dict[str,str], str, Callable[[Any, Any], Any], **Any) -> Generator[WorkflowJob, None, None]\n+ # type: (Dict[unicode, unicode], Callable[[Any, Any], Any], **Any) -> Generator[WorkflowJob, None, None]\n builder = self._init_job(joborder, **kwargs)\n wj = WorkflowJob(self, **kwargs)\n yield wj\n@@ -407,7 +425,7 @@ def visit(self, op):\n class WorkflowStep(Process):\n \n def __init__(self, toolpath_object, pos, **kwargs):\n- # type: (Dict[str, Any], int, **Any) -> None\n+ # type: (Dict[unicode, Any], int, **Any) -> None\n if \"id\" in toolpath_object:\n self.id = toolpath_object[\"id\"]\n else:\n@@ -486,7 +504,7 @@ def __init__(self, toolpath_object, pos, **kwargs):\n self.tool[\"outputs\"] = outputparms\n \n def receive_output(self, output_callback, jobout, processStatus):\n- # type: (Callable[...,Any], Dict[str, str], str) -> None\n+ # type: (Callable[...,Any], Dict[unicode, str], str) -> None\n #_logger.debug(\"WorkflowStep output from run is %s\", jobout)\n output = {}\n for i in self.tool[\"outputs\"]:\n@@ -498,7 +516,7 @@ def receive_output(self, output_callback, jobout, processStatus):\n output_callback(output, processStatus)\n \n def job(self, joborder, output_callback, **kwargs):\n- # type: (Dict[str, Any], str, Callable[...,Any], **Any) -> Generator\n+ # type: (Dict[unicode, Any], Callable[...,Any], **Any) -> Generator\n for i in self.tool[\"inputs\"]:\n p = i[\"id\"]\n field = shortname(p)\n@@ -555,7 +573,7 @@ def setTotal(self, total): # type: (int) -> None\n \n \n def dotproduct_scatter(process, joborder, scatter_keys, output_callback, **kwargs):\n- # type: (WorkflowJobStep, Dict[str, Any], str, List[str], Callable[..., Any], **Any) -> Generator[WorkflowJob, None, None]\n+ # type: (WorkflowJobStep, Dict[unicode, Any], List[str], Callable[..., Any], **Any) -> Generator[WorkflowJob, None, None]\n l = None\n for s in scatter_keys:\n if l is None:\n@@ -581,7 +599,7 @@ def dotproduct_scatter(process, joborder, scatter_keys, output_callback, **kwarg\n \n \n def nested_crossproduct_scatter(process, joborder, scatter_keys, output_callback, **kwargs):\n- # type: (WorkflowJobStep, Dict[str, Any], str, List[str], Callable[..., Any], **Any) -> Generator[WorkflowJob, None, None]\n+ # type: (WorkflowJobStep, Dict[unicode, Any], List[str], Callable[..., Any], **Any) -> Generator[WorkflowJob, None, None]\n scatter_key = scatter_keys[0]\n l = len(joborder[scatter_key])\n output = {} # type: Dict[str,List[str]]\n@@ -610,7 +628,7 @@ def nested_crossproduct_scatter(process, joborder, scatter_keys, output_callback\n \n \n def crossproduct_size(joborder, scatter_keys):\n- # type: (Dict[str, Any], List[str]) -> int\n+ # type: (Dict[unicode, Any], List[str]) -> int\n scatter_key = scatter_keys[0]\n if len(scatter_keys) == 1:\n sum = len(joborder[scatter_key])\n@@ -623,7 +641,7 @@ def crossproduct_size(joborder, scatter_keys):\n return sum\n \n def flat_crossproduct_scatter(process, joborder, scatter_keys, output_callback, startindex, **kwargs):\n- # type: (WorkflowJobStep, Dict[str, Any], str, List[str], Union[ReceiveScatterOutput,Callable[..., Any]], int, **Any) -> Generator[WorkflowJob, None, None]\n+ # type: (WorkflowJobStep, Dict[unicode, Any], List[str], Union[ReceiveScatterOutput,Callable[..., Any]], int, **Any) -> Generator[WorkflowJob, None, None]\n scatter_key = scatter_keys[0]\n l = len(joborder[scatter_key])\n rc = None # type: ReceiveScatterOutput\ndiff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -41,7 +41,7 @@\n 'rdflib >= 4.1.0',\n 'rdflib-jsonld >= 0.3.0',\n 'shellescape',\n- 'schema-salad==1.11.20160524152859',\n+ 'schema-salad==1.12.20160610104117',\n 'typing'\n ],\n test_suite='tests',\n", "test_patch": "diff --git a/cwltool/schemas/draft-4/salad/tests/Process.yml b/cwltool/schemas/draft-4/salad/tests/Process.yml\nnew file mode 100644\n--- /dev/null\n+++ b/cwltool/schemas/draft-4/salad/tests/Process.yml\n@@ -0,0 +1,36 @@\n+$base: \"https://w3id.org/cwl/cwl#\"\n+\n+$namespaces:\n+ cwl: \"https://w3id.org/cwl/cwl#\"\n+ sld: \"https://w3id.org/cwl/salad#\"\n+\n+$graph:\n+\n+- $import: \"../schema_salad/metaschema/metaschema_base.yml\"\n+\n+- name: InputBinding\n+ type: record\n+ abstract: true\n+ fields:\n+ - name: loadContents\n+ type:\n+ - \"null\"\n+ - boolean\n+ jsonldPredicate: \"cwl:loadContents\"\n+ doc: |\n+ Only valid when `type: File` or is an array of `items: File`.\n+\n+ Read up to the first 64 KiB of text from the file and place it in the\n+ \"contents\" field of the file object for use by expressions.\n+\n+- name: InputRecordField\n+ type: record\n+ extends: \"sld:RecordField\"\n+ fields:\n+ - name: inputBinding\n+ type: [ \"null\", \"#InputBinding\" ]\n+ jsonldPredicate: \"cwl:inputBinding\"\n+\n+- name: Blurb\n+ type: record\n+ extends: InputBinding\n\\ No newline at end of file\ndiff --git a/cwltool/schemas/draft-4/salad/tests/test_examples.py b/cwltool/schemas/draft-4/salad/tests/test_examples.py\n--- a/cwltool/schemas/draft-4/salad/tests/test_examples.py\n+++ b/cwltool/schemas/draft-4/salad/tests/test_examples.py\n@@ -3,25 +3,28 @@\n import schema_salad.main\n import schema_salad.schema\n import rdflib\n-import yaml\n+import ruamel.yaml as yaml\n+try:\n+ from ruamel.yaml import CSafeLoader as SafeLoader\n+except ImportError:\n+ from ruamel.yaml import SafeLoader\n+\n \n class TestSchemas(unittest.TestCase):\n def test_schemas(self):\n l = schema_salad.ref_resolver.Loader({})\n \n ra, _ = l.resolve_all({\n- \"$schemas\": [\"tests/EDAM.owl\"],\n- \"$namespaces\": {\"edam\": \"http://edamontology.org/\"},\n- \"edam:has_format\": \"edam:format_1915\"\n- }, \"\")\n-\n- self.assertEqual(ra, {\n- \"$schemas\": [\"tests/EDAM.owl\"],\n- \"$namespaces\": {\"edam\": \"http://edamontology.org/\"},\n- 'http://edamontology.org/has_format': 'http://edamontology.org/format_1915'\n- })\n-\n+ u\"$schemas\": [u\"tests/EDAM.owl\"],\n+ u\"$namespaces\": {u\"edam\": u\"http://edamontology.org/\"},\n+ u\"edam:has_format\": u\"edam:format_1915\"\n+ }, \"\")\n \n+ self.assertEqual({\n+ u\"$schemas\": [u\"tests/EDAM.owl\"],\n+ u\"$namespaces\": {u\"edam\": u\"http://edamontology.org/\"},\n+ u'http://edamontology.org/has_format': u'http://edamontology.org/format_1915'\n+ }, ra)\n \n # def test_domain(self):\n # l = schema_salad.ref_resolver.Loader({})\n@@ -43,12 +46,15 @@ def test_schemas(self):\n # })\n \n def test_self_validate(self):\n- schema_salad.main.main(args=[\"schema_salad/metaschema/metaschema.yml\"])\n- schema_salad.main.main(args=[\"schema_salad/metaschema/metaschema.yml\",\n- \"schema_salad/metaschema/metaschema.yml\"])\n+ self.assertEqual(0, schema_salad.main.main(argsl=[\"schema_salad/metaschema/metaschema.yml\"]))\n+ self.assertEqual(0, schema_salad.main.main(argsl=[\"schema_salad/metaschema/metaschema.yml\",\n+ \"schema_salad/metaschema/metaschema.yml\"]))\n+\n+ def test_avro_regression(self):\n+ self.assertEqual(0, schema_salad.main.main(argsl=[\"tests/Process.yml\"]))\n \n def test_jsonld_ctx(self):\n- ldr, _, _ = schema_salad.schema.load_schema({\n+ ldr, _, _, _ = schema_salad.schema.load_schema({\n \"$base\": \"Y\",\n \"name\": \"X\",\n \"$namespaces\": {\n@@ -80,7 +86,7 @@ def test_idmap(self):\n \"@type\": \"@id\",\n \"identity\": True,\n },\n- \"id\": \"@id\"})\n+ \"id\": \"@id\"})\n \n ra, _ = ldr.resolve_all({\n \"id\": \"stuff\",\n@@ -94,28 +100,154 @@ def test_idmap(self):\n }\n }, \"http://example2.com/\")\n \n- self.assertEqual(ra,\n- {\n- \"id\": \"http://example2.com/#stuff\",\n- 'inputs': [{'a': 2, 'id': 'http://example2.com/#stuff/zing'},\n- {'a': 1, 'id': 'http://example2.com/#stuff/zip'}],\n- 'outputs': ['http://example2.com/#stuff/out'],\n- 'other': {\n- 'n': 9\n- }\n- }\n- )\n+ self.assertEqual(\"http://example2.com/#stuff\", ra[\"id\"])\n+ for item in ra[\"inputs\"]:\n+ if item[\"a\"] == 2:\n+ self.assertEquals('http://example2.com/#stuff/zing', item[\"id\"])\n+ else:\n+ self.assertEquals('http://example2.com/#stuff/zip', item[\"id\"])\n+ self.assertEquals(['http://example2.com/#stuff/out'], ra['outputs'])\n+ self.assertEquals({'n': 9}, ra['other'])\n+\n+ def test_scoped_ref(self):\n+ ldr = schema_salad.ref_resolver.Loader({})\n+ ldr.add_context({\n+ \"scatter\": {\n+ \"@type\": \"@id\",\n+ \"refScope\": 0,\n+ },\n+ \"source\": {\n+ \"@type\": \"@id\",\n+ \"refScope\": 2,\n+ },\n+ \"in\": {\n+ \"mapSubject\": \"id\",\n+ \"mapPredicate\": \"source\"\n+ },\n+ \"out\": {\n+ \"@type\": \"@id\",\n+ \"identity\": True\n+ },\n+ \"inputs\": {\n+ \"mapSubject\": \"id\",\n+ \"mapPredicate\": \"type\"\n+ },\n+ \"outputs\": {\n+ \"mapSubject\": \"id\",\n+ },\n+ \"steps\": {\n+ \"mapSubject\": \"id\"\n+ },\n+ \"id\": \"@id\"})\n+\n+ ra, _ = ldr.resolve_all({\n+ \"inputs\": {\n+ \"inp\": \"string\",\n+ \"inp2\": \"string\"\n+ },\n+ \"outputs\": {\n+ \"out\": {\n+ \"type\": \"string\",\n+ \"source\": \"step2/out\"\n+ }\n+ },\n+ \"steps\": {\n+ \"step1\": {\n+ \"in\": {\n+ \"inp\": \"inp\",\n+ \"inp2\": \"#inp2\",\n+ \"inp3\": [\"inp\", \"inp2\"]\n+ },\n+ \"out\": [\"out\"],\n+ \"scatter\": \"inp\"\n+ },\n+ \"step2\": {\n+ \"in\": {\n+ \"inp\": \"step1/out\"\n+ },\n+ \"scatter\": \"inp\",\n+ \"out\": [\"out\"]\n+ }\n+ }\n+ }, \"http://example2.com/\")\n+\n+ self.assertEquals(\n+ {'inputs': [{\n+ 'id': 'http://example2.com/#inp',\n+ 'type': 'string'\n+ }, {\n+ 'id': 'http://example2.com/#inp2',\n+ 'type': 'string'\n+ }],\n+ 'outputs': [{\n+ 'id': 'http://example2.com/#out',\n+ 'type': 'string',\n+ 'source': 'http://example2.com/#step2/out'\n+ }],\n+ 'steps': [{\n+ 'id': 'http://example2.com/#step1',\n+ 'scatter': 'http://example2.com/#step1/inp',\n+ 'in': [{\n+ 'id': 'http://example2.com/#step1/inp',\n+ 'source': 'http://example2.com/#inp'\n+ }, {\n+ 'id': 'http://example2.com/#step1/inp2',\n+ 'source': 'http://example2.com/#inp2'\n+ }, {\n+ 'id': 'http://example2.com/#step1/inp3',\n+ 'source': ['http://example2.com/#inp', 'http://example2.com/#inp2']\n+ }],\n+ \"out\": [\"http://example2.com/#step1/out\"],\n+ }, {\n+ 'id': 'http://example2.com/#step2',\n+ 'scatter': 'http://example2.com/#step2/inp',\n+ 'in': [{\n+ 'id': 'http://example2.com/#step2/inp',\n+ 'source': 'http://example2.com/#step1/out'\n+ }],\n+ \"out\": [\"http://example2.com/#step2/out\"],\n+ }]\n+ }, ra)\n+\n \n def test_examples(self):\n self.maxDiff = None\n for a in [\"field_name\", \"ident_res\", \"link_res\", \"vocab_res\"]:\n- ldr, _, _ = schema_salad.schema.load_schema(\"schema_salad/metaschema/%s_schema.yml\" % a)\n+ ldr, _, _, _ = schema_salad.schema.load_schema(\n+ \"schema_salad/metaschema/%s_schema.yml\" % a)\n with open(\"schema_salad/metaschema/%s_src.yml\" % a) as src_fp:\n- src = ldr.resolve_all(yaml.load(src_fp), \"\")[0]\n+ src = ldr.resolve_all(\n+ yaml.load(src_fp, Loader=SafeLoader), \"\", checklinks=False)[0]\n with open(\"schema_salad/metaschema/%s_proc.yml\" % a) as src_proc:\n- proc = yaml.load(src_proc)\n+ proc = yaml.load(src_proc, Loader=SafeLoader)\n self.assertEqual(proc, src)\n \n+ def test_yaml_float_test(self):\n+ self.assertEqual(yaml.load(\"float-test: 2e-10\")[\"float-test\"], 2e-10)\n+\n+ def test_typedsl_ref(self):\n+ ldr = schema_salad.ref_resolver.Loader({})\n+ ldr.add_context({\n+ \"File\": \"http://example.com/File\",\n+ \"null\": \"http://example.com/null\",\n+ \"array\": \"http://example.com/array\",\n+ \"type\": {\n+ \"@type\": \"@vocab\",\n+ \"typeDSL\": True\n+ }\n+ })\n+\n+ ra, _ = ldr.resolve_all({\"type\": \"File\"}, \"\")\n+ self.assertEqual({'type': 'File'}, ra)\n+\n+ ra, _ = ldr.resolve_all({\"type\": \"File?\"}, \"\")\n+ self.assertEqual({'type': ['null', 'File']}, ra)\n+\n+ ra, _ = ldr.resolve_all({\"type\": \"File[]\"}, \"\")\n+ self.assertEqual({'type': {'items': 'File', 'type': 'array'}}, ra)\n+\n+ ra, _ = ldr.resolve_all({\"type\": \"File[]?\"}, \"\")\n+ self.assertEqual({'type': ['null', {'items': 'File', 'type': 'array'}]}, ra)\n \n if __name__ == '__main__':\n unittest.main()\n", "problem_statement": "", "hints_text": "", "created_at": "2016-06-08T13:17:05Z"} | |