Datasets:
UD Tools
This package contains Python tools for Universal Dependencies:
- The official UD/CoNLL-U validator
- The official UD parsing scorer from the CoNLL (2017, 2018) and IWPT (2020, 2021) shared tasks
The official UD/CoNLL-U validator
Reads a CoNLL-U file and verifies that it complies with the UD specification. For more details on UD validation, visit the description on the UD website.
The most up-to-date version of the validator always resides in the master branch of the
tools repository on GitHub. It is possible to run the script
validate.py from your local copy of the repository even without installing the udtools package via pip.
Nevertheless, you will need a few third-party modules the validator depends on. You can install them like this:
pip install -r requirements.txt.
If the root folder of the tools repository is in your system PATH, you do not have to be in that folder when
launching the script:
cat la_proiel-ud-train.conllu | python validate.py --lang la --max-err=0
You can run python validate.py --help for a list of available options.
Invoking validation from your Python program
To use the validator from your Python code, first install udtools (possibly after creating and activating a virtual
environment). This should give you access to a fairly recent version of the validator but it will not necessarily be
the authoritative version, as it may lack some modifications of the language-specific data.
pip install --upgrade udtools
from udtools import Validator
validator = Validator(lang='la')
state = validator.validate_files(['la_proiel-ud-train.conllu', 'la_proiel-ud-dev.conllu', 'la_proiel-ud-test.conllu'])
print(state)
The state is an object with various pieces of information collected during the validation run. Its string
representation is a summary of the warnings and errors found, as well as the string "*** PASSED " or
" FAILED ***". You can also use the state in a boolean context (condition), where “passed” evaluates as True and
“failed” as False. Note however that the default behavior of the validator is still to print errors and warnings to
STDERR as soon as they are detected. To suppress printing, add output=None to the arguments when constructing the
validator. (The default value of this argument is sys.stderr. You could also set it to sys.stdout or to a handle
of a file open for writing.)
from udtools import Validator
validator = Validator(lang='la', output=None)
state = validator.validate_files(['la_proiel-ud-train.conllu', 'la_proiel-ud-dev.conllu', 'la_proiel-ud-test.conllu'])
if state:
print('Yay!')
else:
print('Oh no ☹')
Alternatively, you could simulate supplying the --quiet option as if it came from the command line:
import sys
from udtools.argparser import parse_args_validator
from udtools import Validator
sys.argv = ['validate.py', '--lang=la', '--quiet']
args = parse_args_validator()
validator = Validator(lang='la', args=args)
state = validator.validate_files(['la_proiel-ud-train.conllu', 'la_proiel-ud-dev.conllu', 'la_proiel-ud-test.conllu'])
if state:
print('Yay!')
else:
print('Oh no ☹')
Instead of printing the errors to STDERR as soon as they are found, you can have them saved in the validation state
and later process them the way you prefer. Note that if you use the argparser approach from the previous example, the
number of incidents saved (per category) is limited by default. This is to save your memory if you do not need to keep
the errors (some treebanks have hundreds of thousands of errors and warnings). By setting --max-store=0, this limit
is turned off. However, the default limit is set in the argparser, so if you use the simpler approach with
output=None and you do not invoke the argparser for other reasons, no limit will be imposed.
from udtools import Validator
validator = Validator(lang='la', output=None)
state = validator.validate_files(['la_proiel-ud-train.conllu', 'la_proiel-ud-dev.conllu', 'la_proiel-ud-test.conllu'])
# Take only errors, skip warnings.
all_errors = [x for x in state.error_tracker if x.is_error()]
all_errors.sort(key=lambda incident: incident.testid)
for error in all_errors:
print(error)
Entry points
The validator has several other entry points in addition to validate_files():
validate_file()takes just one file name (path), reads that file and tests its validity. If the file name is '-', it is interpreted as reading from STDIN. Note thatvalidate_files()callsvalidate_file()for each file in turn, then it also callsvalidate_end()to perform checks that can only be done after the whole treebank has been read. If you call directlyvalidate_file(), you should take care of callingvalidate_end()yourself.validate_end()takes just the state from the validation performed so far, and checks that the observations saved in the state are not in conflict.
validate_file_handle()takes the object associated with an open file (orsys.stdin). Otherwise it is analogous tovalidate_file()(and is in fact called fromvalidate_file()).validate_sentence()takes a list of CoNLL-U lines corresponding to one sentence, including the sentence-terminating empty line. When called fromvalidate_file_handle(), it will have at most one empty line and this will be the last line in the list, as it is how the file reader detected the sentence end. However, the method is aware that other callers could supply lists with empty lines in the middle, and it will report an error if this happens.
All the validate_*() methods mentioned above return a State object. All of them can optionally take a State from
previous runs as an argument (named state), in which case they will base their decisions on this state, and save
their observations in it, too.
The validator uses data files with specifications of feature values, lemmas of auxiliaries etc. for each language.
These files change more often than the validator code itself, so it is likely that your pip-installed udtools does
not have the most up-to-date version. Therefore, you may want to have a local copy of the tools repository, regularly
update it by calling git pull, and tell the validator where to load the data files from (instead of its installation
location):
validator = Validator(lang='la', datapath='/my/copy/of/ud/tools/data')
Printing incidents in JSON
Instead of prose error messages suitable for human users, you can print the error descriptions in JSON so it can be easily read and processed by an external application.
from udtools import Validator
validator = Validator(lang='la', output=None)
state = validator.validate_files(['la_proiel-ud-train.conllu', 'la_proiel-ud-dev.conllu', 'la_proiel-ud-test.conllu'])
incidents = sorted(state.error_tracker, key=lambda incident: incident.testid)
print('[')
print(',\n'.join([incident.json() for incident in incidents]))
print(']')
Selecting only some tests
UD defines several
levels of validity
of CoNLL-U files. By default, validity on the highest level 5 is required; this is the level that UD treebanks must
pass in order to be released as part of Universal Dependencies. It is possible to request a lower level of validity,
for example, only the backbone file structure can be checked, omitting any linguistic checks of the annotation
guidelines. When invoking validate.py from the command line, the numeric option --level (e.g., --level 1)
tells the validator to skip tests on levels 2 and above. The same argument can be given directly to the constructor
of the Validator class. The lowest level is not specific to individual languages, so we can give the generic
language "ud" instead.
validator = Validator(lang='ud', level=1)
One may want to filter the tests along various other dimensions: errors only (skipping warnings); selected test classes
(FORMAT, MORPHO, SYNTAX, ENHANCED, METADATA etc.); individual test ids (e.g., obl-should-be-nmod). It is always
possible to do what we showed above, i.e., collecting all incidents, then processing them and showing only the selected
ones. However, this approach has its drawbacks: We waste time by running tests whose results we do not want to see;
for large treebanks it is not practical to postpone showing first results until the whole treebank is processed; and
it may be also quite heavy to keep all unnecessary incidents in memory.
You may try to get around this by implementing your own alternative to validate_sentence() and call individual tests
directly. There are some dangers though, which you should consider first:
- The tests are not documented at present, so you have to consult the source code. The relevant functions are methods
of
Validatorand their names start withcheck_(as opposed tovalidate_, which signals the better documented entry points). Note that onecheck_method may generate multiple different incident types, whose ids are not reflected in the name of the method; and a few incidents can even occur outside anycheck_method (e.g., directly in avalidate_method). - The interface is far from stable. Names of methods may change at any time, as well as the types of incidents they generate, the arguments they expect, their return values (if any) or side effects. Some checks only look at individual cells in the CoNLL-U tabular format, others expect the fully built tree structure.
- There are dependencies among the tests. Some
check_methods can be run safely only if othercheck_methods have been run previously and did not encounter errors.
Adding your own tests
You may want to add language-specific consistency tests beyond what the official validator can do (e.g., ensuring that
all personal pronouns have a non-empty value of the Person feature), or even treebank/project-specific tests (e.g.,
all tokens should have a valid Ref attribute in MISC). One way of doing this would be to derive your own validator
class based on udtools.Validator.
from udtools import Validator
from udtools.incident import TestClass, Error
class MyValidator(Validator):
def validate_sentence(self, lines, state=None):
state = super().validate_sentence(lines, state)
self.check_my_own_stuff(state, lines)
return state
def check_my_own_stuff(self, state, lines):
for line in lines:
if re.match(r'40\t', line):
Error(
state=state, config=self.incfg,
level=1,
testclass=TestClass.FORMAT,
testid='id-40',
message="Node ID 40 is not allowed in this treebank."
).confirm()
validator = MyValidator(lang='la')
state = validator.validate_files(['la_proiel-ud-train.conllu', 'la_proiel-ud-dev.conllu', 'la_proiel-ud-test.conllu'])
print(state)
The official UD parsing scorer
Reads two CoNLL-U files: gold standard (annotated manually) and system output (predicted by a parsing model). Both files must be valid at least at level 2, and their underlying text must be compatible, i.e., it can differ in whitespace but not in other characters. The scorer evaluates similarity of the system output to the gold standard by computing several metrics that were defined in the UD parsing shared tasks (CoNLL 2017 & 2018, IWPT 2020 & 2021).
To load two files and evaluate their similarity without enhanced dependencies (i.e., in the style of the CoNLL shared tasks), you can do the following.
from udtools.udeval import load_conllu_file, evaluate, build_evaluation_table
gold_ud = load_conllu_file('gold.conllu')
system_ud = load_conllu_file('system.conllu')
evaluation = evaluate(gold_ud, system_ud)
results = build_evaluation_table(evaluation, verbose=True)
print(results)
To use the command line interface and arguments, you can use parse_args_scorer() as shown below. If you supply
--help as the only argument, you will get the description of the options available.
from udtools.argparser import parse_args_scorer
from udtools.udeval import evaluate_wrapper, build_evaluation_table
args = parse_args_scorer()
evaluation = evaluate_wrapper(args)
results = build_evaluation_table(evaluation, args.verbose, args.counts, args.enhanced)
print(results)