Buckets:
| # Loading methods | |
| Methods for listing and loading evaluation modules: | |
| ## List[[evaluate.list_evaluation_modules]] | |
| #### evaluate.list_evaluation_modules[[evaluate.list_evaluation_modules]] | |
| [Source](https://github.com/huggingface/evaluate/blob/main/src/evaluate/inspect.py#L35) | |
| List all evaluation modules available on the Hugging Face Hub. | |
| Example: | |
| ```py | |
| >>> from evaluate import list_evaluation_modules | |
| >>> list_evaluation_modules(module_type="metric") | |
| ``` | |
| **Parameters:** | |
| module_type (`str`, *optional*, defaults to `None`) : Type of evaluation modules to list. Has to be one of `'metric'`, `'comparison'`, or `'measurement'`. If `None`, all types are listed. | |
| include_community (`bool`, *optional*, defaults to `True`) : Include community modules in the list. | |
| with_details (`bool`, *optional*, defaults to `False`) : Return the full details on the metrics instead of only the ID. | |
| **Returns:** | |
| `List[Union[str, dict]]` | |
| ## Load[[evaluate.load]] | |
| #### evaluate.load[[evaluate.load]] | |
| [Source](https://github.com/huggingface/evaluate/blob/main/src/evaluate/loading.py#L692) | |
| Load a [EvaluationModule](/docs/evaluate/main/en/package_reference/main_classes#evaluate.EvaluationModule). | |
| Example: | |
| ```py | |
| >>> from evaluate import load | |
| >>> accuracy = load("accuracy") | |
| ``` | |
| **Parameters:** | |
| path (`str`) : Path to the evaluation processing script with the evaluation builder. Can be either: - a local path to processing script or the directory containing the script (if the script has the same name as the directory), e.g. `'./metrics/rouge'` or `'./metrics/rouge/rouge.py'` - a evaluation module identifier on the HuggingFace evaluate repo e.g. `'rouge'` or `'bleu'` that are in either `'metrics/'`, `'comparisons/'`, or `'measurements/'` depending on the provided `module_type` | |
| config_name (`str`, *optional*) : Selecting a configuration for the metric (e.g. the GLUE metric has a configuration for each subset). | |
| module_type (`str`, default `'metric'`) : Type of evaluation module, can be one of `'metric'`, `'comparison'`, or `'measurement'`. | |
| process_id (`int`, *optional*) : For distributed evaluation: id of the process. | |
| num_process (`int`, *optional*) : For distributed evaluation: total number of processes. | |
| cache_dir (`str`, *optional*) : Path to store the temporary predictions and references (default to `~/.cache/huggingface/evaluate/`). | |
| experiment_id (`str`) : A specific experiment id. This is used if several distributed evaluations share the same file system. This is useful to compute metrics in distributed setups (in particular non-additive metrics like F1). | |
| keep_in_memory (`bool`) : Whether to store the temporary results in memory (defaults to `False`). | |
| download_config (`~evaluate.DownloadConfig`, *optional*) : Specific download configuration parameters. | |
| download_mode (`DownloadMode`, defaults to `REUSE_DATASET_IF_EXISTS`) : Download/generate mode. | |
| revision (`Union[str, evaluate.Version]`, *optional*) : If specified, the module will be loaded from the datasets repository at this version. By default it is set to the local version of the lib. Specifying a version that is different from your local version of the lib might cause compatibility issues. | |
| **Returns:** | |
| [evaluate.EvaluationModule](/docs/evaluate/main/en/package_reference/main_classes#evaluate.EvaluationModule) | |
Xet Storage Details
- Size:
- 3.32 kB
- Xet hash:
- cb5a5913c3c1385a3386ada7682c93392d6908d2849bc5137a7c28e165f7c017
·
Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.