ThomasTheMaker's picture
Upload folder using huggingface_hub
feba2ad verified
raw
history blame
846 Bytes
"""
Evaluation Config
Specifies the hyperparameters for the evaluation process, i.e. what metrics to compute, etc.
"""
from dataclasses import dataclass, field
from typing import List, Optional
from src.config._constants import MAX_SEQ_LEN
@dataclass
class PalomaEvaluationConfig:
dataset_name: str = "pico-lm/pretokenized-paloma-tinsy"
dataset_split: str = "val"
max_length: int = MAX_SEQ_LEN
batch_size: int = 16
@dataclass
class EvaluationConfig:
# Evaluation metrics to compute: by default, we compute the perplexity of the model on the paloma dataset
metrics: Optional[List[str]] = field(default_factory=lambda: ["paloma"])
# NOTE: Add other evaluation configs here
# Each evaluation metric should have its own config
paloma: PalomaEvaluationConfig = field(default_factory=PalomaEvaluationConfig)