|
|
import abc |
|
|
import os |
|
|
from dataclasses import dataclass, field |
|
|
from typing import Hashable, List, Optional, TypeVar |
|
|
from .consts import (COMPLETE_ALPHABET, WHITESPACE_CHARACTERS, DEFAULT_MAX_CONSECUTIVE_WHITESPACES, |
|
|
DEFAULT_FORCE_JSON_FIELD_ORDER, CONFIG_ENV_VAR_MAX_CONSECUTIVE_WHITESPACES, |
|
|
CONFIG_ENV_VAR_STRICT_JSON_FIELD_ORDER, CONFIG_ENV_VAR_MAX_JSON_ARRAY_LENGTH, |
|
|
DEFAULT_MAX_JSON_ARRAY_LENGTH) |
|
|
|
|
|
|
|
|
def _parse_bool(s: str) -> bool: |
|
|
return s and (s.strip().lower() in ['true', '1']) |
|
|
|
|
|
|
|
|
def _env_or_default_field(env_var: str, default_val): |
|
|
default_val_type = type(default_val) |
|
|
parser_func = _parse_bool if default_val_type == bool else default_val_type |
|
|
def factory_func(): |
|
|
return parser_func(os.environ.get(env_var, str(default_val))) |
|
|
return field(default_factory=factory_func) |
|
|
|
|
|
|
|
|
@dataclass |
|
|
class CharacterLevelParserConfig: |
|
|
alphabet: str = COMPLETE_ALPHABET |
|
|
max_consecutive_whitespaces: int = _env_or_default_field(CONFIG_ENV_VAR_MAX_CONSECUTIVE_WHITESPACES, |
|
|
DEFAULT_MAX_CONSECUTIVE_WHITESPACES) |
|
|
"""How many consective whitespaces the JsonSchemaParser will allow""" |
|
|
force_json_field_order: bool = _env_or_default_field(CONFIG_ENV_VAR_STRICT_JSON_FIELD_ORDER, |
|
|
DEFAULT_FORCE_JSON_FIELD_ORDER) |
|
|
"""Whether the JsonSchemaParser will force fields to appear in the |
|
|
order of the 'required' field in the schema""" |
|
|
max_json_array_length: int = _env_or_default_field(CONFIG_ENV_VAR_MAX_JSON_ARRAY_LENGTH, |
|
|
DEFAULT_MAX_JSON_ARRAY_LENGTH) |
|
|
"""What is the maximum json array length if not specified by the schema. Helps the LLM |
|
|
avoid infinite loops.""" |
|
|
|
|
|
|
|
|
class CharacterLevelParser(abc.ABC): |
|
|
"""CharacterLevelParser is an interface for classes that can parse strings one character at a time, and determine which characters are allowed at any specific time""" |
|
|
|
|
|
def __init__(self, config: Optional[CharacterLevelParserConfig] = None): |
|
|
self._config = config or CharacterLevelParserConfig() |
|
|
|
|
|
@abc.abstractmethod |
|
|
def add_character(self, new_character: str) -> 'CharacterLevelParser': |
|
|
"""Add a character to the parser, and return a new parser that represents the state of the parser after the character has been added. This has to be |
|
|
an immutable operation - the original CharacterLevelParser (self) must not be modified.""" |
|
|
raise NotImplementedError() |
|
|
|
|
|
@abc.abstractmethod |
|
|
def get_allowed_characters(self) -> str: |
|
|
"""Return a string containing all characters that are allowed at the current point in the parsing process.""" |
|
|
raise NotImplementedError() |
|
|
|
|
|
@abc.abstractmethod |
|
|
def can_end(self) -> bool: |
|
|
"""Return True if the parser is in a state where it can end (potentially finished parsing the desired structure), and False otherwise.""" |
|
|
raise NotImplementedError() |
|
|
|
|
|
def shortcut_key(self) -> Optional[Hashable]: |
|
|
"""Optional. Return a key that denotes that this state is a repeating state, full tree traversal should be avoided.""" |
|
|
return None |
|
|
|
|
|
def cache_key(self) -> Optional[Hashable]: |
|
|
"""Optional. Return a key that denotes that this state is a repeating state, and if it is visited again, results can be cached.""" |
|
|
return None |
|
|
|
|
|
@property |
|
|
def config(self) -> CharacterLevelParserConfig: |
|
|
return self._config |
|
|
|
|
|
@config.setter |
|
|
def config(self, new_config: CharacterLevelParserConfig): |
|
|
self._config = new_config |
|
|
return self |
|
|
|
|
|
|
|
|
class StringParser(CharacterLevelParser): |
|
|
"""RegexParser is an example CharacterLevelParser that only allows an exact string. It is a debugging / learning tool |
|
|
to show how CharacterLevelParser works together with TokenizerPrefixTree to filter the allowed tokens (some of whom may contain multiple characters)""" |
|
|
def __init__(self, string: str): |
|
|
self.target_str = string |
|
|
|
|
|
def add_character(self, new_character: str) -> CharacterLevelParser: |
|
|
if self.target_str.startswith(new_character): |
|
|
return StringParser(self.target_str[len(new_character):]) |
|
|
else: |
|
|
raise ValueError(f"Expected '{self.target_str[0]}' but got '{new_character}'") |
|
|
|
|
|
def get_allowed_characters(self) -> str: |
|
|
return self.target_str[0] if self.target_str else "" |
|
|
|
|
|
def can_end(self) -> bool: |
|
|
return not self.target_str |
|
|
|
|
|
|
|
|
class ForceStopParser(CharacterLevelParser): |
|
|
"""A simple parser that forbids any characters except the stop token. Used to force stop LM operation""" |
|
|
def __init__(self, allow_whitespace: bool = False): |
|
|
self.allow_whitespace = allow_whitespace |
|
|
def add_character(self, new_character: str) -> CharacterLevelParser: |
|
|
return self |
|
|
def get_allowed_characters(self) -> str: |
|
|
return WHITESPACE_CHARACTERS if self.allow_whitespace else "" |
|
|
def can_end(self) -> bool: |
|
|
return True |
|
|
|
|
|
|
|
|
class UnionParser(CharacterLevelParser): |
|
|
"""A parser that allows a string that would be allowed by any of several different parsers""" |
|
|
def __init__(self, parsers: List[CharacterLevelParser]): |
|
|
self.parsers = parsers |
|
|
|
|
|
def add_character(self, new_character: str) -> CharacterLevelParser: |
|
|
|
|
|
relevant_parsers = [parser for parser in self.parsers if new_character in parser.get_allowed_characters()] |
|
|
next_parsers = [parser.add_character(new_character) for parser in relevant_parsers] |
|
|
if len(next_parsers) == 1: |
|
|
return next_parsers[0] |
|
|
return UnionParser(next_parsers) |
|
|
|
|
|
def get_allowed_characters(self) -> str: |
|
|
allowed = "".join([parser.get_allowed_characters() for parser in self.parsers]) |
|
|
return "".join(set(allowed)) |
|
|
|
|
|
def can_end(self) -> bool: |
|
|
return any([parser.can_end() for parser in self.parsers]) |
|
|
|
|
|
def shortcut_key(self) -> Optional[Hashable]: |
|
|
unique_shortcut_keys = set(parser.shortcut_key() for parser in self.parsers) |
|
|
if len(unique_shortcut_keys) == 1: |
|
|
return next(iter(unique_shortcut_keys)) |
|
|
return None |
|
|
|
|
|
def cache_key(self) -> Optional[Hashable]: |
|
|
all_cache_keys = tuple(parser.cache_key() for parser in self.parsers) |
|
|
if all(key is not None for key in all_cache_keys): |
|
|
return ('union', all_cache_keys) |
|
|
return None |
|
|
|
|
|
|
|
|
class SequenceParser(CharacterLevelParser): |
|
|
"""A parser that is a sequence of multiple parsers.""" |
|
|
def __init__(self, parsers: List[CharacterLevelParser]): |
|
|
self.parsers = parsers |
|
|
|
|
|
def add_character(self, new_character: str) -> CharacterLevelParser: |
|
|
legal_parsers = [] |
|
|
|
|
|
|
|
|
|
|
|
for idx, parser in enumerate(self.parsers): |
|
|
if new_character in parser.get_allowed_characters(): |
|
|
updated_parser = parser.add_character(new_character) |
|
|
next_parsers = [updated_parser] + self.parsers[idx+1:] |
|
|
if len(next_parsers) == 1: |
|
|
legal_parsers.append(next_parsers[0]) |
|
|
else: |
|
|
legal_parsers.append(SequenceParser(next_parsers)) |
|
|
if not parser.can_end(): |
|
|
break |
|
|
if len(legal_parsers) == 1: |
|
|
return legal_parsers[0] |
|
|
return UnionParser(legal_parsers) |
|
|
|
|
|
def get_allowed_characters(self) -> str: |
|
|
allowed_characters = set() |
|
|
for parser in self.parsers: |
|
|
allowed_characters.update(parser.get_allowed_characters()) |
|
|
if not parser.can_end(): |
|
|
break |
|
|
return "".join(allowed_characters) |
|
|
|
|
|
def can_end(self) -> bool: |
|
|
return all([parser.can_end() for parser in self.parsers]) |
|
|
|
|
|
def shortcut_key(self) -> Optional[str]: |
|
|
return self.parsers[0].shortcut_key() if len(self.parsers) == 1 else None |
|
|
|
|
|
def cache_key(self) -> Optional[Hashable]: |
|
|
all_cache_keys = tuple(parser.cache_key() for parser in self.parsers) |
|
|
if all(key is not None for key in all_cache_keys): |
|
|
return ('sequence', all_cache_keys) |
|
|
return None |
|
|
|
|
|
|
|
|
|