"""Backend module for LLM model loading and inference.""" from .model_loader import load_model from .chat import Conversation, Message from .chat_model import ChatModel from .inference import generate_streaming, generate_simple __all__ = [ # Model loading "load_model", # OOP classes (recomendado) "Conversation", "ChatModel", # Functions (compatibilidade) "generate_streaming", "generate_simple", # Types "Message", ]