File size: 462 Bytes
91c131d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
"""Backend module for LLM model loading and inference."""

from .model_loader import load_model
from .chat import Conversation, Message
from .chat_model import ChatModel
from .inference import generate_streaming, generate_simple

__all__ = [
    # Model loading
    "load_model",
    # OOP classes (recomendado)
    "Conversation",
    "ChatModel",
    # Functions (compatibilidade)
    "generate_streaming",
    "generate_simple",
    # Types
    "Message",
]