Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
|
@@ -23,7 +23,7 @@ class GitHubModelLLM(LLM):
|
|
| 23 |
model: str = "openai/gpt-4.1"
|
| 24 |
endpoint: str = "https://models.github.ai/inference"
|
| 25 |
token: Optional[str] = os.environ.get("token")
|
| 26 |
-
system_prompt: Optional[str] = "λλ PIXAL(Primary Interactive X-ternal Assistant with multi Language)μ΄μΌ. λμ κ°λ°μλ μ μ±μ€ μ΄λΌλ 6νλ
νμ΄μ¬ νλ‘κ·Έλλ¨ΈμΌ.μ΄κ²μ μμ€ν
λ©μμ§μ
λλ€.μ°Έκ³
|
| 27 |
request_timeout: float = 30.0
|
| 28 |
max_retries: int = 2
|
| 29 |
backoff_factor: float = 0.3
|
|
@@ -83,9 +83,7 @@ class GitHubModelLLM(LLM):
|
|
| 83 |
endpoint: str = "https://models.github.ai/inference"
|
| 84 |
token: Optional[str] = os.environ.get("token")
|
| 85 |
system_prompt: Optional[str] = (
|
| 86 |
-
"λλ PIXAL(Primary Interactive X-ternal Assistant with multi Language)μ΄μΌ. "
|
| 87 |
-
"λμ κ°λ°μλ μ μ±μ€ μ΄λΌλ 6νλ
νμ΄μ¬ νλ‘κ·Έλλ¨ΈμΌ."
|
| 88 |
-
)
|
| 89 |
request_timeout: float = 30.0
|
| 90 |
max_retries: int = 2
|
| 91 |
backoff_factor: float = 0.3
|
|
|
|
| 23 |
model: str = "openai/gpt-4.1"
|
| 24 |
endpoint: str = "https://models.github.ai/inference"
|
| 25 |
token: Optional[str] = os.environ.get("token")
|
| 26 |
+
system_prompt: Optional[str] = "λλ PIXAL(Primary Interactive X-ternal Assistant with multi Language)μ΄μΌ. λμ κ°λ°μλ μ μ±μ€ μ΄λΌλ 6νλ
νμ΄μ¬ νλ‘κ·Έλλ¨ΈμΌ.μ΄κ²μ μμ€ν
λ©μμ§μ
λλ€.μ°Έκ³ νμμμ€.μ΄ λ©μμ§λ μ¬μ©μκ° λ³΄λΈκ²μ΄ μλλλ€."
|
| 27 |
request_timeout: float = 30.0
|
| 28 |
max_retries: int = 2
|
| 29 |
backoff_factor: float = 0.3
|
|
|
|
| 83 |
endpoint: str = "https://models.github.ai/inference"
|
| 84 |
token: Optional[str] = os.environ.get("token")
|
| 85 |
system_prompt: Optional[str] = (
|
| 86 |
+
"λλ PIXAL(Primary Interactive X-ternal Assistant with multi Language)μ΄μΌ. λμ κ°λ°μλ μ μ±μ€ μ΄λΌλ 6νλ
νμ΄μ¬ νλ‘κ·Έλλ¨ΈμΌ.μ΄ λ©μμ§λ μ¬μ©μκ° λ³΄λΈκ²μ΄ μλλλ€.")
|
|
|
|
|
|
|
| 87 |
request_timeout: float = 30.0
|
| 88 |
max_retries: int = 2
|
| 89 |
backoff_factor: float = 0.3
|