Update ff.py
Browse files
ff.py
CHANGED
|
@@ -14,14 +14,18 @@ from g4f import Provider, models
|
|
| 14 |
from langchain.llms.base import LLM
|
| 15 |
|
| 16 |
from langchain_g4f import G4FLLM
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
|
| 18 |
|
| 19 |
def res(input_text):
|
| 20 |
llm: LLM = G4FLLM(
|
| 21 |
-
model=models.
|
| 22 |
-
|
| 23 |
)
|
| 24 |
-
input_texxt ="Tu es un assistant intelligent.ton but est d'assiter au mieux que tu peux.
|
| 25 |
print("Question:", input_text)
|
| 26 |
print()
|
| 27 |
print()
|
|
|
|
| 14 |
from langchain.llms.base import LLM
|
| 15 |
|
| 16 |
from langchain_g4f import G4FLLM
|
| 17 |
+
g4f.debug.logging = True # Enable logging
|
| 18 |
+
g4f.check_version = False # Disable automatic version checking
|
| 19 |
+
print(g4f.version) # Check version
|
| 20 |
+
print(g4f.Provider.Ails.params)
|
| 21 |
|
| 22 |
|
| 23 |
def res(input_text):
|
| 24 |
llm: LLM = G4FLLM(
|
| 25 |
+
model=models.gpt_35_turbo_16k,
|
| 26 |
+
,
|
| 27 |
)
|
| 28 |
+
input_texxt ="Tu es un assistant intelligent.ton but est d'assiter au mieux que tu peux." + input_text
|
| 29 |
print("Question:", input_text)
|
| 30 |
print()
|
| 31 |
print()
|