Spaces:
Sleeping
Sleeping
Update tools.py
Browse files
tools.py
CHANGED
|
@@ -8,18 +8,6 @@ from langchain_google_genai import ChatGoogleGenerativeAI
|
|
| 8 |
import base64
|
| 9 |
|
| 10 |
#LLMs
|
| 11 |
-
def get_access_token():
|
| 12 |
-
credential = EnvironmentCredential()
|
| 13 |
-
access_token = credential.get_token("https://cognitiveservices.azure.com/.default")
|
| 14 |
-
return access_token.token
|
| 15 |
-
|
| 16 |
-
llm = AzureChatOpenAI(
|
| 17 |
-
model_name="gpt-4o",
|
| 18 |
-
api_key=get_access_token(),
|
| 19 |
-
azure_endpoint="https://cog-sandbox-dev-eastus2-001.openai.azure.com/",
|
| 20 |
-
api_version="2024-08-01-preview"
|
| 21 |
-
)
|
| 22 |
-
|
| 23 |
google_llm = ChatGoogleGenerativeAI(model='gemini-2.0-flash-lite')
|
| 24 |
|
| 25 |
#IMAGE_TOOLS
|
|
@@ -65,7 +53,7 @@ def extract_text(img_path: str) -> str:
|
|
| 65 |
]
|
| 66 |
|
| 67 |
# Call the vision-capable model
|
| 68 |
-
response =
|
| 69 |
|
| 70 |
# Append extracted text
|
| 71 |
all_text += response.content + "\n\n"
|
|
@@ -119,7 +107,7 @@ def describe_image(img_path: str) -> str:
|
|
| 119 |
]
|
| 120 |
|
| 121 |
# Call the vision-capable model
|
| 122 |
-
response =
|
| 123 |
|
| 124 |
# Append extracted text
|
| 125 |
all_text += response.content + "\n\n"
|
|
|
|
| 8 |
import base64
|
| 9 |
|
| 10 |
#LLMs
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 11 |
google_llm = ChatGoogleGenerativeAI(model='gemini-2.0-flash-lite')
|
| 12 |
|
| 13 |
#IMAGE_TOOLS
|
|
|
|
| 53 |
]
|
| 54 |
|
| 55 |
# Call the vision-capable model
|
| 56 |
+
response = google_llm.invoke(message)
|
| 57 |
|
| 58 |
# Append extracted text
|
| 59 |
all_text += response.content + "\n\n"
|
|
|
|
| 107 |
]
|
| 108 |
|
| 109 |
# Call the vision-capable model
|
| 110 |
+
response = google_llm.invoke(message)
|
| 111 |
|
| 112 |
# Append extracted text
|
| 113 |
all_text += response.content + "\n\n"
|