Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -33,7 +33,6 @@ def speech_to_text(audio_file, language_code):
|
|
| 33 |
recognizer = sr.Recognizer()
|
| 34 |
|
| 35 |
try:
|
| 36 |
-
# Save uploaded file to temporary location
|
| 37 |
with tempfile.NamedTemporaryFile(delete=False, suffix=".wav") as tmp_file:
|
| 38 |
tmp_file.write(audio_file.getvalue())
|
| 39 |
audio_path = tmp_file.name
|
|
@@ -86,7 +85,7 @@ def setup_agents(language='en'):
|
|
| 86 |
goal="Analyze and validate space information",
|
| 87 |
backstory="Expert in multilingual space data analysis with NASA mission experience.",
|
| 88 |
verbose=True,
|
| 89 |
-
llm=LLM_REPO,
|
| 90 |
llm_kwargs={
|
| 91 |
"temperature": 0.4,
|
| 92 |
"max_length": 512
|
|
@@ -99,7 +98,7 @@ def setup_agents(language='en'):
|
|
| 99 |
goal=f"Explain complex concepts in {language} using simple terms",
|
| 100 |
backstory=f"Multilingual science communicator specializing in {language} explanations.",
|
| 101 |
verbose=True,
|
| 102 |
-
llm=LLM_REPO,
|
| 103 |
llm_kwargs={
|
| 104 |
"temperature": 0.5,
|
| 105 |
"max_length": 612
|
|
@@ -171,4 +170,4 @@ if question:
|
|
| 171 |
st.markdown(answer)
|
| 172 |
|
| 173 |
st.markdown("---")
|
| 174 |
-
st.markdown("*Powered by NASA API & Open Source AI*")
|
|
|
|
| 33 |
recognizer = sr.Recognizer()
|
| 34 |
|
| 35 |
try:
|
|
|
|
| 36 |
with tempfile.NamedTemporaryFile(delete=False, suffix=".wav") as tmp_file:
|
| 37 |
tmp_file.write(audio_file.getvalue())
|
| 38 |
audio_path = tmp_file.name
|
|
|
|
| 85 |
goal="Analyze and validate space information",
|
| 86 |
backstory="Expert in multilingual space data analysis with NASA mission experience.",
|
| 87 |
verbose=True,
|
| 88 |
+
llm={"provider": "huggingface", "model": LLM_REPO}, # ✅ Explicit provider fix
|
| 89 |
llm_kwargs={
|
| 90 |
"temperature": 0.4,
|
| 91 |
"max_length": 512
|
|
|
|
| 98 |
goal=f"Explain complex concepts in {language} using simple terms",
|
| 99 |
backstory=f"Multilingual science communicator specializing in {language} explanations.",
|
| 100 |
verbose=True,
|
| 101 |
+
llm={"provider": "huggingface", "model": LLM_REPO}, # ✅ Explicit provider fix
|
| 102 |
llm_kwargs={
|
| 103 |
"temperature": 0.5,
|
| 104 |
"max_length": 612
|
|
|
|
| 170 |
st.markdown(answer)
|
| 171 |
|
| 172 |
st.markdown("---")
|
| 173 |
+
st.markdown("*Powered by NASA API & Open Source AI*")
|