Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
import requests
|
| 2 |
-
import os
|
| 3 |
import json
|
|
|
|
| 4 |
|
| 5 |
api_token = os.environ.get("TOKEN")
|
| 6 |
API_URL = "https://api-inference.huggingface.co/models/meta-llama/Meta-Llama-3-8B-Instruct"
|
|
@@ -51,7 +51,7 @@ assistant
|
|
| 51 |
return 'questions'
|
| 52 |
elif ChatmodelLLM == 2:
|
| 53 |
return 'Chat Model/LLM'
|
| 54 |
-
elif other == 2:
|
| 55 |
return "Other"
|
| 56 |
elif image_generation == 2:
|
| 57 |
return "Image Generation"
|
|
@@ -71,7 +71,7 @@ assistant
|
|
| 71 |
return f"Erreur: Réponse ambiguë - '{response}'"
|
| 72 |
|
| 73 |
# URL de base de l'API
|
| 74 |
-
base_url = "https://huggingface.co/api/posts"
|
| 75 |
|
| 76 |
# Paramètres pour la pagination
|
| 77 |
skip = 0 # Nombre d'éléments à sauter
|
|
@@ -114,16 +114,7 @@ while True:
|
|
| 114 |
# Maintenant, all_posts_with_text contient tous les posts récupérés avec leur texte
|
| 115 |
|
| 116 |
# Initialisation des compteurs
|
| 117 |
-
questions_count = 0
|
| 118 |
-
Chat_ModelLLM_count = 0
|
| 119 |
-
Other_count = 0
|
| 120 |
-
ImageGeneration_count = 0
|
| 121 |
-
Fine_tuning_count = 0
|
| 122 |
-
EthicsandBias_count = 0
|
| 123 |
-
Datasets_count = 0
|
| 124 |
-
ToolsandLibraries_count = 0
|
| 125 |
-
Tutorials_and_Guides_count = 0
|
| 126 |
-
Debugging_count = 0
|
| 127 |
|
| 128 |
# Appliquer votre algorithme d'analyse à tous les posts
|
| 129 |
for i, post in enumerate(all_posts_with_text, 1):
|
|
@@ -156,13 +147,17 @@ for i, post in enumerate(all_posts_with_text, 1):
|
|
| 156 |
Debugging_count += 1
|
| 157 |
|
| 158 |
# Affichage des résultats
|
| 159 |
-
print(
|
| 160 |
-
|
| 161 |
-
|
| 162 |
-
|
| 163 |
-
|
| 164 |
-
|
| 165 |
-
|
| 166 |
-
|
| 167 |
-
|
| 168 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
import requests
|
|
|
|
| 2 |
import json
|
| 3 |
+
import os
|
| 4 |
|
| 5 |
api_token = os.environ.get("TOKEN")
|
| 6 |
API_URL = "https://api-inference.huggingface.co/models/meta-llama/Meta-Llama-3-8B-Instruct"
|
|
|
|
| 51 |
return 'questions'
|
| 52 |
elif ChatmodelLLM == 2:
|
| 53 |
return 'Chat Model/LLM'
|
| 54 |
+
elif other == 2:
|
| 55 |
return "Other"
|
| 56 |
elif image_generation == 2:
|
| 57 |
return "Image Generation"
|
|
|
|
| 71 |
return f"Erreur: Réponse ambiguë - '{response}'"
|
| 72 |
|
| 73 |
# URL de base de l'API
|
| 74 |
+
base_url = "https://huggingface.co/api/posts?skip=20&sort=recent"
|
| 75 |
|
| 76 |
# Paramètres pour la pagination
|
| 77 |
skip = 0 # Nombre d'éléments à sauter
|
|
|
|
| 114 |
# Maintenant, all_posts_with_text contient tous les posts récupérés avec leur texte
|
| 115 |
|
| 116 |
# Initialisation des compteurs
|
| 117 |
+
questions_count, Chat_ModelLLM_count, Other_count, ImageGeneration_count, Fine_tuning_count, EthicsandBias_count, Datasets_count, ToolsandLibraries_count, Tutorials_and_Guides_count,Debugging_count = 0,0,0,0,0,0,0,0,0,0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 118 |
|
| 119 |
# Appliquer votre algorithme d'analyse à tous les posts
|
| 120 |
for i, post in enumerate(all_posts_with_text, 1):
|
|
|
|
| 147 |
Debugging_count += 1
|
| 148 |
|
| 149 |
# Affichage des résultats
|
| 150 |
+
print(
|
| 151 |
+
|
| 152 |
+
f"Questions: {questions_count}",
|
| 153 |
+
f"Chat Model/LLM: {Chat_ModelLLM_count}",
|
| 154 |
+
f"Other: {Other_count}",
|
| 155 |
+
f"Image Generation: {ImageGeneration_count}",
|
| 156 |
+
f"Fine-tuning: {Fine_tuning_count}",
|
| 157 |
+
f"Ethics and Bias: {EthicsandBias_count}",
|
| 158 |
+
f"Datasets: {Datasets_count}",
|
| 159 |
+
f"Tools and Libraries: {ToolsandLibraries_count}",
|
| 160 |
+
f"Tutorials and Guides: {Tutorials_and_Guides_count}",
|
| 161 |
+
f"Debugging: {Debugging_count}"
|
| 162 |
+
|
| 163 |
+
)
|