Spaces:
Sleeping
Sleeping
Update app.py
Browse filesSwitch to google/pegasus-xsum for better summarization
app.py
CHANGED
|
@@ -10,13 +10,12 @@ if not HF_TOKEN:
|
|
| 10 |
print(f"HF_TOKEN loaded: {len(HF_TOKEN) if HF_TOKEN else 0} characters")
|
| 11 |
|
| 12 |
try:
|
| 13 |
-
client = InferenceClient(model="
|
| 14 |
def generate_summary(text):
|
| 15 |
for _ in range(3):
|
| 16 |
try:
|
| 17 |
-
response = client.summarization(text)
|
| 18 |
print(f"API Response: {response}")
|
| 19 |
-
# 优先使用 generated_text,如果为空则尝试 summary_text
|
| 20 |
if hasattr(response, 'generated_text') and response.generated_text:
|
| 21 |
return response.generated_text
|
| 22 |
elif hasattr(response, 'summary_text') and response.summary_text:
|
|
|
|
| 10 |
print(f"HF_TOKEN loaded: {len(HF_TOKEN) if HF_TOKEN else 0} characters")
|
| 11 |
|
| 12 |
try:
|
| 13 |
+
client = InferenceClient(model="google/pegasus-xsum", token=HF_TOKEN)
|
| 14 |
def generate_summary(text):
|
| 15 |
for _ in range(3):
|
| 16 |
try:
|
| 17 |
+
response = client.summarization(text, max_length=50, min_length=10, num_beams=4)
|
| 18 |
print(f"API Response: {response}")
|
|
|
|
| 19 |
if hasattr(response, 'generated_text') and response.generated_text:
|
| 20 |
return response.generated_text
|
| 21 |
elif hasattr(response, 'summary_text') and response.summary_text:
|