Spaces:
Running
Running
default to qwen3 0.6b
Browse files- summarize_transcript.py +7 -6
summarize_transcript.py
CHANGED
|
@@ -56,10 +56,11 @@ def stream_summarize_transcript(llm, transcript):
|
|
| 56 |
|
| 57 |
stream = llm.create_chat_completion(
|
| 58 |
messages=messages,
|
| 59 |
-
max_tokens=
|
| 60 |
-
temperature=0.
|
| 61 |
-
|
| 62 |
-
|
|
|
|
| 63 |
stop=["<|end_of_text|>", "<|eot_id|>", "<|eom_id|>"],
|
| 64 |
stream=True
|
| 65 |
)
|
|
@@ -86,8 +87,8 @@ def main():
|
|
| 86 |
default="./transcripts/short.txt",
|
| 87 |
help="Path to the input transcript file (default: ./transcripts/short.txt)")
|
| 88 |
parser.add_argument("-m", "--model", type=str,
|
| 89 |
-
default="
|
| 90 |
-
help="HuggingFace model in format repo_id:quant (e.g.,
|
| 91 |
parser.add_argument("-c", "--cpu", action="store_true", help="Force CPU only inference")
|
| 92 |
args = parser.parse_args()
|
| 93 |
|
|
|
|
| 56 |
|
| 57 |
stream = llm.create_chat_completion(
|
| 58 |
messages=messages,
|
| 59 |
+
max_tokens=1024,
|
| 60 |
+
temperature=0.6,
|
| 61 |
+
min_p=0.0,
|
| 62 |
+
top_p=0.95,
|
| 63 |
+
top_k=20,
|
| 64 |
stop=["<|end_of_text|>", "<|eot_id|>", "<|eom_id|>"],
|
| 65 |
stream=True
|
| 66 |
)
|
|
|
|
| 87 |
default="./transcripts/short.txt",
|
| 88 |
help="Path to the input transcript file (default: ./transcripts/short.txt)")
|
| 89 |
parser.add_argument("-m", "--model", type=str,
|
| 90 |
+
default="unsloth/Qwen3-0.6B-GGUF:Q4_0",
|
| 91 |
+
help="HuggingFace model in format repo_id:quant (e.g., bartowski/baidu_ERNIE-4.5-0.3B-PT-GGUF:Q4_0)")
|
| 92 |
parser.add_argument("-c", "--cpu", action="store_true", help="Force CPU only inference")
|
| 93 |
args = parser.parse_args()
|
| 94 |
|