New discussion

Update README.md

#26 opened 8 days ago by
mnhasmuiscas

it run good gguf on colab cpu

#25 opened 27 days ago by
asdgad

extract reasoning_content error

👀 2
3
#24 opened about 1 month ago by
zet-yd

Cannot summarize 8000 tokens

2
#22 opened about 1 month ago by
kalle07

llama.cpp support

🚀 7
#21 opened about 1 month ago by
ngxson

VLLM 启动报错了

1
#20 opened about 1 month ago by
qinghuiyyds

Update README.md

#17 opened about 2 months ago by
byjiang1996

it run good in colab t4

10
#16 opened about 2 months ago by
asdgad

4bit

1
#15 opened about 2 months ago by
asdgad

run colab t4 but

5
#14 opened about 2 months ago by
asdgad

not run

👀 1
1
#13 opened about 2 months ago by
asdgad

Question regarding the FP8 version

1
#9 opened about 2 months ago by
thecr7guy

vLLM error

10
#8 opened about 2 months ago by
ccernat

It's really top_k = 2?

👍 2
1
#6 opened about 2 months ago by
CHNtentes

The demo script loads forever.

3
#1 opened about 2 months ago by
AliceThirty