tool calls sometimes fail on vllm
#19 opened about 1 month ago
by
viktara
Not support flash_attn_2๏ผ
#17 opened about 1 month ago
by
bdsqlsz
chatllm.cpp adds support of this model
๐ 7
5
#16 opened about 1 month ago
by
J22
possible bug in processing_step3.py?
2
#15 opened about 1 month ago
by
J22
Quantisation support for this model
โค๏ธ 4
5
#14 opened about 1 month ago
by
dineshananthi
Update processing_step3.py
#13 opened about 1 month ago
by
jakubstrawa
Installation Video and Testing - Step by Step
๐ 6
#12 opened about 1 month ago
by
fahdmirzac
is it possible to provide FP8 version in additional to BF16
๐ฅ 8
1
#10 opened about 1 month ago
by
water258
Incorrect eos_token_id in config causing infinite generation
1
#9 opened about 1 month ago
by
WenyaLi
perfect
๐ค 15
2
#8 opened about 1 month ago
by
ares2324
Doesn't work in OpenAI Streaming Interface
2
#7 opened about 1 month ago
by
pytokusu
Add pipeline tag and library name to metadata
#6 opened about 1 month ago
by
nielsr
vLLM is not working
11
#4 opened about 1 month ago
by
wei01
Typo in paper
1
#3 opened about 1 month ago
by
hgsg
Llama.cpp support
๐ ๐ฅ 13
2
#2 opened about 1 month ago
by
tcpmux
่ฏท้ฎๆๆจ่็ๅค็ฐ่ถ ๅๅ๏ผ
1
#1 opened about 1 month ago
by
JjjjjZzz