Parallel Thinking activation
#17 opened 8 days ago
by
PhilippeEiffel
when using with opencode, tool calls have trouble
👍 1
3
#16 opened 14 days ago
by
rosspanda0
Performance report with 72GB VRAM: 32 t/s
1
#14 opened 17 days ago
by
SlavikF
Better Perplexity Alternative GGUFs
❤️ 👍 5
3
#13 opened 17 days ago
by
ubergarm
chat template is broken
11
#12 opened 18 days ago
by
grapevine-AI
Is it possible to release a version with low bit quantization?
4
#11 opened 19 days ago
by
lan0004
How do I run it using Oobabooga? I'm getting the following error
1
#10 opened 19 days ago
by
TeaDiffusion
What are the benchmarks of the 4 bit model vs the FP8 model?
2
#9 opened 20 days ago
by
Grossor
Make this model more visible on the hub
🚀 5
1
#8 opened 22 days ago
by
victor
INT8 quantization for KVCache on DGX Spark/GB10
3
#6 opened 22 days ago
by
JDWarner
config.json file needed at root?
1
#4 opened 22 days ago
by
pathosethoslogos
cool model !!
👍 1
3
#3 opened 23 days ago
by
gopi87
great job! thanks!
#1 opened 23 days ago
by
semon017