ztz
aabbccddwasd
AI & ML interests
LLM
Recent Activity
new activity
about 6 hours ago
nvidia/Qwen3.5-397B-A17B-NVFP4:Support SM120 new activity
3 days ago
Sehyo/Qwen3.5-397B-A17B-NVFP4:not working on vllm new activity
3 days ago
Sehyo/Qwen3.5-397B-A17B-NVFP4:missing think tag Organizations
None yet
Support SM120
❤️ 👍 10
2
#2 opened about 20 hours ago
by
darkstar3537
not working on vllm
11
#1 opened 4 days ago
by
aabbccddwasd
missing think tag
8
#2 opened 3 days ago
by
fouvy
Anyone try this on 4x RTX 6000 Pro yet?
39
#1 opened 6 days ago
by
zenmagnets
vllm/sglang issue?
1
#1 opened 4 days ago
by
Halbin
vllm 部署oom
5
#22 opened 6 days ago
by
Chris2me
W4A16 quant
👍 2
3
#1 opened 5 days ago
by
timroethig
fake knowledge 假知识
6
#21 opened 6 days ago
by
aabbccddwasd
we still need some Air
🚀 👍 65
15
#1 opened 2 months ago
by
jacek2024
Is it possible to make smaller NVFP4 quant at 340-360GB to fit in 4x96gb?
👍 1
68
#1 opened 3 months ago
by
Fernanda24
Aww Man!
20
#1 opened 3 months ago
by
mtcl
FP8 and FP4 please?
➕ 1
#4 opened 6 months ago
by
aabbccddwasd
An Improvement, But Q3 30b Still Has Very Little General Knowledge
❤️ 👍 3
11
#2 opened 7 months ago
by
phil111
🚀[Fine-tuning] 8x80GiB GPUs LoRA finetuning Qwen3-235B-A22B-Instruct-2507
🤗 4
1
#25 opened 7 months ago
by
study-hjt
刚部署满血deepseek r1 0528版本,推理性能提升这么多嘛?不是架构没变嘛?
12
#75 opened 9 months ago
by
jakyer
quantize deepseek-r1-0528 please
👍 2
3
#14 opened 9 months ago
by
aabbccddwasd
benchmark please?
👍 1
#1 opened 9 months ago
by
aabbccddwasd
In complex reasoning tasks Qwen3 is far behind QwQ
12
#32 opened 10 months ago
by
AdamF92
Qwen/Qwen2.5-VL-72B-Instruct-AWQ and Qwen/Qwen2.5-VL-40<B-Instruct-AWQ please
➕ ❤️ 18
6
#1 opened about 1 year ago
by
devops724