Zheng Han
traphix
·
AI & ML interests
None yet
Recent Activity
new activity 3 days ago
nm-testing/MiniMax-M2.5-W4A16:oneshot vs model_free_ptq? which one has better recovery? new activity 9 days ago
RedHatAI/Qwen3.5-397B-A17B-FP8-dynamic:W4A16 quant new activity 10 days ago
apolo13x/Qwen3.5-35B-A3B-quantized.w4a16:Any creation details?Organizations
None yet
oneshot vs model_free_ptq? which one has better recovery?
1
#1 opened 3 days ago
by
traphix
W4A16 quant
👍 2
5
#1 opened about 2 months ago
by
timroethig
Any creation details?
#2 opened 10 days ago
by
traphix
Creation details?
#8 opened 14 days ago
by
traphix
Creation details?
#2 opened 15 days ago
by
traphix
Which framework was used for FP8 quantization? LLM-compressor?
2
#1 opened 15 days ago
by
traphix
GPTQ quantization
2
#2 opened about 2 months ago
by
ArtemSultanov
Question about weight_observer?
2
#1 opened 26 days ago
by
traphix
INT4 w4a16 quantinization?
➕ 1
#1 opened about 1 month ago
by
traphix
Quantization code for int4(w4a16) ?
#6 opened about 1 month ago
by
traphix
Tokenizer you are loading with an incorrect regex pattern
1
#2 opened 4 months ago
by
traphix
Failed to find a kernel that can implement the WNA16 linear layer
#1 opened 4 months ago
by
traphix
vllm error: Extra inputs are not permitted
#1 opened 4 months ago
by
traphix
Can A100 run Qwen3-235B-A22B-Instruct-2507-NVFP4?
#1 opened 4 months ago
by
traphix
Error on 4 x L40s
➕ 2
1
#4 opened 7 months ago
by
traphix
I got ValueError
👀 2
10
#3 opened 7 months ago
by
spow12
How to run this model via vllm?
11
#2 opened 7 months ago
by
traphix
FP8 please
👀➕ 16
8
#18 opened 7 months ago
by
aliquis-pe