Zheng Han
traphix
·
AI & ML interests
None yet
Recent Activity
new activity 2 days ago
RedHatAI/Qwen3.6-35B-A3B-NVFP4:Regarding the correctness of the int4 quantization script new activity 5 days ago
RedHatAI/Qwen3.6-35B-A3B-NVFP4:Creation details? new activity 5 days ago
apolo13x/Qwen3.5-27B-quantized.w4a16:Any creation details?Organizations
None yet
Regarding the correctness of the int4 quantization script
1
#5 opened 2 days ago
by
traphix
Creation details?
1
#3 opened 5 days ago
by
traphix
Any creation details?
#1 opened 5 days ago
by
traphix
oneshot vs model_free_ptq? which one has better recovery?
1
#1 opened 19 days ago
by
traphix
W4A16 quant
👍 2
5
#1 opened 2 months ago
by
timroethig
Any creation details?
#2 opened 26 days ago
by
traphix
Creation details?
1
#8 opened 29 days ago
by
traphix
Creation details?
#2 opened about 1 month ago
by
traphix
Which framework was used for FP8 quantization? LLM-compressor?
2
#1 opened about 1 month ago
by
traphix
GPTQ quantization
2
#2 opened 2 months ago
by
ArtemSultanov
Which framework was used to quantize this model? llm-compressor? or Can you share the quantization Python script?
#1 opened about 1 month ago
by
traphix
Which framework was used to quantize this model? llm-compressor? or Can you share the quantization Python script?
1
#2 opened about 1 month ago
by
traphix
Question about weight_observer?
2
#1 opened about 1 month ago
by
traphix
INT4 w4a16 quantinization?
➕ 1
#1 opened about 2 months ago
by
traphix
Quantization code for int4(w4a16) ?
#6 opened about 2 months ago
by
traphix
Tokenizer you are loading with an incorrect regex pattern
1
#2 opened 4 months ago
by
traphix
Failed to find a kernel that can implement the WNA16 linear layer
#1 opened 4 months ago
by
traphix
vllm error: Extra inputs are not permitted
#1 opened 4 months ago
by
traphix
Can A100 run Qwen3-235B-A22B-Instruct-2507-NVFP4?
#1 opened 5 months ago
by
traphix
Error on 4 x L40s
➕ 2
1
#4 opened 7 months ago
by
traphix