[Kernel] add triton fused moe kernel for gptq/awq #706
Annotations
2 errors
Ruff (E501):
vllm/model_executor/layers/quantization/moe_wna16.py#L174
vllm/model_executor/layers/quantization/moe_wna16.py:174:81: E501 Line too long (87 > 80)
|
pre-commit
Process completed with exit code 1.
|