The official prequantized EfficientQAT models.
Mengzhao Chen
ChenMnZ
AI & ML interests
model compression
Organizations
None yet
Collections
4
models
129
ChenMnZ/Mistral-Large-Instruct-2407-EfficientQAT-w2g64-GPTQ
Updated
•
21
•
25
ChenMnZ/Llama-3-70b-EfficientQAT-w4g128-BitBLAS
Text Generation
•
Updated
•
1
ChenMnZ/Llama-3-70b-instruct-EfficientQAT-w2g128-BitBLAS
Text Generation
•
Updated
ChenMnZ/Llama-3-8b-instruct-EfficientQAT-w4g128-BitBLAS
Text Generation
•
Updated
ChenMnZ/Llama-3-8b-instruct-EfficientQAT-w2g64-BitBLAS
Text Generation
•
Updated
ChenMnZ/Llama-3-8b-instruct-EfficientQAT-w2g128-BitBLAS
Text Generation
•
Updated
ChenMnZ/Llama-3-8b-EfficientQAT-w4g128-BitBLAS
Text Generation
•
Updated
ChenMnZ/Llama-3-8b-EfficientQAT-w2g64-BitBLAS
Text Generation
•
Updated
ChenMnZ/Llama-3-8b-EfficientQAT-w2g128-BitBLAS
Text Generation
•
Updated
ChenMnZ/Llama-3-70b-instruct-EfficientQAT-w4g128-BitBLAS
Text Generation
•
Updated
datasets
None public yet