You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi @sorasoras, I appreciate the excitement over the new llama.cpp K-bit quants that are slowly rolling in. At the current stage of AutoAWQ, I cannot take on these large things on my own. Although I wish I had the time (and compute+$), I do not. So the priority is to make AutoAWQ smooth to work with and to provide quantization of newer models as they come out.
ggerganov/llama.cpp#4856
ggerganov/llama.cpp#4861
There might be some interesting idea you can use to improve AutoAWQ
It does looks pretty promising
or combine them to generate a quantization.
The text was updated successfully, but these errors were encountered: