We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
See vllm-project/vllm#1002, vllm-project/vllm#5191.
Should be able to set gguf as QUANTIZATION envar, but we also need to provide exact quant. I'm thinking of some MODEL_FILENAME envar containing the exact filename in the model's repository. The model download logic shall be changed, see https://github.com/Isotr0py/vllm/blob/main/examples/gguf_inference.py.
gguf
QUANTIZATION
MODEL_FILENAME
The text was updated successfully, but these errors were encountered:
yeah i guess now we dont need set quantization env var but we need to support it like from this one
Sorry, something went wrong.
@nerdylive123 How could I implement GGUF for my fork?
No branches or pull requests
See vllm-project/vllm#1002, vllm-project/vllm#5191.
Should be able to set
gguf
asQUANTIZATION
envar, but we also need to provide exact quant. I'm thinking of someMODEL_FILENAME
envar containing the exact filename in the model's repository. The model download logic shall be changed, see https://github.com/Isotr0py/vllm/blob/main/examples/gguf_inference.py.The text was updated successfully, but these errors were encountered: