mirror of
https://github.com/vllm-project/vllm.git
synced 2025-10-20 23:03:52 +08:00
[Misc] small improve (#18680)
Signed-off-by: reidliu41 <reid201711@gmail.com> Co-authored-by: reidliu41 <reid201711@gmail.com>
This commit is contained in:
@ -15,7 +15,7 @@ pip install bitsandbytes>=0.45.3
|
|||||||
|
|
||||||
vLLM reads the model's config file and supports both in-flight quantization and pre-quantized checkpoint.
|
vLLM reads the model's config file and supports both in-flight quantization and pre-quantized checkpoint.
|
||||||
|
|
||||||
You can find bitsandbytes quantized models on <https://huggingface.co/models?search=bitsandbytes>.
|
You can find bitsandbytes quantized models on [Hugging Face](https://huggingface.co/models?search=bitsandbytes).
|
||||||
And usually, these repositories have a config.json file that includes a quantization_config section.
|
And usually, these repositories have a config.json file that includes a quantization_config section.
|
||||||
|
|
||||||
## Read quantized checkpoint
|
## Read quantized checkpoint
|
||||||
|
Reference in New Issue
Block a user