mirror of
				https://github.com/vllm-project/vllm.git
				synced 2025-11-04 17:34:34 +08:00 
			
		
		
		
	[Misc] small improve (#18680)
Signed-off-by: reidliu41 <reid201711@gmail.com> Co-authored-by: reidliu41 <reid201711@gmail.com>
This commit is contained in:
		@ -15,7 +15,7 @@ pip install bitsandbytes>=0.45.3
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
vLLM reads the model's config file and supports both in-flight quantization and pre-quantized checkpoint.
 | 
					vLLM reads the model's config file and supports both in-flight quantization and pre-quantized checkpoint.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
You can find bitsandbytes quantized models on <https://huggingface.co/models?search=bitsandbytes>.
 | 
					You can find bitsandbytes quantized models on [Hugging Face](https://huggingface.co/models?search=bitsandbytes).
 | 
				
			||||||
And usually, these repositories have a config.json file that includes a quantization_config section.
 | 
					And usually, these repositories have a config.json file that includes a quantization_config section.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
## Read quantized checkpoint
 | 
					## Read quantized checkpoint
 | 
				
			||||||
 | 
				
			|||||||
		Reference in New Issue
	
	Block a user