Compare commits

...

1 Commits

Author SHA1 Message Date
9cb290cc89 FlashAttention remove Nvidia only GPUs to more generic. 2023-11-15 15:33:09 +01:00

View File

@ -62,7 +62,7 @@ model = AutoModelForCausalLM.from_pretrained(
<Tip>
FlashAttention-2 can only be used when the model's dtype is `fp16` or `bf16`, and it only runs on Nvidia GPUs. Make sure to cast your model to the appropriate dtype and load them on a supported device before using FlashAttention-2.
FlashAttention-2 can only be used when the model's dtype is `fp16` or `bf16`, and is available on both AMD & Nvidia GPUs. Make sure to cast your model to the appropriate dtype and load them on a supported device before using FlashAttention-2.
</Tip>