mirror of
https://github.com/vllm-project/vllm.git
synced 2025-10-20 14:53:52 +08:00
[Misc] Update TritonLanguagePlaceholder to have attributes that are used by Flash Linear Attention ops. (#26853)
Co-authored-by: Xudong Ma <mxd@meta.com>
This commit is contained in:
@ -98,3 +98,6 @@ class TritonLanguagePlaceholder(types.ModuleType):
|
|||||||
self.int64 = None
|
self.int64 = None
|
||||||
self.int32 = None
|
self.int32 = None
|
||||||
self.tensor = None
|
self.tensor = None
|
||||||
|
self.exp = None
|
||||||
|
self.log = None
|
||||||
|
self.log2 = None
|
||||||
|
Reference in New Issue
Block a user