* fix: manual edits
* Apply suggestions from code review
Apply suggestions from code review and make additional revisions
Co-authored-by: HyunSang Jang <tasker.dev103@gmail.com>
* Apply suggestions from code review
Apply suggestions from code review — updated inline links for related text
* Apply suggestions from code review
Apply suggestions from code review - final
* Update docs/source/ko/_toctree.yml
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
---------
Co-authored-by: HyunSang Jang <tasker.dev103@gmail.com>
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
Removed unnecessary checks for key being a torch.fx.Proxy in GQA conditions because fx tracing is no longer supported, and torch.export supports enable_gqa.
* update
* batch update model code
* typos
* too many diffs, dump
* dump again
* another dump
* fix copies
* make `rope_scaling_dict` self attr
* fix a few more tests
* another update
* fix a few more tests, hopefully last ones
* fox copies
* fix copies again
* fix newly added models, I hate rebasing on main
* update config files
* modular files
* fix rope utils test
* docstring has to be indented more, why?
* oops forgot to update some modualr files
* copy from doesn't copy decorators?
* fix overriden test as well
* add a new test
* fix failing tests again
* update docstrings
* fix phi3
* fix two models
* fix copies
* forgot to add
* stupid bug from modular conversion
* fix slow tests
* update to call rotary emb once per model forward
* 3K tests failing?!
* update
* update more models
* fix copies
* fix the rest of tests hopefully
* fix after rebase
* fix the rope tests
* fix docs omni
* change a bit
* models with layer types
* why it was deleted?
* fix a few tests
* fix last test!
* delete extra empty lines
* add a test case
* more changes
* fix models
* typing hint for nested rope params
* missed when resolving conflicts
* delete layer types and fix typo
* fix copies
* fix copies
* update docs text
* docs
* huuge update all models
* fix copies
* rename attr to align with new format
* delete redundant rope tests
* trigger ci
* update the case
* this is why i hate rebasing
* maybe fixed?
* oops
* now fix?
* fix last tests and copies
* fix copies?
* fix minimax and gemma3n
* update typo
* deprecation end version
* final fix copies :fingers-crossed:
* oh my, add the docs in toctree
* oke, this is really the last fix
* fix copies and hope that tests won't start failing again
* use rope scaling if saved
* fix slow tests
* fix cwm and unrelated deepseek
* fix last
* update
* hope it works now, it took so long
* lets keep None for now, I will try to remove after checking tests
* some more fixes, i find and replace does not always find all cases
* last fix of tests
* arthur's comment for extra foreward kwargs
* delete unused code
* fix slow qwen tests
* delete layer types from models
* faulty modular conversion
* fix qwen omni
* fix copies and style
* address my comment
---------
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
* Default implementation - no time improvement
* Improved implementation - apparently 2 times faster with only simple function refactor
* elementary torch first approach, still need further implementation of torch first method
* torch-first approach finished
* refactor processor
* refactor test
* partial doc update
* EfficientLoFTRImageProcessorFast based implementation
* EfficientLoFTRImageProcessorFast based implementation
* Logic checked - Test Passed - Validated execution speed
* use modular for efficientloftr
* fix import
---------
Co-authored-by: yonigozlan <yoni.gozlan@huggingface.co>
Co-authored-by: Yoni Gozlan <74535834+yonigozlan@users.noreply.github.com>
* Add a switch to CB in case of paged cache
* Added paged as a valid cache implem
* Added a fallback on inputs_ids as a name
* Rookie mistake
* Removed paged from cache implems
* Added warning about some beam search args
* Moved up CB warning
* Fix EncoderDecoder cache
* Add the option for the ddp data tuples to have 2 elems
* Modifiy the order of the KV and sliding
* Adapted RAG and Whisper to new EncoderDecoderCache
* A single comma
* Remove kwargs in map
* Fixed order in manual injection cache test
* Slight changes to support legacy format
* Removed Nonnes
This commit addresses a noisy warning and improves the robustness of the base pipeline implementation.
- The device placement message in the pipeline base class has been changed from a `warning` to a `debug` log. This reduces log noise for users who are aware of their device setup, while still providing the information for debugging purposes.
- Additionally, potential `UnboundLocalError` exceptions in the `_pad` and `check_model_type` functions have been prevented by initializing variables before their conditional assignment.
* Add is_causal to KosmosTextAttention
* Move get target_dtype to be imported elsewhere
* Fix fp32 flash attention bug in bark
* Fix is_causal in mllama
* Fix fp32 issue on StableLM
* Fix repo-consistency