🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
🚨 [Cache] Native mamba & hybrid cache (#44950)
* add Cache and test on Mamba * fix * fix * fix * fix * fix * final fix * test hybrid with jamba * fix tests * fixes * fix * fix * fix * combine both types + zambas * add config mapèping * adjust tests * fix * fix * fix * more models * final mambas * config * finalize almost everything * simplify tests * simplify tests further * fix tests * oupsi * fix * fix broken no_split_modules * fix * fixes * fix * fix * fixes * add layer type * oupsi * fix * style * fix * fixes * final fix * forgot those qwens * tests * offloading * much better static shape native design * oupsi * adjustments in generate * allow cudagraphs * small oupsi * start renaming * revert unrelated what are they doing here * more renaming * revert offloading change * add offloading skips * split shapes for tests * comments and renaming
C
Cyril Vallez committed
2dba8e0495974930af02274d75bd182d22cc1686
Parent: a91232a
Committed by GitHub <[email protected]>
on 3/31/2026, 1:09:42 PM