SIGN IN SIGN UP

🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.

158577 0 0 Python

🚨 [Cache] Native mamba & hybrid cache (#44950)

* add Cache and test on Mamba

* fix

* fix

* fix

* fix

* fix

* final fix

* test hybrid with jamba

* fix tests

* fixes

* fix

* fix

* fix

* combine both types + zambas

* add config mapèping

* adjust tests

* fix

* fix

* fix

* more models

* final mambas

* config

* finalize almost everything

* simplify tests

* simplify tests further

* fix tests

* oupsi

* fix

* fix broken no_split_modules

* fix

* fixes

* fix

* fix

* fixes

* add layer type

* oupsi

* fix

* style

* fix

* fixes

* final fix

* forgot those qwens

* tests

* offloading

* much better static shape native design

* oupsi

* adjustments in generate

* allow cudagraphs

* small oupsi

* start renaming

* revert unrelated what are they doing here

* more renaming

* revert offloading change

* add offloading skips

* split shapes for tests

* comments and renaming
C
Cyril Vallez committed
2dba8e0495974930af02274d75bd182d22cc1686
Parent: a91232a
Committed by GitHub <[email protected]> on 3/31/2026, 1:09:42 PM