SIGN IN SIGN UP

🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.

0 0 0 Python

[XPU] Add flash_attn2 support for XPU (#41956)

* Add flash_attention_2 and kernels-community/flash-attn support for XPU

* Add flash-attn-2 support for XPU

* Delete deterministic algorithm for xpu

* Fix code style

* Modify repo_id to match the latest kernels-community/flash-attn2

* Fix code style

* Update

* Make quality

* Use kernels loading

* Update

* Delete invalid import

* Update comment

---------

Co-authored-by: Anton Vlasjuk <[email protected]>
Y
YangKai0616 committed
07bfd2f8ecd0776591b3c051a061fbfd81848052
Parent: 9162e19
Committed by GitHub <[email protected]> on 11/21/2025, 4:43:48 PM