🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
[XPU] Add flash_attn2 support for XPU (#41956)
* Add flash_attention_2 and kernels-community/flash-attn support for XPU * Add flash-attn-2 support for XPU * Delete deterministic algorithm for xpu * Fix code style * Modify repo_id to match the latest kernels-community/flash-attn2 * Fix code style * Update * Make quality * Use kernels loading * Update * Delete invalid import * Update comment --------- Co-authored-by: Anton Vlasjuk <[email protected]>
Y
YangKai0616 committed
07bfd2f8ecd0776591b3c051a061fbfd81848052
Parent: 9162e19
Committed by GitHub <[email protected]>
on 11/21/2025, 4:43:48 PM