HomeAI Tutorial

Aule-Attention

Public

High-performance FlashAttention-2 for AMD, Intel, and Apple GPUs. Drop-in replacement for PyTorch SDPA. Triton backend for ROCm (MI300X, RDNA3), Vulkan backend for consumer GPUs. No CUDA required.

Creat2025-11-28T19:06:23
Update2025-12-09T10:31:28
https://auletechnologies.com
25
Stars
0
Stars Increase

Related projects