flashattention_logo

flash-attn @ <>

Flash Attention 2 Pre-built Wheels

This repository contains a copy of all prebuilt wheels for flash_attn-2.8.3 from Dao-AILab's Flash Attention, organized for better categorization based on PyTorch versions and Python (cp) release versions.

This repository provides pre-built wheels for flash-attn version 2.8.3 for various PyTorch versions, Python versions, and architectures (compiled with CUDA 12). You can install these directly using pip install <url> or add the provided strings directly to your requirements.txt.

The detailed categories and structured view of the strangertoolshf/flash_attention_2_wheelhouse folders and files on the hf-tree shareable link are available here: huggingface-tree

Open this to view the entire directory.
strangertoolshf/flash_attention_2_wheelhouse (main)
+-- wheelhouse-flash_attn-2.8.3
|   +-- linux_aarch64
|   |   +-- torch2.9
|   |       +-- cu12
|   |           +-- abiTRUE
|   |               +-- cp312
|   |                   +-- flash_attn-2.8.3+cu12torch2.9cxx11abiTRUE-cp312-cp312-linux_aarch64.whl (241.8 MB)
|   +-- linux_x86_64
|       +-- torch2.4
|       |   +-- cu12
|       |       +-- abiFALSE
|       |       |   +-- cp310
|       |       |   |   +-- flash_attn-2.8.3+cu12torch2.4cxx11abiFALSE-cp310-cp310-linux_x86_64.whl (244.1 MB)
|       |       |   +-- cp311
|       |       |   |   +-- flash_attn-2.8.3+cu12torch2.4cxx11abiFALSE-cp311-cp311-linux_x86_64.whl (244.1 MB)
|       |       |   +-- cp312
|       |       |   |   +-- flash_attn-2.8.3+cu12torch2.4cxx11abiFALSE-cp312-cp312-linux_x86_64.whl (244.1 MB)
|       |       |   +-- cp39
|       |       |       +-- flash_attn-2.8.3+cu12torch2.4cxx11abiFALSE-cp39-cp39-linux_x86_64.whl (244.1 MB)
|       |       +-- abiTRUE
|       |           +-- cp310
|       |           |   +-- flash_attn-2.8.3+cu12torch2.4cxx11abiTRUE-cp310-cp310-linux_x86_64.whl (244.1 MB)
|       |           +-- cp311
|       |           |   +-- flash_attn-2.8.3+cu12torch2.4cxx11abiTRUE-cp311-cp311-linux_x86_64.whl (244.1 MB)
|       |           +-- cp312
|       |           |   +-- flash_attn-2.8.3+cu12torch2.4cxx11abiTRUE-cp312-cp312-linux_x86_64.whl (244.1 MB)
|       |           +-- cp39
|       |               +-- flash_attn-2.8.3+cu12torch2.4cxx11abiTRUE-cp39-cp39-linux_x86_64.whl (244.1 MB)
|       +-- torch2.5
|       |   +-- cu12
|       |       +-- abiFALSE
|       |       |   +-- cp310
|       |       |   |   +-- flash_attn-2.8.3+cu12torch2.5cxx11abiFALSE-cp310-cp310-linux_x86_64.whl (244.1 MB)
|       |       |   +-- cp311
|       |       |   |   +-- flash_attn-2.8.3+cu12torch2.5cxx11abiFALSE-cp311-cp311-linux_x86_64.whl (244.1 MB)
|       |       |   +-- cp312
|       |       |   |   +-- flash_attn-2.8.3+cu12torch2.5cxx11abiFALSE-cp312-cp312-linux_x86_64.whl (244.1 MB)
|       |       |   +-- cp313
|       |       |   |   +-- flash_attn-2.8.3+cu12torch2.5cxx11abiFALSE-cp313-cp313-linux_x86_64.whl (244.1 MB)
|       |       |   +-- cp39
|       |       |       +-- flash_attn-2.8.3+cu12torch2.5cxx11abiFALSE-cp39-cp39-linux_x86_64.whl (244.1 MB)
|       |       +-- abiTRUE
|       |           +-- cp310
|       |           |   +-- flash_attn-2.8.3+cu12torch2.5cxx11abiTRUE-cp310-cp310-linux_x86_64.whl (244.1 MB)
|       |           +-- cp311
|       |           |   +-- flash_attn-2.8.3+cu12torch2.5cxx11abiTRUE-cp311-cp311-linux_x86_64.whl (244.2 MB)
|       |           +-- cp312
|       |           |   +-- flash_attn-2.8.3+cu12torch2.5cxx11abiTRUE-cp312-cp312-linux_x86_64.whl (244.2 MB)
|       |           +-- cp313
|       |           |   +-- flash_attn-2.8.3+cu12torch2.5cxx11abiTRUE-cp313-cp313-linux_x86_64.whl (244.2 MB)
|       |           +-- cp39
|       |               +-- flash_attn-2.8.3+cu12torch2.5cxx11abiTRUE-cp39-cp39-linux_x86_64.whl (244.1 MB)
|       +-- torch2.6
|       |   +-- cu12
|       |       +-- abiFALSE
|       |       |   +-- cp310
|       |       |   |   +-- flash_attn-2.8.3+cu12torch2.6cxx11abiFALSE-cp310-cp310-linux_x86_64.whl (244.1 MB)
|       |       |   +-- cp311
|       |       |   |   +-- flash_attn-2.8.3+cu12torch2.6cxx11abiFALSE-cp311-cp311-linux_x86_64.whl (244.2 MB)
|       |       |   +-- cp312
|       |       |   |   +-- flash_attn-2.8.3+cu12torch2.6cxx11abiFALSE-cp312-cp312-linux_x86_64.whl (244.2 MB)
|       |       |   +-- cp313
|       |       |   |   +-- flash_attn-2.8.3+cu12torch2.6cxx11abiFALSE-cp313-cp313-linux_x86_64.whl (244.2 MB)
|       |       |   +-- cp39
|       |       |       +-- flash_attn-2.8.3+cu12torch2.6cxx11abiFALSE-cp39-cp39-linux_x86_64.whl (244.1 MB)
|       |       +-- abiTRUE
|       |           +-- cp310
|       |           |   +-- flash_attn-2.8.3+cu12torch2.6cxx11abiTRUE-cp310-cp310-linux_x86_64.whl (244.1 MB)
|       |           +-- cp311
|       |           |   +-- flash_attn-2.8.3+cu12torch2.6cxx11abiTRUE-cp311-cp311-linux_x86_64.whl (244.2 MB)
|       |           +-- cp312
|       |           |   +-- flash_attn-2.8.3+cu12torch2.6cxx11abiTRUE-cp312-cp312-linux_x86_64.whl (244.2 MB)
|       |           +-- cp313
|       |           |   +-- flash_attn-2.8.3+cu12torch2.6cxx11abiTRUE-cp313-cp313-linux_x86_64.whl (244.2 MB)
|       |           +-- cp39
|       |               +-- flash_attn-2.8.3+cu12torch2.6cxx11abiTRUE-cp39-cp39-linux_x86_64.whl (244.1 MB)
|       +-- torch2.7
|       |   +-- cu12
|       |       +-- abiFALSE
|       |       |   +-- cp310
|       |       |   |   +-- flash_attn-2.8.3+cu12torch2.7cxx11abiFALSE-cp310-cp310-linux_x86_64.whl (244.1 MB)
|       |       |   +-- cp311
|       |       |   |   +-- flash_attn-2.8.3+cu12torch2.7cxx11abiFALSE-cp311-cp311-linux_x86_64.whl (244.2 MB)
|       |       |   +-- cp312
|       |       |   |   +-- flash_attn-2.8.3+cu12torch2.7cxx11abiFALSE-cp312-cp312-linux_x86_64.whl (244.2 MB)
|       |       |   +-- cp313
|       |       |   |   +-- flash_attn-2.8.3+cu12torch2.7cxx11abiFALSE-cp313-cp313-linux_x86_64.whl (244.2 MB)
|       |       |   +-- cp39
|       |       |       +-- flash_attn-2.8.3+cu12torch2.7cxx11abiFALSE-cp39-cp39-linux_x86_64.whl (244.1 MB)
|       |       +-- abiTRUE
|       |           +-- cp310
|       |           |   +-- flash_attn-2.8.3+cu12torch2.7cxx11abiTRUE-cp310-cp310-linux_x86_64.whl (244.1 MB)
|       |           +-- cp311
|       |           |   +-- flash_attn-2.8.3+cu12torch2.7cxx11abiTRUE-cp311-cp311-linux_x86_64.whl (244.2 MB)
|       |           +-- cp312
|       |           |   +-- flash_attn-2.8.3+cu12torch2.7cxx11abiTRUE-cp312-cp312-linux_x86_64.whl (244.2 MB)
|       |           +-- cp313
|       |           |   +-- flash_attn-2.8.3+cu12torch2.7cxx11abiTRUE-cp313-cp313-linux_x86_64.whl (244.2 MB)
|       |           +-- cp39
|       |               +-- flash_attn-2.8.3+cu12torch2.7cxx11abiTRUE-cp39-cp39-linux_x86_64.whl (244.1 MB)
|       +-- torch2.8
|       |   +-- cu12
|       |       +-- abiFALSE
|       |       |   +-- cp310
|       |       |   |   +-- flash_attn-2.8.3+cu12torch2.8cxx11abiFALSE-cp310-cp310-linux_x86_64.whl (244.2 MB)
|       |       |   +-- cp311
|       |       |   |   +-- flash_attn-2.8.3+cu12torch2.8cxx11abiFALSE-cp311-cp311-linux_x86_64.whl (244.2 MB)
|       |       |   +-- cp312
|       |       |   |   +-- flash_attn-2.8.3+cu12torch2.8cxx11abiFALSE-cp312-cp312-linux_x86_64.whl (244.2 MB)
|       |       |   +-- cp313
|       |       |   |   +-- flash_attn-2.8.3+cu12torch2.8cxx11abiFALSE-cp313-cp313-linux_x86_64.whl (244.2 MB)
|       |       |   +-- cp39
|       |       |       +-- flash_attn-2.8.3+cu12torch2.8cxx11abiFALSE-cp39-cp39-linux_x86_64.whl (244.2 MB)
|       |       +-- abiTRUE
|       |           +-- cp310
|       |           |   +-- flash_attn-2.8.3+cu12torch2.8cxx11abiTRUE-cp310-cp310-linux_x86_64.whl (244.2 MB)
|       |           +-- cp311
|       |           |   +-- flash_attn-2.8.3+cu12torch2.8cxx11abiTRUE-cp311-cp311-linux_x86_64.whl (244.2 MB)
|       |           +-- cp312
|       |           |   +-- flash_attn-2.8.3+cu12torch2.8cxx11abiTRUE-cp312-cp312-linux_x86_64.whl (244.2 MB)
|       |           +-- cp313
|       |           |   +-- flash_attn-2.8.3+cu12torch2.8cxx11abiTRUE-cp313-cp313-linux_x86_64.whl (244.2 MB)
|       |           +-- cp39
|       |               +-- flash_attn-2.8.3+cu12torch2.8cxx11abiTRUE-cp39-cp39-linux_x86_64.whl (244.2 MB)
|       +-- torch2.9
|           +-- cu12
|               +-- cp312
|                   +-- flash_attn-2.8.3+cu12torch2.9cxx11abiTRUE-cp312-cp312-linux_x86_64.whl (242.0 MB)
+-- .gitattributes (10.1 KB)
+-- flash-attention-2.8.3.zip (6.2 MB)
+-- README.md (16.9 KB)

Quick Start with StrangerTools-HF

flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-<FLASH_ATTN_VERSION>/<ARCHITECTURE>/torch<TORCH_VERSION>/cu<CUDA_VERSION>/abi<ABI_FLAG>/<PYTHON_TAG>/flash_attn-<FLASH_ATTN_VERSION>+cu<CUDA_VERSION>torch<TORCH_VERSION>cxx11abi<ABI_FLAG>-<PYTHON_TAG>-<PYTHON_TAG>-<ARCHITECTURE>.whl

Breakdown of the Variables:

  • <FLASH_ATTN_VERSION>: The version of Flash Attention (e.g., 2.8.3).
  • <ARCHITECTURE>: The OS and system architecture (e.g., linux_x86_64, linux_aarch64).
  • <TORCH_VERSION>: The PyTorch version without the minor patch (e.g., 2.4, 2.5, 2.8).
  • <CUDA_VERSION>: The CUDA major version (e.g., 12).
  • <ABI_FLAG>: The CXX11 ABI boolean flag (e.g., TRUE, FALSE).
  • <PYTHON_TAG>: The specific CPython version tag (e.g., cp39, cp310, cp312).

Find all the provided pre-built wheels below.👇


Linux x86_64

Torch 2.9

ABI: TRUE (Implied)

Python 3.12 (cp312)

flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.9/cu12/cp312/flash_attn-2.8.3+cu12torch2.9cxx11abiTRUE-cp312-cp312-linux_x86_64.whl

End of wheels for Torch 2.9.


Torch 2.8

ABI: FALSE

Python 3.9 (cp39)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.8/cu12/abiFALSE/cp39/flash_attn-2.8.3+cu12torch2.8cxx11abiFALSE-cp39-cp39-linux_x86_64.whl
Python 3.10 (cp310)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.8/cu12/abiFALSE/cp310/flash_attn-2.8.3+cu12torch2.8cxx11abiFALSE-cp310-cp310-linux_x86_64.whl
Python 3.11 (cp311)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.8/cu12/abiFALSE/cp311/flash_attn-2.8.3+cu12torch2.8cxx11abiFALSE-cp311-cp311-linux_x86_64.whl
Python 3.12 (cp312)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.8/cu12/abiFALSE/cp312/flash_attn-2.8.3+cu12torch2.8cxx11abiFALSE-cp312-cp312-linux_x86_64.whl
Python 3.13 (cp313)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.8/cu12/abiFALSE/cp313/flash_attn-2.8.3+cu12torch2.8cxx11abiFALSE-cp313-cp313-linux_x86_64.whl

ABI: TRUE

Python 3.9 (cp39)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.8/cu12/abiTRUE/cp39/flash_attn-2.8.3+cu12torch2.8cxx11abiTRUE-cp39-cp39-linux_x86_64.whl
Python 3.10 (cp310)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.8/cu12/abiTRUE/cp310/flash_attn-2.8.3+cu12torch2.8cxx11abiTRUE-cp310-cp310-linux_x86_64.whl
Python 3.11 (cp311)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.8/cu12/abiTRUE/cp311/flash_attn-2.8.3+cu12torch2.8cxx11abiTRUE-cp311-cp311-linux_x86_64.whl
Python 3.12 (cp312)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.8/cu12/abiTRUE/cp312/flash_attn-2.8.3+cu12torch2.8cxx11abiTRUE-cp312-cp312-linux_x86_64.whl
Python 3.13 (cp313)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.8/cu12/abiTRUE/cp313/flash_attn-2.8.3+cu12torch2.8cxx11abiTRUE-cp313-cp313-linux_x86_64.whl

Torch 2.7

ABI: FALSE

Python 3.9 (cp39)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.7/cu12/abiFALSE/cp39/flash_attn-2.8.3+cu12torch2.7cxx11abiFALSE-cp39-cp39-linux_x86_64.whl
Python 3.10 (cp310)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.7/cu12/abiFALSE/cp310/flash_attn-2.8.3+cu12torch2.7cxx11abiFALSE-cp310-cp310-linux_x86_64.whl
Python 3.11 (cp311)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.7/cu12/abiFALSE/cp311/flash_attn-2.8.3+cu12torch2.7cxx11abiFALSE-cp311-cp311-linux_x86_64.whl
Python 3.12 (cp312)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.7/cu12/abiFALSE/cp312/flash_attn-2.8.3+cu12torch2.7cxx11abiFALSE-cp312-cp312-linux_x86_64.whl
Python 3.13 (cp313)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.7/cu12/abiFALSE/cp313/flash_attn-2.8.3+cu12torch2.7cxx11abiFALSE-cp313-cp313-linux_x86_64.whl

ABI: TRUE

Python 3.9 (cp39)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.7/cu12/abiTRUE/cp39/flash_attn-2.8.3+cu12torch2.7cxx11abiTRUE-cp39-cp39-linux_x86_64.whl
Python 3.10 (cp310)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.7/cu12/abiTRUE/cp310/flash_attn-2.8.3+cu12torch2.7cxx11abiTRUE-cp310-cp310-linux_x86_64.whl
Python 3.11 (cp311)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.7/cu12/abiTRUE/cp311/flash_attn-2.8.3+cu12torch2.7cxx11abiTRUE-cp311-cp311-linux_x86_64.whl
Python 3.12 (cp312)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.7/cu12/abiTRUE/cp312/flash_attn-2.8.3+cu12torch2.7cxx11abiTRUE-cp312-cp312-linux_x86_64.whl
Python 3.13 (cp313)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.7/cu12/abiTRUE/cp313/flash_attn-2.8.3+cu12torch2.7cxx11abiTRUE-cp313-cp313-linux_x86_64.whl

Torch 2.6

ABI: FALSE

Python 3.9 (cp39)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.6/cu12/abiFALSE/cp39/flash_attn-2.8.3+cu12torch2.6cxx11abiFALSE-cp39-cp39-linux_x86_64.whl
Python 3.10 (cp310)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.6/cu12/abiFALSE/cp310/flash_attn-2.8.3+cu12torch2.6cxx11abiFALSE-cp310-cp310-linux_x86_64.whl
Python 3.11 (cp311)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.6/cu12/abiFALSE/cp311/flash_attn-2.8.3+cu12torch2.6cxx11abiFALSE-cp311-cp311-linux_x86_64.whl
Python 3.12 (cp312)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.6/cu12/abiFALSE/cp312/flash_attn-2.8.3+cu12torch2.6cxx11abiFALSE-cp312-cp312-linux_x86_64.whl
Python 3.13 (cp313)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.6/cu12/abiFALSE/cp313/flash_attn-2.8.3+cu12torch2.6cxx11abiFALSE-cp313-cp313-linux_x86_64.whl

ABI: TRUE

Python 3.9 (cp39)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.6/cu12/abiTRUE/cp39/flash_attn-2.8.3+cu12torch2.6cxx11abiTRUE-cp39-cp39-linux_x86_64.whl
Python 3.10 (cp310)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.6/cu12/abiTRUE/cp310/flash_attn-2.8.3+cu12torch2.6cxx11abiTRUE-cp310-cp310-linux_x86_64.whl
Python 3.11 (cp311)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.6/cu12/abiTRUE/cp311/flash_attn-2.8.3+cu12torch2.6cxx11abiTRUE-cp311-cp311-linux_x86_64.whl
Python 3.12 (cp312)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.6/cu12/abiTRUE/cp312/flash_attn-2.8.3+cu12torch2.6cxx11abiTRUE-cp312-cp312-linux_x86_64.whl
Python 3.13 (cp313)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.6/cu12/abiTRUE/cp313/flash_attn-2.8.3+cu12torch2.6cxx11abiTRUE-cp313-cp313-linux_x86_64.whl

Torch 2.5

ABI: FALSE

Python 3.9 (cp39)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.5/cu12/abiFALSE/cp39/flash_attn-2.8.3+cu12torch2.5cxx11abiFALSE-cp39-cp39-linux_x86_64.whl
Python 3.10 (cp310)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.5/cu12/abiFALSE/cp310/flash_attn-2.8.3+cu12torch2.5cxx11abiFALSE-cp310-cp310-linux_x86_64.whl
Python 3.11 (cp311)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.5/cu12/abiFALSE/cp311/flash_attn-2.8.3+cu12torch2.5cxx11abiFALSE-cp311-cp311-linux_x86_64.whl
Python 3.12 (cp312)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.5/cu12/abiFALSE/cp312/flash_attn-2.8.3+cu12torch2.5cxx11abiFALSE-cp312-cp312-linux_x86_64.whl
Python 3.13 (cp313)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.5/cu12/abiFALSE/cp313/flash_attn-2.8.3+cu12torch2.5cxx11abiFALSE-cp313-cp313-linux_x86_64.whl

ABI: TRUE

Python 3.9 (cp39)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.5/cu12/abiTRUE/cp39/flash_attn-2.8.3+cu12torch2.5cxx11abiTRUE-cp39-cp39-linux_x86_64.whl
Python 3.10 (cp310)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.5/cu12/abiTRUE/cp310/flash_attn-2.8.3+cu12torch2.5cxx11abiTRUE-cp310-cp310-linux_x86_64.whl
Python 3.11 (cp311)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.5/cu12/abiTRUE/cp311/flash_attn-2.8.3+cu12torch2.5cxx11abiTRUE-cp311-cp311-linux_x86_64.whl
Python 3.12 (cp312)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.5/cu12/abiTRUE/cp312/flash_attn-2.8.3+cu12torch2.5cxx11abiTRUE-cp312-cp312-linux_x86_64.whl
Python 3.13 (cp313)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.5/cu12/abiTRUE/cp313/flash_attn-2.8.3+cu12torch2.5cxx11abiTRUE-cp313-cp313-linux_x86_64.whl

Torch 2.4

ABI: FALSE

Python 3.9 (cp39)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.4/cu12/abiFALSE/cp39/flash_attn-2.8.3+cu12torch2.4cxx11abiFALSE-cp39-cp39-linux_x86_64.whl
Python 3.10 (cp310)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.4/cu12/abiFALSE/cp310/flash_attn-2.8.3+cu12torch2.4cxx11abiFALSE-cp310-cp310-linux_x86_64.whl
Python 3.11 (cp311)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.4/cu12/abiFALSE/cp311/flash_attn-2.8.3+cu12torch2.4cxx11abiFALSE-cp311-cp311-linux_x86_64.whl
Python 3.12 (cp312)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.4/cu12/abiFALSE/cp312/flash_attn-2.8.3+cu12torch2.4cxx11abiFALSE-cp312-cp312-linux_x86_64.whl

ABI: TRUE

Python 3.9 (cp39)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.4/cu12/abiTRUE/cp39/flash_attn-2.8.3+cu12torch2.4cxx11abiTRUE-cp39-cp39-linux_x86_64.whl
Python 3.10 (cp310)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.4/cu12/abiTRUE/cp310/flash_attn-2.8.3+cu12torch2.4cxx11abiTRUE-cp310-cp310-linux_x86_64.whl
Python 3.11 (cp311)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.4/cu12/abiTRUE/cp311/flash_attn-2.8.3+cu12torch2.4cxx11abiTRUE-cp311-cp311-linux_x86_64.whl
Python 3.12 (cp312)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_x86_64/torch2.4/cu12/abiTRUE/cp312/flash_attn-2.8.3+cu12torch2.4cxx11abiTRUE-cp312-cp312-linux_x86_64.whl

Linux aarch64 (ARM)

Torch 2.9 (Nightly/Pre-release)

ABI: TRUE

Python 3.12 (cp312)
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-2.8.3/linux_aarch64/torch2.9/cu12/abiTRUE/cp312/flash_attn-2.8.3+cu12torch2.9cxx11abiTRUE-cp312-cp312-linux_aarch64.whl

Acknowledgements and Note:

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Space using strangertoolshf/flash_attention_2_wheelhouse 1

Paper for strangertoolshf/flash_attention_2_wheelhouse