{"id":23701295,"url":"https://github.com/gindachen/flexflashattention3","last_synced_at":"2025-07-01T09:05:36.008Z","repository":{"id":258459422,"uuid":"862623747","full_name":"GindaChen/FlexFlashAttention3","owner":"GindaChen","description":"FlexAttention w/ FlashAttention3 Support","archived":false,"fork":false,"pushed_at":"2024-10-05T01:59:56.000Z","size":27,"stargazers_count":26,"open_issues_count":0,"forks_count":2,"subscribers_count":1,"default_branch":"main","last_synced_at":"2025-07-01T09:04:46.842Z","etag":null,"topics":[],"latest_commit_sha":null,"homepage":null,"language":"Python","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":null,"status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/GindaChen.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":null,"funding":null,"license":null,"code_of_conduct":null,"threat_model":null,"audit":null,"citation":null,"codeowners":null,"security":null,"support":null,"governance":null,"roadmap":null,"authors":null,"dei":null,"publiccode":null,"codemeta":null}},"created_at":"2024-09-24T23:08:58.000Z","updated_at":"2025-02-01T13:51:32.000Z","dependencies_parsed_at":"2024-10-19T05:16:01.345Z","dependency_job_id":"2a3d636d-e053-4ef3-8bc6-87cd6f4cbc98","html_url":"https://github.com/GindaChen/FlexFlashAttention3","commit_stats":null,"previous_names":["gindachen/flexflashattention3"],"tags_count":0,"template":false,"template_full_name":null,"purl":"pkg:github/GindaChen/FlexFlashAttention3","repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/GindaChen%2FFlexFlashAttention3","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/GindaChen%2FFlexFlashAttention3/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/GindaChen%2FFlexFlashAttention3/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/GindaChen%2FFlexFlashAttention3/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/GindaChen","download_url":"https://codeload.github.com/GindaChen/FlexFlashAttention3/tar.gz/refs/heads/main","sbom_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/GindaChen%2FFlexFlashAttention3/sbom","host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":262933261,"owners_count":23386778,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2022-07-04T15:15:14.044Z","host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"keywords":[],"created_at":"2024-12-30T09:36:29.833Z","updated_at":"2025-07-01T09:05:35.977Z","avatar_url":"https://github.com/GindaChen.png","language":"Python","readme":"⚠️WIP: We have not put the flexattention/FA3 logic here yet! Coming soon...\n\n\nSupport FlexAttention in FlashAttention3\n\n\n## Build environment\n\nThese instructions currently only tested in CUDA 12.4. We eventually need CUDA 12.3 for the build, and this is still under dev.\n\n\nInstall mamba\n```bash\n# https://mamba.readthedocs.io/en/latest/installation/micromamba-installation.html\n\"${SHELL}\" \u003c(curl -L micro.mamba.pm/install.sh)\n```\n\nBuild environment to benchmark FlashAttention3 with nightly torch\n```bash\nmkdir -p envs\nmamba create -p ~/envs/flashattn_3 python=3.10 -y\nln -s ~/envs/flashattn_3 ./envs/flashattn_3\nmamba activate ./envs/flashattn_3\nmamba install pytorch torchvision torchaudio pytorch-cuda=12.4 -c pytorch-nightly -c nvidia -y\nmamba install ipython nvitop ninja cmake -y\npip install packaging\n# Install FlashAttention3\nMAX_JOBS=80 python setup.py install\ncd hopper\nMAX_JOBS=80 python setup.py install\n```\n\n\nBuild environment to benchmark FlexAttention in nightly build torch\n```bash\nmkdir -p envs\nmamba create -p ~/envs/flexattn-nightly-built python=3.10 -y\nmamba activate ~/envs/flexattn-nightly-built\nmamba install pytorch torchvision torchaudio pytorch-cuda=12.4 -c pytorch-nightly -c nvidia -y\nmamba install ipython nvitop -y\n```\n\nBuild environment for Flex attention gym\n```bash\nmamba activate ~/envs/flexattn-nightly-built\ngit clone https://github.com/pytorch-labs/attention-gym.git\ncd attention-gym\npip install .\n```\n","funding_links":[],"categories":[],"sub_categories":[],"project_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fgindachen%2Fflexflashattention3","html_url":"https://awesome.ecosyste.ms/projects/github.com%2Fgindachen%2Fflexflashattention3","lists_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fgindachen%2Fflexflashattention3/lists"}