(fork https://github.com/ROCm/flash-attention). Contribute to LunNova/flash-attention development by creating an account on GitHub.
recursive-include flash_attn *.cpp Makefile +9Lines changed: 9 additions & 0 deletions Original file line numberDiff line numberDiff line change @@ -0,0 +1,9 @@ 1+ 2+ clean_dist: 3+ rm -rf dist/* 4+ 5+ create_dist: clean_dist ...
recursive-include flash_attn *.cpp 9 changes: 9 additions & 0 deletions 9 Makefile Original file line numberDiff line numberDiff line change @@ -0,0 +1,9 @@ clean_dist: rm -rf dist/* create_dist: clean_dist python setup.py sdist upload_package: create_dist twine upload dist/* 4...
recursive-include flash_attn *.h recursive-include flash_attn *.cpp 9 changes: 9 additions & 0 deletions 9 Makefile Original file line numberDiff line numberDiff line change @@ -0,0 +1,9 @@ clean_dist: rm -rf dist/* create_dist: clean_dist python setup.py sdist upload_package: ...
7 + recursive-include flash_attn *.cpp Makefile +9 Original file line numberDiff line numberDiff line change @@ -0,0 +1,9 @@ 1 + 2 + clean_dist: 3 + rm -rf dist/* 4 + 5 + create_dist: clean_dist 6 + python setup.py sdist 7 + 8 + upload_package: cre...