Skip to content

Conversation

@DanFu09
Copy link

@DanFu09 DanFu09 commented Nov 10, 2022

Integration of FlashAttention into OPT fine-tuning, starting with causal self-attention only.

To test this, install FlashAttention (requires CUDA 11, NVCC, and a Turing or Ampere GPU):

git clone https://github.com/HazyResearch/flash-attention.git
cd flash-attention
python setup.py install

@DanFu09 DanFu09 changed the title Integrate FlashAttention into OPT fine-tuning [WIP, do not merge] Integrate FlashAttention into OPT fine-tuning Nov 10, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant