* Flash Attention, an efficient attention module which significantly speeds up training, only works on Ampere GPUs [1]
* Even if I bought a 3090, I would have to get a computer to go with it, along with a PSU and some cooling. Don't know where to start with that.
[1] https://github.com/Dao-AILab/flash-attention/issues/190