-
-
Notifications
You must be signed in to change notification settings - Fork 827
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Running Example on Free T4 GPU through Google Colab #1905
Labels
bug
Something isn't working
Comments
I can confirm this. Here is the full report with the error message as well:
|
can confirm this. @winglian could you kindly take a look and update the colab? thanks |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Please check that this issue hasn't been reported before.
Expected Behavior
The setup or first 2 cells in the notebook can install the environment. This set of installation code should run all subsequent code to successfully recreate the training and inference steps of LLM's through Axolotl. https://github.com/axolotl-ai-cloud/axolotl/blob/main/examples/colab-notebooks/colab-axolotl-example.ipynb
Current behaviour
The first cell is successful,
The second cell never completes and even if we try to change some of the code, the
accelerate
command doesn't work. It hangs at flash attention.Here's the tail of the installation output from the cell,
Steps to reproduce
axolotl-ai-cloud
for organization and selectaxolotl
for repositoryConfig yaml
Possible solution
This might be because the T4 GPU is not supported by this library. Here's the documentation about which GPU's are supported, https://github.com/Dao-AILab/flash-attention?tab=readme-ov-file#nvidia-cuda-support
I used the instructions in the README instead to install and train TinyLLAMA using Axolotl instead. I also changed the config to set the flash attention layer to false. This still installs the flash-attn somehow but we get
RuntimeError: FlashAttention only supports Ampere GPUs or newer
error at runtime instead.This cell will install it,
Then in the config.yaml,
This does run on the free T4 GPU but still takes hours to finish and may need a different config.
Which Operating Systems are you using?
Python Version
3.10
axolotl branch-commit
main
Acknowledgements
The text was updated successfully, but these errors were encountered: