Install Axoltol on Ubuntu 24 with Nvidia Cuda 12.

Warning: Turing cards are not supported on Ubuntu 24 due to compatibility with flash-attn 1.x. flash-attn 2.x plans to support Turing cards in the future but for now are Ampere and up.

pip3 install flash-attn==1.0.9 <– Latest 1.x release

Install for Turing cards on Ubuntu 22

Configuration documentation for .yaml files:

https://axolotl-ai-cloud.github.io/axolotl/docs/config.html –

https://modal.com/docs/examples/llm-finetuning

HTTP configuration app for .yaml files:

https://axolotl-ui.vercel.app

Note: While training I had to modify a few configs as the 4096 sequence_len was much to large to train the model timely. This does has an impact on the context length and its response so this just be too large of a job for dual RTX 8000 cards. I moved the sequence length to 2048 and was able to up the micro batch reate from 2->6 which cut time down greatly. I also updated a few other params including the use of fp16 (Turing cards support it).

The amount of ram used is about 27 Gigs and range from 40 – 90 Gigs of VRAM during training.

Leave a comment

Your email address will not be published. Required fields are marked *