It's easy to finetue Llama 3 on the question & answer dataset we just created.
./scripts/train.sh
The python is as follows
lamini-examples/06_ift/train.py
Lines 1 to 48 in 03f35d3
After you submit a training job, it is scheduled on the cluster. You can monitor the progress of the job by visiting the link provided in the output of the training script.
python3 /app/lamini-earnings-sdk/06_fine_tuning/scripts/../train.py
Uploading data....
Upload to blob completed for data.
Data pairs uploaded to blob.
Your dataset id is: 0713f8cfa5746a0897079e7f249a048deb653cf7e849d6fc26f3d2dacc5722d0 . Consider using this in the future to train using the same data.
Eg: llm.train(dataset_id='0713f8cfa5746a0897079e7f249a048deb653cf7e849d6fc26f3d2dacc5722d0')
Training job submitted! Check status of job 6367 here: https://app.lamini.ai/train/6367
The page lets you monitor all of your jobs, view eval results, view loss curves, and logs. Jobs will automatically use all of the GPUs in your Lamini cluster.