From c0094ae54b7426c6007690636c1538b0b096a0e3 Mon Sep 17 00:00:00 2001 From: jfrery Date: Fri, 20 Dec 2024 14:45:29 +0100 Subject: [PATCH] chore: downgrade accelerate --- .../lora_finetuning/GPT2FineTuneHybrid.ipynb | 6 +++--- .../lora_finetuning/LLamaFineTuning.ipynb | 10 +++++++++- use_case_examples/lora_finetuning/README.md | 2 ++ use_case_examples/lora_finetuning/requirements.txt | 2 +- 4 files changed, 15 insertions(+), 5 deletions(-) diff --git a/use_case_examples/lora_finetuning/GPT2FineTuneHybrid.ipynb b/use_case_examples/lora_finetuning/GPT2FineTuneHybrid.ipynb index 5013dbb4f..b0729905c 100644 --- a/use_case_examples/lora_finetuning/GPT2FineTuneHybrid.ipynb +++ b/use_case_examples/lora_finetuning/GPT2FineTuneHybrid.ipynb @@ -19,8 +19,6 @@ "source": [ "# Import necessary libraries\n", "import math\n", - "import shutil\n", - "from pathlib import Path\n", "\n", "import matplotlib.pyplot as plt\n", "import torch\n", @@ -261,7 +259,6 @@ " learning_rate=2e-3,\n", " lr_scheduler_type=\"linear\",\n", " seed=SEED,\n", - " data_seed=SEED,\n", " warmup_steps=10,\n", " weight_decay=0.01,\n", " prediction_loss_only=True,\n", @@ -1299,6 +1296,9 @@ "metadata": { "execution": { "timeout": 10800 + }, + "language_info": { + "name": "python" } }, "nbformat": 4, diff --git a/use_case_examples/lora_finetuning/LLamaFineTuning.ipynb b/use_case_examples/lora_finetuning/LLamaFineTuning.ipynb index 2bf12280a..e86cd43a2 100644 --- a/use_case_examples/lora_finetuning/LLamaFineTuning.ipynb +++ b/use_case_examples/lora_finetuning/LLamaFineTuning.ipynb @@ -175,7 +175,6 @@ " learning_rate=2e-4,\n", " lr_scheduler_type=\"linear\",\n", " seed=SEED,\n", - " data_seed=SEED,\n", " warmup_steps=10,\n", " weight_decay=0.01,\n", " prediction_loss_only=True,\n", @@ -351,6 +350,15 @@ "metadata": { "execution": { "timeout": 10800 + }, + "kernelspec": { + "display_name": ".venv", + "language": "python", + "name": "python3" + }, + "language_info": { + "name": "python", + "version": "3.10.11" } }, "nbformat": 4, diff --git a/use_case_examples/lora_finetuning/README.md b/use_case_examples/lora_finetuning/README.md index a1513298f..a4b0872ef 100644 --- a/use_case_examples/lora_finetuning/README.md +++ b/use_case_examples/lora_finetuning/README.md @@ -19,6 +19,8 @@ Fine-tuning large language models typically requires access to sensitive data, w ### Installation +This project requires Python 3.9 or higher. + Install the required packages: diff --git a/use_case_examples/lora_finetuning/requirements.txt b/use_case_examples/lora_finetuning/requirements.txt index da6495fef..86d35cb02 100644 --- a/use_case_examples/lora_finetuning/requirements.txt +++ b/use_case_examples/lora_finetuning/requirements.txt @@ -4,6 +4,6 @@ peft==0.12.0 Jinja2==3.1.4 matplotlib==3.7.5 datasets==3.1.0 -accelerate==1.2.0 +accelerate==1.0.1 jupyter==1.1.1 tqdm==4.67.1 \ No newline at end of file