You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We finetune [Mistral-7B](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) using [LoRA](https://arxiv.org/abs/2106.09685) and [DeepSpeed](https://github.com/microsoft/DeepSpeed). We ran LoRA on two 40 GB A100 GPUs utilizing DeepSpeed.
4
+
5
+
To get started, first install Determined on your local machine:
6
+
```bash
7
+
pip install determined
8
+
```
9
+
10
+
Then finetune with LoRA:
11
+
```bash
12
+
det e create lora.yaml .
13
+
```
14
+
15
+
You can view the actual training code in `finetune.py`.
16
+
17
+
18
+
## Configuration
19
+
20
+
Change configuration options in `lora.yaml`. Some important options are:
21
+
-`slots_per_trial`: the number of GPUs to use.
22
+
-`dataset_subset`: the difficulty subset to train on.
23
+
-`per_device_train_batch_size`: the batch size per GPU.
24
+
25
+
26
+
DeepSpeed configuration files are in the `ds_configs` folder.
0 commit comments