emessy commited on
Commit
c0d4fa2
1 Parent(s): 6afc8ab

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -0
README.md CHANGED
@@ -10,6 +10,8 @@ tags:
10
  - llama
11
  - trl
12
  - sft
 
 
13
  ---
14
 
15
  # Uploaded model
@@ -21,3 +23,26 @@ tags:
21
  This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
22
 
23
  [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  - llama
11
  - trl
12
  - sft
13
+ datasets:
14
+ - emessy/flash_fiction_1
15
  ---
16
 
17
  # Uploaded model
 
23
  This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
24
 
25
  [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
26
+
27
+ # Configure LoRA
28
+ lora_config = LoraConfig(
29
+ r=16,
30
+ lora_alpha=16,
31
+ target_modules=["q_proj", "k_proj", "v_proj", "o_proj"],
32
+ lora_dropout=0.05,
33
+ bias="none",
34
+ task_type="CAUSAL_LM"
35
+ )
36
+
37
+ # Training arguments
38
+ training_args = TrainingArguments(
39
+ output_dir="./results",
40
+ num_train_epochs=5,
41
+ per_device_train_batch_size=4,
42
+ gradient_accumulation_steps=4,
43
+ learning_rate=2e-4,
44
+ fp16=True, # Use half-precision
45
+ logging_steps=10,
46
+ save_steps=50,
47
+ eval_steps=50,
48
+ )