fearlessdots commited on
Commit
cc0099a
1 Parent(s): 45e55ce

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +56 -3
README.md CHANGED
@@ -1,3 +1,56 @@
1
- ---
2
- license: llama3
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: llama3
3
+ datasets:
4
+ - NobodyExistsOnTheInternet/ToxicQAFinal
5
+ ---
6
+
7
+ # Llama-3-Alpha-Centauri-experiment-1-LoRA
8
+
9
+ ---
10
+
11
+ ## Disclaimer
12
+
13
+ **Note:** All models and LoRAs from the **Centaurus** series were created with the sole purpose of research. The usage of this model and/or its related LoRA implies agreement with the following terms:
14
+
15
+ - The user is responsible for what they might do with it, including how the output of the model is interpreted and used;
16
+ - The user should not use the model and its outputs for any illegal purposes;
17
+ - The user is the only one resposible for any misuse or negative consequences from using this model and/or its related LoRA.
18
+
19
+ I do not endorse any particular perspectives presented in the training data.
20
+
21
+ ---
22
+
23
+ ## Base
24
+
25
+ This model and its related LoRA was fine-tuned on [https://huggingface.co/failspy/Meta-Llama-3-8B-Instruct-abliterated-v3](https://huggingface.co/failspy/Meta-Llama-3-8B-Instruct-abliterated-v3).
26
+
27
+ ## Datasets
28
+
29
+ - [https://huggingface.co/datasets/NobodyExistsOnTheInternet/ToxicQAFinal](https://huggingface.co/datasets/NobodyExistsOnTheInternet/ToxicQAFinal)
30
+
31
+ ## Fine Tuning
32
+
33
+ ### - PEFT Parameters
34
+
35
+ - lora_alpha=64,
36
+ - lora_dropout=0.05,
37
+ - r=128,
38
+ - bias="none",
39
+
40
+ ### - Training Arguments
41
+
42
+ - num_train_epochs=1,
43
+ - per_device_train_batch_size=1,
44
+ - gradient_accumulation_steps=4,
45
+ - optim="adamw_bnb_8bit",
46
+ - save_steps=25,
47
+ - logging_steps=25,
48
+ - learning_rate=2e-4,
49
+ - weight_decay=0.001,
50
+ - fp16=False,
51
+ - bf16=False,
52
+ - max_grad_norm=0.3,
53
+ - max_steps=-1,
54
+ - warmup_ratio=0.03,
55
+ - group_by_length=True,
56
+ - lr_scheduler_type="constant",