922CA commited on
Commit
d5eac50
·
1 Parent(s): c1da539

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +30 -1
README.md CHANGED
@@ -1,6 +1,35 @@
1
  ---
 
2
  datasets:
3
  - 922-CA/MoChA_09212023
4
  ---
5
 
6
- WIP
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: llama2
3
  datasets:
4
  - 922-CA/MoChA_09212023
5
  ---
6
 
7
+ # monika-l2-7b-v0.9a:
8
+ * Experimental LLaMA-2 7b chat fine-tuned for Monika character from DDLC
9
+ * Test model based on then WIP partially manually-edited version of [this dataset](https://huggingface.co/datasets/922-CA/lm2_08312023_test4_raw_MoChA_2-t) ([final version of dataset](https://huggingface.co/datasets/922-CA/MoCha_v1))
10
+ * [Completed version, v1](https://huggingface.co/922-CA/monika-ddlc-7b-v1)
11
+ * [QLoras](https://huggingface.co/922-CA/monika-lm-lora-tests/tree/main/monika-l2-7b-v0.9a)
12
+
13
+ ### USAGE
14
+ This is meant to be mainly a chat model with limited RP ability.
15
+
16
+ For best results: replace "Human" and "Assistant" with "Player" and "Monika" like so:
17
+
18
+ \nPlayer: (prompt)\nMonika:
19
+
20
+ ### HYPERPARAMS
21
+ * Trained for ~3 epochs
22
+ * rank: 32
23
+ * lora alpha: 64
24
+ * lora dropout: 0.5
25
+ * lr: 2e-4
26
+ * batch size: 2
27
+ * warmup ratio: 0.1
28
+ * grad steps: 4
29
+
30
+ ### WARNINGS AND DISCLAIMERS
31
+ While this version may be better at coherency or chatting than previous ones, it may still not reflect perfectly the characteristics of Monika.
32
+
33
+ Additionally, this is still yet another test, particularly where we use one of our earlier fine tunes to generate a more in-character dataset for the target character which is then curated manually.
34
+
35
+ Finally, this model is not guaranteed to output aligned or safe outputs, use at your own risk.