Casual-Autopsy commited on
Commit
9cbb2ac
1 Parent(s): 528961f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +102 -27
README.md CHANGED
@@ -3,45 +3,120 @@ tags:
3
  - merge
4
  - mergekit
5
  - lazymergekit
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  ---
7
 
8
- # Umbra-Test011
9
 
10
- Umbra-Test011 is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
11
 
12
- ## 🧩 Configuration
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
 
14
  ```yaml
15
- models:
16
- - model: Casual-Autopsy/Umbra-Test011-1
17
- - model: Casual-Autopsy/Umbra-Test011-2
18
- - model: Casual-Autopsy/Umbra-Test11-3
19
- merge_method: model_stock
20
- base_model: Casual-Autopsy/Umbra-Test011-1
 
 
 
 
21
  dtype: bfloat16
 
22
  ```
23
 
24
- ## 💻 Usage
25
 
26
- ```python
27
- !pip install -qU transformers accelerate
 
 
 
 
 
 
 
 
 
 
 
28
 
29
- from transformers import AutoTokenizer
30
- import transformers
31
- import torch
32
 
33
- model = "Casual-Autopsy/Umbra-Test011"
34
- messages = [{"role": "user", "content": "What is a large language model?"}]
 
 
 
 
 
 
 
 
 
 
 
35
 
36
- tokenizer = AutoTokenizer.from_pretrained(model)
37
- prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
38
- pipeline = transformers.pipeline(
39
- "text-generation",
40
- model=model,
41
- torch_dtype=torch.float16,
42
- device_map="auto",
43
- )
44
 
45
- outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
46
- print(outputs[0]["generated_text"])
 
 
 
 
 
 
47
  ```
 
3
  - merge
4
  - mergekit
5
  - lazymergekit
6
+ - not-for-all-audiences
7
+ - rp
8
+ - nsfw
9
+ license: llama3
10
+ language:
11
+ - en
12
+ library_name: transformers
13
+ pipeline_tag: text-generation
14
+ base_model:
15
+ - Sao10K/L3-8B-Stheno-v3.2
16
+ - Hastagaras/Jamet-8B-L3-MK.V-Blackroot
17
+ - grimjim/Llama-3-Oasis-v1-OAS-8B
18
+ - failspy/Llama-3-8B-Instruct-MopeyMule
19
  ---
20
 
21
+ # merge
22
 
23
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
24
 
25
+ ## Merge Details
26
+
27
+ The goal of this merge was to make an RP model better suited for role-plays with heavy themes such as but not limited to:
28
+ - Mental illness
29
+ - Self-harm
30
+ - Trauma
31
+ - Suicide
32
+
33
+ I hated how RP models tended to be overly positive and hopeful with role-plays involving such themes,
34
+ but thanks to [failspy/Llama-3-8B-Instruct-MopeyMule](https://huggingface.co/failspy/Llama-3-8B-Instruct-MopeyMule) this problem has been lessened considerably.
35
+
36
+ I haven't done extensive testing yet, but the bit of testing I did do has shown noticeably good results in lessening how positive the bot is in its writing style.
37
+
38
+ If you're an enjoyer of savior/reverse savior type role-plays like myself, then this bot is for you.
39
+
40
+ ### Merge Method
41
+
42
+ This model was merged using several Task Arithmetic merges and then tied together with a Model Stock merge.
43
+
44
+ ### Models Merged
45
+
46
+ The following models were included in the merge:
47
+ * Casual-Autopsy/Umbral-1
48
+ * [Sao10K/L3-8B-Stheno-v3.2](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2)
49
+ * [failspy/Llama-3-8B-Instruct-MopeyMule](https://huggingface.co/failspy/Llama-3-8B-Instruct-MopeyMule)
50
+
51
+ * Casual-Autopsy/Umbral-2
52
+ * [Hastagaras/Jamet-8B-L3-MK.V-Blackroot](https://huggingface.co/Hastagaras/Jamet-8B-L3-MK.V-Blackroot)
53
+ * [failspy/Llama-3-8B-Instruct-MopeyMule](https://huggingface.co/failspy/Llama-3-8B-Instruct-MopeyMule)
54
+
55
+ * Casual-Autopsy/Umbral-3
56
+ * [grimjim/Llama-3-Oasis-v1-OAS-8B](https://huggingface.co/grimjim/Llama-3-Oasis-v1-OAS-8B)
57
+ * [failspy/Llama-3-8B-Instruct-MopeyMule](https://huggingface.co/failspy/Llama-3-8B-Instruct-MopeyMule)
58
+
59
+ ## Secret Sauce
60
+
61
+ The following YAML configurations were used to produce this model:
62
+
63
+ ### Umbral-1
64
 
65
  ```yaml
66
+ slices:
67
+ - sources:
68
+ - model: Sao10K/L3-8B-Stheno-v3.2
69
+ layer_range: [0, 32]
70
+ - model: failspy/Llama-3-8B-Instruct-MopeyMule
71
+ layer_range: [0, 32]
72
+ parameters:
73
+ weight: 0.25
74
+ merge_method: task_arithmetic
75
+ base_model: Sao10K/L3-8B-Stheno-v3.2
76
  dtype: bfloat16
77
+
78
  ```
79
 
80
+ ### Umbral-2
81
 
82
+ ```yaml
83
+ slices:
84
+ - sources:
85
+ - model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot
86
+ layer_range: [0, 32]
87
+ - model: failspy/Llama-3-8B-Instruct-MopeyMule
88
+ layer_range: [0, 32]
89
+ parameters:
90
+ weight: 0.15
91
+ merge_method: task_arithmetic
92
+ base_model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot
93
+ dtype: bfloat16
94
+ ```
95
 
96
+ ### Umbral-3
 
 
97
 
98
+ ```yaml
99
+ slices:
100
+ - sources:
101
+ - model: grimjim/Llama-3-Oasis-v1-OAS-8B
102
+ layer_range: [0, 32]
103
+ - model: failspy/Llama-3-8B-Instruct-MopeyMule
104
+ layer_range: [0, 32]
105
+ parameters:
106
+ weight: 0.35
107
+ merge_method: task_arithmetic
108
+ base_model: grimjim/Llama-3-Oasis-v1-OAS-8B
109
+ dtype: bfloat16
110
+ ```
111
 
112
+ ### Umbral-Mind-RP-8B
 
 
 
 
 
 
 
113
 
114
+ ```yaml
115
+ models:
116
+ - model: Casual-Autopsy/Umbral-1
117
+ - model: Casual-Autopsy/Umbral-2
118
+ - model: Casual-Autopsy/Umbral-3
119
+ merge_method: model_stock
120
+ base_model: Casual-Autopsy/Umbral-1
121
+ dtype: bfloat16
122
  ```