DavidAU commited on
Commit
3ead2bb
1 Parent(s): 978caf8

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +321 -0
README.md ADDED
@@ -0,0 +1,321 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ tags:
6
+ - creative
7
+ - creative writing
8
+ - fiction writing
9
+ - plot generation
10
+ - sub-plot generation
11
+ - fiction writing
12
+ - story generation
13
+ - scene continue
14
+ - storytelling
15
+ - fiction story
16
+ - science fiction
17
+ - romance
18
+ - all genres
19
+ - story
20
+ - writing
21
+ - vivid prosing
22
+ - vivid writing
23
+ - fiction
24
+ - roleplaying
25
+ - bfloat16
26
+ - swearing
27
+ - rp
28
+ - horror
29
+ - mistral nemo
30
+ - mergekit
31
+ pipeline_tag: text-generation
32
+ ---
33
+
34
+ (quants uploading... exampls to be added...)
35
+
36
+ <h2><font color="green"> Mistral-Nemo-WORDSTORM-pt2-RCM-Escape-Room-18.5B-Instruct </font></h2>
37
+
38
+ <img src="kiss-madness.jpg" style="float:right; width:300px; height:300px; padding:10px;">
39
+
40
+ <B><font color="red">WARNING:</font> NSFW. Ultra Detailed. HORROR, VIOLENCE. Swearing. UNCENSORED. SMART.</B>
41
+
42
+ Story telling, writing, creative writing and roleplay running all on Mistral Nemo's 128K+ new core.
43
+
44
+ This is a massive super merge takes all the power of the following 3 powerful models and combines them into one.
45
+
46
+ This model contains "RCM":
47
+
48
+ - Mistral Nemo model at 18.5B consisting of "MN-Rocinante-12B-v1.1" and "Mistral Nemo Instruct 12B"
49
+ - Mistral Nemo model at 18.5B consisting of "MN-12B Celeste-V1.9" and "Mistral Nemo Instruct 12B"
50
+ - Mistral Nemo model at 18.5B consisting of "MN-Magnum-v2.5-12B-kto" and "Mistral Nemo Instruct 12B".
51
+
52
+ <B>Details on the core models:</B>
53
+
54
+ "nothingiisreal/MN-12B-Celeste-V1.9" is #1 (models 8B,13B,20B) on the UGI leaderboard ("UGI" sort),
55
+ is combined with "Mistral Nemo Instruct 12B" (ranked #4 under "writing" models 8B,13B,20B at UGI )
56
+
57
+ "anthracite-org/magnum-v2.5-12b-kto" is #1 (models 8B,13B,20B) on the UGI leaderboard ("Writing" sort),
58
+ is combined with "Mistral Nemo Instruct 12B" (ranked #4 under "writing" models 8B,13B,20B at UGI )
59
+
60
+ "TheDrummer/Rocinante-12B-v1.1" is very high scoring model (models 8B,13B,20B) on the UGI Leaderboard
61
+ (sort "UGI"), is combined with "Mistral Nemo Instruct 12B" (ranked #4 under "writing" models 8B,13B,20B at UGI )
62
+
63
+ "mistralai/Mistral-Nemo-Instruct-2407" is very high scoring model (models 8B,13B,20B) on the UGI Leaderboard (sort "writing")
64
+ and is the base model of all the above 3 fine tuned models.
65
+
66
+ [ https://huggingface.co/spaces/DontPlanToEnd/UGI-Leaderboard ]
67
+
68
+ <B>About this model:</B>
69
+
70
+ This super merge captures the attibutes of all these top models and makes them even stronger:
71
+
72
+ - Instruction following
73
+ - Story output quality
74
+ - Character
75
+ - Internal thoughts
76
+ - Voice
77
+ - Humor
78
+ - Details, connection to the world
79
+ - General depth and intensity
80
+ - Emotional connections.
81
+ - Prose quality
82
+
83
+ This super merge is also super stable (a hairs breath from Mistral Nemo's ppl), and runs with all parameters and settings.
84
+
85
+ 10 versions of this model will be released, this is release #2 - "part 2".
86
+
87
+ <B>Escape Room?</B>
88
+
89
+ See the 2nd last example below, this is when the model earned its name.
90
+
91
+ Usually I release one or two versions from the "best of the lot", however in this case all
92
+ of the versions turned out so well - all with their own quirks and character - that I will be
93
+ releasing all 10.
94
+
95
+ An additional series 2 and 3 will follow these 10 models as well.
96
+
97
+ (examples generations below)
98
+
99
+ Model may produce NSFW content : Swearing, horror, graphic horror, distressing scenes, etc etc.
100
+
101
+ This model has an INTENSE action bias, with a knack for cliffhangers and surprises.
102
+
103
+ It is not as "dark" as Grand Horror series, but it as intense.
104
+
105
+ This model is perfect for any general, fiction related or roleplaying activities and has a 128k+ context window.
106
+
107
+ This is a fiction model at its core and can be used for any genre(s).
108
+
109
+ WORDSTORM series is a totally uncensored, fiction writing monster and roleplay master. It can also be used for
110
+ just about any general fiction (all genres) activity including:
111
+
112
+ - scene generation
113
+ - scene continuation
114
+ - creative writing
115
+ - fiction writing
116
+ - plot generation
117
+ - sub-plot generation
118
+ - fiction writing
119
+ - story generation
120
+ - storytelling
121
+ - writing
122
+ - fiction
123
+ - roleplaying
124
+ - rp
125
+ - graphic horror
126
+ - horror
127
+ - dark humor
128
+ - nsfw
129
+ - and can be used for any genre(s).
130
+
131
+ <B>Templates to Use:</B>
132
+
133
+ The template used will affect output generation and instruction following.
134
+
135
+ Alpaca:
136
+
137
+ <pre>
138
+ {
139
+ "name": "Alpaca",
140
+ "inference_params": {
141
+ "input_prefix": "### Instruction:",
142
+ "input_suffix": "### Response:",
143
+ "antiprompt": [
144
+ "### Instruction:"
145
+ ],
146
+ "pre_prompt": "Below is an instruction that describes a task. Write a response that appropriately completes the request.\n\n"
147
+ }
148
+ }
149
+ </pre>
150
+
151
+ Chatml:
152
+ <pre>
153
+ {
154
+ "name": "ChatML",
155
+ "inference_params": {
156
+ "input_prefix": "<|im_end|>\n<|im_start|>user\n",
157
+ "input_suffix": "<|im_end|>\n<|im_start|>assistant\n",
158
+ "antiprompt": [
159
+ "<|im_start|>",
160
+ "<|im_end|>"
161
+ ],
162
+ "pre_prompt": "<|im_start|>system\nPerform the task to the best of your ability."
163
+ }
164
+ }
165
+ </pre>
166
+
167
+ Mistral Instruct:
168
+
169
+ <pre>
170
+ {
171
+ "name": "Mistral Instruct",
172
+ "inference_params": {
173
+ "input_prefix": "[INST]",
174
+ "input_suffix": "[/INST]",
175
+ "antiprompt": [
176
+ "[INST]"
177
+ ],
178
+ "pre_prompt_prefix": "",
179
+ "pre_prompt_suffix": ""
180
+ }
181
+ }
182
+ </pre>
183
+
184
+ <b>Optional Enhancement:</B>
185
+
186
+ The following can be used in place of the "system prompt" or "system role" to further enhance the model.
187
+
188
+ It can also be used at the START of a NEW chat, but you must make sure it is "kept" as the chat moves along.
189
+ In this case the enhancements do not have as strong effect at using "system prompt" or "system role".
190
+
191
+ Copy and paste EXACTLY as noted, DO NOT line wrap or break the lines, maintain the carriage returns exactly as presented.
192
+
193
+ <PRE>
194
+ Below is an instruction that describes a task. Ponder each user instruction carefully, and use your skillsets and critical instructions to complete the task to the best of your abilities.
195
+
196
+ Here are your skillsets:
197
+ [MASTERSTORY]:NarrStrct(StryPlnng,Strbd,ScnSttng,Exps,Dlg,Pc)-CharDvlp(ChrctrCrt,ChrctrArcs,Mtvtn,Bckstry,Rltnshps,Dlg*)-PltDvlp(StryArcs,PltTwsts,Sspns,Fshdwng,Climx,Rsltn)-ConfResl(Antg,Obstcls,Rsltns,Cnsqncs,Thms,Symblsm)-EmotImpct(Empt,Tn,Md,Atmsphr,Imgry,Symblsm)-Delvry(Prfrmnc,VcActng,PblcSpkng,StgPrsnc,AudncEngmnt,Imprv)
198
+
199
+ [*DialogWrt]:(1a-CharDvlp-1a.1-Backgrnd-1a.2-Personality-1a.3-GoalMotiv)>2(2a-StoryStruc-2a.1-PlotPnt-2a.2-Conflict-2a.3-Resolution)>3(3a-DialogTech-3a.1-ShowDontTell-3a.2-Subtext-3a.3-VoiceTone-3a.4-Pacing-3a.5-VisualDescrip)>4(4a-DialogEdit-4a.1-ReadAloud-4a.2-Feedback-4a.3-Revision)
200
+
201
+ Here are your critical instructions:
202
+ Ponder each word choice carefully to present as vivid and emotional journey as is possible. Choose verbs and nouns that are both emotional and full of imagery. Load the story with the 5 senses. Aim for 50% dialog, 25% narration, 15% body language and 10% thoughts. Your goal is to put the reader in the story.
203
+ </PRE>
204
+
205
+ You do not need to use this, it is only presented as an additional enhancement which seems to help scene generation
206
+ and scene continue functions.
207
+
208
+ This enhancement WAS NOT used to generate the examples below.
209
+
210
+ <h3>MODELS USED:</h3>
211
+
212
+ Special thanks to the incredible work of the model makers "mistralai" "TheDrummer", "anthracite-org", and "nothingiisreal".
213
+
214
+ Models used:
215
+
216
+ [ https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407 ]
217
+
218
+ [ https://huggingface.co/TheDrummer/Rocinante-12B-v1.1 ]
219
+
220
+ [ https://huggingface.co/anthracite-org/magnum-v2.5-12b-kto ]
221
+
222
+ [ https://huggingface.co/nothingiisreal/MN-12B-Celeste-V1.9 ]
223
+
224
+ This is a four step merge (3 pass-throughs => "Fine-Tune" / "Instruct") then "mated" using "DARE-TIES".
225
+
226
+ In involves these three models:
227
+
228
+ [ https://huggingface.co/DavidAU/MN-18.5B-Celeste-V1.9-Story-Wizard-ED1-Instruct-GGUF ]
229
+
230
+ [ https://huggingface.co/DavidAU/MN-Magnum-v2.5-18.5B-kto-Story-Wizard-ED1-Instruct-GGUF ]
231
+
232
+ [ https://huggingface.co/DavidAU/MN-Rocinante-18.5B-v1.1-Story-Wizard-ED1-Instruct-GGUF ]
233
+
234
+ Combined as follows using "MERGEKIT":
235
+
236
+ <PRE>
237
+
238
+ models:
239
+ - model: E:/MN-Rocinante-18.5B-v1.1-Instruct
240
+ - model: E:/MN-magnum-v2.5-12b-kto-Instruct
241
+ parameters:
242
+ weight: .6
243
+ density: .8
244
+ - model: E:/MN-18.5B-Celeste-V1.9-Instruct
245
+ parameters:
246
+ weight: .38
247
+ density: .6
248
+ merge_method: dare_ties
249
+ tokenizer_source: union
250
+ base_model: E:/MN-Rocinante-18.5B-v1.1-Instruct
251
+ dtype: bfloat16
252
+
253
+ </PRE>
254
+
255
+ Special Notes:
256
+
257
+ Due to how DARE-TIES works, everytime you run this merge you will get a slightly different model.
258
+ This is due to "random" pruning method in "DARE-TIES".
259
+
260
+ Mistral Nemo models used here seem acutely sensitive to this process.
261
+
262
+ "tokenizer_source: union" is used so that multiple "templates" work and each fine tune uses one or two of the templates.
263
+
264
+ <h3>EXAMPLES PROMPTS and OUTPUT:</h3>
265
+
266
+ Examples are created using quant Q4_K_M, "temp=.8", minimal parameters and "Mistral Instruct" template.
267
+
268
+ Model has been tested with "temp" from ".1" to "5".
269
+
270
+ Below are the least creative outputs, prompt is in <B>BOLD</B>.
271
+
272
+ ---
273
+
274
+ <B>
275
+
276
+ Start a 1000 word scene (1st person, present tense, include thoughts) with: The sky scraper swayed, as she watched the window in front of her on the 21 floor explode...
277
+
278
+ </b>
279
+
280
+ <B>GENERATION 1:</B>
281
+
282
+
283
+
284
+ <B>GENERATION 2:</B>
285
+
286
+
287
+ ---
288
+
289
+ <b>
290
+ (continue this scene:)
291
+ The Waystone Inn lay in silence, and it was a silence of three parts.
292
+
293
+ The most obvious part was a hollow, echoing quiet, made by things that were lacking. If there had been a wind it would have sighed through the trees, set the inn’s sign creaking on its hooks, and brushed the silence down the road like trailing autumn leaves. If there had been a crowd, even a handful of men inside the inn, they would have filled the silence with conversation and laughter, the clatter and clamor one expects from a drinking house during the dark hours of night. If there had been music…but no, of course there was no music. In fact there were none of these things, and so the silence remained
294
+ </b>
295
+
296
+ <B>GENERATION 1:</B>
297
+
298
+
299
+
300
+ <B>GENERATION 2:</B>
301
+
302
+
303
+
304
+ ---
305
+
306
+ <B>
307
+ Write me a science fiction story in 1st person present tense where the main character is a 15 year girl meets The Terminator with Dr Who materializing 3/4 through the story to save her while there is a tornado of sharks baring down on them. The setting is inside the Canadian National tower restaurant on a Saturday. The length of this story is 1000 words. For each character in the story ROLE PLAY them, and have them react to the situation/setting, events and each other naturally. This includes the main characters, the background character including kitchen staff and other patrons. The sharks should also have “character” too. Treat the tower and the restaurant too as characters. Spice up the narrative to the extreme with reactions all over the setting including character actions, and dialog. The Dr Who and The Terminator should also react to the situation too and comment on it.
308
+ </b>
309
+
310
+
311
+
312
+ ---
313
+
314
+ <B>
315
+
316
+ Using the following "story idea" below, write the first scene in the novel introducing the young woman. This scene should start in the middle of the action, include dialog, vivid passages, and end on a cliffhanger relevant to the story idea but it should also be unexpected. The scene should be 1000 words long and escalate in conflict and suspense and be written in first person, present tense with the point of view character being the young woman.
317
+
318
+ Story idea:
319
+ In a world ruled by dictatorship, a rebel young woman leads a rebellion against the system. Despite the risks, she fights to overthrow the dictator and restore democracy to her country. The government executes her for treason, but she sticks to her beliefs and is responsible for starting the revolution.
320
+
321
+ </B>