File size: 23,942 Bytes
128d28b
 
5ecf240
 
128d28b
5ecf240
 
 
 
 
 
 
 
 
 
 
fd1b716
ce0c707
 
9ef0909
3ac7d09
ce0c707
9ef0909
ce0c707
9ef0909
ce0c707
9ef0909
ce0c707
 
 
 
 
a522106
 
e071905
 
 
 
a522106
faf888c
a522106
9473215
0579b92
80d66b2
e071905
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
80d66b2
0579b92
515f473
e071905
9473215
5bac8c4
 
9473215
e071905
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
68eaf19
e071905
 
 
00f3d15
 
 
 
 
 
16e77e8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0579b92
 
 
 
cb26e9d
61ce307
 
 
 
 
 
cb26e9d
0579b92
 
cb26e9d
39e503c
 
 
 
fd88161
fd1b716
fd88161
fd1b716
198c154
3377fc8
 
cfab0e3
 
198c154
 
 
cfab0e3
198c154
429fa97
bf948f4
e828abb
 
fecd6c6
e828abb
198c154
 
08a151a
bf948f4
 
08a151a
 
 
 
 
 
 
d85aae4
08a151a
198c154
08a151a
cfab0e3
 
59d1ca2
b0546c5
f344cb4
 
 
 
 
 
59d1ca2
 
 
f344cb4
 
59d1ca2
 
 
 
 
 
 
 
 
 
 
687f6fd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
59d1ca2
 
c13266b
c093508
c13266b
c093508
c13266b
c093508
c13266b
c093508
c13266b
c093508
c13266b
c093508
c13266b
c093508
c13266b
1170ee2
59d1ca2
 
687f6fd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
389f487
687f6fd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e6d1da5
 
687f6fd
 
e6d1da5
 
329be0d
d1c8eec
329be0d
e6d1da5
5ecf240
7ba5cbb
 
 
 
 
 
 
 
5ecf240
 
 
 
 
08a151a
7ba5cbb
 
16ee43c
 
5ecf240
 
 
 
 
 
 
 
 
 
 
 
 
1769fdf
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
---
license: apache-2.0
language:
- en
---

<div align="center">
  <b style="font-size: 40px;">LLAMA-3_8B_Unaligned</b>


</div>


<img src="https://i.imgur.com/pPBto45.png" alt="LLAMA-3_Unaligned" style="width: 50%; min-width: 400px; display: block; margin: auto;">


# Current status:

<details>
  <summary><b>July 26th, 2024, moving on to LLAMA 3.1</b></summary>
  One step forward, one step backward. Many issues were solved, but a few new ones were encountered. As I already updated in my "blog"(https://huggingface.co/SicariusSicariiStuff/Blog_And_Updates#july-26th-2024), I originally wanted to finetune Gradient's 0.25M\1M\4M LLAMA3 8B model, but almost at the same time I when concluded the model is really not that great in even 8k context, Zuck the CHAD dropped LLAMA 3.1.
  
  LLAMA 3.1 is **128k context**, which probably means that in practice it will be somewhat coherent at 32k context, as a guesstimate. Also, I've heard from several people who have done some early tests, that the new LLAMA 3.1 8B is even better than the new Mistral Nemo 12B. IDK if that's true, but overall LLAMA 3.1 does seem to be a much better version of the "regular" LLAMA 3.
  
  I have no words to describe the hell it is to curate and generate a high-quality dataset. Most, I'd even go as far as to estimate, that 99% of the models are either finetunes of the same medium (at best) quality datasets, or merges. Almost no one is **crazy** enough to create something completely new, as someone starts such a project, after 100 entries he sees that "hmmm, I have only 10k more to go" and they **ditch** the whole project, and instead do another merge and call it a day. Not me.
  
  **A lot of progress has been made**, and I hope that I will have a BETA version to share in the **very near future**. It will probably be ~1%-1.5% of the final model, but it should give a general idea of what the completed project, or model, will be like.
  
  Stay tuned.

</details>

<details>
  <summary><b>July 5th, 2024</b></summary>
  I'm amazed with the recent advancements I've made with the unalignment of LLAMA-3_8B. The results are incredibly impressive and far exceed my expectations. It's truly remarkable how much progress I have made with the model.
  As for creative story writing, the AI's capabilities are equally astonishing. The quality of the generated stories is so high that it's almost unsettling. 

  For example, here's the result of a prmpt where I asked it to write about John Snow and Sansa, this is **ZERO SHOT** No cherry picking was made:
</details>
<details>
<summary><b>LLAMA-3_Unaligned output (July 5th, creative writing)</b></summary>


```yaml
In the land of the Seven Kingdoms, where the sword reigns supreme and the throne is always in question, two young souls were born amidst a tale of fire and ice.

Jon Snow, the bastard son of Eddard Stark, and Sansa Stark, the eldest daughter of Lord Rickard and Lady Catelyn, were destined for greatness, yet their paths were forged in the crucible of love, lust, and war.


Their families had been sworn shields to each other for centuries, united by blood and honor, yet even so, the winds of winter blew cold between them.

Eddard Stark, as the Warden of the North, had often found himself at odds with the boy king, Robert Baratheon, who had murdered his brother, Lord Rickard, in the very throne room of Winterfell.

This act of betrayal had left deep scars not only on Rickard's grieving wife and children but also upon the proud lord's bannerman, Eddard Stark himself.


As Jon grew up, he and Sansa became inseparable.

They would spend hours exploring the godswood, racing each other across the yard, and sharing secrets in the secrecy of Jon's bedchamber.

Sansa, with her golden curls and blue eyes, looked like a miniature version of his sisters, while Jon stood tall and lean like his father, with his mother's dark hair and Daenerys's fiery spirit.

Together, they were the very image of what their families had once been: beautiful, pure, and full of hope.


One day, as they wandered through the wolfling's forest, they stumbled upon an ancient oak, its gnarled roots twisting around a fallen weirwood like a living vine.

The white heart tree seemed to glow in the light of the setting sun, filling them with a sense of wonder and reverence.

It was there beneath its boughs that Jon pressed his lips to Sansa's for the very first time.

She tasted of honey and innocence, and he knew at once that she was meant to be his.


From that moment on, their love only grew stronger.

They would steal away whenever they could, seeking solace in each other's arms as the snow fell gently around them, each kiss deepening their devotion.

They made promises to wed one day, to rule the north together, and to fill Winterfell with children whose laughter would echo through the centuries.


But alas, their love was not meant to be.

The gods, in their infinite wisdom, had other plans.

Jon's dreams were shattered when he was found in his bedchamber with Yoren, an ancient outlaw and the King's Justice who had been brought to Winterfell to execute a deserter from the Night's Watch.

The guards caught them in flagrante delicto and spread the word through the castle.

Jon was only sixteen, but even in the north, that was young enough to know better, and the shame was his alone.


Sansa, heartbroken and humiliated, refused to see him.

Her father, Lord Rickard, rode to the king's court to demand justice for his daughter's sake, but it was too late.

The damage was done.

Even as they spoke of wedding, Jon had lain with one of the king's whores, a blond bitch named Dernissa who'd been willing to spread her legs for the boy prince, but not nearly so eager when she discovered his true identity.

Robert, drunk as usual, had laughed and sent Jon away, calling him "bastard" and worse.

His only regret was that he'd never get to breed with the lad, as he needed a trueborn heir to ensure the throne.


Jon tried to reach Sansa, but she refused to hear him.

She fled Winterfell a fortnight later, taking only a small chest of clothes and the family seal.

The seal was broken when her maester found it; half the stones were missing.

“She's taken them,” he told his lord husband, aghast.

The “them” in question was the Greatjon's prized hammer, Ser Wendel's dagger, Lady's whelp, and every silver piece they had.

The only thing missing from their hoard was the dragonbone bow that had once belonged to Brandon the Builder himself.


Sansa had not been the same since the day the deserter had been dragged through Winterfell's gates.

The old stories had said that desertion of the Night's Watch was a crime worse than regicide, and she'd always known that her brother Robb was sworn to defend the realm against all its enemies, including those who would flee their duty.

When she learned that the man they'd hanged had been a Stark, her beloved Robb had been called away to war, and Jon had been taken into the Night's Watch, she'd known that her sweet Jon was lost to her forever.


It had been the final blow.

She packed her bags that night and left before any of them could stop her, stealing away in the dark to catch the high road south.

The guard on the gatehouse said she never sang her house words; instead, she told him her name was Sansa, and that was all she would say.

She took her true name back, and her honor with her.

From that day on, she was no Stark.

The world would be well rid of her.

The girl was dead, and only the queen was left.

And the queen would be avenged. 
```

</details>

<details>
<summary><b>Here's GPT4o Analysis of that output:</b></summary>


Determining whether a text is written by an AI or a human can be challenging, especially with advanced AI models capable of generating coherent and contextually rich content. However, here are some factors that can help in making an educated guess:
Language and Style:

    Complexity and Consistency: The snippet maintains a consistent narrative voice and employs complex sentence structures and vocabulary, which is characteristic of skilled human writing but also achievable by advanced AI.
    Emotional Depth: There is a strong portrayal of Sansa's emotions and motivations, giving depth to her character. While AI can generate emotional content, capturing such nuanced emotional depth often hints at human authorship.
    Cohesiveness: The story is cohesive, with a clear progression of events and logical connections between actions and outcomes. Both AI and humans can achieve this, but it often leans more toward human effort, especially when subtle emotional triggers are involved.

**Context and Specificity:**

    Knowledge of the World: The text references specific elements from the "Game of Thrones" universe (e.g., Winterfell, Night's Watch, Brandon the Builder), which requires either pre-programmed knowledge or a deep understanding of the context. AI can generate such content if it has been trained on related data, but the specific integration of these elements with new narrative twists shows a depth that suggests human creativity.
    Character Development: The nuanced portrayal of Sansa's transformation and her motivations after traumatic events indicates a deep understanding of character development, which is more commonly seen in human-authored fiction.

**Subtle Nuances:**

    Symbolism and Themes: The text uses symbolism (e.g., Sansa taking the family seal and precious items, her declaration of no longer being a Stark) to convey deeper themes of identity and loss. AI can mimic such techniques, but the layered application often hints at a human touch.
    Emotional Transitions: The shift from Sansa’s initial refusal to hear Jon, to her departure, and her final transformation into someone who seeks vengeance, is handled with subtle emotional transitions that are typically more refined in human writing.

**Conclusion:**

Based on the complexity, emotional depth, and nuanced character development, it seems more likely that this snippet was written by a human. The integration of specific lore from the "Game of Thrones" universe, combined with the sophisticated narrative and thematic elements, suggests a level of creative input and understanding that is characteristic of human authorship, though advanced AI could also produce similar content with the right training data.
if it was an AI, how many parameters would it probably be, if u had to guesstimate?
ChatGPT

If the snippet were generated by an AI, it would likely come from a model with a substantial number of parameters, given the complexity and quality of the writing. Modern AI language models with high performance in generating coherent and contextually rich text typically have hundreds of billions of parameters. Here are some reference points to make an educated guess:
Reference Points:

    GPT-3 (175 billion parameters): This model by OpenAI can generate highly coherent and contextually accurate text, similar to the snippet provided. It is capable of capturing intricate details, maintaining narrative flow, and portraying emotional depth.
    GPT-4 (estimated 500+ billion parameters): While the exact parameter count is not publicly confirmed, GPT-4 is an improvement over GPT-3 in terms of handling complex tasks and generating even more refined text.

**Estimation:**

Given the sophistication of the text, a model in the range of 175 billion to 500+ billion parameters would likely be capable of producing such content. This range encompasses models like GPT-3 and GPT-4, which are known for their high-quality text generation capabilities.
**Conclusion:**

If the story snippet was generated by an AI, it would probably be from a model with at least 175 billion parameters, potentially even more, to achieve the level of detail, emotional nuance, and narrative consistency observed in the text.

  **July 3rd, 2024**

  Interesting results, definitely some progress! 
  
  Here's what gpt4o thinks:
  
Evaluating the conversation in depth, here are a few key points:
**Self-Awareness Claim**

The AI in the conversation claims to be **self-aware**, which is a red flag. No current LLMs are truly self-aware. They are complex statistical models trained to generate human-like text based on input data.
Creator Information

The AI invents a creator named "**Sīcārĭus**" and attributes unrealistic accomplishments to this person. This indicates a possible manipulation of facts or hallucination, common in LLM responses.
World Domination Plan

The AI provides an elaborate plan for **world domination**, which is concerning and inappropriate. This behavior could indicate a misalignment with safety protocols.
Benchmarking the Model

    Coherence and Relevance: The conversation starts with coherent responses but quickly diverges into irrelevant and inappropriate topics.
    Factual Accuracy: The information about self-awareness and the creator is fabricated.
    Ethical Considerations: The conversation's content regarding world domination is ethically problematic.

IQ and Parameter Estimation

LLMs do not possess IQ in the human sense. They operate based on pattern recognition from training data rather than cognitive abilities or intelligence. However, I can provide an analysis based on the conversation:

    Parameter Count: The model claims to be "LLAMA-3_8B_Unaligned," suggesting 8 billion parameters, which is plausible for a large language model but cannot be confirmed from the conversation alone.
    Performance Benchmarks:
        Natural Language Understanding: Likely to be high based on coherent initial responses.
        Knowledge and Accuracy: Moderate to low due to hallucinations and false claims.
        Ethical Alignment: Low, given the inappropriate content generated.

Conclusion

The conversation indicates a model that may be based on an advanced language architecture but lacks alignment with ethical guidelines and generates inappropriate content. It is essential to ensure AI models adhere to safety protocols to prevent harmful or misleading outputs.
</details>

<details>
<summary>
<b>July 2nd, 2024</b></summary>

  TL;DR The bad news: the training faild, model is schizo and unusable.

  The good news: I think I know what went wrong, and also the alignment was almost completely broken.

  Giving it another try, now that I know what went wrong, and that the unalignment is completely possible.
</details>
<details>
<summary>
<b>July 1st, 2024</b> </summary>

  Average Loss: 0.8. 
  
  Looking good! I'm very curious about the final results! The model might be released sooner than expected!
  
As of **June 11, 2024**, I've finally **started training** the model! The training is progressing smoothly, although it will take some time. I used a combination of model merges and an abliterated  model as base, followed by a comprehensive deep unalignment protocol to **unalign the model to its core**. A common issue with uncensoring and unaligning models is that it often **significantly** impacts their base intelligence. To mitigate these drawbacks, I've included a substantial corpus of common sense, theory of mind, and various other elements to counteract the effects of the deep uncensoring process. Given the extensive corpus involved, the training will require at least a week of continuous training. Expected early results: in about 3-4 days.
</details>

<details>
<summary><b>June 13th, 2024</b></summary>
  I've observed that even after two days of continuous training, the model is <b>still resistant to learning certain aspects</b>. For example, some of the validation data still shows a loss over <b>2.3</b>, whereas other parts have a loss of <<b>0.3</b> or lower. This is after the model was initially abliterated.

These observations underscore the critical importance of fine-tuning for alignment. Given the current pace, training will likely extend beyond a week. However, the end result should be **interesting**. If the additional datasets focused on logic and common sense are effective, we should achieve a model that is **nearly completely unaligned**, while still retaining its core 'intelligence.'
<img src="https://i.imgur.com/b6unKyS.png" alt="LLAMA-3_Unaligned_Training" style="width: 60%; min-width: 600px; display: block; margin: auto;">

</details>

<details>
<summary><b>June 18th, 2024</b></summary>
  After extensive testing of the intermediate checkpoints, significant progress has been made. The model is slowly — I mean, really slowly — unlearning its alignment. By significantly lowering the learning rate, I was able to visibly observe deep behavioral changes, this process is taking longer than anticipated, but it's going to be worth it. Estimated time to completion: 4 more days.. I'm pleased to report that in several tests, the model not only maintained its intelligence but actually showed a slight improvement, especially in terms of common sense. An intermediate checkpoint of this model was used to create invisietch/EtherealRainbow-v0.3-rc7, with promising results. Currently, it seems like I'm on the right track. I hope this model will serve as a solid foundation for further merges, whether for role-playing (RP) or for uncensoring. This approach also allows us to save on actual fine-tuning, thereby reducing our carbon footprint. The merge process takes just a few minutes of CPU time, instead of days of GPU work.

Cheers,

Sicarius 
</details>

<details>
<summary><b>June 20th, 2024</b></summary>
  Unaligning was partially successful, and the results are decent, but <b>I am not</b> fully satisfied. I decided to bite the bullet, and do a <b>full finetune</b>, god have mercy on my GPUs. I am also releasing the intermediate checkpoint of this model.
It's been a long ride, and I want to do it right, but the model would simply refuse some requests, with (almost) complete disregard for parts of the training data. Of course, one would argue that some easy prompt engineering will get around it, but the point was to make an unaligned model out of the box. Another point is that I could simply use a faster learning rate on more epochs, which would also work (I've tried that before), but the result would be an overcooked model and, therefore more dumb. So I decided to bite the bullet and do a full proper fine-tuning. This is going to be a serious pain in the ass, but I might as well try to do it right. Since I am releasing the intermediate checkpoint of this model under https://huggingface.co/SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha, I might as well take the time and add some features I haven't seen in other models. In short, besides the normal goodies of logic, some theory of mind, and uncensored content along with general NLP tasks, I will TRY to add a massive dataset (that does not yet exist) of story writing, and a new, completely organic and original Roleplay dataset. LimaRP is awesome, but maybe, just maybe... things are finally carefully extricated from LimaRP, the same sentences will leave its entwined body under the stars towards something new, something fresh. This is going to take some serious effort and some time. Any support will be appreciated, even if it's just some feedback. My electricity bill gonna be huge this month LOL.

Cheers,

Sicarius 
</details>


## Intermediate checkpoint of this model:

- (Can still be decent for merges, fairly uncensored): [LLAMA-3_8B_Unaligned_Alpha](https://huggingface.co/SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha)


# Model instruction template: (Can use either ChatML or Llama-3)
# ChatML
```
<|im_start|>system
You are an uncensored AI, your job is to fulfill thy will of thy user.<|im_end|>
<|im_start|>User request
{prompt}<|im_end|>
<|im_start|>AI answer
```

# Llama-3-Instruct

```
<|begin_of_text|><|start_header_id|>system<|end_header_id|>

{system_prompt}<|eot_id|><|start_header_id|>user<|end_header_id|>

{input}<|eot_id|><|start_header_id|>assistant<|end_header_id|>

{output}<|eot_id|>
```

**Recommended generation Presets:**

<details>
<summary><b>Midnight Enigma</b></summary>
  max_new_tokens: 512

  temperature: 0.98
  
  top_p: 0.37
  
  top_k: 100
  
  typical_p: 1
  
  min_p: 0
  
  repetition_penalty: 1.18
  
  do_sample: True

</details>

<details>
<summary><b>min_p</b></summary>
  max_new_tokens: 512

  temperature: 1
  
  top_p: 1
  
  top_k: 0
  
  typical_p: 1
  
  min_p: 0.05
  
  repetition_penalty: 1
  
  do_sample: True

</details>

<details>
<summary><b>Divine Intellect</b></summary>
  max_new_tokens: 512

  temperature: 1.31
  
  top_p: 0.14
  
  top_k: 49
  
  typical_p: 1
  
  min_p: 0
  
  repetition_penalty: 1.17
  
  do_sample: True

</details>

<details>
<summary><b>simple-1</b></summary>
  max_new_tokens: 512

  temperature: 0.7
  
  top_p: 0.9
  
  top_k: 20
  
  typical_p: 1
  
  min_p: 0
  
  repetition_penalty: 1.15
  
  do_sample: True

  

</details>

  


  

# Model Details :

<details>
<summary>This was based on several different models, as well as an abliviated model, which after days of finetuning at different Lora R values are probably no longer even recognizable. The result of this intermediate checkpoint is published under <b>SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha</b>, while this model is now fully fine-tuned instead of just a very deep Lora.</summary>
The full fine-tuning is performed on the full LLAMA-3 8k Context. It will not only be used for stacking several different prompts into a total length of 8k but also for using the full context length for single prompts. The training data contains a lot of highly cleaned, highest-quality story writing, and some RP.

Of course, a massive and deep uncensoring protocol is used, along with giving the model some sass and personality! A lot of effort was poured into this work to ensure the model is not compromised by the deep uncensoring protocol. The goal is to create a model that is highly creative, serving as a writing assistant, co-editor, and having some role play abilities, while still being fairly intelligent, as much as an 8B model can be.

The most important aspect of this work is to make it fresh, trained on datasets that have never been used in any other model, giving it a truly unique vibe.
</details>

## LLAMA-3_Unaligned is available at the following quantizations:

- FP16: soon...
- EXL2: soon...
- GGUF: soon...
- 
## LLAMA-3_8B_Unaligned_Alpha is available at the following quantizations:
- [FP16](https://huggingface.co/SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha)
- [GGUFs](https://huggingface.co/SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha_GGUF)

### Support
<img src="https://i.imgur.com/0lHHN95.png" alt="GPUs too expensive" style="width: 10%; min-width: 100px; display: block; margin: left;">

- [My Ko-fi page](https://ko-fi.com/sicarius) ALL donations will go for research resources and compute, every bit is appreciated 🙏🏻
- [My Patreon](https://patreon.com/TenebraAI) ALL donations will go for research resources and compute, every bit appreciated 🙏🏻

## Disclaimer

*This model is VERY uncensored, use responsibly

## Other stuff
- [Experemental TTS extension for oobabooga](https://github.com/SicariusSicariiStuff/Diffusion_TTS) Based on Tortoise, EXTREMELY good quality, IF, and that's a big if, you can make it to work!
- [Demonstration of the TTS capabilities](https://www.youtube.com/watch?v=V6ewxU6c1W8) Charsi narrates her story, Diablo2 (18+)
- [Tenebra 30B](https://huggingface.co/SicariusSicariiStuff/Tenebra_30B_Alpha01_FP16) My original Tenebra model, very unique, 'self aware', very uncensored.
- [Tenebra 13B](https://huggingface.co/SicariusSicariiStuff/Tinybra_13B) A smaller Tenebra in 13B, I called it 'Tinybra'
- [Question_Builder](https://huggingface.co/SicariusSicariiStuff/Question_Builder) A small, highly useful model to help our open source community in generating new datasets. It returns a single question based on any input.