File size: 1,478 Bytes
1d80d12
 
 
 
 
cf09ee5
1d80d12
f17ece7
1d80d12
cf09ee5
1d80d12
cf09ee5
1d80d12
cf09ee5
1d80d12
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
---
license: llama2
language:
- en
---
# Daddy Dave's stamp of approval 👍

4-bit GPTQ quants of the writer version of [Sao10K](https://huggingface.co/Sao10K)'s fantastic [Stheno model (Stheno model collection link)](https://huggingface.co/collections/Sao10K/stheno-6536a20823c9d18c09288fb1)

The main branch contains 4-bit groupsize of 128 and no act_order.

The other branches contain groupsizes of 128, 64, and 32 all with act_order.

## **⬇︎** Original card **⬇︎**

My GGUF Quants: https://huggingface.co/Sao10K/Stheno-1.10-L2-13B-GGUF

*** 

Oh, you thought there'd be a 2.0? Nope. Not yet.

A recreation of Stheno with Updated versions of the same models and Merging Values. Feels more coherent, and is uncensored (zero context) at least according to my tests. 

Is somewhat smarter, I think? Atleast it passes 4/5 times in my own test suites.

Feel free to try it out, I'd appreciate Feedback.

Most formats could work, but my tests have all been done in Alpaca format and it works well.
```
### Instruction:
Your instruction or question here.
For roleplay purposes, I suggest the following - Write <CHAR NAME>'s next reply in a chat between <YOUR NAME> and <CHAR NAME>. Write a single reply only.

### Response:
```

support me [here](https://ko-fi.com/sao10k) :)

Once Again, thanks to [Chargoddard](https://huggingface.co/chargoddard) for his amazing and simple [mergekit](https://github.com/cg123/mergekit) script. Thanks to the original model creators too!