File size: 3,396 Bytes
96a23e1
 
d42dc68
 
 
 
 
 
 
 
 
96a23e1
e52a963
d42dc68
 
 
5251164
d42dc68
 
 
 
 
 
 
 
 
 
 
 
9d2d73c
 
3ffc02b
 
 
9d2d73c
 
 
d42dc68
 
 
 
 
 
 
 
 
 
e52a963
d42dc68
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
---
license: cc-by-nc-4.0
base_model:
- Netrve/Miqu-PlayMaid-70B-v0.1
- ShinojiResearch/Senku-70B
library_name: transformers
tags:
- not-for-all-audiences
- nsfw
- mergekit
- merge
---
# aranea-tenebris-120b-v1.0
**aka Netrve/Miqu-PlayMaid-70B-v0.1 + ShinojiResearch/Senku-70B**  
Model merge for uncensored creative writing and rp  

![image/png](https://huggingface.co/divinetaco/aranea-tenebris-120b-v1.0/resolve/main/aranea-tenebris.png)

A [mergekit](https://github.com/arcee-ai/mergekit) frankenmerge based on [Netrve/Miqu-PlayMaid-70B-v0.1](https://huggingface.co/Netrve/Miqu-PlayMaid-70B-v0.1) with interleaved layers of [ShinojiResearch/Senku-70B](https://huggingface.co/ShinojiResearch/Senku-70B).  
This was the top performing model from a second series of merge experiments to create a highly coherant creative writing and rp model.  
Tests consisted of a series of private DnD scenario benchmarks, with manual comparison of the most promising merges.  
  
A number of different base models, interleave models and layer offsets were compared.  
This model outperformed a number of other popular 70B+ models and merges in both creativity and coherancy tests. It was (briefly) compared to Mixtral 8x22B running 2/3/4 experts.  

- Usable context: ~32768
- Recommended prompt format: Alpaca
- Layers: 137


### Quantization
  
llama.cpp [imatrix.dat](./imatrix.dat)  
exllamav2 [measurement.json](./measurement.json)  

Will upload a few quants when bandwidth permits.

### Testing

Two different writing styles were considered for each testing scenario:
- Completions for 3rd person narration. No character role was assumed.
- Completions for 1st and 2nd person turn based (out-of-order) rp. A character role was assumed by the model, but narration of minor characters and events was encouraged.

Tests assumed a mature audience, but a range of scenarios were constructed.  
Thematic inconsistancy or bias in character behaviour was penalized heavily.  

Models showing the following were penalized during manual comparison:
- Consistently short responses.
- Laziness or readily gave up on solving a character problem.
- Overly malleable, where characters could not hold opinions or beliefs.
- Passiveness or an inability to drive the narrative.
- Persistent repeats. Bad merges tend to latch onto and reuse specific keywords.
- Ignoring or missing obvious scenario solutions.
- Impersonating other major characters out of turn during rp tests.
- Faliure to follow a character's description. This criteria is pretty broad, and could include things like character skills, refusals etc.
- Major inconsistencies in scenes or recall. Note - invention of thematically consistant detail was encouraged.

### Interesting observations from benchmarking

- 10 layer interleave stride with a 20 layer interleave width consistently outperformed alternative combinations for coherancy.
- 8 layer interleave stride with a 16 layer interleave width consistantly outperformed alternative combinations for creativity whilst remaining reasonably coherant.
- Regular stride intervals are not optimal. In particular offsetting the first or last set of base models offets often improved metrics.
- Goliath-120B is still a good standard for coherancy below 4096 context. A few miqu-1 merges are comparable, but testing found a small amount coherancy could be sacrificed for notable creativity improvements.