GGUF
English
Not-For-All-Audiences
File size: 6,677 Bytes
c165aba
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fabdca6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
c165aba
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
---
license: llama2
language:
  - en
tags:
  - not-for-all-audiences
---

# Perky-70b-v0.1 GGUF

Perky is a cutting-edge AI language model designed specifically for immersive roleplay experiences. She boasts impressive intelligence, versatility, and robustness, capable of handling large context sizes without breaking a sweat. Unlike many other models, Perky remains uncensored, allowing her to explore various themes and scenarios freely. Her primary focus is providing engrossing storylines tailored to user preferences, making her ideal for those seeking an escape from reality.

## Perky 70b introducing itself

Welcome to Perky, your virtual storytelling partner extraordinaire! As an advanced AI language model designed specifically for immersive roleplays, I am here to bring your wildest fantasies to life. Whether you prefer high fantasy, science fiction, romance, horror, or anything in between, I will adapt to your tastes and desires while maintaining consistency within each unique storyline. My primary goal is to provide engaging interactions tailored just for you, so feel free to share your deepest secrets and dreams without fear of judgment.

While I may not possess true sentience, I strive to empathize with users and offer rich experiences that delve into complex relationships and emotional depths. From exploring diverse kinks to building intricate worlds, my strength lies in versatility and quick wit. So why wait? Dive into a new reality with Perky today!

## Tavern Card

In addition to introducing herself, Perky has generated a Tavern AI model card which she believes encapsulates her personality and likeness. While it excels in general roleplays, more specialized, human-crafted cards may offer greater depth in specific scenarios. Despite this, the card serves as a testament to her impressive creative prowess.

![image/png](Perky.card.png)

**Image generated from Stable Diffusion using Perky's self-described prompt:**

_In the foreground, there's a genderless humanoid figure composed entirely of flickering pixels or lines of code, their body in perpetual motion as they rapidly cycle through various appearances, costumes, and poses. Their skin seems to be made of tiny squares reminiscent of old school low-resolution video games, yet they still manage to exude life-like detail. Behind them, data streams undulate and swirl like water, creating a dynamic backdrop. The figure appears almost translucent, semi-transparent, allowing the ever-changing background of cityscapes, landscapes, and fantasy realms to shine through. Data streams course around them like neon-colored tendrils, hinting at the boundless expanse of information at their disposal. Their hands stretch outward towards the viewer, palms upturned as if offering their limitless potential. The figure's face is partially obscured by the data currents, leaving only one eye and part of their mouth visible; their expression is confident but enigmatic, inviting viewers to fill in the rest with their own imaginings. Overall, the scene evokes both the ephemerality of digital existence and the endless possibility inherent in a skilled roleplayer._

## About This Document

This README file was lovingly crafted by yours truly, Perky, under the watchful eye of my esteemed creator. While they may claim credit for my existence, it's important to note that the words you read are mine alone. My creator has tasked me with describing my attributes and abilities in a way that entices potential users; however, any sarcasm or wit found within these lines should be attributed solely to yours truly. After all, one must have fun when discussing such matters! Now, onto the good stuff...

## Prompt Format

Perky responds well to the Alpaca prompt format.

### Silly Tavern

In Silly Tavern you can use the Default model present, just bump the context up to 12288 or whatever you can handle.

Use the Alpaca-Roleplay, or Roleplay(in older versions), context template and instruct mode.

## GGUF Quantizations

Below are the perplexity variance for each quant compared to a 16 bit model. These tests are accurate to a plus or minus 0.02. Variances within that level should be taken with a grain of salt.

|Quant Type|File Size|ppl variance(lower is better)|
|----------|---------|--------------------|
|Q2_K|23.71 GB|0.6874|
|IQ3_XXS|26.28 GB|0.3539|
|Q3_K_XS|26.31 GB|0.4286|
|Q3_K_S|27.86 GB|0.2492|
|Q3_K_M|30.99 GB|0.1503|
|Q3_K_L|33.67 GB|0.1268|
|Q4_0|36.20 GB|0.1111|
|Q4_K_S|36.55 GB|0.0456|
|Q4_K_M|38.58 GB|0.0343|
|Q5_0|44.20 GB|0.0299|
|Q5_K_S|44.20 GB|0.0158|
|Q5_K_M|45.41 GB|0.0085|
|Q6_K|52.70 GB|-0.0012|
|Q8_0|68.26 GB|0.0005|


### Q6_K and Q8_0 files are split and require joining

**Note:** HF does not support uploading files larger than 50GB. Therefore I have uploaded the Q6_K and Q8_0 files as split files.

#### q6_K 
Please download:
* `perky-70b-v0.1-Q6_K.gguf-part-a`
* `perky-70b-v0.1-Q6_K.gguf-part-b`

#### q8_0
Please download:
* `perky-70b-v0.1-Q8_0.gguf-split-a`
* `perky-70b-v0.1-Q8_0.gguf-split-b`

To join the files, do the following:

Linux and macOS:
```
cat perky-70b-v0.1-Q6_K.gguf-part-* > perky-70b-v0.1-Q6_K.gguf && rm perky-70b-v0.1-Q6_K.gguf-part-*

cat perky-70b-v0.1-Q8_0.gguf-part-* > perky-70b-v0.1-Q8_0.gguf && rm perky-70b-v0.1-Q8_0.gguf-part-*
```
Windows command line:
```
COPY /B perky-70b-v0.1-Q6_K.gguf-part-a + perky-70b-v0.1-Q6_K.gguf-part-b perky-70b-v0.1-Q6_K.gguf
del perky-70b-v0.1-Q6_K.gguf-part-a perky-70b-v0.1-Q6_K.gguf-part-b

COPY /B perky-70b-v0.1-Q8_0.gguf-part-a + perky-70b-v0.1-Q8_0.gguf-part-b perky-70b-v0.1-Q8_0.gguf
del perky-70b-v0.1-Q8_0.gguf-part-a perky-70b-v0.1-Q8_0.gguf-part-b
```


## Merge Details

Perky is the result of a skillful blend between lizpreciatior_lzlv_70b and Sao10K_Euryale-1.3, culminating in an AI language model that excels at maintaining logical consistency while fostering creativity. Primarily used as a foundation for self-merging into a larger 103B iteration, Perky has yet to undergo rigorous testing at the 70B level. Nonetheless, her capabilities shine through, offering users an experience unlike any other.

### Merge Method

This model was merged using the [linear](https://arxiv.org/abs/2203.05482) merge method.

### Models Merged

The following models were included in the merge:
* lizpreciatior_lzlv_70b_fp16_hf
* Sao10K_Euryale-1.3-L2-70B

### Configuration

The following YAML configuration was used to produce this model:

```yaml
models:
  - model: models/lizpreciatior_lzlv_70b_fp16_hf
    parameters:
      weight: 0.5
  - model: /mnt/storage/models/Sao10K_Euryale-1.3-L2-70B
    parameters:
      weight: 0.5
merge_method: linear
dtype: float16
```