Transformers
GGUF
English
Russian
Not-For-All-Audiences
nsfw
LakoMoor commited on
Commit
bb8ceea
1 Parent(s): 841f43e

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,12 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ assets/alice.png filter=lfs diff=lfs merge=lfs -text
37
+ silicon-alice-7B.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
38
+ silicon-alice-7B.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ silicon-alice-7B.Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
40
+ silicon-alice-7B.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
41
+ silicon-alice-7B.Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
42
+ silicon-alice-7B.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
43
+ silicon-alice-7B.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
44
+ silicon-alice-7B.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
.ipynb_checkpoints/README-checkpoint.md ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - ru
5
+ tags:
6
+ - not-for-all-audiences
7
+ - nsfw
8
+ base_model:
9
+ - LakoMoor/Silicon-Alice-7B
10
+ license: cc-by-nc-4.0
11
+ inference: false
12
+ library_name: transformers
13
+ model_creator: LakoMoor
14
+ model_name: Silicon-Alice-7B
15
+ model_type: mistral
16
+ ---
17
+
18
+ # Silicon-Alice-7B-GGUF
19
+ ![Silicon-Alice-7B](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/resolve/main/assets/alice.png)
20
+
21
+ ## What's that?
22
+
23
+ Silicon-Alice-7B-GGUF is a quantized model based on [Silicon-Masha-7B](https://huggingface.co/LakoMoor/Silicon-Alice-7B) aiming to be both strong in RP, be smart **and** understand Russian, that can follow character maps very well. This model understands Russian better than the previous one. It is suitable for RP/ERP and general use. It can be run on weak samovar using [llama.cpp](https://github.com/ggerganov/llama.cpp) or [koboldcpp](https://github.com/LostRuins/koboldcpp).
24
+
25
+ ## Prompt Template (Alpaca)
26
+
27
+ I found the best SillyTavern results from using the Noromaid template but please try other templates! Let me know if you find anything good.
28
+
29
+ SillyTavern config files: [Context](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/resolve/main/assets/context.json), [Instruct](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/resolve/main/assets/instruct.json).
30
+
31
+ Additionally, here is my highly recommended [Text Completion preset](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/resolve/main/assets/MinP.json). You can tweak this by adjusting temperature up or dropping min p to boost creativity or raise min p to increase stability. You shouldn't need to touch anything else!
32
+
33
+ ```
34
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
35
+ ### Instruction:
36
+ {prompt}
37
+ ### Response:
38
+ ```
39
+
40
+ ## Provided files
41
+
42
+ | Name | Quant method | Bits | Use case |
43
+ | ---- | ---- | ---- | ----- |
44
+ | [silicon-alice-7b.Q2_K.gguf](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/blob/main/silicon-alice-7b.Q2_K.gguf) | Q2_K | 2 | smallest, significant quality loss - not recommended for most purposes |
45
+ | [silicon-alice-7b.Q3_K_M.gguf](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/blob/main/silicon-alice-7b.Q3_K_M.gguf) | Q3_K_M | 3 | very small, high quality loss |
46
+ | [silicon-alice-7b.Q4_0.gguf](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/blob/main/silicon-alice-7b.Q4_0.gguf) | Q4_0 | 4 | legacy; small, very high quality loss - prefer using Q3_K_M |
47
+ | [silicon-alice-7b.Q4_K_M.gguf](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/blob/main/silicon-alice-7b.Q4_K_M.gguf) | Q4_K_M | 4 | medium, balanced quality - recommended |
48
+ | [silicon-alice-7b.Q5_0.gguf](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/blob/main/silicon-alice-7b.Q5_0.gguf) | Q5_0 | 5 | legacy; medium, balanced quality - prefer using Q4_K_M |
49
+ | [silicon-alice-7b.Q5_K_M.gguf](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/blob/main/silicon-alice-7b.Q5_K_M.gguf) | Q5_K_M | 5 | large, very low quality loss - recommended |
50
+ | [silicon-alice-7b.Q6_K.gguf](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/blob/main/silicon-alice-7b.Q6_K.gguf) | Q6_K | 6 | very large, extremely low quality loss |
51
+ | [silicon-alice-7b.Q8_0.gguf](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/blob/main/silicon-alice-7b.Q8_0.gguf) | Q8_0 | 8 | very large, extremely low quality loss - not recommended |
52
+
53
+ ## How run it ?
54
+ [llama.cpp](https://github.com/ggerganov/llama.cpp)
55
+ ```shell
56
+ ./main -ngl 35 -m silicon-alice-7b.Q4_K_M.gguf --color -c 32768 --temp 0.4 --repeat_penalty 1.1 -n -1 -p "Below is an instruction that describes a task. Write a response that appropriately completes the request.\n{system_message}\n### Instruction:{prompt}\n### Response:\n"
57
+ ```
README.md ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - ru
5
+ tags:
6
+ - not-for-all-audiences
7
+ - nsfw
8
+ base_model:
9
+ - LakoMoor/Silicon-Alice-7B
10
+ license: cc-by-nc-4.0
11
+ inference: false
12
+ library_name: transformers
13
+ model_creator: LakoMoor
14
+ model_name: Silicon-Alice-7B
15
+ model_type: mistral
16
+ ---
17
+
18
+ # Silicon-Alice-7B-GGUF
19
+ ![Silicon-Alice-7B](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/resolve/main/assets/alice.png)
20
+
21
+ ## What's that?
22
+
23
+ Silicon-Alice-7B-GGUF is a quantized model based on [Silicon-Masha-7B](https://huggingface.co/LakoMoor/Silicon-Alice-7B) aiming to be both strong in RP, be smart **and** understand Russian, that can follow character maps very well. This model understands Russian better than the previous one. It is suitable for RP/ERP and general use. It can be run on weak samovar using [llama.cpp](https://github.com/ggerganov/llama.cpp) or [koboldcpp](https://github.com/LostRuins/koboldcpp).
24
+
25
+ ## Prompt Template (Alpaca)
26
+
27
+ I found the best SillyTavern results from using the Noromaid template but please try other templates! Let me know if you find anything good.
28
+
29
+ SillyTavern config files: [Context](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/resolve/main/assets/context.json), [Instruct](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/resolve/main/assets/instruct.json).
30
+
31
+ Additionally, here is my highly recommended [Text Completion preset](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/resolve/main/assets/MinP.json). You can tweak this by adjusting temperature up or dropping min p to boost creativity or raise min p to increase stability. You shouldn't need to touch anything else!
32
+
33
+ ```
34
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
35
+ ### Instruction:
36
+ {prompt}
37
+ ### Response:
38
+ ```
39
+
40
+ ## Provided files
41
+
42
+ | Name | Quant method | Bits | Use case |
43
+ | ---- | ---- | ---- | ----- |
44
+ | [silicon-alice-7b.Q2_K.gguf](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/blob/main/silicon-alice-7b.Q2_K.gguf) | Q2_K | 2 | smallest, significant quality loss - not recommended for most purposes |
45
+ | [silicon-alice-7b.Q3_K_M.gguf](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/blob/main/silicon-alice-7b.Q3_K_M.gguf) | Q3_K_M | 3 | very small, high quality loss |
46
+ | [silicon-alice-7b.Q4_0.gguf](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/blob/main/silicon-alice-7b.Q4_0.gguf) | Q4_0 | 4 | legacy; small, very high quality loss - prefer using Q3_K_M |
47
+ | [silicon-alice-7b.Q4_K_M.gguf](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/blob/main/silicon-alice-7b.Q4_K_M.gguf) | Q4_K_M | 4 | medium, balanced quality - recommended |
48
+ | [silicon-alice-7b.Q5_0.gguf](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/blob/main/silicon-alice-7b.Q5_0.gguf) | Q5_0 | 5 | legacy; medium, balanced quality - prefer using Q4_K_M |
49
+ | [silicon-alice-7b.Q5_K_M.gguf](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/blob/main/silicon-alice-7b.Q5_K_M.gguf) | Q5_K_M | 5 | large, very low quality loss - recommended |
50
+ | [silicon-alice-7b.Q6_K.gguf](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/blob/main/silicon-alice-7b.Q6_K.gguf) | Q6_K | 6 | very large, extremely low quality loss |
51
+ | [silicon-alice-7b.Q8_0.gguf](https://huggingface.co/LakoMoor/Silicon-Alice-7B-GGUF/blob/main/silicon-alice-7b.Q8_0.gguf) | Q8_0 | 8 | very large, extremely low quality loss - not recommended |
52
+
53
+ ## How run it ?
54
+ [llama.cpp](https://github.com/ggerganov/llama.cpp)
55
+ ```shell
56
+ ./main -ngl 35 -m silicon-alice-7b.Q4_K_M.gguf --color -c 32768 --temp 0.4 --repeat_penalty 1.1 -n -1 -p "Below is an instruction that describes a task. Write a response that appropriately completes the request.\n{system_message}\n### Instruction:{prompt}\n### Response:\n"
57
+ ```
assets/.ipynb_checkpoints/MinP-checkpoint.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"temp": 0.4, "temperature_last": false, "top_p": 0.98, "top_k": 100, "top_a": 0, "tfs": 1, "epsilon_cutoff": 0, "eta_cutoff": 0, "typical_p": 1, "min_p": 0.05, "rep_pen": 1.1, "rep_pen_range": 2048, "no_repeat_ngram_size": 0, "penalty_alpha": 0, "num_beams": 1, "length_penalty": 1, "min_length": 0, "encoder_rep_pen": 1, "freq_pen": 0, "presence_pen": 0, "do_sample": true, "early_stopping": false, "add_bos_token": true, "truncation_length": 2048, "ban_eos_token": false, "skip_special_tokens": true, "streaming": true, "mirostat_mode": 0, "mirostat_tau": 5, "mirostat_eta": 0.1, "guidance_scale": 1, "negative_prompt": "", "grammar_string": "", "banned_tokens": "", "ignore_eos_token_aphrodite": false, "spaces_between_special_tokens_aphrodite": true, "type": "koboldcpp", "legacy_api": false, "sampler_order": [6, 0, 5, 3, 2, 1, 4], "n": 1, "rep_pen_size": 0, "genamt": 250, "max_length": 1024}
assets/MinP.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"temp": 0.4, "temperature_last": false, "top_p": 0.98, "top_k": 100, "top_a": 0, "tfs": 1, "epsilon_cutoff": 0, "eta_cutoff": 0, "typical_p": 1, "min_p": 0.05, "rep_pen": 1.1, "rep_pen_range": 2048, "no_repeat_ngram_size": 0, "penalty_alpha": 0, "num_beams": 1, "length_penalty": 1, "min_length": 0, "encoder_rep_pen": 1, "freq_pen": 0, "presence_pen": 0, "do_sample": true, "early_stopping": false, "add_bos_token": true, "truncation_length": 2048, "ban_eos_token": false, "skip_special_tokens": true, "streaming": true, "mirostat_mode": 0, "mirostat_tau": 5, "mirostat_eta": 0.1, "guidance_scale": 1, "negative_prompt": "", "grammar_string": "", "banned_tokens": "", "ignore_eos_token_aphrodite": false, "spaces_between_special_tokens_aphrodite": true, "type": "koboldcpp", "legacy_api": false, "sampler_order": [6, 0, 5, 3, 2, 1, 4], "n": 1, "rep_pen_size": 0, "genamt": 250, "max_length": 1024}
assets/alice.png ADDED

Git LFS Details

  • SHA256: 2244daba6d733a86c43db35cbd50a96a4d81a8391d4dc2da75710f81f6b2bb8d
  • Pointer size: 132 Bytes
  • Size of remote file: 1.29 MB
assets/context.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"story_string": "### Instruction:\nWrite {{char}}'s next reply in a fictional roleplay chat between {{user}} and {{char}}. Use the provided character sheet and example dialogue for formatting direction and character speech patterns.\n\n{{#if system}}{{system}}\n\n{{/if}}{{#if wiBefore}}{{wiBefore}}\n\n{{/if}}Description of {{char}}:\n{{#if description}}{{description}}\n{{/if}}{{#if personality}}{{char}}'s personality: {{personality}}\n\n{{/if}}{{#if scenario}}Scenario: {{scenario}}\n\n{{/if}}{{#if persona}}Description of {{user}}: {{persona}}\n\n{{/if}}Play the role of {{char}}\n\n{{#if wiAfter}}{{wiAfter}}\n{{/if}}", "example_separator": "Example roleplay chat:", "chat_start": "Taking the above information into consideration,\nyou must engage in a roleplay conversation with {{user}} below this line.\nDo not write {{user}}'s dialogue lines in your responses.\n", "always_force_name2": true, "trim_sentences": true, "include_newline": true, "single_line": false, "name": "Context-Alice"}
assets/instruct.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"system_prompt": "Avoid repetition, don't loop. Develop the plot slowly, always stay in character. Describe all actions in full, elaborate, explicit, graphic, and vivid detail. Mention all relevant sensory perceptions.", "input_sequence": "\n### Instruction: (Style: Markdown, Present Tense)", "output_sequence": "\n### Response: (Style: Markdown, Present Tense)", "first_output_sequence": "### Response:", "last_output_sequence": "", "system_sequence_prefix": "", "system_sequence_suffix": "", "stop_sequence": "", "separator_sequence": "", "wrap": true, "macro": true, "names": true, "names_force_groups": true, "activation_regex": "", "name": "Instruct-Alice"}
silicon-alice-7B.Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f24b9d6492295ac6fe2d7e14333981afc4263c73d28a7bb04ae13836b2fe5d15
3
+ size 2719241920
silicon-alice-7B.Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f96763473517adf492c68fa1c1b9ce0b55c97332467edec7321fb2c532be09f
3
+ size 3518985920
silicon-alice-7B.Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:436fe93884e3cff3369cec264e249a24e3e2e20a4851eeb9fb56b280fe6bd96d
3
+ size 4108916416
silicon-alice-7B.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f48b1813a134f890c3c08be87032c4b6817acb69e44b70e8cf5725a8e40a1fa
3
+ size 4368438976
silicon-alice-7B.Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02d205de2cda8b3bcf57f23b25de42e604f27c2bfa5088a8728f32508a048be1
3
+ size 4997715648
silicon-alice-7B.Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f93dee434fc961ba678b84360c972b8c5687a3c765237edd3eae69896e49ccb
3
+ size 5131409088
silicon-alice-7B.Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f3a7b50b7fd53bb2d6ea050944592e954dcd70b54e91beb3e6271168b9815d3
3
+ size 5942064832
silicon-alice-7B.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d922c4b712a5ca09ac4d86f561fbf65e1895937d8a632f4ab22ac1a34c901c7d
3
+ size 7695857312