RichardErkhov commited on
Commit
f9a015f
1 Parent(s): 2fd4178

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +225 -0
README.md ADDED
@@ -0,0 +1,225 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ llama-3-experiment-v1-9B - GGUF
11
+ - Model creator: https://huggingface.co/grimjim/
12
+ - Original model: https://huggingface.co/grimjim/llama-3-experiment-v1-9B/
13
+
14
+
15
+ | Name | Quant method | Size |
16
+ | ---- | ---- | ---- |
17
+ | [llama-3-experiment-v1-9B.Q2_K.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.Q2_K.gguf) | Q2_K | 3.26GB |
18
+ | [llama-3-experiment-v1-9B.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.IQ3_XS.gguf) | IQ3_XS | 3.61GB |
19
+ | [llama-3-experiment-v1-9B.IQ3_S.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.IQ3_S.gguf) | IQ3_S | 3.78GB |
20
+ | [llama-3-experiment-v1-9B.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.Q3_K_S.gguf) | Q3_K_S | 3.76GB |
21
+ | [llama-3-experiment-v1-9B.IQ3_M.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.IQ3_M.gguf) | IQ3_M | 3.88GB |
22
+ | [llama-3-experiment-v1-9B.Q3_K.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.Q3_K.gguf) | Q3_K | 4.13GB |
23
+ | [llama-3-experiment-v1-9B.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.Q3_K_M.gguf) | Q3_K_M | 4.13GB |
24
+ | [llama-3-experiment-v1-9B.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.Q3_K_L.gguf) | Q3_K_L | 4.45GB |
25
+ | [llama-3-experiment-v1-9B.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.IQ4_XS.gguf) | IQ4_XS | 4.61GB |
26
+ | [llama-3-experiment-v1-9B.Q4_0.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.Q4_0.gguf) | Q4_0 | 4.8GB |
27
+ | [llama-3-experiment-v1-9B.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.IQ4_NL.gguf) | IQ4_NL | 4.84GB |
28
+ | [llama-3-experiment-v1-9B.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.Q4_K_S.gguf) | Q4_K_S | 4.83GB |
29
+ | [llama-3-experiment-v1-9B.Q4_K.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.Q4_K.gguf) | Q4_K | 5.07GB |
30
+ | [llama-3-experiment-v1-9B.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.Q4_K_M.gguf) | Q4_K_M | 5.07GB |
31
+ | [llama-3-experiment-v1-9B.Q4_1.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.Q4_1.gguf) | Q4_1 | 5.29GB |
32
+ | [llama-3-experiment-v1-9B.Q5_0.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.Q5_0.gguf) | Q5_0 | 5.77GB |
33
+ | [llama-3-experiment-v1-9B.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.Q5_K_S.gguf) | Q5_K_S | 5.77GB |
34
+ | [llama-3-experiment-v1-9B.Q5_K.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.Q5_K.gguf) | Q5_K | 5.91GB |
35
+ | [llama-3-experiment-v1-9B.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.Q5_K_M.gguf) | Q5_K_M | 5.91GB |
36
+ | [llama-3-experiment-v1-9B.Q5_1.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.Q5_1.gguf) | Q5_1 | 6.26GB |
37
+ | [llama-3-experiment-v1-9B.Q6_K.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.Q6_K.gguf) | Q6_K | 6.81GB |
38
+ | [llama-3-experiment-v1-9B.Q8_0.gguf](https://huggingface.co/RichardErkhov/grimjim_-_llama-3-experiment-v1-9B-gguf/blob/main/llama-3-experiment-v1-9B.Q8_0.gguf) | Q8_0 | 8.82GB |
39
+
40
+
41
+
42
+
43
+ Original model description:
44
+ ---
45
+ language:
46
+ - en
47
+ base_model:
48
+ - meta-llama/Meta-Llama-3-8B-Instruct
49
+ library_name: transformers
50
+ tags:
51
+ - meta
52
+ - llama-3
53
+ - pytorch
54
+ - mergekit
55
+ - merge
56
+ license: llama3
57
+ license_link: LICENSE
58
+ pipeline_tag: text-generation
59
+ widget:
60
+ - example_title: Hello
61
+ messages:
62
+ - role: user
63
+ content: Hey my name is Corwin! How are you?
64
+ - example_title: Hellriding out of Amber
65
+ messages:
66
+ - role: system
67
+ content: You are a helpful and honest assistant. Please, respond concisely and truthfully.
68
+ - role: user
69
+ content: Can you recommend a good destination for a hellride out of Amber?
70
+ inference:
71
+ parameters:
72
+ max_new_tokens: 300
73
+ stop:
74
+ - <|end_of_text|>
75
+ - <|eot_id|>
76
+ model-index:
77
+ - name: grimjim/grimjim/llama-3-experiment-v1-9B
78
+ results:
79
+ - task:
80
+ type: text-generation
81
+ name: Text Generation
82
+ dataset:
83
+ name: AI2 Reasoning Challenge (25-Shot)
84
+ type: ai2_arc
85
+ config: ARC-Challenge
86
+ split: test
87
+ args:
88
+ num_few_shot: 25
89
+ metrics:
90
+ - type: acc_norm
91
+ value: 66.41
92
+ name: normalized accuracy
93
+ source:
94
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=grimjim/grimjim/llama-3-experiment-v1-9B
95
+ name: Open LLM Leaderboard
96
+ - task:
97
+ type: text-generation
98
+ name: Text Generation
99
+ dataset:
100
+ name: HellaSwag (10-Shot)
101
+ type: hellaswag
102
+ split: validation
103
+ args:
104
+ num_few_shot: 10
105
+ metrics:
106
+ - type: acc_norm
107
+ value: 78.56
108
+ name: normalized accuracy
109
+ source:
110
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=grimjim/llama-3-experiment-v1-9B
111
+ name: Open LLM Leaderboard
112
+ - task:
113
+ type: text-generation
114
+ name: Text Generation
115
+ dataset:
116
+ name: MMLU (5-Shot)
117
+ type: cais/mmlu
118
+ config: all
119
+ split: test
120
+ args:
121
+ num_few_shot: 5
122
+ metrics:
123
+ - type: acc
124
+ value: 66.71
125
+ name: accuracy
126
+ source:
127
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=grimjim/llama-3-experiment-v1-9B
128
+ name: Open LLM Leaderboard
129
+ - task:
130
+ type: text-generation
131
+ name: Text Generation
132
+ dataset:
133
+ name: TruthfulQA (0-shot)
134
+ type: truthful_qa
135
+ config: multiple_choice
136
+ split: validation
137
+ args:
138
+ num_few_shot: 0
139
+ metrics:
140
+ - type: mc2
141
+ value: 50.7
142
+ source:
143
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=grimjim/llama-3-experiment-v1-9B
144
+ name: Open LLM Leaderboard
145
+ - task:
146
+ type: text-generation
147
+ name: Text Generation
148
+ dataset:
149
+ name: Winogrande (5-shot)
150
+ type: winogrande
151
+ config: winogrande_xl
152
+ split: validation
153
+ args:
154
+ num_few_shot: 5
155
+ metrics:
156
+ - type: acc
157
+ value: 75.93
158
+ name: accuracy
159
+ source:
160
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=grimjim/llama-3-experiment-v1-9B
161
+ name: Open LLM Leaderboard
162
+ - task:
163
+ type: text-generation
164
+ name: Text Generation
165
+ dataset:
166
+ name: GSM8k (5-shot)
167
+ type: gsm8k
168
+ config: main
169
+ split: test
170
+ args:
171
+ num_few_shot: 5
172
+ metrics:
173
+ - type: acc
174
+ value: 65.88
175
+ name: accuracy
176
+ source:
177
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=grimjim/llama-3-experiment-v1-9B
178
+ name: Open LLM Leaderboard
179
+ ---
180
+ # llama-3-experiment-v1-9B
181
+
182
+ This is an experimental merge, replicating additional layers to the model without post-merge healing.
183
+ There is damage to the model, but it appears to be tolerable as is; the performance difference in benchmarks from the original 8B Instruct model does not appear to be significant.
184
+ The resulting impact on narrative text completion may also be of interest.
185
+
186
+ Light testing performed with instruct prompting and the following sampler settings:
187
+ - temp=1 and minP=0.02
188
+ - temp=1 and smoothing factor=0.33
189
+
190
+ Full weights: [grimjim/llama-3-experiment-v1-9B](https://huggingface.co/grimjim/llama-3-experiment-v1-9B)
191
+
192
+ GGUF quants: [grimjim/llama-3-experiment-v1-9B-GGUF](https://huggingface.co/grimjim/llama-3-experiment-v1-9B-GGUF)
193
+
194
+ This is a merge of pre-trained language model meta-llama/Meta-Llama-3-8B-Instruct created using [mergekit](https://github.com/cg123/mergekit).
195
+
196
+ Built with Meta Llama 3.
197
+
198
+ ## Merge Details
199
+ ### Merge Method
200
+
201
+ This model was merged using the passthrough merge method.
202
+
203
+ ### Models Merged
204
+
205
+ The following models were included in the merge:
206
+ * meta-llama/Meta-Llama-3-8B-Instruct
207
+
208
+ ### Configuration
209
+
210
+ The following YAML configuration was used to produce this model:
211
+
212
+ ```yaml
213
+ slices:
214
+ - sources:
215
+ - model: meta-llama/Meta-Llama-3-8B-Instruct
216
+ layer_range: [0, 12]
217
+ - sources:
218
+ - model: meta-llama/Meta-Llama-3-8B-Instruct
219
+ layer_range: [8, 32]
220
+ merge_method: passthrough
221
+ dtype: bfloat16
222
+
223
+ ```
224
+
225
+