Add data sources to the README
Browse files- README.md +15 -0
- config.json +1 -1
README.md
CHANGED
@@ -53,3 +53,18 @@ It is supported by:
|
|
53 |
{prompt}<|im_end|>
|
54 |
<|im_start|>assistant
|
55 |
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
53 |
{prompt}<|im_end|>
|
54 |
<|im_start|>assistant
|
55 |
```
|
56 |
+
|
57 |
+
## Training data
|
58 |
+
|
59 |
+
- [no_robots dataset](https://huggingface.co/Undi95/Llama2-13B-no_robots-alpaca-lora) let the model have more human behavior, enhances the output.
|
60 |
+
- [Aesir Private RP dataset] New data from a new and never used before dataset, add fresh data, no LimaRP spam, this is 100% new. Thanks to the [MinvervaAI Team](https://huggingface.co/MinervaAI) and, in particular, [Gryphe](https://huggingface.co/Gryphe) for letting us use it!
|
61 |
+
- [Another private Aesir dataset]
|
62 |
+
- [Another private Aesir dataset]
|
63 |
+
- [limarp](https://huggingface.co/datasets/lemonilia/LimaRP)
|
64 |
+
|
65 |
+
## DPO training data
|
66 |
+
|
67 |
+
- [Intel/orca_dpo_pairs](https://huggingface.co/datasets/Intel/orca_dpo_pairs)
|
68 |
+
- [NobodyExistsOnTheInternet/ToxicDPOqa](https://huggingface.co/datasets/NobodyExistsOnTheInternet/ToxicDPOqa)
|
69 |
+
- [Undi95/toxic-dpo-v0.1-NoWarning](https://huggingface.co/datasets/Undi95/toxic-dpo-v0.1-NoWarning)
|
70 |
+
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "/srv/huggingface/hub/models--NeverSleep--Noromaid-7B-0.4-DPO/snapshots/47a417a2167064112038e71f2be30d7293eb485d",
|
3 |
"architectures": [
|
4 |
"MistralForCausalLM"
|
5 |
],
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "/srv/home/shaun/.cache/huggingface/hub/models--NeverSleep--Noromaid-7B-0.4-DPO/snapshots/47a417a2167064112038e71f2be30d7293eb485d",
|
3 |
"architectures": [
|
4 |
"MistralForCausalLM"
|
5 |
],
|