JeongwonChoi commited on
Commit
abf0a66
β€’
1 Parent(s): b61b9bf

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +68 -36
README.md CHANGED
@@ -1,77 +1,109 @@
1
  ---
2
  tags:
3
- - text-generation
4
  license: cc-by-nc-sa-4.0
5
  language:
6
- - ko
7
  base_model: yanolja/KoSOLAR-10.7B-v0.1
8
  pipeline_tag: text-generation
 
 
 
9
  ---
10
 
11
  # **DataVortexS-10.7B-v0.2**
12
- <img src="./DataVortex.png" alt="DataVortex" style="height: 8em;">
13
-
14
- ## **License**
15
 
16
- [cc-by-nc-sa-4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/)
17
 
18
  ## **Model Details**
19
 
20
  ### **Base Model**
21
- [yanolja/KoSOLAR-10.7B-v0.1](https://huggingface.co/yanolja/KoSOLAR-10.7B-v0.1)
 
22
 
23
  ### **Trained On**
24
- H100 80GB 1ea
 
 
 
 
 
 
 
 
25
 
26
  ### **Instruction format**
27
 
28
- It follows **(No Input) Alpaca** format.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
 
30
  ## **Model Benchmark**
31
 
32
- ### **Ko-LLM-Leaderboard**
 
 
33
 
34
- On Benchmarking...
 
 
 
 
 
 
 
 
 
35
 
36
- # **Implementation Code**
37
 
38
- Since, chat_template already contains insturction format above.
39
  You can use the code below.
40
 
41
  ```python
42
  from transformers import AutoModelForCausalLM, AutoTokenizer
43
 
44
- device = "cuda"
45
 
46
- model = AutoModelForCausalLM.from_pretrained("Edentns/DataVortexS-10.7B-v0.2", device_map=device)
47
  tokenizer = AutoTokenizer.from_pretrained("Edentns/DataVortexS-10.7B-v0.2")
48
 
49
  messages = [
50
- { "role": "user", "content": "λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” μ–΄λ””μ•Ό?" }
 
 
 
51
  ]
52
 
53
- encoded = tokenizer.apply_chat_template(
54
- messages,
55
- add_generation_prompt=True,
56
- return_tensors="pt",
57
- return_token_type_ids=False
58
- ).to(device)
59
-
60
- decoded = model.generate(
61
- input_ids=encoded,
62
- temperature=0.2,
63
- top_p=0.9,
64
- repetition_penalty=1.2,
65
- do_sample=True,
66
- max_length=4096,
67
- eos_token_id=tokenizer.eos_token_id,
68
- pad_token_id=tokenizer.eos_token_id
69
- )
70
- decoded = decoded[0][encoded.shape[1]:decoded[0].shape[-1]]
71
- decoded_text = tokenizer.decode(decoded, skip_special_tokens=True)
72
- print(decoded_text)
73
  ```
74
 
 
 
 
 
75
  <div align="center">
76
  <a href="https://edentns.com/">
77
  <img src="./Logo.png" alt="Logo" style="height: 3em;">
 
1
  ---
2
  tags:
3
+ - text-generation
4
  license: cc-by-nc-sa-4.0
5
  language:
6
+ - ko
7
  base_model: yanolja/KoSOLAR-10.7B-v0.1
8
  pipeline_tag: text-generation
9
+ datasets:
10
+ - beomi/KoAlpaca-v1.1a
11
+ - Edentns/Worktronics-FAQ
12
  ---
13
 
14
  # **DataVortexS-10.7B-v0.2**
 
 
 
15
 
16
+ <img src="./DataVortex.png" alt="DataVortex" style="height: 8em;">
17
 
18
  ## **Model Details**
19
 
20
  ### **Base Model**
21
+
22
+ [yanolja/KoSOLAR-10.7B-v0.1](https://huggingface.co/yanolja/KoSOLAR-10.7B-v0.1)
23
 
24
  ### **Trained On**
25
+
26
+ - **OS**: Ubuntu 20.04
27
+ - **GPU**: H100 80GB 1ea
28
+ - **transformers**: v4.36.2
29
+
30
+ ### **Dataset**
31
+
32
+ - [beomi/KoAlpaca-v1.1a](https://huggingface.co/datasets/beomi/KoAlpaca-v1.1a)
33
+ - Edentns/Worktronics-FAQ - private
34
 
35
  ### **Instruction format**
36
 
37
+ It follows **Alpaca** format.
38
+
39
+ E.g.
40
+
41
+ ```python
42
+ text = """\
43
+ 당신은 μ‚¬λžŒλ“€μ΄ 정보λ₯Ό 찾을 수 μžˆλ„λ‘ λ„μ™€μ£ΌλŠ” 인곡지λŠ₯ λΉ„μ„œμž…λ‹ˆλ‹€.
44
+
45
+ ### Instruction:
46
+ λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” μ–΄λ””μ•Ό?
47
+
48
+ ### Response:
49
+ λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” μ„œμšΈμž…λ‹ˆλ‹€.
50
+
51
+ ### Instruction:
52
+ μ„œμšΈ μΈκ΅¬λŠ” 총 λͺ‡ λͺ…이야?
53
+ """
54
+ ```
55
 
56
  ## **Model Benchmark**
57
 
58
+ ### **[Ko-LLM-Leaderboard](https://huggingface.co/spaces/upstage/open-ko-llm-leaderboard)**
59
+
60
+ On Benchmarking ...
61
 
62
+ | Model | Average | Ko-ARC | Ko-HellaSwag | Ko-MMLU | Ko-TruthfulQA | Ko-CommonGen V2 |
63
+ | ---------------------------- | ------- | ------ | ------------ | ------- | ------------- | --------------- |
64
+ | DataVortexM-7B-Instruct-v0.1 | 39.81 | 34.13 | 42.35 | 38.73 | 45.46 | 38.37 |
65
+ | DataVortexS-10.7B-v0.1 | 0 | 0 | 0 | 0 | 0 | 0 |
66
+ | **DataVortexS-10.7B-v0.2** | **0** | **0** | **0** | **0** | **0** | **0** |
67
+ | DataVortexS-10.7B-v0.3 | 0 | 0 | 0 | 0 | 0 | 0 |
68
+ | DataVortexS-10.7B-v0.4 | 0 | 0 | 0 | 0 | 0 | 0 |
69
+ | DataVortexS-10.7B-v0.5 | 0 | 0 | 0 | 0 | 0 | 0 |
70
+ | DataVortexTL-1.1B-v0.1 | 0 | 0 | 0 | 0 | 0 | 0 |
71
+ | DataVortexS-10.7B-dpo-v0.1 | 0 | 0 | 0 | 0 | 0 | 0 |
72
 
73
+ ## **Implementation Code**
74
 
75
+ This model contains the chat_template instruction format.
76
  You can use the code below.
77
 
78
  ```python
79
  from transformers import AutoModelForCausalLM, AutoTokenizer
80
 
81
+ device = "cuda" # the device to load the model onto
82
 
83
+ model = AutoModelForCausalLM.from_pretrained("Edentns/DataVortexS-10.7B-v0.2")
84
  tokenizer = AutoTokenizer.from_pretrained("Edentns/DataVortexS-10.7B-v0.2")
85
 
86
  messages = [
87
+ {"role": "system", "content": "당신은 μ‚¬λžŒλ“€μ΄ 정보λ₯Ό 찾을 수 μžˆλ„λ‘ λ„μ™€μ£ΌλŠ” 인곡지λŠ₯ λΉ„μ„œμž…λ‹ˆλ‹€."},
88
+ {"role": "user", "content": "λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” μ–΄λ””μ•Ό?"},
89
+ {"role": "assistant", "content": "λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” μ„œμšΈμž…λ‹ˆλ‹€."},
90
+ {"role": "user", "content": "μ„œμšΈ μΈκ΅¬λŠ” 총 λͺ‡ λͺ…이야?"}
91
  ]
92
 
93
+ encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt")
94
+
95
+ model_inputs = encodeds.to(device)
96
+ model.to(device)
97
+
98
+ generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True)
99
+ decoded = tokenizer.batch_decode(generated_ids)
100
+ print(decoded[0])
 
 
 
 
 
 
 
 
 
 
 
 
101
  ```
102
 
103
+ ## **License**
104
+
105
+ The model is licensed under the [cc-by-nc-sa-4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/) license, which allows others to copy, modify, and share the work non-commercially, as long as they give appropriate credit and distribute any derivative works under the same license.
106
+
107
  <div align="center">
108
  <a href="https://edentns.com/">
109
  <img src="./Logo.png" alt="Logo" style="height: 3em;">