Edit model card

"We must sleep, but AI Never Sleeps!"

 

Prompt Template

A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.
Human: {prompt}
Assistant:

Simple-Usage

from transformers import AutoTokenizer
from transformers import AutoModelForCausalLM

model = AutoModelForCausalLM.from_pretrained("yanolja/EEVE-Korean-Instruct-2.8B-v1.0", trust_remote_code=True)
tokenizer = AutoTokenizer.from_pretrained("yanolja/EEVE-Korean-Instruct-2.8B-v1.0", trust_remote_code=True)

prompt_template = "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.\nHuman: {prompt}\nAssistant:\n"
text = '๋‹ค์ด์–ดํŠธ์‹ ๋ฉ”๋‰ด๋ฅผ ์ถ”์ฒœํ•ด์ฃผ์„ธ์š”.\n\n(A) ์ƒ๋Ÿฌ๋“œ\n(B) ์น˜ํ‚จ\n(C) ํ”ผ์ž\n(D) ํŒŒ์Šคํƒ€'
model_inputs = tokenizer(prompt_template.format(prompt=text), return_tensors='pt')

outputs = model.generate(**model_inputs, max_new_tokens=256)
output_text = tokenizer.batch_decode(outputs, skip_special_tokens=True)[0]
print(output_text)

Example Output

A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.
Human: ๋‹ค์ด์–ดํŠธ์‹ ๋ฉ”๋‰ด๋ฅผ ์ถ”์ฒœํ•ด์ฃผ์„ธ์š”.

(A) ์ƒ๋Ÿฌ๋“œ
(B) ์น˜ํ‚จ
(C) ํ”ผ์ž
(D) ํŒŒ์Šคํƒ€
Assistant:
(A) ์ƒ๋Ÿฌ๋“œ๋ฅผ ์ถ”์ฒœ๋“œ๋ฆฝ๋‹ˆ๋‹ค. ์ƒ๋Ÿฌ๋“œ๋Š” ์ €์นผ๋กœ๋ฆฌ์ด๋ฉด์„œ๋„ ์˜์–‘์†Œ๊ฐ€ ํ’๋ถ€ํ•ด ๋‹ค์ด์–ดํŠธ์‹์œผ๋กœ ์ ํ•ฉํ•ฉ๋‹ˆ๋‹ค. ๋‹ค์–‘ํ•œ ์ฑ„์†Œ์™€ ๋‹จ๋ฐฑ์งˆ์„ ์ถ”๊ฐ€ํ•˜์—ฌ ๊ท ํ˜• ์žกํžŒ ์‹์‚ฌ๋ฅผ ๋งŒ๋“œ์‹ค ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.

About the Model

First of all, Overwhelming gratitude to 'yanolja/EEVE' Model & Team! This model is a fine-tuned version of crimsonjoo/Neversleep-3B-v0.1, which is a Korean vocabulary-extended version of microsoft/phi-2. Specifically, we utilized Direct Preference Optimization (DPO) through the use of Axolotl.

For more details, please refer to our technical report: Efficient and Effective Vocabulary Expansion Towards Multilingual Large Language Models.

Training Data

Downloads last month
8
Safetensors
Model size
2.82B params
Tensor type
BF16
ยท
Inference Examples
Inference API (serverless) is not available, repository is disabled.

Model tree for crimsonjoo/Neversleep-3B-Instruct-v0.1

Base model

microsoft/phi-2
Finetuned
this model