JeongwonChoi's picture
Update README.md
a2fb05b
|
raw
history blame
No virus
1.81 kB
metadata
tags:
  - text-generation
license: cc-by-nc-sa-4.0
language:
  - ko
base_model: mistralai/Mistral-7B-Instruct-v0.2
pipeline_tag: text-generation

DataVortexM-7B-Instruct-v0.2

DataVortex

License

cc-by-nc-sa-4.0

Model Details

Base Model

mistralai/Mistral-7B-Instruct-v0.2

Trained On

H100 80GB 4ea

Instruction format

It follows Alpaca format.

Model Benchmark

Ko-LLM-Leaderboard

On Benchmarking...

Implementation Code

Since, chat_template already contains insturction format above. You can use the code below.

from transformers import AutoModelForCausalLM, AutoTokenizer

device = "cuda"

model = AutoModelForCausalLM.from_pretrained("Edentns/DataVortexM-7B-Instruct-v0.2", device_map=device)
tokenizer = AutoTokenizer.from_pretrained("Edentns/DataVortexM-7B-Instruct-v0.2")

messages = [
    { "role": "user", "content": "๋Œ€ํ•œ๋ฏผ๊ตญ์˜ ์ˆ˜๋„๋Š” ์–ด๋””์•ผ?" }
]

encoded = tokenizer.apply_chat_template(
    messages,
    add_generation_prompt=True,
    return_tensors="pt",
    return_token_type_ids=False
).to(device)

decoded = model.generate(
    input_ids=encoded,
    temperature=0.2,
    top_p=0.9,
    repetition_penalty=1.2,
    do_sample=True,
    max_length=4096,
    eos_token_id=tokenizer.eos_token_id,
    pad_token_id=tokenizer.eos_token_id
)
decoded = decoded[0][encoded.shape[1]:decoded[0].shape[-1]]
decoded_text = tokenizer.decode(decoded, skip_special_tokens=True)
print(decoded_text)