--- base_model: jpacifico/Chocolatine-3B-Instruct-DPO-v1.2 datasets: - jpacifico/french-orca-dpo-pairs-revised language: - fr - en library_name: transformers license: mit pipeline_tag: text-generation tags: - french - chocolatine - llama-cpp --- # Chocolatine-3B-Instruct-DPO-v1.2-Q4_K_M-GGUF Quantized q4_k_m GGUF version of the original model [`jpacifico/Chocolatine-3B-Instruct-DPO-v1.2`](https://huggingface.co/jpacifico/Chocolatine-3B-Instruct-DPO-v1.2) can be used on a CPU device, compatible [llama.cpp](https://github.com/ggerganov/llama.cpp) now supported architecture by [LM Studio](https://lmstudio.ai/). Also ready for Raspberry Pi 5 8Gb. ### Ollama [jpacifico/chocolatine-3b](https://ollama.com/jpacifico/chocolatine-3b) Usage: ```bash ollama run jpacifico/chocolatine-3b ``` Ollama *Modelfile* example : ```bash FROM ./chocolatine-3b-instruct-dpo-v1.2-q4_k_m.gguf TEMPLATE """{{ if .System }}<|system|> {{ .System }}<|end|> {{ end }}{{ if .Prompt }}<|user|> {{ .Prompt }}<|end|> {{ end }}<|assistant|> {{ .Response }}<|end|> """ PARAMETER stop """{"stop": ["<|end|>","<|user|>","<|assistant|>"]}""" SYSTEM """You are a friendly assistant called Chocolatine.""" ``` ### Limitations The Chocolatine model is a quick demonstration that a base model can be easily fine-tuned to achieve compelling performance. It does not have any moderation mechanism. - **Developed by:** Jonathan Pacifico, 2024 - **Model type:** LLM - **Language(s) (NLP):** French, English - **License:** MIT