Edit model card

Chocolatine-3B-Instruct-DPO-v1.2-Q4_K_M-GGUF

Quantized q4_k_m GGUF version of the original model jpacifico/Chocolatine-3B-Instruct-DPO-v1.2
can be used on a CPU device, compatible llama.cpp
now supported architecture by LM Studio.
Also ready for Raspberry Pi 5 8Gb.

Ollama

jpacifico/chocolatine-3b

Usage:

ollama run jpacifico/chocolatine-3b

Ollama Modelfile example :

FROM ./chocolatine-3b-instruct-dpo-v1.2-q4_k_m.gguf
TEMPLATE """{{ if .System }}<|system|>
{{ .System }}<|end|>
{{ end }}{{ if .Prompt }}<|user|>
{{ .Prompt }}<|end|>
{{ end }}<|assistant|>
{{ .Response }}<|end|>
"""
PARAMETER stop """{"stop": ["<|end|>","<|user|>","<|assistant|>"]}"""
SYSTEM """You are a friendly assistant called Chocolatine."""

Limitations

The Chocolatine model is a quick demonstration that a base model can be easily fine-tuned to achieve compelling performance.
It does not have any moderation mechanism.

  • Developed by: Jonathan Pacifico, 2024
  • Model type: LLM
  • Language(s) (NLP): French, English
  • License: MIT
Downloads last month
24
GGUF
Model size
3.82B params
Architecture
phi3

4-bit

Inference Examples
Inference API (serverless) is not available, repository is disabled.

Model tree for jpacifico/Chocolatine-3B-Instruct-DPO-v1.2-Q4_K_M-GGUF

Quantized
this model

Dataset used to train jpacifico/Chocolatine-3B-Instruct-DPO-v1.2-Q4_K_M-GGUF

Collection including jpacifico/Chocolatine-3B-Instruct-DPO-v1.2-Q4_K_M-GGUF