Text Generation
Transformers
Safetensors
English
olmo
conversational
Inference Endpoints
hamishivi commited on
Commit
4a6943b
1 Parent(s): 6553c84

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -30,8 +30,8 @@ They show the performance gain that OLMo base models can achieve with existing f
30
  We release two adapted model versions:
31
  | Model | Training Method(s) | Datasets | Context Length |
32
  |------|--------|---------|--|
33
- | [OLMo 7B July 2024 SFT](https://huggingface.co/allenai/OLMo-1.7-7B-Nitro-SFT-hf) | SFT | [Tulu 2 SFT Mix](https://huggingface.co/datasets/allenai/tulu-v2-sft-mixture-olmo-4096) | 2048 |
34
- | [OLMo 7B July 2024 Instruct](https://huggingface.co/allenai/OLMo-1.7-7B-Nitro-Instruct-hf) | SFT + DPO | [Tulu 2 SFT Mix](https://huggingface.co/datasets/allenai/tulu-v2-sft-mixture-olmo-4096) + [Ultrafeedback Cleaned](https://huggingface.co/datasets/allenai/ultrafeedback_binarized_cleaned) | 2048 |
35
 
36
  These models are both trained on top of OLMo 7b July 2024:
37
  | Size | Training Tokens | Layers | Hidden Size | Attention Heads | Context Length |
 
30
  We release two adapted model versions:
31
  | Model | Training Method(s) | Datasets | Context Length |
32
  |------|--------|---------|--|
33
+ | [OLMo 7B July 2024 SFT](https://huggingface.co/allenai/OLMo-1.7-7B-Nitro-SFT-hf) | SFT | [Tulu 2 SFT Mix](https://huggingface.co/datasets/allenai/tulu-v2-sft-mixture-olmo-4096) | 4096 |
34
+ | [OLMo 7B July 2024 Instruct](https://huggingface.co/allenai/OLMo-1.7-7B-Nitro-Instruct-hf) | SFT + DPO | [Tulu 2 SFT Mix](https://huggingface.co/datasets/allenai/tulu-v2-sft-mixture-olmo-4096) + [Ultrafeedback Cleaned](https://huggingface.co/datasets/allenai/ultrafeedback_binarized_cleaned) | 4096 |
35
 
36
  These models are both trained on top of OLMo 7b July 2024:
37
  | Size | Training Tokens | Layers | Hidden Size | Attention Heads | Context Length |