charioteer commited on
Commit
9de5aac
1 Parent(s): e32e32e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -9,16 +9,16 @@ pipeline_tag: text-generation
9
  ---
10
  # Model Card: Neural-phi2
11
 
12
- ![Poster Image](path/to/poster.png)
13
 
14
  ## Model Details
15
 
16
  - **Model Name**: Neural-phi2
17
  - **Model Type**: Large Language Model (LLM)
18
  - **Model Architecture**: A finetuned version of the Phi2 model from Microsoft, utilizing Direct Preference Optimization (DPO) on the `distilabel-intel-orca-dpo-pairs` dataset.
19
- - **Model Size**: Approximately 7B parameters
20
  - **Training Data**: The model was finetuned on the `distilabel-intel-orca-dpo-pairs` dataset, which consists of chat-like prompts and responses.
21
- - **Training Procedure**: The Phi2 model was finetuned using the DPO technique as described in the Jupyter notebook. The training process involved:
22
  - Loading and formatting the `distilabel-intel-orca-dpo-pairs` dataset
23
  - Defining the training configuration, including batch size, learning rate, and number of epochs
24
  - Initializing the DPO Trainer and training the model
 
9
  ---
10
  # Model Card: Neural-phi2
11
 
12
+ ![Poster Image](poster.png)
13
 
14
  ## Model Details
15
 
16
  - **Model Name**: Neural-phi2
17
  - **Model Type**: Large Language Model (LLM)
18
  - **Model Architecture**: A finetuned version of the Phi2 model from Microsoft, utilizing Direct Preference Optimization (DPO) on the `distilabel-intel-orca-dpo-pairs` dataset.
19
+ - **Model Size**: Approximately 2B parameters
20
  - **Training Data**: The model was finetuned on the `distilabel-intel-orca-dpo-pairs` dataset, which consists of chat-like prompts and responses.
21
+ - **Training Procedure**: The Phi2 model was finetuned using the DPO technique. The training process involved:
22
  - Loading and formatting the `distilabel-intel-orca-dpo-pairs` dataset
23
  - Defining the training configuration, including batch size, learning rate, and number of epochs
24
  - Initializing the DPO Trainer and training the model