anujsahani01 commited on
Commit
935737d
1 Parent(s): 6bd72d9

End of training

Browse files
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
- license: apache-2.0
3
- base_model: codeparrot/codeparrot-small
4
  tags:
5
  - generated_from_trainer
6
  model-index:
@@ -13,7 +13,7 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # PyLoomer_CodeParrot
15
 
16
- This model is a fine-tuned version of [codeparrot/codeparrot-small](https://huggingface.co/codeparrot/codeparrot-small) on an unknown dataset.
17
 
18
  ## Model description
19
 
@@ -33,8 +33,8 @@ More information needed
33
 
34
  The following hyperparameters were used during training:
35
  - learning_rate: 0.0001
36
- - train_batch_size: 6
37
- - eval_batch_size: 6
38
  - seed: 42
39
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
  - lr_scheduler_type: linear
 
1
  ---
2
+ license: bsd-3-clause
3
+ base_model: Salesforce/codegen-350M-mono
4
  tags:
5
  - generated_from_trainer
6
  model-index:
 
13
 
14
  # PyLoomer_CodeParrot
15
 
16
+ This model is a fine-tuned version of [Salesforce/codegen-350M-mono](https://huggingface.co/Salesforce/codegen-350M-mono) on an unknown dataset.
17
 
18
  ## Model description
19
 
 
33
 
34
  The following hyperparameters were used during training:
35
  - learning_rate: 0.0001
36
+ - train_batch_size: 2
37
+ - eval_batch_size: 2
38
  - seed: 42
39
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
  - lr_scheduler_type: linear
adapter_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "auto_mapping": null,
3
- "base_model_name_or_path": "codeparrot/codeparrot-small",
4
  "bias": "none",
5
  "fan_in_fan_out": false,
6
  "inference_mode": true,
@@ -14,7 +14,7 @@
14
  "r": 8,
15
  "revision": null,
16
  "target_modules": [
17
- "c_attn"
18
  ],
19
  "task_type": "CAUSAL_LM"
20
  }
 
1
  {
2
  "auto_mapping": null,
3
+ "base_model_name_or_path": "Salesforce/codegen-350M-mono",
4
  "bias": "none",
5
  "fan_in_fan_out": false,
6
  "inference_mode": true,
 
14
  "r": 8,
15
  "revision": null,
16
  "target_modules": [
17
+ "qkv_proj"
18
  ],
19
  "task_type": "CAUSAL_LM"
20
  }
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:90bfa4ff95e678a3605aa40c2b742757b30e0f87a64cd36aa89c2ef065777237
3
- size 1188025
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7fbfe3de320fee92686dae4ea51384102e7601e2b2c0241ebb5194094d82261
3
+ size 2635433
added_tokens.json CHANGED
@@ -1,7 +1,45 @@
1
  {
2
- "<ASSISTANT_TASK:>": 32771,
3
- "<END_TASK>": 32772,
4
- "<SYSTEM_TASK:>": 32769,
5
- "<USER_TASK:>": 32770,
6
- "<|PAD|>": 32768
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  }
 
1
  {
2
+ "\t\t": 50294,
3
+ "\t\t\t": 50293,
4
+ "\t\t\t\t": 50292,
5
+ "\t\t\t\t\t": 50291,
6
+ "\t\t\t\t\t\t": 50290,
7
+ "\t\t\t\t\t\t\t": 50289,
8
+ "\t\t\t\t\t\t\t\t": 50288,
9
+ "\t\t\t\t\t\t\t\t\t": 50287,
10
+ " ": 50286,
11
+ " ": 50285,
12
+ " ": 50284,
13
+ " ": 50283,
14
+ " ": 50282,
15
+ " ": 50281,
16
+ " ": 50280,
17
+ " ": 50279,
18
+ " ": 50278,
19
+ " ": 50277,
20
+ " ": 50276,
21
+ " ": 50275,
22
+ " ": 50274,
23
+ " ": 50273,
24
+ " ": 50272,
25
+ " ": 50271,
26
+ " ": 50270,
27
+ " ": 50269,
28
+ " ": 50268,
29
+ " ": 50267,
30
+ " ": 50266,
31
+ " ": 50265,
32
+ " ": 50264,
33
+ " ": 50263,
34
+ " ": 50262,
35
+ " ": 50261,
36
+ " ": 50260,
37
+ " ": 50259,
38
+ " ": 50258,
39
+ " ": 50257,
40
+ "<ASSISTANT_TASK:>": 50298,
41
+ "<END_TASK>": 50299,
42
+ "<SYSTEM_TASK:>": 50296,
43
+ "<USER_TASK:>": 50297,
44
+ "<|PAD|>": 50295
45
  }
merges.txt CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -9,8 +9,8 @@
9
  "bos_token": "<|endoftext|>",
10
  "clean_up_tokenization_spaces": true,
11
  "eos_token": "<|endoftext|>",
12
- "model_max_length": 5000,
13
  "pad_token": "<|PAD|>",
14
- "tokenizer_class": "GPT2Tokenizer",
15
  "unk_token": "<|endoftext|>"
16
  }
 
9
  "bos_token": "<|endoftext|>",
10
  "clean_up_tokenization_spaces": true,
11
  "eos_token": "<|endoftext|>",
12
+ "model_max_length": 2048,
13
  "pad_token": "<|PAD|>",
14
+ "tokenizer_class": "CodeGenTokenizer",
15
  "unk_token": "<|endoftext|>"
16
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c972704e61d011e855ba280a4c3b5b3b8cab5d2aaf47fbbb2c732b43b8c895a0
3
  size 3963
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:514277c1178436a679fc1c176f6553af2198dc0889010b1b9d9306673a3e8302
3
  size 3963
vocab.json CHANGED
The diff for this file is too large to render. See raw diff