PranavBP525 commited on
Commit
5101ee9
·
verified ·
1 Parent(s): 43a7d3e

phi-2-storygen-reward-model-v2

Browse files
README.md CHANGED
@@ -36,7 +36,7 @@ More information needed
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 5e-05
39
- - train_batch_size: 8
40
  - eval_batch_size: 8
41
  - seed: 42
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
@@ -49,8 +49,8 @@ The following hyperparameters were used during training:
49
 
50
  ### Framework versions
51
 
52
- - PEFT 0.10.0
53
- - Transformers 4.39.3
54
- - Pytorch 2.1.2
55
  - Datasets 2.18.0
56
- - Tokenizers 0.15.2
 
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 5e-05
39
+ - train_batch_size: 32
40
  - eval_batch_size: 8
41
  - seed: 42
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
 
49
 
50
  ### Framework versions
51
 
52
+ - PEFT 0.10.1.dev0
53
+ - Transformers 4.41.0.dev0
54
+ - Pytorch 2.2.2+cu121
55
  - Datasets 2.18.0
56
+ - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -14,14 +14,17 @@
14
  "lora_dropout": 0.05,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
- "modules_to_save": null,
 
 
 
18
  "peft_type": "LORA",
19
  "r": 4,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "v_proj",
24
- "q_proj"
25
  ],
26
  "task_type": "SEQ_CLS",
27
  "use_dora": false,
 
14
  "lora_dropout": 0.05,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
+ "modules_to_save": [
18
+ "classifier",
19
+ "score"
20
+ ],
21
  "peft_type": "LORA",
22
  "r": 4,
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "q_proj",
27
+ "v_proj"
28
  ],
29
  "task_type": "SEQ_CLS",
30
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ee531d9d73100221be4efc2d5ae4a30e97fe5128661fd7e327977b93bd449c5d
3
  size 5270056
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce151f599bbead7df3101c683be03ff97216b70384daedaae270261f21f5b5e0
3
  size 5270056
tokenizer.json CHANGED
@@ -396,6 +396,7 @@
396
  "end_of_word_suffix": "",
397
  "fuse_unk": false,
398
  "byte_fallback": false,
 
399
  "vocab": {
400
  "!": 0,
401
  "\"": 1,
 
396
  "end_of_word_suffix": "",
397
  "fuse_unk": false,
398
  "byte_fallback": false,
399
+ "ignore_merges": false,
400
  "vocab": {
401
  "!": 0,
402
  "\"": 1,
tokenizer_config.json CHANGED
@@ -319,6 +319,7 @@
319
  "eos_token": "<|endoftext|>",
320
  "model_max_length": 2048,
321
  "pad_token": "<|endoftext|>",
 
322
  "tokenizer_class": "CodeGenTokenizer",
323
  "unk_token": "<|endoftext|>"
324
  }
 
319
  "eos_token": "<|endoftext|>",
320
  "model_max_length": 2048,
321
  "pad_token": "<|endoftext|>",
322
+ "return_token_type_ids": false,
323
  "tokenizer_class": "CodeGenTokenizer",
324
  "unk_token": "<|endoftext|>"
325
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:75be32db41c635c9a45d80f0b0471d79a1a517e6623d2269b4bb11e0a6ac3ee7
3
  size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6908f52ca677ab5f42c1a3faa15a34b053fe7a0d439cc8b29a1ad37726023d3
3
  size 4984