annay3294 commited on
Commit
7da757c
·
verified ·
1 Parent(s): 591c437

annay3294/llm-course-hw2-reward-model

Browse files
README.md CHANGED
@@ -2,15 +2,13 @@
2
  base_model: HuggingFaceTB/SmolLM-135M-Instruct
3
  datasets: HumanLLMs/Human-Like-DPO-Dataset
4
  library_name: transformers
5
- model_name: annay3294/llm-course-hw2-reward-model
6
  tags:
7
  - generated_from_trainer
8
- - trl
9
- - reward-trainer
10
  licence: license
11
  ---
12
 
13
- # Model Card for annay3294/llm-course-hw2-reward-model
14
 
15
  This model is a fine-tuned version of [HuggingFaceTB/SmolLM-135M-Instruct](https://huggingface.co/HuggingFaceTB/SmolLM-135M-Instruct) on the [HumanLLMs/Human-Like-DPO-Dataset](https://huggingface.co/datasets/HumanLLMs/Human-Like-DPO-Dataset) dataset.
16
  It has been trained using [TRL](https://github.com/huggingface/trl).
@@ -31,7 +29,7 @@ print(output["generated_text"])
31
 
32
 
33
 
34
- This model was trained with Reward.
35
 
36
  ### Framework versions
37
 
@@ -43,7 +41,16 @@ This model was trained with Reward.
43
 
44
  ## Citations
45
 
 
46
 
 
 
 
 
 
 
 
 
47
 
48
  Cite TRL as:
49
 
 
2
  base_model: HuggingFaceTB/SmolLM-135M-Instruct
3
  datasets: HumanLLMs/Human-Like-DPO-Dataset
4
  library_name: transformers
5
+ model_name: trainer_output
6
  tags:
7
  - generated_from_trainer
 
 
8
  licence: license
9
  ---
10
 
11
+ # Model Card for trainer_output
12
 
13
  This model is a fine-tuned version of [HuggingFaceTB/SmolLM-135M-Instruct](https://huggingface.co/HuggingFaceTB/SmolLM-135M-Instruct) on the [HumanLLMs/Human-Like-DPO-Dataset](https://huggingface.co/datasets/HumanLLMs/Human-Like-DPO-Dataset) dataset.
14
  It has been trained using [TRL](https://github.com/huggingface/trl).
 
29
 
30
 
31
 
32
+ This model was trained with PPO, a method introduced in [Fine-Tuning Language Models from Human Preferences](https://huggingface.co/papers/1909.08593).
33
 
34
  ### Framework versions
35
 
 
41
 
42
  ## Citations
43
 
44
+ Cite PPO as:
45
 
46
+ ```bibtex
47
+ @article{mziegler2019fine-tuning,
48
+ title = {{Fine-Tuning Language Models from Human Preferences}},
49
+ author = {Daniel M. Ziegler and Nisan Stiennon and Jeffrey Wu and Tom B. Brown and Alec Radford and Dario Amodei and Paul F. Christiano and Geoffrey Irving},
50
+ year = 2019,
51
+ eprint = {arXiv:1909.08593}
52
+ }
53
+ ```
54
 
55
  Cite TRL as:
56
 
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[PAD]": 49152
3
+ }
config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "architectures": [
3
- "LlamaForSequenceClassification"
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
@@ -9,14 +9,8 @@
9
  "head_dim": 64,
10
  "hidden_act": "silu",
11
  "hidden_size": 576,
12
- "id2label": {
13
- "0": "LABEL_0"
14
- },
15
  "initializer_range": 0.02,
16
  "intermediate_size": 1536,
17
- "label2id": {
18
- "LABEL_0": 0
19
- },
20
  "max_position_embeddings": 2048,
21
  "mlp_bias": false,
22
  "model_type": "llama",
 
1
  {
2
  "architectures": [
3
+ "LlamaForCausalLM"
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
 
9
  "head_dim": 64,
10
  "hidden_act": "silu",
11
  "hidden_size": 576,
 
 
 
12
  "initializer_range": 0.02,
13
  "intermediate_size": 1536,
 
 
 
14
  "max_position_embeddings": 2048,
15
  "mlp_bias": false,
16
  "model_type": "llama",
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "max_new_tokens": 40,
5
+ "pad_token_id": 2,
6
+ "transformers_version": "4.50.0"
7
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:65533307bfe7df64c014f52a265af4e616bacff0fc092ec1d58b095ff6148c7e
3
- size 538092792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc5f0f95f81a5d84f7e1b024f23fd5cc4e994ba5ceef9d5fb7c14a00a6da1a7c
3
+ size 538090408
runs/Mar30_13-05-30_b018ffff2fe4/events.out.tfevents.1743339932.b018ffff2fe4.1160.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b1ae5c36689ff0b9c2e9d13012ae8938272c094952aac3a9aadf4a01992827a
3
+ size 662995
runs/Mar30_13-43-25_b018ffff2fe4/events.out.tfevents.1743342208.b018ffff2fe4.1160.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18e6ddbcc7110de8aaac3486803c691c961d3e03c20e967e7db945ee85267987
3
+ size 4184
runs/Mar30_13-45-16_b018ffff2fe4/events.out.tfevents.1743342317.b018ffff2fe4.12110.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:102c2b510b23bb699afd8ba968156d4f7ad9085383d6439f1273e5ad17968ced
3
+ size 59102
special_tokens_map.json CHANGED
@@ -17,7 +17,13 @@
17
  "rstrip": false,
18
  "single_word": false
19
  },
20
- "pad_token": "<|im_end|>",
 
 
 
 
 
 
21
  "unk_token": {
22
  "content": "<|endoftext|>",
23
  "lstrip": false,
 
17
  "rstrip": false,
18
  "single_word": false
19
  },
20
+ "pad_token": {
21
+ "content": "[PAD]",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false
26
+ },
27
  "unk_token": {
28
  "content": "<|endoftext|>",
29
  "lstrip": false,
tokenizer.json CHANGED
@@ -155,6 +155,15 @@
155
  "rstrip": false,
156
  "normalized": false,
157
  "special": true
 
 
 
 
 
 
 
 
 
158
  }
159
  ],
160
  "normalizer": null,
 
155
  "rstrip": false,
156
  "normalized": false,
157
  "special": true
158
+ },
159
+ {
160
+ "id": 49152,
161
+ "content": "[PAD]",
162
+ "single_word": false,
163
+ "lstrip": false,
164
+ "rstrip": false,
165
+ "normalized": false,
166
+ "special": true
167
  }
168
  ],
169
  "normalizer": null,
tokenizer_config.json CHANGED
@@ -136,6 +136,14 @@
136
  "rstrip": false,
137
  "single_word": false,
138
  "special": true
 
 
 
 
 
 
 
 
139
  }
140
  },
141
  "additional_special_tokens": [
@@ -148,7 +156,8 @@
148
  "eos_token": "<|im_end|>",
149
  "extra_special_tokens": {},
150
  "model_max_length": 2048,
151
- "pad_token": "<|im_end|>",
 
152
  "tokenizer_class": "GPT2Tokenizer",
153
  "unk_token": "<|endoftext|>",
154
  "vocab_size": 49152
 
136
  "rstrip": false,
137
  "single_word": false,
138
  "special": true
139
+ },
140
+ "49152": {
141
+ "content": "[PAD]",
142
+ "lstrip": false,
143
+ "normalized": false,
144
+ "rstrip": false,
145
+ "single_word": false,
146
+ "special": true
147
  }
148
  },
149
  "additional_special_tokens": [
 
156
  "eos_token": "<|im_end|>",
157
  "extra_special_tokens": {},
158
  "model_max_length": 2048,
159
+ "pad_token": "[PAD]",
160
+ "padding_side": "left",
161
  "tokenizer_class": "GPT2Tokenizer",
162
  "unk_token": "<|endoftext|>",
163
  "vocab_size": 49152
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bea79283a2116ad3a2e77acd224a666ac52ded02dacf45e2c68f140f8a100355
3
- size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b23dcc5d19af379e6d711c48a1eb61cecbac5215bfe019cb5ec865fe7b89ebc7
3
+ size 6136