davanstrien HF Staff commited on
Commit
dfb9522
·
verified ·
1 Parent(s): e3bfdd4

Training in progress, step 50

Browse files
Files changed (4) hide show
  1. README.md +4 -4
  2. config.json +16 -16
  3. model.safetensors +2 -2
  4. training_args.bin +1 -1
README.md CHANGED
@@ -1,17 +1,17 @@
1
  ---
2
- base_model: LiquidAI/LFM2-VL-1.6B
3
  library_name: transformers
4
  model_name: lfm2-vl-iconclass
5
  tags:
6
  - generated_from_trainer
7
- - trl
8
  - sft
 
9
  licence: license
10
  ---
11
 
12
  # Model Card for lfm2-vl-iconclass
13
 
14
- This model is a fine-tuned version of [LiquidAI/LFM2-VL-1.6B](https://huggingface.co/LiquidAI/LFM2-VL-1.6B).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
@@ -27,7 +27,7 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/davanstrien/huggingface/runs/9vfvugcn)
31
 
32
 
33
  This model was trained with SFT.
 
1
  ---
2
+ base_model: LiquidAI/LFM2-VL-450M
3
  library_name: transformers
4
  model_name: lfm2-vl-iconclass
5
  tags:
6
  - generated_from_trainer
 
7
  - sft
8
+ - trl
9
  licence: license
10
  ---
11
 
12
  # Model Card for lfm2-vl-iconclass
13
 
14
+ This model is a fine-tuned version of [LiquidAI/LFM2-VL-450M](https://huggingface.co/LiquidAI/LFM2-VL-450M).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/davanstrien/huggingface/runs/0da6dcsh)
31
 
32
 
33
  This model was trained with SFT.
config.json CHANGED
@@ -12,7 +12,7 @@
12
  "image_token_index": 396,
13
  "max_image_tokens": 256,
14
  "max_num_patches": 1024,
15
- "max_pixels_tolerance": 1.5,
16
  "max_tiles": 10,
17
  "min_image_tokens": 64,
18
  "min_tiles": 2,
@@ -21,13 +21,13 @@
21
  "projector_hidden_act": "gelu",
22
  "projector_hidden_size": 2560,
23
  "text_config": {
24
- "_name_or_path": "LiquidAI/LFM2-1.2B",
25
  "architectures": [
26
  "Lfm2ForCausalLM"
27
  ],
28
  "block_auto_adjust_ff_dim": true,
29
- "block_dim": 2048,
30
- "block_ff_dim": 12288,
31
  "block_ffn_dim_multiplier": 1.0,
32
  "block_mlp_init_scale": 1.0,
33
  "block_multiple_of": 256,
@@ -37,13 +37,13 @@
37
  "block_use_xavier_init": true,
38
  "conv_L_cache": 3,
39
  "conv_bias": false,
40
- "conv_dim": 2048,
41
- "conv_dim_out": 2048,
42
  "conv_use_xavier_init": true,
43
  "eos_token_id": 7,
44
- "hidden_size": 2048,
45
  "initializer_range": 0.02,
46
- "intermediate_size": 12288,
47
  "layer_types": [
48
  "conv",
49
  "conv",
@@ -65,8 +65,8 @@
65
  "max_position_embeddings": 128000,
66
  "model_type": "lfm2",
67
  "norm_eps": 1e-05,
68
- "num_attention_heads": 32,
69
- "num_heads": 32,
70
  "num_hidden_layers": 16,
71
  "num_key_value_heads": 8,
72
  "rope_theta": 1000000.0,
@@ -79,21 +79,21 @@
79
  "torch_dtype": "bfloat16",
80
  "transformers_version": "4.55.0",
81
  "use_image_special_tokens": true,
82
- "use_thumbnail": true,
83
  "vision_config": {
84
  "attention_dropout": 0.0,
85
  "hidden_act": "gelu_pytorch_tanh",
86
- "hidden_size": 1152,
87
- "intermediate_size": 4304,
88
  "layer_norm_eps": 1e-06,
89
  "model_type": "siglip2_vision_model",
90
- "num_attention_heads": 16,
91
  "num_channels": 3,
92
- "num_hidden_layers": 27,
93
  "num_patches": 256,
94
  "patch_size": 16,
95
  "torch_dtype": "bfloat16",
96
  "vision_use_head": false
97
  },
98
- "vision_feature_layer": -2
99
  }
 
12
  "image_token_index": 396,
13
  "max_image_tokens": 256,
14
  "max_num_patches": 1024,
15
+ "max_pixels_tolerance": 2.0,
16
  "max_tiles": 10,
17
  "min_image_tokens": 64,
18
  "min_tiles": 2,
 
21
  "projector_hidden_act": "gelu",
22
  "projector_hidden_size": 2560,
23
  "text_config": {
24
+ "_name_or_path": "LiquidAI/LFM2-350M",
25
  "architectures": [
26
  "Lfm2ForCausalLM"
27
  ],
28
  "block_auto_adjust_ff_dim": true,
29
+ "block_dim": 1024,
30
+ "block_ff_dim": 6656,
31
  "block_ffn_dim_multiplier": 1.0,
32
  "block_mlp_init_scale": 1.0,
33
  "block_multiple_of": 256,
 
37
  "block_use_xavier_init": true,
38
  "conv_L_cache": 3,
39
  "conv_bias": false,
40
+ "conv_dim": 1024,
41
+ "conv_dim_out": 1024,
42
  "conv_use_xavier_init": true,
43
  "eos_token_id": 7,
44
+ "hidden_size": 1024,
45
  "initializer_range": 0.02,
46
+ "intermediate_size": 6656,
47
  "layer_types": [
48
  "conv",
49
  "conv",
 
65
  "max_position_embeddings": 128000,
66
  "model_type": "lfm2",
67
  "norm_eps": 1e-05,
68
+ "num_attention_heads": 16,
69
+ "num_heads": 16,
70
  "num_hidden_layers": 16,
71
  "num_key_value_heads": 8,
72
  "rope_theta": 1000000.0,
 
79
  "torch_dtype": "bfloat16",
80
  "transformers_version": "4.55.0",
81
  "use_image_special_tokens": true,
82
+ "use_thumbnail": false,
83
  "vision_config": {
84
  "attention_dropout": 0.0,
85
  "hidden_act": "gelu_pytorch_tanh",
86
+ "hidden_size": 768,
87
+ "intermediate_size": 3072,
88
  "layer_norm_eps": 1e-06,
89
  "model_type": "siglip2_vision_model",
90
+ "num_attention_heads": 12,
91
  "num_channels": 3,
92
+ "num_hidden_layers": 12,
93
  "num_patches": 256,
94
  "patch_size": 16,
95
  "torch_dtype": "bfloat16",
96
  "vision_use_head": false
97
  },
98
+ "vision_feature_layer": -1
99
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ae00d8b556f6fac6ab9f616bd576008cc2d349d5fa02e497ff899a0b2b4d4351
3
- size 3169688328
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29b5cf0aa9e685c13492826cbb781caef44eeab998657271d12e2ea33297f169
3
+ size 901692416
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:63bf6315f8bebb183d9c173d0795709f60f13031db9034ecf9c1a14bb3c849d4
3
  size 6161
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c049c52fd40bedd7b5fabd7c160997c99db0c09fef4e049b41c801ff07e6dd0d
3
  size 6161