danelcsb commited on
Commit
efeda1d
·
verified ·
1 Parent(s): a7af02c

Upload model

Browse files
Files changed (2) hide show
  1. config.json +175 -0
  2. model.safetensors +3 -0
config.json ADDED
@@ -0,0 +1,175 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_all_frames_to_correct_as_cond": false,
3
+ "architectures": [
4
+ "Sam2Model"
5
+ ],
6
+ "backbone_stride": 16,
7
+ "binarize_mask_from_pts_for_mem_enc": true,
8
+ "enable_occlusion_spatial_embedding": true,
9
+ "enable_temporal_pos_encoding_for_object_pointers": true,
10
+ "fill_hole_area": 8,
11
+ "image_encoder_config": {
12
+ "backbone_channel_list": [
13
+ 896,
14
+ 448,
15
+ 224,
16
+ 112
17
+ ],
18
+ "backbone_feature_sizes": [
19
+ [
20
+ 256,
21
+ 256
22
+ ],
23
+ [
24
+ 128,
25
+ 128
26
+ ],
27
+ [
28
+ 64,
29
+ 64
30
+ ]
31
+ ],
32
+ "dim_mul": 2.0,
33
+ "drop_path_rate": 0.0,
34
+ "fpn_hidden_size": 256,
35
+ "fpn_interpolation_mode": "nearest",
36
+ "fpn_kernel_size": 1,
37
+ "fpn_padding": 0,
38
+ "fpn_stride": 1,
39
+ "fpn_top_down_levels": [
40
+ 2,
41
+ 3
42
+ ],
43
+ "fuse_type": "sum",
44
+ "global_attention_blocks": [
45
+ 12,
46
+ 16,
47
+ 20
48
+ ],
49
+ "head_mul": 2.0,
50
+ "hidden_act": "gelu",
51
+ "hidden_size": 112,
52
+ "image_size": 1024,
53
+ "layer_norm_eps": 1e-06,
54
+ "model_type": "",
55
+ "num_channels": 3,
56
+ "num_feature_levels": 3,
57
+ "num_heads": 2,
58
+ "patch_kernel_size": 7,
59
+ "patch_padding": 3,
60
+ "patch_stride": 4,
61
+ "q_pool": 3,
62
+ "q_stride": [
63
+ 2,
64
+ 2
65
+ ],
66
+ "stages": [
67
+ 2,
68
+ 3,
69
+ 16,
70
+ 3
71
+ ],
72
+ "window_positional_embedding_background_size": [
73
+ 14,
74
+ 14
75
+ ],
76
+ "window_spec": [
77
+ 8,
78
+ 4,
79
+ 14,
80
+ 7
81
+ ]
82
+ },
83
+ "image_size": 1024,
84
+ "initializer_range": 0.02,
85
+ "iou_prediction_use_sigmoid": true,
86
+ "mask_decoder_config": {
87
+ "dynamic_multimask_stability_delta": 0.05,
88
+ "dynamic_multimask_stability_thresh": 0.98,
89
+ "dynamic_multimask_via_stability": true,
90
+ "feed_forward_hidden_act": "relu",
91
+ "hidden_act": "gelu",
92
+ "hidden_size": 256,
93
+ "iou_head_depth": 3,
94
+ "iou_head_hidden_dim": 256,
95
+ "iou_prediction_use_sigmoid": true,
96
+ "model_type": "",
97
+ "num_multimask_outputs": 3,
98
+ "two_way_transformer_activation": "relu",
99
+ "two_way_transformer_attention_downsample_rate": 2,
100
+ "two_way_transformer_depth": 2,
101
+ "two_way_transformer_embedding_dim": 256,
102
+ "two_way_transformer_mlp_dim": 2048,
103
+ "two_way_transformer_num_heads": 8,
104
+ "use_multimask_token_for_object_pointer": true
105
+ },
106
+ "max_cond_frames_in_attn": -1,
107
+ "max_object_pointers_in_encoder": 16,
108
+ "memory_attention_config": {
109
+ "apply_pe_at_cross_attn_keys": true,
110
+ "apply_pe_at_cross_attn_queries": false,
111
+ "apply_pe_at_self_attn": false,
112
+ "dim_feedforward": 2048,
113
+ "dropout": 0.1,
114
+ "hidden_act": "relu",
115
+ "hidden_size": 256,
116
+ "model_type": "",
117
+ "num_layers": 4,
118
+ "rope_downsample_rate": 1,
119
+ "rope_dropout": 0.1,
120
+ "rope_embedding_dim": 256,
121
+ "rope_feat_sizes": [
122
+ 32,
123
+ 32
124
+ ],
125
+ "rope_num_heads": 1,
126
+ "rope_theta": 10000
127
+ },
128
+ "memory_encoder_config": {
129
+ "hidden_size": 256,
130
+ "mask_downsampler_embed_dim": 256,
131
+ "mask_downsampler_hidden_act": "gelu",
132
+ "mask_downsampler_kernel_size": 3,
133
+ "mask_downsampler_padding": 1,
134
+ "mask_downsampler_stride": 2,
135
+ "mask_downsampler_total_stride": 16,
136
+ "memory_fuser_embed_dim": 256,
137
+ "memory_fuser_hidden_act": "gelu",
138
+ "memory_fuser_kernel_size": 7,
139
+ "memory_fuser_layer_scale_init_value": 1e-06,
140
+ "memory_fuser_num_layers": 2,
141
+ "memory_fuser_padding": 3,
142
+ "memory_fuser_use_depthwise_conv": true,
143
+ "model_type": "",
144
+ "output_channels": 64
145
+ },
146
+ "memory_temporal_stride_for_eval": 1,
147
+ "model_type": "sam2",
148
+ "multimask_max_pt_num": 1,
149
+ "multimask_min_pt_num": 0,
150
+ "multimask_output_for_tracking": true,
151
+ "multimask_output_in_sam": true,
152
+ "non_overlap_masks": false,
153
+ "non_overlap_masks_for_mem_enc": false,
154
+ "num_maskmem": 7,
155
+ "preserve_temporal_direction_in_object_pointers": true,
156
+ "project_temporal_pos_encoding_in_object_pointers": true,
157
+ "prompt_encoder_config": {
158
+ "hidden_act": "gelu",
159
+ "hidden_size": 256,
160
+ "image_size": 1024,
161
+ "layer_norm_eps": 1e-06,
162
+ "mask_input_channels": 16,
163
+ "model_type": "",
164
+ "num_point_embeddings": 4,
165
+ "patch_size": 16,
166
+ "scale": 1
167
+ },
168
+ "sigmoid_bias_for_mem_enc": -10.0,
169
+ "sigmoid_scale_for_mem_enc": 20.0,
170
+ "torch_dtype": "float32",
171
+ "transformers_version": "4.53.0.dev0",
172
+ "use_mask_input_as_output_without_sam": true,
173
+ "use_multimask_token_for_object_pointer": true,
174
+ "use_object_pointers_in_encoder": true
175
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ab1d245fe351ff34d14022acd13a6e3dc0bc0fe18287a7aa09a740fcb3b2ac0
3
+ size 323473016