farjvd commited on
Commit
64023d3
·
1 Parent(s): e0556d2

added the checkpoints

Browse files
.gitattributes CHANGED
@@ -33,3 +33,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ 1b-model filter=lfs diff=lfs merge=lfs -text
37
+ 3b-model filter=lfs diff=lfs merge=lfs -text
38
+ 70m-model/ filter=lfs diff=lfs merge=lfs -text
1b-model/best-model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d7b790ce2c915b22fe7175c5d15bae67bb1e8b121147f0f9c6f38f37288304a
3
+ size 5415539456
1b-model/collator_config.yml ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ do_mlm: true
2
+ num_bins: 51
3
+ sampling: true
4
+ pad_value: -2
5
+ data_style: both
6
+ do_binning: true
7
+ do_padding: true
8
+ mask_value: -1
9
+ max_length: 1024
10
+ right_binning: false
11
+ mlm_probability: 0.5
12
+ use_junk_tokens: false
13
+ pad_token_id: 0
1b-model/model_config.yml ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ mvc:
2
+ arch_style: inner product
3
+ query_activation: sigmoid
4
+ scaled_dot_product: true
5
+ name: vevo_scgpt
6
+ d_model: 2048
7
+ n_heads: 16
8
+ use_glu: false
9
+ n_layers: 24
10
+ attn_config:
11
+ attn_impl: triton
12
+ attn_type: grouped_query_attention
13
+ kv_nheads: 16
14
+ attn_pdrop: 0
15
+ init_device: cpu
16
+ norm_config:
17
+ eps: 1.0e-05
18
+ norm_type: layernorm
19
+ norm_scheme: pre
20
+ gene_encoder:
21
+ use_norm: true
22
+ cell_emb_style: cls
23
+ expansion_ratio: 4
24
+ expression_decoder:
25
+ n_layers: 1
26
+ n_outputs: 1
27
+ activation: leaky_relu
28
+ expression_encoder:
29
+ dropout: 0.1
30
+ use_norm: true
31
+ max_value: 512
32
+ activation: relu
33
+ input_emb_style: continuous
34
+ standard_scale_outputs: false
35
+ transformer_activation: relu
36
+ use_generative_training: false
37
+ use_cell_conditioned_generation: false
38
+ vocab_size: 62720
39
+ precision: amp_bf16
40
+ wandb_id: vevotx/vevo-scgpt/26iormxc
1b-model/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
3b-model/best-model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4c94fcc1f970688c44543888b9edc86fb26e0c9173e8c5591e07c99ccc04a64
3
+ size 10868338326
3b-model/collator_config.yml ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ do_padding: true
2
+ pad_value: -2
3
+ do_mlm: true
4
+ do_binning: true
5
+ mlm_probability: 0.5
6
+ mask_value: -1
7
+ max_length: 2048
8
+ sampling: true
9
+ data_style: both
10
+ num_bins: 51
11
+ right_binning: false
12
+ use_junk_tokens: false
13
+ use_chem_token: true
14
+ drug_to_id_path:
15
+ remote: s3://vevo-ml-datasets/mosaicfm_v2/datasets/drug_to_id_pad.json
16
+ local: drug_to_id_pad.json
17
+ keep_first_n_tokens: 2
18
+ pad_token_id: 0
3b-model/model_config.yml ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: mosaicfm
2
+ d_model: 2560
3
+ n_layers: 32
4
+ init_device: cpu
5
+ expansion_ratio: 4
6
+ standard_scale_outputs: false
7
+ transformer_activation: gelu
8
+ n_heads: 20
9
+ norm_scheme: pre
10
+ use_generative_training: false
11
+ use_cell_conditioned_generation: false
12
+ use_glu: false
13
+ cell_emb_style: cls
14
+ attn_config:
15
+ attn_impl: flash
16
+ use_attn_mask: false
17
+ attn_type: grouped_query_attention
18
+ kv_nheads: 20
19
+ attn_pdrop: 0.0
20
+ norm_config:
21
+ norm_type: layernorm
22
+ eps: 1.0e-05
23
+ expression_encoder:
24
+ input_emb_style: continuous
25
+ dropout: 0.1
26
+ max_value: 512
27
+ activation: gelu
28
+ use_norm: true
29
+ gene_encoder:
30
+ use_norm: true
31
+ mvc:
32
+ arch_style: inner product
33
+ query_activation: sigmoid
34
+ scaled_dot_product: true
35
+ expression_decoder:
36
+ n_outputs: 1
37
+ n_layers: 1
38
+ activation: gelu
39
+ chemical_encoder:
40
+ drug_fps_path:
41
+ remote: s3://vevo-ml-datasets/mosaicfm_v2/datasets/drug_fps_pad.npy
42
+ local: drug_fps_pad.npy
43
+ activation: gelu
44
+ padding_idx: 0
45
+ freeze: false
46
+ vocab_size: 62721
47
+ precision: amp_bf16
48
+ wandb_id: vevotx/vevo-MFM-v2/cdk744ih
3b-model/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
70m-model/best-model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0452ea29a2d81dbdac83ccb05a30ff327dc947b471248f3f5bddb9566661c152
3
+ size 284050604
70m-model/collator_config.yml ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ do_mlm: true
2
+ num_bins: 51
3
+ sampling: true
4
+ pad_value: -2
5
+ data_style: both
6
+ do_binning: true
7
+ do_padding: true
8
+ mask_value: -1
9
+ max_length: 1024
10
+ right_binning: false
11
+ mlm_probability: 0.5
12
+ use_junk_tokens: false
13
+ pad_token_id: 0
70m-model/model_config.yml ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ mvc:
2
+ arch_style: inner product
3
+ query_activation: sigmoid
4
+ scaled_dot_product: true
5
+ name: vevo_scgpt
6
+ d_model: 512
7
+ n_heads: 8
8
+ use_glu: false
9
+ n_layers: 12
10
+ attn_config:
11
+ attn_impl: triton
12
+ attn_type: grouped_query_attention
13
+ kv_nheads: 8
14
+ attn_pdrop: 0
15
+ init_device: cpu
16
+ norm_config:
17
+ eps: 1.0e-05
18
+ norm_type: layernorm
19
+ norm_scheme: pre
20
+ gene_encoder:
21
+ use_norm: true
22
+ cell_emb_style: cls
23
+ expansion_ratio: 4
24
+ expression_decoder:
25
+ n_layers: 1
26
+ n_outputs: 1
27
+ activation: leaky_relu
28
+ expression_encoder:
29
+ dropout: 0.1
30
+ use_norm: true
31
+ max_value: 512
32
+ activation: relu
33
+ input_emb_style: continuous
34
+ standard_scale_outputs: false
35
+ transformer_activation: relu
36
+ use_generative_training: false
37
+ use_cell_conditioned_generation: false
38
+ vocab_size: 62720
39
+ precision: amp_bf16
40
+ wandb_id: vevotx/vevo-scgpt/ftb65le8
70m-model/vocab.json ADDED
The diff for this file is too large to render. See raw diff