added the checkpoints
Browse files- .gitattributes +3 -0
- 1b-model/best-model.pt +3 -0
- 1b-model/collator_config.yml +13 -0
- 1b-model/model_config.yml +40 -0
- 1b-model/vocab.json +0 -0
- 3b-model/best-model.pt +3 -0
- 3b-model/collator_config.yml +18 -0
- 3b-model/model_config.yml +48 -0
- 3b-model/vocab.json +0 -0
- 70m-model/best-model.pt +3 -0
- 70m-model/collator_config.yml +13 -0
- 70m-model/model_config.yml +40 -0
- 70m-model/vocab.json +0 -0
.gitattributes
CHANGED
|
@@ -33,3 +33,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
1b-model filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
3b-model filter=lfs diff=lfs merge=lfs -text
|
| 38 |
+
70m-model/ filter=lfs diff=lfs merge=lfs -text
|
1b-model/best-model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2d7b790ce2c915b22fe7175c5d15bae67bb1e8b121147f0f9c6f38f37288304a
|
| 3 |
+
size 5415539456
|
1b-model/collator_config.yml
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
do_mlm: true
|
| 2 |
+
num_bins: 51
|
| 3 |
+
sampling: true
|
| 4 |
+
pad_value: -2
|
| 5 |
+
data_style: both
|
| 6 |
+
do_binning: true
|
| 7 |
+
do_padding: true
|
| 8 |
+
mask_value: -1
|
| 9 |
+
max_length: 1024
|
| 10 |
+
right_binning: false
|
| 11 |
+
mlm_probability: 0.5
|
| 12 |
+
use_junk_tokens: false
|
| 13 |
+
pad_token_id: 0
|
1b-model/model_config.yml
ADDED
|
@@ -0,0 +1,40 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
mvc:
|
| 2 |
+
arch_style: inner product
|
| 3 |
+
query_activation: sigmoid
|
| 4 |
+
scaled_dot_product: true
|
| 5 |
+
name: vevo_scgpt
|
| 6 |
+
d_model: 2048
|
| 7 |
+
n_heads: 16
|
| 8 |
+
use_glu: false
|
| 9 |
+
n_layers: 24
|
| 10 |
+
attn_config:
|
| 11 |
+
attn_impl: triton
|
| 12 |
+
attn_type: grouped_query_attention
|
| 13 |
+
kv_nheads: 16
|
| 14 |
+
attn_pdrop: 0
|
| 15 |
+
init_device: cpu
|
| 16 |
+
norm_config:
|
| 17 |
+
eps: 1.0e-05
|
| 18 |
+
norm_type: layernorm
|
| 19 |
+
norm_scheme: pre
|
| 20 |
+
gene_encoder:
|
| 21 |
+
use_norm: true
|
| 22 |
+
cell_emb_style: cls
|
| 23 |
+
expansion_ratio: 4
|
| 24 |
+
expression_decoder:
|
| 25 |
+
n_layers: 1
|
| 26 |
+
n_outputs: 1
|
| 27 |
+
activation: leaky_relu
|
| 28 |
+
expression_encoder:
|
| 29 |
+
dropout: 0.1
|
| 30 |
+
use_norm: true
|
| 31 |
+
max_value: 512
|
| 32 |
+
activation: relu
|
| 33 |
+
input_emb_style: continuous
|
| 34 |
+
standard_scale_outputs: false
|
| 35 |
+
transformer_activation: relu
|
| 36 |
+
use_generative_training: false
|
| 37 |
+
use_cell_conditioned_generation: false
|
| 38 |
+
vocab_size: 62720
|
| 39 |
+
precision: amp_bf16
|
| 40 |
+
wandb_id: vevotx/vevo-scgpt/26iormxc
|
1b-model/vocab.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
3b-model/best-model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b4c94fcc1f970688c44543888b9edc86fb26e0c9173e8c5591e07c99ccc04a64
|
| 3 |
+
size 10868338326
|
3b-model/collator_config.yml
ADDED
|
@@ -0,0 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
do_padding: true
|
| 2 |
+
pad_value: -2
|
| 3 |
+
do_mlm: true
|
| 4 |
+
do_binning: true
|
| 5 |
+
mlm_probability: 0.5
|
| 6 |
+
mask_value: -1
|
| 7 |
+
max_length: 2048
|
| 8 |
+
sampling: true
|
| 9 |
+
data_style: both
|
| 10 |
+
num_bins: 51
|
| 11 |
+
right_binning: false
|
| 12 |
+
use_junk_tokens: false
|
| 13 |
+
use_chem_token: true
|
| 14 |
+
drug_to_id_path:
|
| 15 |
+
remote: s3://vevo-ml-datasets/mosaicfm_v2/datasets/drug_to_id_pad.json
|
| 16 |
+
local: drug_to_id_pad.json
|
| 17 |
+
keep_first_n_tokens: 2
|
| 18 |
+
pad_token_id: 0
|
3b-model/model_config.yml
ADDED
|
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
name: mosaicfm
|
| 2 |
+
d_model: 2560
|
| 3 |
+
n_layers: 32
|
| 4 |
+
init_device: cpu
|
| 5 |
+
expansion_ratio: 4
|
| 6 |
+
standard_scale_outputs: false
|
| 7 |
+
transformer_activation: gelu
|
| 8 |
+
n_heads: 20
|
| 9 |
+
norm_scheme: pre
|
| 10 |
+
use_generative_training: false
|
| 11 |
+
use_cell_conditioned_generation: false
|
| 12 |
+
use_glu: false
|
| 13 |
+
cell_emb_style: cls
|
| 14 |
+
attn_config:
|
| 15 |
+
attn_impl: flash
|
| 16 |
+
use_attn_mask: false
|
| 17 |
+
attn_type: grouped_query_attention
|
| 18 |
+
kv_nheads: 20
|
| 19 |
+
attn_pdrop: 0.0
|
| 20 |
+
norm_config:
|
| 21 |
+
norm_type: layernorm
|
| 22 |
+
eps: 1.0e-05
|
| 23 |
+
expression_encoder:
|
| 24 |
+
input_emb_style: continuous
|
| 25 |
+
dropout: 0.1
|
| 26 |
+
max_value: 512
|
| 27 |
+
activation: gelu
|
| 28 |
+
use_norm: true
|
| 29 |
+
gene_encoder:
|
| 30 |
+
use_norm: true
|
| 31 |
+
mvc:
|
| 32 |
+
arch_style: inner product
|
| 33 |
+
query_activation: sigmoid
|
| 34 |
+
scaled_dot_product: true
|
| 35 |
+
expression_decoder:
|
| 36 |
+
n_outputs: 1
|
| 37 |
+
n_layers: 1
|
| 38 |
+
activation: gelu
|
| 39 |
+
chemical_encoder:
|
| 40 |
+
drug_fps_path:
|
| 41 |
+
remote: s3://vevo-ml-datasets/mosaicfm_v2/datasets/drug_fps_pad.npy
|
| 42 |
+
local: drug_fps_pad.npy
|
| 43 |
+
activation: gelu
|
| 44 |
+
padding_idx: 0
|
| 45 |
+
freeze: false
|
| 46 |
+
vocab_size: 62721
|
| 47 |
+
precision: amp_bf16
|
| 48 |
+
wandb_id: vevotx/vevo-MFM-v2/cdk744ih
|
3b-model/vocab.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
70m-model/best-model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0452ea29a2d81dbdac83ccb05a30ff327dc947b471248f3f5bddb9566661c152
|
| 3 |
+
size 284050604
|
70m-model/collator_config.yml
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
do_mlm: true
|
| 2 |
+
num_bins: 51
|
| 3 |
+
sampling: true
|
| 4 |
+
pad_value: -2
|
| 5 |
+
data_style: both
|
| 6 |
+
do_binning: true
|
| 7 |
+
do_padding: true
|
| 8 |
+
mask_value: -1
|
| 9 |
+
max_length: 1024
|
| 10 |
+
right_binning: false
|
| 11 |
+
mlm_probability: 0.5
|
| 12 |
+
use_junk_tokens: false
|
| 13 |
+
pad_token_id: 0
|
70m-model/model_config.yml
ADDED
|
@@ -0,0 +1,40 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
mvc:
|
| 2 |
+
arch_style: inner product
|
| 3 |
+
query_activation: sigmoid
|
| 4 |
+
scaled_dot_product: true
|
| 5 |
+
name: vevo_scgpt
|
| 6 |
+
d_model: 512
|
| 7 |
+
n_heads: 8
|
| 8 |
+
use_glu: false
|
| 9 |
+
n_layers: 12
|
| 10 |
+
attn_config:
|
| 11 |
+
attn_impl: triton
|
| 12 |
+
attn_type: grouped_query_attention
|
| 13 |
+
kv_nheads: 8
|
| 14 |
+
attn_pdrop: 0
|
| 15 |
+
init_device: cpu
|
| 16 |
+
norm_config:
|
| 17 |
+
eps: 1.0e-05
|
| 18 |
+
norm_type: layernorm
|
| 19 |
+
norm_scheme: pre
|
| 20 |
+
gene_encoder:
|
| 21 |
+
use_norm: true
|
| 22 |
+
cell_emb_style: cls
|
| 23 |
+
expansion_ratio: 4
|
| 24 |
+
expression_decoder:
|
| 25 |
+
n_layers: 1
|
| 26 |
+
n_outputs: 1
|
| 27 |
+
activation: leaky_relu
|
| 28 |
+
expression_encoder:
|
| 29 |
+
dropout: 0.1
|
| 30 |
+
use_norm: true
|
| 31 |
+
max_value: 512
|
| 32 |
+
activation: relu
|
| 33 |
+
input_emb_style: continuous
|
| 34 |
+
standard_scale_outputs: false
|
| 35 |
+
transformer_activation: relu
|
| 36 |
+
use_generative_training: false
|
| 37 |
+
use_cell_conditioned_generation: false
|
| 38 |
+
vocab_size: 62720
|
| 39 |
+
precision: amp_bf16
|
| 40 |
+
wandb_id: vevotx/vevo-scgpt/ftb65le8
|
70m-model/vocab.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|