Is the config.json wrong?

#4
by kkyeer - opened

The 4b-pt config.json is much same as 4b-it, and diffrent from 1b-pt version.For example,no vocab_size prop in 4b but can be found in 1b.Is this intended or a mistake?

Same problem here

Google org

Hi ,

Thank you for pointing this out! You're correct — the config.json for the 4b-pt and 4b-it versions appears to omit the vocab_size field, while it's present in the 1b-pt version.

This could be intentional if the tokenizer is assumed to be loaded independently and the model config relies on tokenizer metadata. In many recent transformer-based models, the vocab_size is not strictly required in config.json during inference. Thank you.

Sign up or log in to comment