MANZANO: A Simple and Scalable Unified Multimodal Model with a Hybrid Vision Tokenizer Paper • 2509.16197 • Published Sep 19 • 54
Granite 2.0 Code Models Collection A series of code models trained by IBM licensed under Apache 2.0 license. We release both the base pretrained and instruct models. • 23 items • Updated 11 days ago • 201
view article Article Saving Memory Using Padding-Free Transformer Layers during Finetuning Jun 11, 2024 • 20
view article Article DS-MoE: Making MoE Models More Efficient and Less Memory-Intensive Apr 9, 2024 • 30