gpt-oss-20b-Malaysian-Reasoning-SFT-v0.1

LoRA SFT openai/gpt-oss-20b on initial mesolitica/Malaysian-Reasoning

  1. Use kernels-community/vllm-flash-attn3 for Flash Attention 3 with Sink.
  2. All linear layers with rank 16 alpha 32.

Source code

Source code at https://github.com/Scicom-AI-Enterprise-Organization/small-ablation/blob/main/malaysian-reasoning/20b.sh

Acknowledgement

Special thanks to https://www.scitix.ai/ for H100 Node!

Downloads last month
81
Safetensors
Model size
2B params
Tensor type
BF16
·
U8
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Scicom-intl/gpt-oss-20b-Malaysian-Reasoning-SFT-v0.1

Base model

openai/gpt-oss-20b
Quantized
(142)
this model
Quantizations
2 models

Dataset used to train Scicom-intl/gpt-oss-20b-Malaysian-Reasoning-SFT-v0.1