Model Card for hacer201145/Hasex0.2-0.6B

Model Details

Description

Here comes Hasex0.2-0.6B. I deleted the previous version because it was worse. I also decided to change the base model from DialoGPT-medium (this is not an assistant, but rather a meme generator). But, okay. It was trained on ~9500 qualitative examples (including algebra, geometry, etc.). In the results, it shows better speed and responses, unlike its base model Qwen3-0.6B. Although it was trained in only one era, reasoning has become shorter in it, and it is more of an immediate essence than endless "hmm". I started collecting the dataset on Friday and finished it yesterday. After that, I was still tormented by how to get it, but it was definitely worth the result. To use it, I recommend using the example from the Qwen3-0.6B model card. It is very well represented there. But okay, I decided to take a break from it a bit, so expect Hasex0.3 in at least 3-4 months. It will already use this model as the main one. And I want to make a dataset of ~20,000 examples for it. Okay, thanks to everyone who read this.

Model Sources

  • Repository: Maybe soon?
  • Paper [optional]: Soon
  • Demo [optional]: Maybe soon?
Downloads last month
19
Safetensors
Model size
0.6B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for hacer201145/Hasex0.2-0.6B

Finetuned
Qwen/Qwen3-0.6B
Finetuned
(677)
this model
Quantizations
1 model