--- base_model: Qwen/Qwen2.5-Coder-0.5B datasets: None language: - en license: apache-2.0 tags: - text-generation-inference - transformers - torch - trl - unsloth - llama - gguf --- # Uploaded model - **Developed by:** student-abdullah - **License:** apache-2.0 - **Quantized from model:** Qwen2.5-Coder-0.5B - **Created on:** 06th July, 2025 --- # Acknowledgement
Qwen2ForCausalLM(
(model): Qwen2Model(
(embed_tokens): Embedding(151936, 896, padding_idx=151665)
(layers): ModuleList(
(0-23): 24 x Qwen2DecoderLayer(
(self_attn): Qwen2Attention(
(q_proj): Linear(in_features=896, out_features=896, bias=True)
(k_proj): Linear(in_features=896, out_features=128, bias=True)
(v_proj): Linear(in_features=896, out_features=128, bias=True)
(o_proj): Linear(in_features=896, out_features=896, bias=False)
(rotary_emb): LlamaRotaryEmbedding()
)
(mlp): Qwen2MLP(
(gate_proj): Linear(in_features=896, out_features=4864, bias=False)
(up_proj): Linear(in_features=896, out_features=4864, bias=False)
(down_proj): Linear(in_features=4864, out_features=896, bias=False)
(act_fn): SiLU()
)
(input_layernorm): Qwen2RMSNorm((896,), eps=1e-06)
(post_attention_layernorm): Qwen2RMSNorm((896,), eps=1e-06)
)
)
(norm): Qwen2RMSNorm((896,), eps=1e-06)
(rotary_emb): LlamaRotaryEmbedding()
)
(lm_head): Linear(in_features=896, out_features=151936, bias=False)
)
---
# Performance & Limitations
- YET TO BE EXAMINED
---
# Model Performace Evaluation:
- YET TO BE EVALUATED