Darwin-Qwen3-4B
evolutionary algorithm 'Darwin A2AP' 🤔

openfree/Darwin-Qwen3-4B

This model is automatically merged using evolutionary algorithm 'Darwin A2AP' v3.2

Overview

This study introduces a new paradigm of AI model fusion. Traditional "model merging" techniques have been restricted to models of the same family (e.g., transformer-based LLMs). We transcend this limitation by proposing a method to directly collide and fuse the core representational structures (DNA) of entirely different species — such as transformers and diffusion models. This approach acts as an "AI particle accelerator," colliding fundamentally distinct elements of intelligence to uncover new possibilities. The paper and source code (to be released on GitHub and Hugging Face) are currently under preparation and will be made publicly available soon. They will be released in a reproducible and extensible form for anyone to explore.

Contribution

Breaking the Species Barrier Fusion of fundamentally different models such as transformers and diffusion architectures. Realization of cross-species model merging once deemed impossible.

AI Embryo Creation

Formation of an initial “AI embryo” based on fused DNA. The embryo is not confined to a single domain or function but serves as the foundation for multi-capability intelligence.

Virtual Evolutionary Environment

AI embryos are placed into a simulated environment spanning thousands of generations. Through survival and adaptation, natural selection drives evolution beyond the limitations of parent models, producing new offspring models.

Merge Information

Father Model 1: Qwen/Qwen3-4B-Instruct-2507 Mother Model 2: Qwen/Qwen3-4B-Thinking-2507 Validation Task Accuracy: 88.56% Note: The above accuracy is a proxy metric used for merge ratio optimization. Algorithm Version: Darwin A2AP Enhanced v3.2

⚠️ Notice

The actual language generation performance of this model requires separate evaluation. The validation score above is not an LLM benchmark score.

⚠️ Benchmarking Test Results

Darwin-Qwen3-4B BenchMark Result

Use_Example

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("openfree/Darwin-Qwen3-4B")
tokenizer = AutoTokenizer.from_pretrained("openfree/Darwin-Qwen3-4B")

# 추론 예시
inputs = tokenizer("Hello, how are you?", return_tensors="pt")
outputs = model.generate(**inputs)

Strengths & Features

Cross-Domain Intelligence

Example: Legal LLM × Medical LLM → instantly produces a “Forensic LLM.” This is not mere knowledge aggregation but the creation of new intelligence at the intersection of domains.

Extreme Efficiency

Achieves results at roughly 1/10,000 of the time and cost compared to training a new foundation model. Accessible via a simple click-based process.

Unified Intelligence

Escapes confinement to a single domain by organically merging multiple expertises. Provides an experimental basis for integrated reasoning and creativity with AGI-like qualities.

Reproducibility & Openness

Source code and models will be fully released on GitHub and Hugging Face. Researchers and developers can freely reproduce, experiment, and expand.

Outlook

This research opens the door to a new generation of model creation, expressed as “Foundation a + Foundation b = Foundation abXc.” It represents far more than a reduction in training costs, serving as a critical turning point for future studies on the evolution and fusion of AI intelligence.

Downloads last month
27
Safetensors
Model size
4B params
Tensor type
F16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for openfree/Darwin-Qwen3-4B

Spaces using openfree/Darwin-Qwen3-4B 2

Collection including openfree/Darwin-Qwen3-4B