Datasets:
The dataset viewer is not available for this dataset.
Error code: RetryableConfigNamesError
Exception: ConnectionError
Message: Couldn't reach 'DATA-MASK/FineWeb-Mask' on the Hub (LocalEntryNotFoundError)
Traceback: Traceback (most recent call last):
File "/src/services/worker/src/worker/job_runners/dataset/config_names.py", line 66, in compute_config_names_response
config_names = get_dataset_config_names(
^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/inspect.py", line 161, in get_dataset_config_names
dataset_module = dataset_module_factory(
^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/load.py", line 1031, in dataset_module_factory
raise e1 from None
File "/usr/local/lib/python3.12/site-packages/datasets/load.py", line 953, in dataset_module_factory
raise ConnectionError(f"Couldn't reach '{path}' on the Hub ({e.__class__.__name__})") from e
ConnectionError: Couldn't reach 'DATA-MASK/FineWeb-Mask' on the Hub (LocalEntryNotFoundError)Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
FineWeb-Mask
π DATAMASK Paper | π» GitHub Repository | π¦ Fineweb-Mask Dataset
π Introduction
FineWeb-Mask is a 1.5 trillion token, high-efficiency pre-training dataset curated using the DATAMASK framework. Developed by the ByteDance Seed team, DATAMASK addresses the fundamental tension in large-scale data selection: the trade-off between high quality and high diversity.
By modeling data selection as a Mask Learning problem, we provide a derivative of the original FineWeb corpus. FineWeb-Mask is designed to eliminate semantic redundancy while preserving the highest quality samples, allowing models to achieve superior performance with significantly less data.
π― The Problem: The Quality-Diversity Trap
In large language model (LLM) pre-training, developers usually face two suboptimal choices:
- The Quality Trap: Filtering solely by quality scores leads to "diminishing returns." Samples become highly clustered, resulting in severe semantic redundancy.
- The Diversity Trap: Filtering solely for diversity often discards high-value quality samples, leading to worse performance than the original raw dataset.
- The Compute Bottleneck: Traditional diversity algorithms (like greedy selection) are computationally prohibitive for trillion-token datasets.
π‘ Highlights: The DATAMASK Framework
DATAMASK breaks this deadlock through a "joint harvesting" strategy:
- Joint Optimization: Uses Policy Gradient algorithms to optimize both quality and diversity metrics within a unified framework.
- Extreme Acceleration: Through probability relaxation and specialized optimization techniques, DATAMASK reduces computation time by 98.9% compared to traditional greedy algorithms, making trillion-token selection feasible.
- The "Balancer": Includes a tunable parameter that allows developers to define the "Golden Ratio" between quality and diversity for their specific needs.
- Semantic De-redundancy: Visual analysis shows that FineWeb-Mask samples are distributed evenly across high-quality regions rather than being rigidly clustered.
π Evaluation Results
FineWeb-Mask demonstrates that 1+1 > 2. By selecting a subset that represents only ~10% of the original scale in specific experiments, we observed:
- Dense Models: A 3.2% average improvement across 12 benchmarks for 1.5B dense models.
- MoE Models: A 1.9% improvement for 7B Mixture-of-Experts (MoE) models.
- Length Bias Correction: While quality filters favor long text and diversity filters favor short text, DATAMASK finds a scientific middle ground.
| Model Size | Dataset | Avg. Score (12 Benchmarks) | Improvement |
|---|---|---|---|
| 1.5B Dense | FineWeb (Original) | Baseline | - |
| 1.5B Dense | FineWeb-Mask | +3.2% | π |
| 7B MoE | FineWeb (Original) | Baseline | - |
| 7B MoE | FineWeb-Mask | +1.9% | π |
β€οΈ Acknowledgements
FineWeb-Mask is built upon the incredible foundational work of the HuggingFace FineWeb team. We are grateful to the open-source community for providing the raw corpora that made this optimization possible.
π Citation
If you find our dataset or the DATAMASK framework useful, please cite our work:
@misc{fan2025jointselectionlargescalepretraining,
title={Joint Selection for Large-Scale Pre-Training Data via Policy Gradient-based Mask Learning},
author={Ziqing Fan and Yuqiao Xian and Yan Sun and Li Shen},
year={2025},
eprint={2512.24265},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2512.24265},
}
π³ License
This dataset is released under the Apache 2.0 license. Users should also adhere to the original license terms of the FineWeb dataset and its constituent sources.
π§ Contact
- Ziqing Fan: [email protected]
- Yuqiao Xian: [email protected]
Would you like me to help you draft the "How to Use" section for loading this dataset via the Hugging Face datasets library?
- Downloads last month
- 14