Dataset Viewer
The dataset viewer is not available for this dataset.
Cannot get the config names for the dataset.
Error code:   RetryableConfigNamesError
Exception:    ConnectionError
Message:      Couldn't reach 'DATA-MASK/FineWeb-Mask' on the Hub (LocalEntryNotFoundError)
Traceback:    Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/dataset/config_names.py", line 66, in compute_config_names_response
                  config_names = get_dataset_config_names(
                                 ^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/inspect.py", line 161, in get_dataset_config_names
                  dataset_module = dataset_module_factory(
                                   ^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/load.py", line 1031, in dataset_module_factory
                  raise e1 from None
                File "/usr/local/lib/python3.12/site-packages/datasets/load.py", line 953, in dataset_module_factory
                  raise ConnectionError(f"Couldn't reach '{path}' on the Hub ({e.__class__.__name__})") from e
              ConnectionError: Couldn't reach 'DATA-MASK/FineWeb-Mask' on the Hub (LocalEntryNotFoundError)

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

FineWeb-Mask

πŸ“œ DATAMASK Paper | πŸ’» GitHub Repository | πŸ“¦ Fineweb-Mask Dataset

πŸ“š Introduction

FineWeb-Mask is a 1.5 trillion token, high-efficiency pre-training dataset curated using the DATAMASK framework. Developed by the ByteDance Seed team, DATAMASK addresses the fundamental tension in large-scale data selection: the trade-off between high quality and high diversity.

By modeling data selection as a Mask Learning problem, we provide a derivative of the original FineWeb corpus. FineWeb-Mask is designed to eliminate semantic redundancy while preserving the highest quality samples, allowing models to achieve superior performance with significantly less data.

🎯 The Problem: The Quality-Diversity Trap

In large language model (LLM) pre-training, developers usually face two suboptimal choices:

  1. The Quality Trap: Filtering solely by quality scores leads to "diminishing returns." Samples become highly clustered, resulting in severe semantic redundancy.
  2. The Diversity Trap: Filtering solely for diversity often discards high-value quality samples, leading to worse performance than the original raw dataset.
  3. The Compute Bottleneck: Traditional diversity algorithms (like greedy selection) are computationally prohibitive for trillion-token datasets.

πŸ’‘ Highlights: The DATAMASK Framework

DATAMASK breaks this deadlock through a "joint harvesting" strategy:

  • Joint Optimization: Uses Policy Gradient algorithms to optimize both quality and diversity metrics within a unified framework.
  • Extreme Acceleration: Through probability relaxation and specialized optimization techniques, DATAMASK reduces computation time by 98.9% compared to traditional greedy algorithms, making trillion-token selection feasible.
  • The "Balancer": Includes a tunable parameter that allows developers to define the "Golden Ratio" between quality and diversity for their specific needs.
  • Semantic De-redundancy: Visual analysis shows that FineWeb-Mask samples are distributed evenly across high-quality regions rather than being rigidly clustered.

πŸ“ˆ Evaluation Results

FineWeb-Mask demonstrates that 1+1 > 2. By selecting a subset that represents only ~10% of the original scale in specific experiments, we observed:

  • Dense Models: A 3.2% average improvement across 12 benchmarks for 1.5B dense models.
  • MoE Models: A 1.9% improvement for 7B Mixture-of-Experts (MoE) models.
  • Length Bias Correction: While quality filters favor long text and diversity filters favor short text, DATAMASK finds a scientific middle ground.
Model Size Dataset Avg. Score (12 Benchmarks) Improvement
1.5B Dense FineWeb (Original) Baseline -
1.5B Dense FineWeb-Mask +3.2% πŸš€
7B MoE FineWeb (Original) Baseline -
7B MoE FineWeb-Mask +1.9% πŸš€

❀️ Acknowledgements

FineWeb-Mask is built upon the incredible foundational work of the HuggingFace FineWeb team. We are grateful to the open-source community for providing the raw corpora that made this optimization possible.

🌟 Citation

If you find our dataset or the DATAMASK framework useful, please cite our work:

@misc{fan2025jointselectionlargescalepretraining,
      title={Joint Selection for Large-Scale Pre-Training Data via Policy Gradient-based Mask Learning}, 
      author={Ziqing Fan and Yuqiao Xian and Yan Sun and Li Shen},
      year={2025},
      eprint={2512.24265},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2512.24265}, 
}

πŸ’³ License

This dataset is released under the Apache 2.0 license. Users should also adhere to the original license terms of the FineWeb dataset and its constituent sources.

πŸ“§ Contact


Would you like me to help you draft the "How to Use" section for loading this dataset via the Hugging Face datasets library?

Downloads last month
14

Paper for DATA-MASK/FineWeb-Mask