--- license: apache-2.0 language: - en pretty_name: Tiny QA Benchmark (Original EN Core for TQB++) size_categories: - n<1K tags: - question-answering - evaluation - benchmark - toy-dataset - tqb++-core task_categories: - question-answering task_ids: - extractive-qa - closed-book-qa arxiv: 2505.12058 datasets: - vincentkoc/tiny_qa_benchmark_pp --- # Tiny QA Benchmark (Original English Core for TQB++) **This dataset (`vincentkoc/tiny_qa_benchmark`) is the original 52-item English Question-Answering set. It now serves as the immutable "gold standard" core for the expanded [Tiny QA Benchmark++ (TQB++)](https://huggingface.co/datasets/vincentkoc/tiny_qa_benchmark_pp) project.** The TQB++ project builds upon this core dataset by introducing a powerful synthetic generation toolkit, pre-built multilingual datasets, and a comprehensive framework for rapid LLM smoke testing. **For the full TQB++ toolkit, the latest research paper, multilingual datasets, and the synthetic generator, please visit:** * **TQB++ Hugging Face Dataset Collection & Toolkit:** [vincentkoc/tiny_qa_benchmark_pp](https://huggingface.co/datasets/vincentkoc/tiny_qa_benchmark_pp) * **TQB++ GitHub Repository (Code, Paper & Toolkit):** [vincentkoc/tiny_qa_benchmark_pp](https://github.com/vincentkoc/tiny_qa_benchmark_pp) This original dataset (`vincentkoc/tiny_qa_benchmark`) contains 52 hand-crafted general-knowledge QA pairs covering geography, history, math, science, literature, and more. It remains ideal for quick sanity checks, pipeline smoke-tests, and as a foundational component of TQB++. Each example includes: - **text**: the question prompt - **label**: the “gold” answer - **metadata.context**: a one-sentence fact - **tags**: additional annotations (`category`, `difficulty`) It’s intentionally tiny (<100 KB) so you can iterate on data loading, evaluation scripts, or CI steps in under a second using these specific 52 items. ## Supported Tasks and Formats (for this core dataset) - **Tasks**: - Extractive QA - Generative QA - **Format**: JSON - **Splits**: - `train` (all 52 examples) ## Languages (for this core dataset) - English (`en`) ## Dataset Structure ### Data Fields Each example in `data/train.json` (as loaded by `datasets`) has: | field | type | description | |---------------------|--------|----------------------------------------------| | `text` | string | The question prompt. | | `label` | string | The correct answer. | | `metadata` | object | Additional info. | | `metadata.context` | string | A one-sentence fact supporting the answer. | | `tags.category` | string | Broad question category (e.g. `geography`). | | `tags.difficulty` | string | Rough difficulty level (e.g. `easy`). | ## Data Example ```json [ { "text": "What is the capital of France?", "label": "Paris", "metadata": { "context": "France is a country in Europe. Its capital is Paris." }, "tags": { "category": "geography", "difficulty": "easy" } }, { "text": "What is 2 + 2?", "label": "4", "metadata": { "context": "Basic arithmetic: 2 + 2 equals 4." }, "tags": { "category": "math", "difficulty": "easy" } } ] ``` *(Note: The actual file on the Hub might be a `.jsonl` file where each line is a JSON object, but `load_dataset` handles this.)* ## Data Splits Only one split for this core dataset: - **train**: 52 examples, used for development, quick evaluation, and as the TQB++ core. ## Data Creation ### Curation Rationale The "Tiny QA Benchmark" (this 52-item set) was originally created to: 1. Smoke-test QA pipelines (loading, preprocessing, evaluation). 2. Demo Hugging Face Datasets integration in tutorials. 3. Verify model–eval loops run without downloading large corpora. 4. **Serve as the immutable "gold standard" English core for the [Tiny QA Benchmark++ (TQB++)](https://huggingface.co/datasets/vincentkoc/tiny_qa_benchmark_pp) project.** ### Source Data Hand-crafted by the dataset creator from well-known, public-domain facts. It was initially developed as a dataset for sample projects to demonstrate [Opik](https://github.com/comet-ml/opik/) and now forms the foundational English core of TQB++. ### Annotations Self-annotated. Each `metadata.context` and `tags` field is manually created for these 52 items. ## Usage Load this specific 52-item core dataset with: ```python from datasets import load_dataset ds = load_dataset("vincentkoc/tiny_qa_benchmark") print(ds["train"][0]) # Expected output: # { # "text": "What is the capital of France?", # "label": "Paris", # "metadata": { # "context": "France is a country in Europe. Its capital is Paris." # }, # "tags": { # "category": "geography", # "difficulty": "easy" # } # } ``` For accessing the full TQB++ suite, including multilingual packs and the synthetic generator, refer to the [TQB++ Hugging Face Dataset Collection](https://huggingface.co/datasets/vincentkoc/tiny_qa_benchmark_pp). ## Considerations for Use * **Immutable Core for TQB++:** This dataset is the stable, hand-curated English core of the TQB++ project. Its 52 items are not intended to change. * **Not a Comprehensive Benchmark (on its own):** While excellent for quick checks, these 52 items are too few for statistically significant model ranking. For broader evaluation, use in conjunction with the TQB++ synthetic generator and its multilingual capabilities found at [vincentkoc/tiny_qa_benchmark_pp](https://huggingface.co/datasets/vincentkoc/tiny_qa_benchmark_pp). * **Do Not Train:** Primarily intended for evaluation, smoke-tests, or demos. * **No Sensitive Data:** All facts are public domain. ## Licensing Apache-2.0. See the `LICENSE` file in the [TQB++ GitHub repository](https://github.com/vincentkoc/tiny_qa_benchmark_pp) for details (as this dataset is now part of that larger project). ## Citation If you use this specific 52-item core English dataset, please cite it. You can use the following BibTeX entry, which has been updated to reflect its role: ```bibtex @misc{koctinyqabenchmark_original_core, author = { Vincent Koc }, title = { Tiny QA Benchmark (Original 52-item English Core for TQB++) }, year = 2025, url = { https://huggingface.co/datasets/vincentkoc/tiny_qa_benchmark }, doi = { 10.57967/hf/5417 }, publisher = { Hugging Face } } ``` For the complete **Tiny QA Benchmark++ (TQB++)** project (which includes this core set, the synthetic generator, multilingual packs, and the associated research paper), please refer to and cite the TQB++ project directly: ```bibtex @misc{koctinyqabenchmark_pp_dataset, author = {Vincent Koc}, title = {Tiny QA Benchmark++ (TQB++) Datasets and Toolkit}, year = {2025}, publisher = {Hugging Face & GitHub}, doi = {10.57967/hf/5531}, /* DOI for the TQB++ collection */ howpublished = {\\url{https://huggingface.co/datasets/vincentkoc/tiny_qa_benchmark_pp}}, note = {See also: \\url{https://github.com/vincentkoc/tiny_qa_benchmark_pp}} } ```