rpv2-subsamples (#26)
Browse files- subsamples 10B, 100B, 1T (7f1cb5c88ab57ae523f1a540f3529219f59462f8)
- encoding (eb143a1e7e4429ce608c9e4046cb12584f3c7d22)
- .gitignore +32 -0
- RedPajama-Data-V2.py +94 -60
    	
        .gitignore
    ADDED
    
    | @@ -0,0 +1,32 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            # Byte-compiled / optimized / DLL files
         | 
| 2 | 
            +
            __pycache__/
         | 
| 3 | 
            +
            *.py[cod]
         | 
| 4 | 
            +
            *$py.class
         | 
| 5 | 
            +
            *.pyc
         | 
| 6 | 
            +
            *.DS_Store
         | 
| 7 | 
            +
             | 
| 8 | 
            +
            # data folders
         | 
| 9 | 
            +
            data/*
         | 
| 10 | 
            +
            !data/.gitkeep
         | 
| 11 | 
            +
             | 
| 12 | 
            +
            # notebooks
         | 
| 13 | 
            +
            notebooks/*
         | 
| 14 | 
            +
            .ipynb_checkpoints
         | 
| 15 | 
            +
             | 
| 16 | 
            +
            # Environments
         | 
| 17 | 
            +
            .env
         | 
| 18 | 
            +
            .venv
         | 
| 19 | 
            +
            env/
         | 
| 20 | 
            +
            venv/
         | 
| 21 | 
            +
            ENV/
         | 
| 22 | 
            +
            env.bak/
         | 
| 23 | 
            +
            venv.bak/
         | 
| 24 | 
            +
             | 
| 25 | 
            +
            # ides
         | 
| 26 | 
            +
            .idea/
         | 
| 27 | 
            +
            .vscode/
         | 
| 28 | 
            +
             | 
| 29 | 
            +
            # distribution
         | 
| 30 | 
            +
            *.egg-info/
         | 
| 31 | 
            +
            dist/
         | 
| 32 | 
            +
            build/
         | 
    	
        RedPajama-Data-V2.py
    CHANGED
    
    | @@ -15,13 +15,12 @@ | |
| 15 | 
             
            # Lint as: python3
         | 
| 16 | 
             
            """RedPajama V2: Quality annotated Web Text Documents."""
         | 
| 17 |  | 
|  | |
| 18 | 
             
            import json
         | 
| 19 | 
            -
             | 
| 20 | 
            -
            import datasets
         | 
| 21 | 
             
            import traceback
         | 
| 22 | 
            -
            import gzip
         | 
| 23 | 
             
            from typing import List
         | 
| 24 |  | 
|  | |
| 25 | 
             
            import pyarrow.parquet as pq
         | 
| 26 |  | 
| 27 | 
             
            logger = datasets.logging.get_logger(__name__)
         | 
| @@ -30,10 +29,11 @@ _DESCRIPTION = """\ | |
| 30 | 
             
            RedPajama V2: an Open Dataset for Training Large Language Models
         | 
| 31 | 
             
            """
         | 
| 32 |  | 
| 33 | 
            -
            _URL_BASE =  | 
| 34 | 
             
            _LANGUAGES = ("en", "de", "fr", "es", "it")
         | 
| 35 | 
             
            _MISSING_FILES_PATTERN = "urls/missing-{component}.txt"
         | 
| 36 | 
             
            _NUM_SHARDS = 5000
         | 
|  | |
| 37 |  | 
| 38 | 
             
            _CC_SNAPSHOT_IDS = (
         | 
| 39 | 
             
                "2014-15",
         | 
| @@ -119,7 +119,7 @@ _CC_SNAPSHOT_IDS = ( | |
| 119 | 
             
                "2022-40",
         | 
| 120 | 
             
                "2022-49",
         | 
| 121 | 
             
                "2023-06",
         | 
| 122 | 
            -
                "2023-14"
         | 
| 123 | 
             
            )
         | 
| 124 |  | 
| 125 |  | 
| @@ -138,19 +138,34 @@ class RedPajamaDataV2Config(datasets.BuilderConfig): | |
| 138 |  | 
| 139 |  | 
| 140 | 
             
            class RedPajamaV2(datasets.GeneratorBasedBuilder):
         | 
| 141 | 
            -
                """ | 
| 142 |  | 
| 143 | 
             
                BUILDER_CONFIGS = [
         | 
| 144 | 
             
                    RedPajamaDataV2Config(
         | 
| 145 | 
            -
                        name= | 
| 146 | 
             
                        version=datasets.Version("1.0.0", ""),
         | 
| 147 | 
             
                        description=f"RedPajamaV2 Sample",
         | 
| 148 | 
             
                    ),
         | 
| 149 | 
             
                    RedPajamaDataV2Config(
         | 
| 150 | 
            -
                        name= | 
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
| 151 | 
             
                        version=datasets.Version("1.0.0", ""),
         | 
| 152 | 
             
                        description=f"RedPajamaV2",
         | 
| 153 | 
            -
                    )
         | 
| 154 | 
             
                ]
         | 
| 155 |  | 
| 156 | 
             
                def _info(self):
         | 
| @@ -161,7 +176,7 @@ class RedPajamaV2(datasets.GeneratorBasedBuilder): | |
| 161 | 
             
                                "raw_content": datasets.Value("string"),
         | 
| 162 | 
             
                                "doc_id": datasets.Value("string"),
         | 
| 163 | 
             
                                "meta": datasets.Value("string"),
         | 
| 164 | 
            -
                                "quality_signals": datasets.Value("string")
         | 
| 165 | 
             
                            }
         | 
| 166 | 
             
                        ),
         | 
| 167 | 
             
                        supervised_keys=None,
         | 
| @@ -177,24 +192,30 @@ class RedPajamaV2(datasets.GeneratorBasedBuilder): | |
| 177 |  | 
| 178 | 
             
                    # fetch documents
         | 
| 179 | 
             
                    logger.info(f"Downloading {len(sample_base_tags)} documents files.")
         | 
| 180 | 
            -
                    documents_files = dl_manager.download( | 
| 181 | 
            -
                         | 
| 182 | 
            -
             | 
| 183 | 
            -
             | 
|  | |
|  | |
| 184 |  | 
| 185 | 
             
                    # fetch quality signals
         | 
| 186 | 
             
                    logger.info(f"Downloading {len(sample_base_tags)} quality signals files.")
         | 
| 187 | 
            -
                    quality_signals_files = dl_manager.download( | 
| 188 | 
            -
                         | 
| 189 | 
            -
             | 
| 190 | 
            -
             | 
|  | |
|  | |
| 191 |  | 
| 192 | 
             
                    # fetch ids of duplicates
         | 
| 193 | 
             
                    logger.info(f"Downloading {len(sample_base_tags)} duplicates ids files.")
         | 
| 194 | 
            -
                    duplicates_ids_files = dl_manager.download( | 
| 195 | 
            -
                         | 
| 196 | 
            -
             | 
| 197 | 
            -
             | 
|  | |
|  | |
| 198 |  | 
| 199 | 
             
                    return [
         | 
| 200 | 
             
                        datasets.SplitGenerator(
         | 
| @@ -203,35 +224,45 @@ class RedPajamaV2(datasets.GeneratorBasedBuilder): | |
| 203 | 
             
                                "base_tags": sample_base_tags,
         | 
| 204 | 
             
                                "documents_files": documents_files,
         | 
| 205 | 
             
                                "quality_signals_files": quality_signals_files,
         | 
| 206 | 
            -
                                "duplicates_ids_files": duplicates_ids_files
         | 
| 207 | 
            -
                            }
         | 
| 208 | 
             
                        )
         | 
| 209 | 
             
                    ]
         | 
| 210 |  | 
| 211 | 
             
                def _split_generators_full(self, dl_manager):
         | 
| 212 | 
            -
                    snapshots = getattr(self.config,  | 
| 213 | 
            -
                    languages = getattr(self.config,  | 
| 214 | 
            -
                    partition = getattr(self.config,  | 
| 215 | 
            -
             | 
| 216 | 
            -
                    if  | 
| 217 | 
            -
                         | 
| 218 | 
            -
             | 
| 219 | 
            -
                         | 
| 220 | 
            -
             | 
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
| 221 | 
             
                        partitions = [partition]
         | 
| 222 | 
             
                    else:
         | 
| 223 | 
            -
                        raise ValueError(f | 
| 224 |  | 
| 225 | 
             
                    # fetch list of missing files (e.g., missing duplicates or corrupted documents and
         | 
| 226 | 
             
                    # quality signal files)
         | 
| 227 | 
            -
                    missing_files_paths = dl_manager.download_and_extract( | 
| 228 | 
            -
                         | 
| 229 | 
            -
             | 
| 230 | 
            -
             | 
|  | |
|  | |
| 231 |  | 
| 232 | 
             
                    missing_files = {}
         | 
| 233 | 
             
                    for component, missing_file in missing_files_paths.items():
         | 
| 234 | 
            -
                        with open(missing_file) as f:
         | 
| 235 | 
             
                            missing_files[component] = set(line.strip() for line in f)
         | 
| 236 |  | 
| 237 | 
             
                    # build list of urls to fetch
         | 
| @@ -243,7 +274,7 @@ class RedPajamaV2(datasets.GeneratorBasedBuilder): | |
| 243 | 
             
                    for lang in languages:
         | 
| 244 | 
             
                        for snapshot in snapshots:
         | 
| 245 | 
             
                            for part in partitions:
         | 
| 246 | 
            -
                                for n in range( | 
| 247 | 
             
                                    base_tag = f"{snapshot}/{n:04d}/{lang}_{part}"
         | 
| 248 | 
             
                                    base_tags.append(base_tag)
         | 
| 249 |  | 
| @@ -281,20 +312,19 @@ class RedPajamaV2(datasets.GeneratorBasedBuilder): | |
| 281 | 
             
                                "base_tags": base_tags,
         | 
| 282 | 
             
                                "documents_files": documents_files,
         | 
| 283 | 
             
                                "quality_signals_files": quality_signals_files,
         | 
| 284 | 
            -
                                "duplicates_ids_files": duplicates_ids_files
         | 
| 285 | 
            -
                            }
         | 
| 286 | 
             
                        )
         | 
| 287 | 
             
                    ]
         | 
| 288 |  | 
| 289 | 
             
                def _split_generators(self, dl_manager):
         | 
| 290 | 
            -
                    if self.config.name | 
| 291 | 
             
                        return self._split_generators_sample(dl_manager)
         | 
| 292 |  | 
| 293 | 
             
                    return self._split_generators_full(dl_manager)
         | 
| 294 |  | 
| 295 | 
             
                def _generate_examples(
         | 
| 296 | 
            -
             | 
| 297 | 
            -
                        duplicates_ids_files
         | 
| 298 | 
             
                ):
         | 
| 299 | 
             
                    key = 0
         | 
| 300 | 
             
                    for base_tag in base_tags:
         | 
| @@ -305,9 +335,7 @@ class RedPajamaV2(datasets.GeneratorBasedBuilder): | |
| 305 | 
             
                        if doc_file is None:
         | 
| 306 | 
             
                            continue
         | 
| 307 |  | 
| 308 | 
            -
                        for sample in self.__get_generator(
         | 
| 309 | 
            -
                                base_tag, doc_file, qs_file, dupe_file
         | 
| 310 | 
            -
                        ):
         | 
| 311 | 
             
                            yield key, sample
         | 
| 312 | 
             
                            key += 1
         | 
| 313 |  | 
| @@ -325,19 +353,17 @@ class RedPajamaV2(datasets.GeneratorBasedBuilder): | |
| 325 | 
             
                                try:
         | 
| 326 | 
             
                                    yield self.handle_record("tail", doc_id, doc, None, None)
         | 
| 327 | 
             
                                except Exception as e:
         | 
| 328 | 
            -
                                    logger.warning(f | 
| 329 | 
             
                                    traceback.print_exc()
         | 
| 330 | 
             
                                    continue
         | 
| 331 |  | 
| 332 | 
             
                    except gzip.BadGzipFile as e:
         | 
| 333 | 
             
                        # skip broken gzip files
         | 
| 334 | 
            -
                        print(f | 
| 335 | 
             
                        traceback.print_exc()
         | 
| 336 | 
             
                        return
         | 
| 337 |  | 
| 338 | 
            -
                def _handle_head_middle(
         | 
| 339 | 
            -
                        self, base_tag, doc_file, qs_file, dupe_file
         | 
| 340 | 
            -
                ):
         | 
| 341 | 
             
                    if qs_file is None:
         | 
| 342 | 
             
                        yield from self._handle_tail(base_tag, doc_file, None, None)
         | 
| 343 | 
             
                        return
         | 
| @@ -345,11 +371,13 @@ class RedPajamaV2(datasets.GeneratorBasedBuilder): | |
| 345 | 
             
                    # load duplicates
         | 
| 346 | 
             
                    try:
         | 
| 347 | 
             
                        with open(dupe_file, "rb") as df:
         | 
| 348 | 
            -
                            duplicates = set( | 
| 349 | 
            -
                                df, columns=["doc_id"], use_pandas_metadata=False
         | 
| 350 | 
            -
             | 
|  | |
|  | |
| 351 | 
             
                    except Exception as e:
         | 
| 352 | 
            -
                        logger.warning(f | 
| 353 | 
             
                        duplicates = set()
         | 
| 354 |  | 
| 355 | 
             
                    try:
         | 
| @@ -360,16 +388,22 @@ class RedPajamaV2(datasets.GeneratorBasedBuilder): | |
| 360 |  | 
| 361 | 
             
                                    try:
         | 
| 362 | 
             
                                        yield self.handle_record(
         | 
| 363 | 
            -
                                            "head_middle", | 
|  | |
|  | |
|  | |
|  | |
| 364 | 
             
                                        )
         | 
| 365 | 
             
                                    except Exception as e:
         | 
| 366 | 
            -
                                        logger.warning( | 
|  | |
|  | |
| 367 | 
             
                                        traceback.print_exc()
         | 
| 368 | 
             
                                        continue
         | 
| 369 |  | 
| 370 | 
             
                    except gzip.BadGzipFile as e:
         | 
| 371 | 
             
                        # skip broken gzip files
         | 
| 372 | 
            -
                        print(f | 
| 373 | 
             
                        traceback.print_exc()
         | 
| 374 | 
             
                        return
         | 
| 375 |  | 
|  | |
| 15 | 
             
            # Lint as: python3
         | 
| 16 | 
             
            """RedPajama V2: Quality annotated Web Text Documents."""
         | 
| 17 |  | 
| 18 | 
            +
            import gzip
         | 
| 19 | 
             
            import json
         | 
|  | |
|  | |
| 20 | 
             
            import traceback
         | 
|  | |
| 21 | 
             
            from typing import List
         | 
| 22 |  | 
| 23 | 
            +
            import datasets
         | 
| 24 | 
             
            import pyarrow.parquet as pq
         | 
| 25 |  | 
| 26 | 
             
            logger = datasets.logging.get_logger(__name__)
         | 
|  | |
| 29 | 
             
            RedPajama V2: an Open Dataset for Training Large Language Models
         | 
| 30 | 
             
            """
         | 
| 31 |  | 
| 32 | 
            +
            _URL_BASE = "https://data.together.xyz/redpajama-data-v2/v1.0.0"
         | 
| 33 | 
             
            _LANGUAGES = ("en", "de", "fr", "es", "it")
         | 
| 34 | 
             
            _MISSING_FILES_PATTERN = "urls/missing-{component}.txt"
         | 
| 35 | 
             
            _NUM_SHARDS = 5000
         | 
| 36 | 
            +
            _SUBSAMPLE_FILE_COUNTS = {"sample-10B": 1, "sample-100B": 10, "sample-1T": 100}
         | 
| 37 |  | 
| 38 | 
             
            _CC_SNAPSHOT_IDS = (
         | 
| 39 | 
             
                "2014-15",
         | 
|  | |
| 119 | 
             
                "2022-40",
         | 
| 120 | 
             
                "2022-49",
         | 
| 121 | 
             
                "2023-06",
         | 
| 122 | 
            +
                "2023-14",
         | 
| 123 | 
             
            )
         | 
| 124 |  | 
| 125 |  | 
|  | |
| 138 |  | 
| 139 |  | 
| 140 | 
             
            class RedPajamaV2(datasets.GeneratorBasedBuilder):
         | 
| 141 | 
            +
                """RedPajama V2: Quality annotated Web Text Documents."""
         | 
| 142 |  | 
| 143 | 
             
                BUILDER_CONFIGS = [
         | 
| 144 | 
             
                    RedPajamaDataV2Config(
         | 
| 145 | 
            +
                        name="sample",
         | 
| 146 | 
             
                        version=datasets.Version("1.0.0", ""),
         | 
| 147 | 
             
                        description=f"RedPajamaV2 Sample",
         | 
| 148 | 
             
                    ),
         | 
| 149 | 
             
                    RedPajamaDataV2Config(
         | 
| 150 | 
            +
                        name="sample-10B",
         | 
| 151 | 
            +
                        version=datasets.Version("1.0.0", ""),
         | 
| 152 | 
            +
                        description=f"RedPajamaV2 Sample with 10B tokens",
         | 
| 153 | 
            +
                    ),
         | 
| 154 | 
            +
                    RedPajamaDataV2Config(
         | 
| 155 | 
            +
                        name="sample-100B",
         | 
| 156 | 
            +
                        version=datasets.Version("1.0.0", ""),
         | 
| 157 | 
            +
                        description=f"RedPajamaV2 Sample with 100B tokens",
         | 
| 158 | 
            +
                    ),
         | 
| 159 | 
            +
                    RedPajamaDataV2Config(
         | 
| 160 | 
            +
                        name="sample-1T",
         | 
| 161 | 
            +
                        version=datasets.Version("1.0.0", ""),
         | 
| 162 | 
            +
                        description=f"RedPajamaV2 Sample with 1T tokens",
         | 
| 163 | 
            +
                    ),
         | 
| 164 | 
            +
                    RedPajamaDataV2Config(
         | 
| 165 | 
            +
                        name="default",
         | 
| 166 | 
             
                        version=datasets.Version("1.0.0", ""),
         | 
| 167 | 
             
                        description=f"RedPajamaV2",
         | 
| 168 | 
            +
                    ),
         | 
| 169 | 
             
                ]
         | 
| 170 |  | 
| 171 | 
             
                def _info(self):
         | 
|  | |
| 176 | 
             
                                "raw_content": datasets.Value("string"),
         | 
| 177 | 
             
                                "doc_id": datasets.Value("string"),
         | 
| 178 | 
             
                                "meta": datasets.Value("string"),
         | 
| 179 | 
            +
                                "quality_signals": datasets.Value("string"),
         | 
| 180 | 
             
                            }
         | 
| 181 | 
             
                        ),
         | 
| 182 | 
             
                        supervised_keys=None,
         | 
|  | |
| 192 |  | 
| 193 | 
             
                    # fetch documents
         | 
| 194 | 
             
                    logger.info(f"Downloading {len(sample_base_tags)} documents files.")
         | 
| 195 | 
            +
                    documents_files = dl_manager.download(
         | 
| 196 | 
            +
                        {
         | 
| 197 | 
            +
                            base_tag: f"sample/documents/{base_tag}.json.gz"
         | 
| 198 | 
            +
                            for base_tag in sample_base_tags
         | 
| 199 | 
            +
                        }
         | 
| 200 | 
            +
                    )
         | 
| 201 |  | 
| 202 | 
             
                    # fetch quality signals
         | 
| 203 | 
             
                    logger.info(f"Downloading {len(sample_base_tags)} quality signals files.")
         | 
| 204 | 
            +
                    quality_signals_files = dl_manager.download(
         | 
| 205 | 
            +
                        {
         | 
| 206 | 
            +
                            base_tag: f"sample/quality_signals/{base_tag}.signals.json.gz"
         | 
| 207 | 
            +
                            for base_tag in sample_base_tags
         | 
| 208 | 
            +
                        }
         | 
| 209 | 
            +
                    )
         | 
| 210 |  | 
| 211 | 
             
                    # fetch ids of duplicates
         | 
| 212 | 
             
                    logger.info(f"Downloading {len(sample_base_tags)} duplicates ids files.")
         | 
| 213 | 
            +
                    duplicates_ids_files = dl_manager.download(
         | 
| 214 | 
            +
                        {
         | 
| 215 | 
            +
                            base_tag: f"sample/duplicates/{base_tag}.duplicates.parquet"
         | 
| 216 | 
            +
                            for base_tag in sample_base_tags
         | 
| 217 | 
            +
                        }
         | 
| 218 | 
            +
                    )
         | 
| 219 |  | 
| 220 | 
             
                    return [
         | 
| 221 | 
             
                        datasets.SplitGenerator(
         | 
|  | |
| 224 | 
             
                                "base_tags": sample_base_tags,
         | 
| 225 | 
             
                                "documents_files": documents_files,
         | 
| 226 | 
             
                                "quality_signals_files": quality_signals_files,
         | 
| 227 | 
            +
                                "duplicates_ids_files": duplicates_ids_files,
         | 
| 228 | 
            +
                            },
         | 
| 229 | 
             
                        )
         | 
| 230 | 
             
                    ]
         | 
| 231 |  | 
| 232 | 
             
                def _split_generators_full(self, dl_manager):
         | 
| 233 | 
            +
                    snapshots = getattr(self.config, "snapshots", _CC_SNAPSHOT_IDS)
         | 
| 234 | 
            +
                    languages = getattr(self.config, "languages", _LANGUAGES)
         | 
| 235 | 
            +
                    partition = getattr(self.config, "partition", "all")
         | 
| 236 | 
            +
             | 
| 237 | 
            +
                    if self.config.name in ("sample-10B", "sample-100B", "sample-1T"):
         | 
| 238 | 
            +
                        partition = "head_middle"
         | 
| 239 | 
            +
                        languages = _LANGUAGES
         | 
| 240 | 
            +
                        snapshots = _CC_SNAPSHOT_IDS
         | 
| 241 | 
            +
                        num_shards = _SUBSAMPLE_FILE_COUNTS[self.config.name]
         | 
| 242 | 
            +
                    else:
         | 
| 243 | 
            +
                        num_shards = _NUM_SHARDS
         | 
| 244 | 
            +
             | 
| 245 | 
            +
                    if partition == "all":
         | 
| 246 | 
            +
                        partitions = ["head", "middle", "tail"]
         | 
| 247 | 
            +
                    elif partition == "head_middle":
         | 
| 248 | 
            +
                        partitions = ["head", "middle"]
         | 
| 249 | 
            +
                    elif partition == "tail":
         | 
| 250 | 
             
                        partitions = [partition]
         | 
| 251 | 
             
                    else:
         | 
| 252 | 
            +
                        raise ValueError(f"invalid partition: {partition}")
         | 
| 253 |  | 
| 254 | 
             
                    # fetch list of missing files (e.g., missing duplicates or corrupted documents and
         | 
| 255 | 
             
                    # quality signal files)
         | 
| 256 | 
            +
                    missing_files_paths = dl_manager.download_and_extract(
         | 
| 257 | 
            +
                        {
         | 
| 258 | 
            +
                            component: _MISSING_FILES_PATTERN.format(component=component)
         | 
| 259 | 
            +
                            for component in ("documents", "signals", "duplicates")
         | 
| 260 | 
            +
                        }
         | 
| 261 | 
            +
                    )
         | 
| 262 |  | 
| 263 | 
             
                    missing_files = {}
         | 
| 264 | 
             
                    for component, missing_file in missing_files_paths.items():
         | 
| 265 | 
            +
                        with open(missing_file, "r", encoding="utf-8") as f:
         | 
| 266 | 
             
                            missing_files[component] = set(line.strip() for line in f)
         | 
| 267 |  | 
| 268 | 
             
                    # build list of urls to fetch
         | 
|  | |
| 274 | 
             
                    for lang in languages:
         | 
| 275 | 
             
                        for snapshot in snapshots:
         | 
| 276 | 
             
                            for part in partitions:
         | 
| 277 | 
            +
                                for n in range(num_shards):
         | 
| 278 | 
             
                                    base_tag = f"{snapshot}/{n:04d}/{lang}_{part}"
         | 
| 279 | 
             
                                    base_tags.append(base_tag)
         | 
| 280 |  | 
|  | |
| 312 | 
             
                                "base_tags": base_tags,
         | 
| 313 | 
             
                                "documents_files": documents_files,
         | 
| 314 | 
             
                                "quality_signals_files": quality_signals_files,
         | 
| 315 | 
            +
                                "duplicates_ids_files": duplicates_ids_files,
         | 
| 316 | 
            +
                            },
         | 
| 317 | 
             
                        )
         | 
| 318 | 
             
                    ]
         | 
| 319 |  | 
| 320 | 
             
                def _split_generators(self, dl_manager):
         | 
| 321 | 
            +
                    if self.config.name == "sample":
         | 
| 322 | 
             
                        return self._split_generators_sample(dl_manager)
         | 
| 323 |  | 
| 324 | 
             
                    return self._split_generators_full(dl_manager)
         | 
| 325 |  | 
| 326 | 
             
                def _generate_examples(
         | 
| 327 | 
            +
                    self, base_tags, documents_files, quality_signals_files, duplicates_ids_files
         | 
|  | |
| 328 | 
             
                ):
         | 
| 329 | 
             
                    key = 0
         | 
| 330 | 
             
                    for base_tag in base_tags:
         | 
|  | |
| 335 | 
             
                        if doc_file is None:
         | 
| 336 | 
             
                            continue
         | 
| 337 |  | 
| 338 | 
            +
                        for sample in self.__get_generator(base_tag, doc_file, qs_file, dupe_file):
         | 
|  | |
|  | |
| 339 | 
             
                            yield key, sample
         | 
| 340 | 
             
                            key += 1
         | 
| 341 |  | 
|  | |
| 353 | 
             
                                try:
         | 
| 354 | 
             
                                    yield self.handle_record("tail", doc_id, doc, None, None)
         | 
| 355 | 
             
                                except Exception as e:
         | 
| 356 | 
            +
                                    logger.warning(f"failed handling row {row} in {doc_file}")
         | 
| 357 | 
             
                                    traceback.print_exc()
         | 
| 358 | 
             
                                    continue
         | 
| 359 |  | 
| 360 | 
             
                    except gzip.BadGzipFile as e:
         | 
| 361 | 
             
                        # skip broken gzip files
         | 
| 362 | 
            +
                        print(f"BadGzipFile: {doc_file, qs_file}")
         | 
| 363 | 
             
                        traceback.print_exc()
         | 
| 364 | 
             
                        return
         | 
| 365 |  | 
| 366 | 
            +
                def _handle_head_middle(self, base_tag, doc_file, qs_file, dupe_file):
         | 
|  | |
|  | |
| 367 | 
             
                    if qs_file is None:
         | 
| 368 | 
             
                        yield from self._handle_tail(base_tag, doc_file, None, None)
         | 
| 369 | 
             
                        return
         | 
|  | |
| 371 | 
             
                    # load duplicates
         | 
| 372 | 
             
                    try:
         | 
| 373 | 
             
                        with open(dupe_file, "rb") as df:
         | 
| 374 | 
            +
                            duplicates = set(
         | 
| 375 | 
            +
                                pq.read_table(df, columns=["doc_id"], use_pandas_metadata=False)[
         | 
| 376 | 
            +
                                    "doc_id"
         | 
| 377 | 
            +
                                ].to_pylist()
         | 
| 378 | 
            +
                            )
         | 
| 379 | 
             
                    except Exception as e:
         | 
| 380 | 
            +
                        logger.warning(f"no duplicate ids found for {base_tag}")
         | 
| 381 | 
             
                        duplicates = set()
         | 
| 382 |  | 
| 383 | 
             
                    try:
         | 
|  | |
| 388 |  | 
| 389 | 
             
                                    try:
         | 
| 390 | 
             
                                        yield self.handle_record(
         | 
| 391 | 
            +
                                            part="head_middle",
         | 
| 392 | 
            +
                                            doc_id=doc_id,
         | 
| 393 | 
            +
                                            doc=doc,
         | 
| 394 | 
            +
                                            qs=qs,
         | 
| 395 | 
            +
                                            is_duplicate=doc_id in duplicates,
         | 
| 396 | 
             
                                        )
         | 
| 397 | 
             
                                    except Exception as e:
         | 
| 398 | 
            +
                                        logger.warning(
         | 
| 399 | 
            +
                                            f"failed handling row {row} in {doc_file} ({qs_file})"
         | 
| 400 | 
            +
                                        )
         | 
| 401 | 
             
                                        traceback.print_exc()
         | 
| 402 | 
             
                                        continue
         | 
| 403 |  | 
| 404 | 
             
                    except gzip.BadGzipFile as e:
         | 
| 405 | 
             
                        # skip broken gzip files
         | 
| 406 | 
            +
                        print(f"BadGzipFile: {doc_file, qs_file}")
         | 
| 407 | 
             
                        traceback.print_exc()
         | 
| 408 | 
             
                        return
         | 
| 409 |  | 

