Datasets:
Tasks:
Text Generation
Modalities:
Text
Formats:
parquet
Sub-tasks:
language-modeling
Languages:
code
Size:
100M - 1B
License:
| # coding=utf-8 | |
| # Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor. | |
| # | |
| # Licensed under the Apache License, Version 2.0 (the "License"); | |
| # you may not use this file except in compliance with the License. | |
| # You may obtain a copy of the License at | |
| # | |
| # http://www.apache.org/licenses/LICENSE-2.0 | |
| # | |
| # Unless required by applicable law or agreed to in writing, software | |
| # distributed under the License is distributed on an "AS IS" BASIS, | |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |
| # See the License for the specific language governing permissions and | |
| # limitations under the License. | |
| """GitHub Code dataset.""" | |
| import os | |
| import pyarrow as pa | |
| import pyarrow.parquet as pq | |
| import datasets | |
| _REPO_NAME = "codeparrot/github-code" | |
| _LANG_TO_EXTENSION = { | |
| "Assembly": [".asm"], | |
| "Batchfile": [".bat", ".cmd"], | |
| "C": [".c", ".h"], | |
| "C#": [".cs"], | |
| "C++": [".cpp", ".hpp", ".c++", ".h++", ".cc", ".hh", ".C", ".H"], | |
| "CMake": [".cmake"], | |
| "CSS": [".css"], | |
| "Dockerfile": [".dockerfile", "Dockerfile"], | |
| "FORTRAN": ['.f90', '.f', '.f03', '.f08', '.f77', '.f95', '.for', '.fpp'], | |
| "GO": [".go"], | |
| "Haskell": [".hs"], | |
| "HTML":[".html"], | |
| "Java": [".java"], | |
| "JavaScript": [".js"], | |
| "Julia": [".jl"], | |
| "Lua": [".lua"], | |
| "Makefile": ["Makefile"], | |
| "Markdown": [".md", ".markdown"], | |
| "PHP": [".php", ".php3", ".php4", ".php5", ".phps", ".phpt"], | |
| "Perl": [".pl", ".pm", ".pod", ".perl"], | |
| "PowerShell": ['.ps1', '.psd1', '.psm1'], | |
| "Python": [".py"], | |
| "Ruby": [".rb"], | |
| "Rust": [".rs"], | |
| "SQL": [".sql"], | |
| "Scala": [".scala"], | |
| "Shell": [".sh", ".bash", ".command", ".zsh"], | |
| "TypeScript": [".ts", ".tsx"], | |
| "TeX": [".tex"], | |
| "Visual Basic": [".vb"] | |
| } | |
| _LICENSES = ['mit', | |
| 'apache-2.0', | |
| 'gpl-3.0', | |
| 'gpl-2.0', | |
| 'bsd-3-clause', | |
| 'agpl-3.0', | |
| 'lgpl-3.0', | |
| 'lgpl-2.1', | |
| 'bsd-2-clause', | |
| 'cc0-1.0', | |
| 'epl-1.0', | |
| 'mpl-2.0', | |
| 'unlicense', | |
| 'isc', | |
| 'artistic-2.0'] | |
| _DESCRIPTION = """\ | |
| The GitHub Code dataest consists of 115M code files from GitHub in 32 programming \ | |
| languages with 60 extensions totalling in 1TB of text data. The dataset was created \ | |
| from the GitHub dataset on BiqQuery. | |
| """ | |
| _HOMEPAGE = "https://cloud.google.com/blog/topics/public-datasets/github-on-bigquery-analyze-all-the-open-source-code/" | |
| _EXTENSION_TO_LANG = {} | |
| for lang in _LANG_TO_EXTENSION: | |
| for extension in _LANG_TO_EXTENSION[lang]: | |
| _EXTENSION_TO_LANG[extension] = lang | |
| _LANG_CONFIGS = ["all"] + list(_LANG_TO_EXTENSION.keys()) | |
| _LICENSE_CONFIGS = ["all"] + _LICENSES | |
| class GithubCodeConfig(datasets.BuilderConfig): | |
| """BuilderConfig for the GitHub Code dataset.""" | |
| def __init__(self, *args, languages=["all"], licenses=["all"], **kwargs): | |
| """BuilderConfig for the GitHub Code dataset. | |
| Args: | |
| languages (:obj:`List[str]`): List of languages to load. | |
| licenses (:obj:`List[str]`): List of licenses to load. | |
| **kwargs: keyword arguments forwarded to super. | |
| """ | |
| super().__init__( | |
| *args, | |
| name="+".join(languages)+"-"+"+".join(licenses), | |
| **kwargs, | |
| ) | |
| languages = set(languages) | |
| licenses = set(licenses) | |
| assert all([language in _LANG_CONFIGS for language in languages]), f"Language not in {_LANG_CONFIGS}." | |
| assert all([license in _LICENSE_CONFIGS for license in licenses]), f"License not in {_LICENSE_CONFIGS}." | |
| if "all" in languages: | |
| assert len(languages)==1, "Passed 'all' together with other languages." | |
| self.filter_languages = False | |
| else: | |
| self.filter_languages = True | |
| if "all" in licenses: | |
| assert len(licenses)==1, "Passed 'all' together with other licenses." | |
| self.filter_licenses = False | |
| else: | |
| self.filter_licenses = True | |
| self.languages = set(languages) | |
| self.licenses = set(licenses) | |
| class GithubCode(datasets.GeneratorBasedBuilder): | |
| """GitHub Code dataset.""" | |
| VERSION = datasets.Version("1.0.0") | |
| BUILDER_CONFIG_CLASS = GithubCodeConfig | |
| BUILDER_CONFIGS = [GithubCodeConfig(languages=[lang], licenses=[license]) for lang in _LANG_CONFIGS | |
| for license in _LICENSE_CONFIGS] | |
| DEFAULT_CONFIG_NAME = "all-all" | |
| def _info(self): | |
| return datasets.DatasetInfo( | |
| description=_DESCRIPTION, | |
| features=datasets.Features({"code": datasets.Value("string"), | |
| "repo_name": datasets.Value("string"), | |
| "path": datasets.Value("string"), | |
| "language": datasets.Value("string"), | |
| "license": datasets.Value("string"), | |
| "size": datasets.Value("int32")}), | |
| supervised_keys=None, | |
| homepage=_HOMEPAGE, | |
| license="Multiple: see the 'license' field of each sample.", | |
| ) | |
| def _split_generators(self, dl_manager): | |
| num_shards = 1126 | |
| data_files = [ | |
| f"data/train-{_index:05d}-of-{num_shards:05d}.parquet" | |
| for _index in range(num_shards) | |
| ] | |
| files = dl_manager.download(data_files) | |
| return [ | |
| datasets.SplitGenerator( | |
| name=datasets.Split.TRAIN, | |
| gen_kwargs={ | |
| "files": files, | |
| }, | |
| ), | |
| ] | |
| def _generate_examples(self, files): | |
| key = 0 | |
| for file_idx, file in enumerate(files): | |
| with open(file, "rb") as f: | |
| parquet_file = pq.ParquetFile(f) | |
| for batch_idx, record_batch in enumerate(parquet_file.iter_batches(batch_size=10_000)): | |
| pa_table = pa.Table.from_batches([record_batch]) | |
| for row_index in range(pa_table.num_rows): | |
| row = pa_table.slice(row_index, 1).to_pydict() | |
| lang = lang_from_name(row['path'][0]) | |
| license = row["license"][0] | |
| if self.config.filter_languages and not lang in self.config.languages: | |
| continue | |
| if self.config.filter_licenses and not license in self.config.licenses: | |
| continue | |
| yield key, {"code": row['content'][0], | |
| "repo_name": row['repo_name'][0], | |
| "path": row['path'][0], | |
| "license": license, | |
| "language": lang, | |
| "size": int(row['size'][0])} | |
| key += 1 | |
| def lang_from_name(name): | |
| for extension in _EXTENSION_TO_LANG: | |
| if name.endswith(extension): | |
| return _EXTENSION_TO_LANG[extension] |