import datasets _DESCRIPTION = "Cleaned Hong Kong Cantonese / Traditional Chinese corpora from multiple sources." _HOMEPAGE = "https://huggingface.co/datasets/SolarisCipher/hk_content_corpus" _CORPUS_FILES = { "appledaily_article": "appledaily_article_dedup.txt", "hkcitizenmedia_article": "hkcitizenmedia_article_dedup.txt", "hkcnews_article": "hkcnews_article_dedup.txt", "inmedia_article": "inmedia_article_dedup.txt", "lihkg_posts": "lihkg_posts_dedup_demoji_128.txt", "openrice_review": "openrice_review_dedup_demoji.txt", "thestandnews_article": "thestandnews_article_dedup.txt", "wiki_hk": "wiki_hk.txt", } class HKContentCorpus(datasets.GeneratorBasedBuilder): def _info(self): return datasets.DatasetInfo( description=_DESCRIPTION, features=datasets.Features({"text": datasets.Value("string")}), supervised_keys=None, homepage=_HOMEPAGE, citation="", ) def _split_generators(self, dl_manager): return [ datasets.SplitGenerator( name=split_name, gen_kwargs={"filepath": filepath}, ) for split_name, filepath in _CORPUS_FILES.items() ] def _generate_examples(self, filepath): with open(filepath, encoding="utf-8") as f: buffer = [] example_id = 0 for line in f: line = line.strip() if line: buffer.append(line) else: if buffer: yield example_id, {"text": "\n".join(buffer)} example_id += 1 buffer = [] if buffer: yield example_id, {"text": "\n".join(buffer)}