|
import os |
|
import re |
|
|
|
import datasets |
|
|
|
|
|
logger = datasets.logging.get_logger(__name__) |
|
|
|
_DESCRIPTION = """\ |
|
Common Crawl - Malayalam. |
|
""" |
|
|
|
_CITATION = """\ |
|
@article{qburst, |
|
title={Common Crawl - Malayalam}, |
|
author={n.d}, |
|
year={2020}, |
|
journal={n.d}, |
|
} |
|
""" |
|
|
|
_URLs = { |
|
"malayalam_wiki_2020": "https://huggingface.co/datasets/rajeshradhakrishnan/malayalam_2020_wiki/resolve/main/", |
|
"checksum_url": "https://huggingface.co/datasets/rajeshradhakrishnan/malayalam_2020_wiki/resolve/main/ml_sha256.txt" |
|
|
|
} |
|
|
|
|
|
class MalayalamWikiConfig(datasets.BuilderConfig): |
|
"""BuilderConfig for MalayalamWiki.""" |
|
|
|
def __init__(self, **kwargs): |
|
"""BuilderConfig for MalayalamWiki. |
|
|
|
Args: |
|
**kwargs: keyword arguments forwarded to super. |
|
""" |
|
super(MalayalamWikiConfig, self).__init__(**kwargs) |
|
|
|
class MalayalamWiki(datasets.GeneratorBasedBuilder): |
|
"""Malayalam News topic classification dataset.""" |
|
|
|
VERSION = datasets.Version("1.0.0") |
|
|
|
BUILDER_CONFIGS = [ |
|
MalayalamWikiConfig( |
|
name="malayalam_wiki_2020", version=VERSION, description="Common Crawl - Malayalam." |
|
), |
|
] |
|
|
|
|
|
def remove_special_characters(self, txt): |
|
chars_to_ignore_regex = '[\,\?\.\!\-\;\:\"\“\%\‘\”\�Utrnle\_]' |
|
unicode_ignore_regex = r'[\u200e\u200c\u200d]' |
|
english_ignore_regex = r'[a-zA-Z]' |
|
|
|
txt = txt.strip() |
|
txt = re.sub(chars_to_ignore_regex, '',txt) |
|
txt = re.sub(unicode_ignore_regex, '',txt) + " " |
|
txt = re.sub(english_ignore_regex, '',txt) + " " |
|
return txt |
|
|
|
def _info(self): |
|
return datasets.DatasetInfo( |
|
description=_DESCRIPTION, |
|
features=datasets.Features( |
|
{ |
|
"text": datasets.Value("string") |
|
} |
|
), |
|
supervised_keys=None, |
|
homepage="https://github.com/qburst/common-crawl-malayalam", |
|
citation=_CITATION |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
"""Returns SplitGenerators.""" |
|
checksum_url = _URLs["checksum_url"] |
|
checksum_file = dl_manager.download(checksum_url) |
|
with open(checksum_file, encoding="utf-8") as f: |
|
data_filenames = [line.strip() for line in f if line] |
|
data_urls = [_URLs["malayalam_wiki_2020"] + data_filename for data_filename in data_filenames[1:2]] |
|
downloaded_files = dl_manager.download(data_urls) |
|
return [ |
|
datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepaths": downloaded_files}), |
|
] |
|
|
|
def _generate_examples(self, filepaths): |
|
"""This function returns the examples in the raw (text) form by iterating on all the files.""" |
|
for file_id,filepath in enumerate(filepaths): |
|
logger.info("generating examples from = %s", filepath) |
|
with open(filepath, encoding="utf-8") as f: |
|
for row_id, row in enumerate(f): |
|
yield f"{file_id}_{row_id}", {"text": self.remove_special_characters(row).strip()} |
|
|