| | import json |
| |
|
| | import datasets |
| |
|
| |
|
| | _CITATION = """\ |
| | @article{scialom2020mlsum, |
| | title={MLSUM: The Multilingual Summarization Corpus}, |
| | author={Scialom, Thomas and Dray, Paul-Alexis and Lamprier, Sylvain and Piwowarski, Benjamin and Staiano, Jacopo}, |
| | journal={arXiv preprint arXiv:2004.14900}, |
| | year={2020} |
| | } |
| | """ |
| |
|
| | _DESCRIPTION = """\ |
| | We present MLSUM, the first large-scale MultiLingual SUMmarization dataset. |
| | Obtained from online newspapers, it contains 1.5M+ article/summary pairs in five different languages -- namely, French, German, Spanish, Russian, Turkish. |
| | Together with English newspapers from the popular CNN/Daily mail dataset, the collected data form a large scale multilingual dataset which can enable new research directions for the text summarization community. |
| | We report cross-lingual comparative analyses based on state-of-the-art systems. |
| | These highlight existing biases which motivate the use of a multi-lingual dataset. |
| | """ |
| |
|
| | _URL = "https://gitlab.lip6.fr/scialom/mlsum_data/-/raw/master/MLSUM" |
| | _LANG = ["de", "es", "fr", "ru", "tu"] |
| |
|
| |
|
| | class Mlsum(datasets.GeneratorBasedBuilder): |
| |
|
| | BUILDER_CONFIGS = [ |
| | datasets.BuilderConfig( |
| | name=lang, |
| | version=datasets.Version("1.0.0"), |
| | description="", |
| | ) |
| | for lang in _LANG |
| | ] |
| |
|
| | def _info(self): |
| | return datasets.DatasetInfo( |
| | |
| | description=_DESCRIPTION, |
| | |
| | features=datasets.Features( |
| | { |
| | "text": datasets.Value("string"), |
| | "summary": datasets.Value("string"), |
| | "topic": datasets.Value("string"), |
| | "url": datasets.Value("string"), |
| | "title": datasets.Value("string"), |
| | "date": datasets.Value("string") |
| | |
| | } |
| | ), |
| | |
| | |
| | |
| | supervised_keys=None, |
| | |
| | homepage="", |
| | citation=_CITATION, |
| | ) |
| |
|
| | def _split_generators(self, dl_manager): |
| | """Returns SplitGenerators.""" |
| | |
| | |
| |
|
| | lang = self.config.name |
| | urls_to_download = { |
| | "train": f"{_URL}/{lang}_train.jsonl?inline=false", |
| | "validation": f"{_URL}/{lang}_val.jsonl?inline=false", |
| | "test": f"{_URL}/{lang}_test.jsonl?inline=false", |
| | } |
| | downloaded_files = dl_manager.download(urls_to_download) |
| |
|
| | return [ |
| | datasets.SplitGenerator( |
| | name=split, |
| | gen_kwargs={ |
| | "filepath": downloaded_files[split], |
| | }, |
| | ) |
| | for split in [datasets.Split.TRAIN, datasets.Split.VALIDATION, datasets.Split.TEST] |
| | ] |
| |
|
| | def _generate_examples(self, filepath): |
| | """Yields examples.""" |
| | with open(filepath, encoding="utf-8") as f: |
| | for id_, line in enumerate(f): |
| | data = json.loads(line) |
| | yield id_, { |
| | "text": data["text"], |
| | "summary": data["summary"], |
| | "topic": data["topic"], |
| | "url": data["url"], |
| | "title": data["title"], |
| | "date": data["date"], |
| | } |
| |
|