File size: 5,851 Bytes
cac2b44 49a2cd7 cac2b44 539f7e5 be45867 e375324 539f7e5 7cd9b3e 539f7e5 cac2b44 3444049 cac2b44 d6c7af4 0232cef 988efa8 d22b169 4d1740e 1473267 d74e10b 4d1740e d22b169 4d1740e dd86e0a b75bf4a d22b169 e375324 d22b169 1473267 12f7d20 e375324 8b2f982 78ed4a0 1473267 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 |
# coding=utf-8
# Copyright 2020 The TensorFlow Datasets Authors and the HuggingFace Datasets Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# Lint as: python3
"""Arabic Poetry Metric dataset."""
import os
import datasets
import pandas as pd
_DESCRIPTION = """\
Masader is the largest public catalogue for Arabic NLP datasets, which consists of more than 200 datasets annotated with 25 attributes.
"""
_CITATION = """\
@misc{alyafeai2021masader,
title={Masader: Metadata Sourcing for Arabic Text and Speech Data Resources},
author={Zaid Alyafeai and Maraim Masoud and Mustafa Ghaleb and Maged S. Al-shaibani},
year={2021},
eprint={2110.06744},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
"""
class MasaderConfig(datasets.BuilderConfig):
"""BuilderConfig for Masader."""
def __init__(self, **kwargs):
"""BuilderConfig for MetRec.
Args:
**kwargs: keyword arguments forwarded to super.
"""
super(MasaderConfig, self).__init__(version=datasets.Version("1.0.0", ""), **kwargs)
class Masader(datasets.GeneratorBasedBuilder):
"""Masaderdataset."""
BUILDER_CONFIGS = [
MasaderConfig(
name="plain_text",
description="Plain text",
)
]
def _info(self):
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=datasets.Features(
{
'Name': datasets.Value("string"),
'Subsets': [{'Name':datasets.Value("string"),
'Dialect':datasets.Value("string") ,
'Volume':datasets.Value("string") ,
'Unit':datasets.Value("string")}],
'HF Link': datasets.Value("string"),
'Link': datasets.Value("string"),
'License': datasets.Value("string"),
'Year': datasets.Value("int32"),
'Language': datasets.Value("string"),
'Dialect': datasets.Value("string"),
'Domain': datasets.Value("string"),
'Form': datasets.Value("string"),
'Collection Style': datasets.Value("string"),
'Description': datasets.Value("string"),
'Volume': datasets.Value("string"),
'Unit': datasets.Value("string"),
'Ethical Risks': datasets.Value("string"),
'Provider': datasets.Value("string"),
'Derived From': datasets.Value("string"),
'Paper Title': datasets.Value("string"),
'Paper Link': datasets.Value("string"),
'Script': datasets.Value("string"),
'Tokenized': datasets.Value("string"),
'Host': datasets.Value("string"),
'Access': datasets.Value("string"),
'Cost': datasets.Value("string"),
'Test Split': datasets.Value("string"),
'Tasks': datasets.Value("string"),
'Venue Title': datasets.Value("string"),
'Citations': datasets.Value("string"),
'Venue Type': datasets.Value("string"),
'Venue Name': datasets.Value("string"),
'Authors': datasets.Value("string"),
'Affiliations': datasets.Value("string"),
'Abstract': datasets.Value("string"),
'Added By': datasets.Value("string"),
}
),
supervised_keys=None,
homepage="https://github.com/arbml/Masader",
citation=_CITATION,)
def _split_generators(self, dl_manager):
sheet_id = "1YO-Vl4DO-lnp8sQpFlcX1cDtzxFoVkCmU1PVw_ZHJDg"
sheet_name = "filtered_clean"
url = f"https://docs.google.com/spreadsheets/d/{sheet_id}/gviz/tq?tqx=out:csv&sheet={sheet_name}"
return [
datasets.SplitGenerator(
name=datasets.Split.TRAIN, gen_kwargs={"url":url }
),
]
def _generate_examples(self, url):
"""Generate examples."""
# For labeled examples, extract the label from the path.
df = pd.read_csv(url, usecols=range(35))
df.columns.values[0] = "No."
df.columns.values[1] = "Name"
subsets = {}
entry_list = []
i = 0
idx = 0
while i < len(df.values) - 1:
next_entry = df.values[i+1]
curr_entry = df.values[i]
i+= 1
if str(curr_entry[0]) != "nan":
entry_list = curr_entry
subsets = []
if str(next_entry[0]) == "nan":
subsets.append({'Name': next_entry[2], 'Dialect':next_entry[8], 'Volume':next_entry[13], 'Unit':next_entry[14]})
continue
idx += 1
masader_entry = {col:entry_list[j+1] for j,col in enumerate(df.columns[1:]) if j != 1}
masader_entry['Year'] = int(entry_list[6])
masader_entry['Subsets'] = subsets
yield idx, masader_entry
|