|
"""Korean Balanced Evaluation of Significant Tasks""" |
|
|
|
|
|
import csv |
|
|
|
import pandas as pd |
|
|
|
import datasets |
|
|
|
|
|
_CITATAION = """\ |
|
TBD |
|
""" |
|
|
|
_DESCRIPTION = """\ |
|
The dataset contains data for KoBEST dataset |
|
""" |
|
|
|
_URL = "https://github.com/SKT-LSL/KoBEST_datarepo" |
|
|
|
_DATA_URLS = { |
|
"boolq": { |
|
"train": _URL + "/v1.0/BoolQ/train.tsv", |
|
"dev": _URL + "/v1.0/BoolQ/dev.tsv", |
|
"test": _URL + "/v1.0/BoolQ/test.tsv", |
|
}, |
|
"copa": { |
|
"train": _URL + "/v1.0/COPA/train.tsv", |
|
"dev": _URL + "/v1.0/COPA/dev.tsv", |
|
"test": _URL + "/v1.0/COPA/test.tsv", |
|
}, |
|
"sentineg": { |
|
"train": _URL + "/v1.0/SentiNeg/train.tsv", |
|
"dev": _URL + "/v1.0/SentiNeg/dev.tsv", |
|
"test": _URL + "/v1.0/SentiNeg/test.tsv", |
|
}, |
|
"hellaswag": { |
|
"train": _URL + "/v1.0/HellaSwag/train.tsv", |
|
"dev": _URL + "/v1.0/HellaSwag/dev.tsv", |
|
"test": _URL + "/v1.0/HellaSwag/test.tsv", |
|
}, |
|
"wic": { |
|
"train": _URL + "/v1.0/WiC/train.tsv", |
|
"dev": _URL + "/v1.0/WiC/dev.tsv", |
|
"test": _URL + "/v1.0/WiC/test.tsv", |
|
}, |
|
} |
|
|
|
|
|
class KoBESTConfig(datasets.BuilderConfig): |
|
"""Config for building KoBEST""" |
|
|
|
def __init__(self, description, data_url, citation, url, **kwargs): |
|
""" |
|
Args: |
|
description: `string`, brief description of the dataset |
|
data_url: `dictionary`, dict with url for each split of data. |
|
citation: `string`, citation for the dataset. |
|
url: `string`, url for information about the dataset. |
|
**kwrags: keyword arguments frowarded to super |
|
""" |
|
super(KoBESTConfig, self).__init__(version=datasets.Version("1.0.0", ""), **kwargs) |
|
self.description = description |
|
self.data_url = data_url |
|
self.citation = citation |
|
self.url = url |
|
|
|
|
|
class KoBEST(datasets.GeneratorBasedBuilder): |
|
BUILDER_CONFIGS = [ |
|
KoBESTConfig(name=name, description=_DESCRIPTION, data_url=_DATA_URLS[name], citation=_CITATAION, url=_URL) |
|
for name in ["boolq", "copa", 'sentineg', 'hellaswag', 'wic'] |
|
] |
|
BUILDER_CONFIG_CLASS = KoBESTConfig |
|
|
|
def _info(self): |
|
features = {} |
|
if self.config.name == "boolq": |
|
labels = ["True", "False"] |
|
features["paragraph"] = datasets.Value("string") |
|
features["question"] = datasets.Value("string") |
|
features["label"] = datasets.features.ClassLabel(names=labels) |
|
|
|
if self.config.name == "copa": |
|
labels = ["alternative_1", "alternative_2"] |
|
features["premise"] = datasets.Value("string") |
|
features["question"] = datasets.Value("string") |
|
features["alternative_1"] = datasets.Value("string") |
|
features["alternative_2"] = datasets.Value("string") |
|
features["label"] = datasets.features.ClassLabel(names=labels) |
|
|
|
if self.config.name == "wic": |
|
labels = ["True", "False"] |
|
features["word"] = datasets.Value("string") |
|
features["context_1"] = datasets.Value("string") |
|
features["context_2"] = datasets.Value("string") |
|
features["label"] = datasets.features.ClassLabel(names=labels) |
|
|
|
if self.config.name == "hellaswag": |
|
labels = ["ending_1", "ending_2", "ending_3", "ending_4"] |
|
|
|
features["context"] = datasets.Value("string") |
|
features["ending_1"] = datasets.Value("string") |
|
features["ending_2"] = datasets.Value("string") |
|
features["ending_3"] = datasets.Value("string") |
|
features["ending_4"] = datasets.Value("string") |
|
features["label"] = datasets.features.ClassLabel(names=labels) |
|
|
|
if self.config.name == "sentineg": |
|
labels = ["negative", "positive"] |
|
features["sentence"] = datasets.Value("string") |
|
features["label"] = datasets.features.ClassLabel(names=labels) |
|
|
|
return datasets.DatasetInfo( |
|
description=_DESCRIPTION, features=datasets.Features(features), homepage=_URL, citation=_CITATAION |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
|
|
train = dl_manager.download_and_extract(self.config.data_url["train"]) |
|
dev = dl_manager.download_and_extract(self.config.data_url["dev"]) |
|
test = dl_manager.download_and_extract(self.config.data_url["test"]) |
|
|
|
return [ |
|
datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepath": train, "split": "train"}), |
|
datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={"filepath": dev, "split": "dev"}), |
|
datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={"filepath": test, "split": "test"}), |
|
] |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def _generate_examples(self, filepath, split): |
|
if self.config.name == "boolq": |
|
df = pd.read_csv(filepath, sep="\t") |
|
df = df.dropna() |
|
|
|
for id_, row in df.iterrows(): |
|
yield id_, { |
|
"paragraph": str(row["Text"]), |
|
"question": str(row["Question"]), |
|
"label": str(int(row["Answer"])), |
|
} |
|
|
|
if self.config.name == "copa": |
|
df = pd.read_csv(filepath, sep="\t") |
|
df = df.dropna() |
|
|
|
for id_, row in df.iterrows(): |
|
yield id_, { |
|
"premise": str(row["sentence"]), |
|
"question": str(row["question"]), |
|
"alternative_1": str(int(row["1"])), |
|
"alternative_2": str(int(row["2"])), |
|
"label": str(row["Answer"]-1), |
|
} |
|
|
|
if self.config.name == "wic": |
|
df = pd.read_csv(filepath, sep="\t") |
|
df = df.dropna() |
|
|
|
for id_, row in df.iterrows(): |
|
yield id_, { |
|
"word": str(row["Target"]), |
|
"context_1": str(row["SENTENCE1"]), |
|
"context_2": str(int(row["SENTENCE2"])), |
|
"label": str(int(row["Answer"])), |
|
} |
|
|
|
if self.config.name == "hellaswag": |
|
df = pd.read_csv(filepath, sep="\t") |
|
df = df.dropna() |
|
|
|
for id_, row in df.iterrows(): |
|
yield id_, { |
|
"context": str(row["context"]), |
|
"ending_1": str(row["choice1"]), |
|
"ending_2": str(int(row["choice2"])), |
|
"ending_3": str(int(row["choice3"])), |
|
"ending_4": str(int(row["choice4"])), |
|
"label": str(row["label"]), |
|
} |
|
|
|
if self.config.name == "sentineg": |
|
df = pd.read_csv(filepath, sep="\t") |
|
df = df.dropna() |
|
|
|
for id_, row in df.iterrows(): |
|
yield id_, { |
|
"sentence": str(row["Text"]), |
|
"label": str(int(row["Label"])), |
|
} |
|
|
|
|