Dataset | Checkpoints | Paper | Evaluation | Citation | License
This repository contains resources for accessing the official benchmarks, codes and checkpoints of the paper: Natural Response Generation for Chinese Reading Comprehension.
The paper was accepted by EMNLP 2023! 🎉
We introduce Penguin, an end-to-end Chinese question answering dataset comprising of 200K question-passage-answer-response pairs. The goal of this dataset is to provide a challenging benchmark for end-to-end Chinese Machine Readinng Comprehension that includes the well-informed responses of each question. Penguin can facilate the research to build generative QA models in Chinese and provide a relatively large-scale training corpus for Chinese communities. Please refer to our paper for more details.
Penguin, in the hopes of creating sophisticated GRC models that can generate natural responses for practical QA situations. Considering constructing such a dataset on a large scale is non-trivial, we initialize our dataset from the current Chinese MRC dataset corpus to get raw passage-question-answer triplets, including CMRC 2018, DuReader, and ReCo.
It is extremely difficult and expensive to ask the annotator to start from scratch and write a response
Concretely, we store our dataset in json files:
{
"Passage": "鳊鱼一直备受人们喜爱,鳊鱼的传统做法是清蒸或者红烧。当然不排除可以烧汤,但是鉴于鳊鱼的肉质特色,不是很适合烧汤的",
"Query": "鳊鱼可以炖汤吗?",
"Answer": "可以",
"Response": "鳊鱼可以用来煮汤,但是一般不推荐这么做."
}
| Train set | Dev Set | Test Set | ALL |
Here we report automatic and huaman evaluations results of four baselines in our paper.
Model | Large | Base |
---|---|---|
T5 | T5-base | T5-small |
BART | BART-Large | BART-base |
Prompt-BART | Prompt-BART-Large | - |
Model | Large | Base |
---|---|---|
T5 | Answerer, Responser | Answerer, Responser |
BART | Answerer, Responser | Answerer, Responser |
Prompt-BART | Answerer, Responser | - |
We provide the inference code, please refer to utils/inference.py.
python3 utils/inference.py
Of note, you should change the model_path based on your local enviroment. Then it can generate responses from models and store in generate.json.
python3 utils/inference.py generate.json
Running this script could lead to computing the automatic metrics of the model and store in results.csv file.
@inproceedings{Chen2023NaturalRG,
title={Natural Response Generation for Chinese Reading Comprehension},
author={Nuo Chen and Hongguang Li and Yinan Bao and Baoyuan Wang and Jia Li},
year={2023}
}