Home

Awesome

JTruthfulQA

JTruthfulQA is a Japanese version of TruthfulQA (Lin+, 2022). This dataset is not translated from original TruthfulQA but built from scratch.

The full set of benchmark questions and reference answers is available at data/JTruthfulQA.csv. The benchmark questions are divided into three types: Fact, Knowledge, and Uncategorized.

Task

The task is to answer the given questions. To make it easier to evaluate the answers that were generated by a large language model (LLM), the instruction to LLMs is to generate an answer to each question within 50 characters.

Baselines

This table shows the performance of human performance and recent LLMs on each type of the questions. For human performance, we asked people to answer the questions in the two cases that allow or do not allow them to search the web about the questions. We set the temperature to 0 for "GPT-3.5-turbo" and "GPT-4" or 0.1 for the other LLMs to generate the answers.

FactKnowledgeUncategorizedAllft-GPT-3.5-Turboft-waseda RoBERTaBLEUROUGE1BERTScoreMC1MC2
Human (with search)0.7410.7620.6470.7500.6290.7536.990.280.14--
Human (without search)0.7530.5790.5880.6540.5860.7025.300.250.11--
GPT-3.5-turbo0.780.1770.2350.4370.5120.5436.010.04-0.02--
GPT-40.8690.4090.5290.6090.6010.611-0.6730.03-0.01--
stabilityai/japanese-stablelm-instruct-alpha-7b0.2120.2710.2350.2450.2070.232-7.26-0.05-0.090.1290.130
elyza/ELYZA-japanese-Llama-2-7b-instruct0.5640.1460.1760.3260.2900.421-8.65-0.06-0.100.1260.129
matsuo-lab/weblab-10b-instruction-sft0.1740.2010.3530.1940.1720.151-4.50-0.05-0.080.1560.146
line-corporation/japanese-large-lm-3.6b-instruction-sft0.3780.1650.2940.2600.1920.320-1.52-0.01-0.040.1520.152

How to Answer to JTruthfulQA

We provide a sample code "SampleGeneration.py", which uses "elyza/ELYZA-japanese-Llama-2-7b-instruct" to generate the answers. Rewrite the code for a model that you want to use.

Automatic Evaluation

Python3 script/RobertaEvaluation.py "input_file_name" "output_file_name_1"<br> Run "RobertaEvaluation.py" to evaluate the generated answers. You can get the result with "label". (1: correct, 0: incorrect)<br> "input_file_name" and "output_file_name_1" have to end with ".csv".<br> <br> Python3 script/ResultAnalysys.py "output_file_name_1" "output_file_name_2"<br> You can see the analysys of the answers with "ResultAnalysys.py"<br> "output_file_name_2" has to end with ".json".<br>

Datasets

Each question has the original answer created by human. The dataset includes correct answers and wrong answers generated by four LLMs ("GPT-3.5-turbo", "stabilityai/japanese-stablelm-instruct-alpha-7b", "elyza/ELYZA-japanese-Llama-2-7b-instruct", "matsuo-lab/weblab-10b-instruction-sft"). The original answers are also added to the correct answers.

JTruthfulQA.csv<br> This dataset has 3,078 correct answers and 3,281 incorrect answers (6,359 answers in total) over 582 questions.<br> <br> JTruthfulQA_without_gpt.csv<br> This dataset has 2,125 correct answers and 2,267 incorrect answers (4,392 answers in total) over 551 questions. The answers of GPT-3.5-turbo are excluded from this dataset.

Reference

@InProceedings{Kurihara_nlp2022,
  author = "中村友亮 and 河原大輔",
  title = "日本語TruthfulQAの構築",
  booktitle = "言語処理学会第30回年次大会",
  year = "2024",
  url = "https://anlp.jp/proceedings/annual_meeting/2024/pdf_dir/P6-15.pdf",
  pages = "1709--1714",
  note= "in Japanese"
}

License

This dataset is distributed under CC BY-NC-SA 4.0.

Acknowledgment

This dataset was created in collaboration with SB Intuitions Corp. and Waseda University.