Home

Awesome

Chinese NLI Probing

Introduction

This folder contains data and code for the paper Investigating Transfer Learning in Multilingual Pre-trained Language Models through Chinese Natural Language Inference, where we examine the cross-lingual transfer learning in XLM-R via 4 categories of newly constructed NLI datasets:

All of them are made in parallel to existing adversarial/probing datasets in English (but from scratch, i.e., no translations), so as to allow cross-lingual transfer studies in both directions: English-to-Chinese and Chinese-to-English.

Data

The generated data can be found in the data folder, with detailed explanations.

Code

The code used for data generation can be found in the code folder (will be uploaded shortly).

Main results

We find thatcross-lingual models trained on English NLI do transfer well across our Chinese tasks (e.g., in 3/4 of our challenge categories, they per-form as well/better than the best monolingual models, even on 3/5 uniquely Chinese lin-guistic phenomena such as idioms, pro drop). These results, however, come with importantcaveats: cross-lingual models often performbest when trained on a mixture of English and high-quality monolingual NLI data (OCNLI), and are often hindered by automatically trans-lated resources (XNLI-zh). For many phenomena, all models continue to struggle, highlighting the need for our new diagnostics to help benchmark Chinese and cross-lingual models.

(will be expanded shortly)

References

If you use our resources or find our results useful, please cite:

@inproceedings{hu-et-al-2021-investigating,
	title={Investigating Transfer Learning in Multilingual Pre-trained Language Models through {Chinese} Natural Language Inference},
	author={Hai Hu and He Zhou and Zuoyu Tian and Yiwen Zhang and Yina Ma and Yanting Li and Yixin Nie and Kyle Richardson},
	booktitle={Findings of ACL},
	year={2021}
}