Home

Awesome

XFUND: A Multilingual Form Understanding Benchmark

Introduction

XFUND is a multilingual form understanding benchmark dataset that includes human-labeled forms with key-value pairs in 7 languages (Chinese, Japanese, Spanish, French, Italian, German, Portuguese).

Download

Statistics

langsplitheaderquestionanswerothertotal
ZHtraining4413,2662,8088967,411
testing1221,0778213122,332
JAtraining2293,6924,6411,66610,228
testing581,2531,7325863,629
EStraining2533,0134,2543,92911,449
testing909091,2181,1963,413
FRtraining1832,4973,4272,7098,816
testing661,0231,2811,1313,501
ITtraining1663,7624,9323,35512,215
testing651,2301,5991,1354,029
DEtraining1552,6093,9921,8768,632
testing598581,3226502,889
PTtraining1853,5105,4282,53111,654
testing591,2881,9408824,169

Baselines

For the code example, please refer to the LayoutXLM repository.

Results

Language-specific Finetuning

ModelFUNSDZHJAESFRITDEPTAvg.
Semantic Entity Recognitionxlm-roberta-base0.6670.87740.77610.61050.67430.66870.68140.68180.7047
infoxlm-base0.68520.88680.78650.62300.70150.67510.70630.70080.7207
layoutxlm-base0.7940.89240.79210.75500.79020.80820.82220.79030.8056
Relation Extractionxlm-roberta-base0.26590.51050.58000.52950.49650.53050.50410.39820.4769
infoxlm-base0.29200.52140.60000.55160.49130.52810.52620.41700.4910
layoutxlm-base0.54830.70730.69630.68960.63530.64150.65510.57180.6432

Zero-shot Transfer Learning

ModelFUNSDZHJAESFRITDEPTAvg.
SERxlm-roberta-base0.6670.41440.30230.30550.3710.27670.32860.39360.3824
infoxlm-base0.68520.44080.36030.31020.40210.28800.35870.45020.4119
layoutxlm-base0.7940.60190.47150.45650.57570.48460.52520.5390.5561
RExlm-roberta-base0.26590.16010.26110.24400.22400.23740.22880.19960.2276
infoxlm-base0.29200.24050.28510.24810.24540.21930.20270.20490.2423
layoutxlm-base0.54830.44940.44080.47080.44160.40900.38200.36850.4388

Multitask Fine-tuning

ModelFUNSDZHJAESFRITDEPTAvg.
SERxlm-roberta-base0.66330.8830.77860.62230.70350.68140.71460.67260.7149
infoxlm-base0.65380.87410.78550.59790.70570.68260.70550.67960.7106
layoutxlm-base0.79240.89730.79640.77980.81730.8210.83220.82410.8201
RExlm-roberta-base0.36380.67970.68290.68280.67270.69370.68870.60820.6341
infoxlm-base0.36990.64930.64730.68280.68310.66900.63840.57630.6145
layoutxlm-base0.66710.82410.81420.81040.82210.83100.78540.70440.7823

Citation

If you find XFUND useful in your research, please cite the following paper:

@inproceedings{xu-etal-2022-xfund,
    title = "{XFUND}: A Benchmark Dataset for Multilingual Visually Rich Form Understanding",
    author = "Xu, Yiheng  and
      Lv, Tengchao  and
      Cui, Lei  and
      Wang, Guoxin  and
      Lu, Yijuan  and
      Florencio, Dinei  and
      Zhang, Cha  and
      Wei, Furu",
    booktitle = "Findings of the Association for Computational Linguistics: ACL 2022",
    month = may,
    year = "2022",
    address = "Dublin, Ireland",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2022.findings-acl.253",
    doi = "10.18653/v1/2022.findings-acl.253",
    pages = "3214--3224",
    abstract = "Multimodal pre-training with text, layout, and image has achieved SOTA performance for visually rich document understanding tasks recently, which demonstrates the great potential for joint learning across different modalities. However, the existed research work has focused only on the English domain while neglecting the importance of multilingual generalization. In this paper, we introduce a human-annotated multilingual form understanding benchmark dataset named XFUND, which includes form understanding samples in 7 languages (Chinese, Japanese, Spanish, French, Italian, German, Portuguese). Meanwhile, we present LayoutXLM, a multimodal pre-trained model for multilingual document understanding, which aims to bridge the language barriers for visually rich document understanding. Experimental results show that the LayoutXLM model has significantly outperformed the existing SOTA cross-lingual pre-trained models on the XFUND dataset. The XFUND dataset and the pre-trained LayoutXLM model have been publicly available at https://aka.ms/layoutxlm.",
}

License

The content of this project itself is licensed under the Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0) Portions of the source code are based on the transformers project. Microsoft Open Source Code of Conduct

Contact Information

For help or issues using XFUND, please submit a GitHub issue.

For other communications related to XFUND, please contact Lei Cui (lecu@microsoft.com), Furu Wei (fuwei@microsoft.com).