Home

Awesome

Low-resource Information Extraction πŸš€

🍎 The repository is a paper set on low-resource information extraction (IE), mainly including NER, RE and EE, which is generally categorized into two paradigms:

πŸ€— We strongly encourage the researchers who want to promote their fantastic work for the community to make pull request and update their papers in this repository!

πŸ“– Survey Paper: Information Extraction in Low-Resource Scenarios: Survey and Perspective (ICKG 2024) [paper]

πŸ—‚οΈ Slides:

Content

Preliminaries

🍎Traditional Methods🍎

🍏LLM-Based Methods🍏

How to Cite

<!-- * [Fine-Tuning LLM](#Fine-Tuning-LLM) --> <!-- * [Retrieval-Augmented Prompting](#Retrieval-Augmented-Prompting)-->

Preliminaries

πŸ› οΈ Low-Resource IE Toolkits

Traditional Toolkits

LLM-Based Toolkits

<!--- CollabKG: A Learnable Human-Machine-Cooperative Information Extraction Toolkit for (Event) Knowledge Graph Construction [[paper](https://arxiv.org/abs/2307.00769), [project](https://github.com/cocacola-lab/CollabKG)]-->

πŸ“Š Low-Resource IE Datasets

Low-Resource NER

Low-Resource RE

Low-Resource EE

πŸ“– Related Surveys and Analysis on Low-Resource IE

Information Extraction

NER

RE

EE

<!--* Low Resource Event Extraction: A Survey (2022) [[paper](https://www.cs.uoregon.edu/Reports/AREA-202210-Lai.pdf)\]-->

General IE

Traditional IE

LLM-based IE

<!--* Knowledge Extraction from Survey Data Using Neural Networks (Procedia Computer Science, 2013) \[[paper](https://www.sciencedirect.com/science/article/pii/S1877050913010995)\]-->

Low-Resource NLP Learning

🍎 Traditional Methods 🍎

1 Exploiting Higher-Resource Data

Weakly Supervised Augmentation

<!--* Distant Supervision for Relation Extraction with Sentence-Level Attention and Entity Descriptions (AAAI 2017) \[[paper](https://ojs.aaai.org/index.php/AAAI/article/view/10953)\]--> <!--* Reinforcement Learning for Relation Classification From Noisy Data (AAAI 2018) \[[paper](https://dl.acm.org/doi/abs/10.5555/3504035.3504744)\]--> <!--* Robust Distant Supervision Relation Extraction via Deep Reinforcement Learning (ACL 2018) \[[paper](https://aclanthology.org/P18-1199.pdf)\]--> <!--* Learning Named Entity Tagger using Domain-Specific Dictionary (EMNLP 2018) \[[paper](https://aclanthology.org/D18-1230.pdf)\]-->

Multimodal Augmentation

Multi-Lingual Augmentation

Auxiliary Knowledge Enhancement

(1) Textual Knowledge (Type-related Knowledge & Synthesized Data)

<!--The last four work are LLM-based DA-->

(2) Structured Knowledge (KG & Ontology & Logical Rules)

<!--* Neuralizing Regular Expressions for Slot Filling (EMNLP 2021) \[[paper](https://aclanthology.org/2021.emnlp-main.747.pdf)\]-->

2 Developing Stronger Data-Efficient Models

Meta Learning

For Low-Resource NER

For Low-Resource RE

For Low-Resource EE

Transfer Learning

Fine-Tuning PLM

3 Optimizing Data and Models Together

Multi-Task Learning

(1) IE & IE-Related Tasks

NER, Named Entity Normalization (NEN)

Word Sense Disambiguation (WSD), Event Detection (ED)

(2) Joint IE & Other Structured Prediction Tasks

NER, RE

NER, RE, EE

NER, RE, EE & Other Structured Prediction Tasks

Task Reformulation

Prompt-Tuning PLM

(1) Vanilla Prompt-Tuning

(2) Augmented Prompt-Tuning

🍏 LLM-Based Methods 🍏

Direct Inference Without Tuning

Instruction Prompting

Code Prompting

In-Context Learning

<!--### Retrieval-Augmented Prompting-->

Model Specialization With Tuning

Prompt-Tuning LLM

Fine-Tuning LLM

How to Cite

πŸ“‹ Thank you very much for your interest in our survey work. If you use or extend our survey, please cite the following paper:

@misc{2023_LowResIE,
    author    = {Shumin Deng and
                 Yubo Ma and
                 Ningyu Zhang and
                 Yixin Cao and
                 Bryan Hooi},
    title     = {Information Extraction in Low-Resource Scenarios: Survey and Perspective}, 
    journal   = {CoRR},
    volume    = {abs/2202.08063},
    year      = {2023},
    url       = {https://arxiv.org/abs/2202.08063}
}