Home

Awesome

<br/> <p align="center"> <h1 align="center">Neural-Hidden-CRF: A Robust Weakly-Supervised Sequence Labeler (KDD-2023)</h1> <p align="center"> <a href="https://scholar.google.com/citations?user=3slpkWAAAAAJ&hl=zh-CN">Zhijun Chen</a>, <a href="https://hsun2022.github.io/">Hailong Sun</a>, <a >Wanhao Zhang</a>, <a >Chunyi Xu</a>, <a href="https://scholar.google.com/citations?hl=zh-CN&user=PnDqlPkAAAAJ&view_op=list_works&sortby=pubdate">Qianren Mao</a>, <a href="https://scholar.google.com/citations?user=mzXg1s8AAAAJ&hl=zh-CN">Pengpeng Chen</a> </p> <p align="center"> <a href='https://browse.arxiv.org/pdf/2309.05086.pdf'> <img src='https://img.shields.io/badge/Paper-PDF-red?style=flat&logo=arXiv&logoColor=red' alt='Paper PDF'> </a> <a href='https://www.youtube.com/watch?v=e9Z8tqzdgJk&t=7s' style='padding-left: 0.5rem;'> <img src='https://img.shields.io/badge/Short-Video-blue?style=flat&logo=Google%20chrome&logoColor=blue' alt='Short Video'> </a> <a href='https://github.com/junchenzhi/Neural-Hidden-CRF/blob/master/tutorial/Neural-Hidden-CRF.pdf' style='padding-left: 0.5rem;'> <img src='https://img.shields.io/badge/Tutorial-English-r?style=flat&logo=Google%20chrome&logoColor=red' alt='Short Video'> </a> <a href='https://github.com/junchenzhi/Neural-Hidden-CRF/blob/master/tutorial/Neural-Hidden-CRF(Chinese).pdf' style='padding-left: 0.5rem;'> <img src='https://img.shields.io/badge/Tutorial-Chinese-yellow?style=flat&logo=Google%20chrome&logoColor=red' alt='Short Video'> </a> </p> <p align="center"> <img src="fig/model.png" alt="Logo" width="55%"> </p> </p>

(Note that: Kindly refer to our arXiv version paper, as it contains additional sections that enhance the clarity of our proposed method when compared to the original version.)

1 Introduction

We propose a neuralized undirected graphical model called Neural-Hidden-CRF to solve the weakly-supervised sequence labeling problem. Under the umbrella of undirected graphical theory, the proposed Neural-Hidden-CRF embedded with a hidden CRF layer models the variables of word sequence, latent ground truth sequence, and weak label sequence with the global perspective that undirected graphical models particularly enjoy. In Neural-Hidden-CRF, we can capitalize on the powerful language model BERT or other deep models to provide rich contextual semantic knowledge to the latent ground truth sequence, and use the hidden CRF layer to capture the internal label dependencies. Neural-Hidden-CRF is conceptually simple and empirically powerful. It obtains new state-of-the-art results on one crowdsourcing benchmark and three weak-supervision benchmarks, including outperforming the recent advanced model CHMM by 2.80 F1 points and 2.23 F1 points in average generalization and inference performance, respectively.

In addition, since our learning context is weak supervision without making use of truth labels for learning, we can also refer to such a learning task as ``weak supervision zero-shot sequence labeling learning’’.

2 Code

2.1 Environment

2.2 Training/Evaluation

2.3 Others: Suggestions for Setting Hyperparameters

Note that: When applying our Neural-Hidden-CRF to other datasets, in most cases, we recommend considering the following suggestions for setting hyperparameters.

3 Some helpful materials

</p>

4 Citation

If you find our work useful, please consider citing:

@inproceedings{chen2023neural,
  title={Neural-Hidden-CRF: A Robust Weakly-Supervised Sequence Labeler},
  author={Chen, Zhijun and Sun, Hailong and Zhang, Wanhao and Xu, Chunyi and Mao, Qianren and Chen, Pengpeng},
  booktitle={Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining},
  pages={274--285},
  year={2023}
}

Acknowledgement

Our code is largely based on Wrench (https://github.com/JieyuZ2/wrench). Thanks for their contributions.

Contributors