Home

Awesome

Transformer Feed-Forward Layers Build Predictions by Promoting Concepts in the Vocabulary Space

This repository includes the accompanying code for the paper "Transformer Feed-Forward Layers Build Predictions by Promoting Concepts in the Vocabulary Space ". Mor Geva*, Avi Caciularu*, Kevin Ro Wang, and Yoav Goldberg.

In this repository, we provide the main scripts and notebooks used to reproduce our experiments.

<p align="center"><img width="40%" src="img/overview-1.png" /></p>

Structure

The repository contains:

For additional data or code please contact Avi, Mor or Kevin.

Citation

If you find our work useful, please cite the paper as:

@article{geva2022transformer,
      title={Transformer Feed-Forward Layers Build Predictions by Promoting Concepts in the Vocabulary Space},
      author={Geva, Mor and Caciularu, Avi and Wang, Kevin Ro and Goldberg, Yoav},
      journal={arXiv preprint arXiv:2203.14680},
      year={2022},
}