Transformer feed-forward layers build predictions by promoting concepts in the vocabulary space
Authors
Venue
arXiv preprint arXiv:2203.14680
BibTeX
Local Entry
@article{geva2022,
title = {Transformer feed-forward layers build predictions by promoting concepts in the vocabulary space},
author = {Mor Geva and Avi Caciularu and Kevin R Wang and Yoav Goldberg},
year = {2022},
journal = {arXiv preprint arXiv:2203.14680}
} From AUTO:OPENALEX
@article{geva2022,
title = {Transformer Feed-Forward Layers Build Predictions by Promoting Concepts in the Vocabulary Space},
author = {Mor Geva and Avi Caciularu and Kevin I‐Kai Wang and Yoav Goldberg},
year = {2022},
doi = {10.18653/v1/2022.emnlp-main.3}
}