Interpreting Pretrained Contextualized Representations via Reductions to Static Embeddings

Published in ACL 2020, 2020

Contextualized representations (e.g. ELMo, BERT) have become the default pretrained representations for downstream NLP. We introduce simple and fully general methods for converting from contextualized representations to static lookup-table embeddings, applied to 5 popular pretrained models and 9 sets of pretrained weights. Our analysis reveals that pooling over many contexts significantly improves representational quality under intrinsic evaluation. We also consider social biases encoded in pretrained representations with respect to gender, race/ethnicity, and religion and find dramatic inconsistencies between social bias estimators for word embeddings.

Recommended citation: Rishi Bommasani, Kelly Davis, Claire Cardie. (2020). "Interpreting Pretrained Contextualized Representations via Reductions to Static Embeddings." ACL 2020.
Download Paper