Skip to content

Commit

Permalink
Update author name for 2023.emnlp-main.534 (acl-org#2943)
Browse files Browse the repository at this point in the history
Co-authored-by: Lim Jia Peng
  • Loading branch information
jararap authored Dec 12, 2023
1 parent 267d715 commit 0de5abe
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion data/xml/2023.emnlp.xml
Original file line number Diff line number Diff line change
Expand Up @@ -6441,7 +6441,7 @@
</paper>
<paper id="534">
<title>Disentangling Transformer Language Models as Superposed Topic Models</title>
<author><first>Jia</first><last>Lim</last></author>
<author><first>Jia Peng</first><last>Lim</last></author>
<author><first>Hady</first><last>Lauw</last></author>
<pages>8646-8666</pages>
<abstract>Topic Modelling is an established research area where the quality of a given topic is measured using coherence metrics. Often, we infer topics from Neural Topic Models (NTM) by interpreting their decoder weights, consisting of top-activated words projected from individual neurons. Transformer-based Language Models (TLM) similarly consist of decoder weights. However, due to its hypothesised superposition properties, the final logits originating from the residual path are considered uninterpretable. Therefore, we posit that we can interpret TLM as superposed NTM by proposing a novel weight-based, model-agnostic and corpus-agnostic approach to search and disentangle decoder-only TLM, potentially mapping individual neurons to multiple coherent topics. Our results show that it is empirically feasible to disentangle coherent topics from GPT-2 models using the Wikipedia corpus. We validate this approach for GPT-2 models using Zero-Shot Topic Modelling. Finally, we extend the proposed approach to disentangle and analyse LLaMA models.</abstract>
Expand Down

0 comments on commit 0de5abe

Please sign in to comment.