Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Regarding one-word summary #21

Open
LezJ opened this issue Mar 14, 2025 · 1 comment
Open

Regarding one-word summary #21

LezJ opened this issue Mar 14, 2025 · 1 comment

Comments

@LezJ
Copy link

LezJ commented Mar 14, 2025

Hi there,

Thank you for publishing this interesting work.

The one-word summary's embedding seems meaningful from the paper, as suggested by your Figure 1, which shows that the text/image/interleaved embedding is close to the relevant word.
I wonder how did you plot the Figure 1, specifically, the inputs for it. From your code the text/image/interleaved embedding refers to the last hidden state generated with one-word summary prompt. But how do you get corresponding embeddings for those single words (e.g. 'Dog', 'Cat') in this case? I assume it can't simply be the word embedding. Or is it just an illustrative figure?

Best

@kongds
Copy link
Owner

kongds commented Mar 15, 2025

Thank you for your interest in our work.

The single words refer to the embeddings from LM heads. Figure 1 is drawn based on the similarity matrix between embeddings and these single word embeddings (which correspond to next token probabilities for these words).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants