Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About ROPE in sample process #54

Open
Leedonus opened this issue Aug 6, 2024 · 6 comments
Open

About ROPE in sample process #54

Leedonus opened this issue Aug 6, 2024 · 6 comments

Comments

@Leedonus
Copy link

Leedonus commented Aug 6, 2024

Hi, thanks for the interesting work. I want to know why the PE of the text token in generating process is set to zero?
image

@daiyixiang666

This comment was marked as abuse.

@zxduan90
Copy link

zxduan90 commented Aug 6, 2024

@daiyixiang666 But these zeros will make the q and k of the text_token zero, according to

xq = apply_rotary_emb(xq, freqs_cis)
, then the attention of the output token and the text token will be the same

@daiyixiang666
Copy link

Oh, I see, so the text only make effect via xv?

@zxduan90
Copy link

zxduan90 commented Aug 6, 2024

@daiyixiang666 yes, i think there is a problem in it

@zxduan90
Copy link

zxduan90 commented Aug 6, 2024

The text token has less relevance with the image compared to stable diffusion method
Uploading Screenshot from 2024-08-06 18-03-56.png…

@daiyixiang666
Copy link

And it will also mean the text attention mask is useless?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants