Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

请问实现luong attn的时候为什么是用context来计算的 #5

Open
kwz219 opened this issue Jan 11, 2020 · 0 comments
Open

请问实现luong attn的时候为什么是用context来计算的 #5

kwz219 opened this issue Jan 11, 2020 · 0 comments

Comments

@kwz219
Copy link

kwz219 commented Jan 11, 2020

论文中写的是用target的当前hiddenstate和整个Encoder_outputs来计算,但是第七课seq2seq的代码中是这样算的:
context_in = self.linear_in(context.view(batch_size*input_len, -1)).view( batch_size, input_len, -1)
attn = torch.bmm(output, context_in.transpose(1,2))
这里的context是不是应该更换成当前时刻的hiddenstate

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant