Skip to content
This repository has been archived by the owner on Aug 18, 2021. It is now read-only.

seq2seq-translation-batched: Bahdanau attention does not work #82

Open
juditacs opened this issue Dec 28, 2017 · 2 comments
Open

seq2seq-translation-batched: Bahdanau attention does not work #82

juditacs opened this issue Dec 28, 2017 · 2 comments

Comments

@juditacs
Copy link

__init__.py fails with AttributeError, max_length does not exist. Fixing this results in a concat error in the Attn class:

    45         elif self.method == 'concat':
---> 46             energy = self.attn(torch.cat((hidden, encoder_output), 1))
     47             energy = self.v.dot(energy)
     48             return energy

RuntimeError: dimension out of range (expected to be in range of [-1, 0], but got 1)

replacing the dimension in line 46 to 0 results in this error:

    44 
     45         elif self.method == 'concat':
---> 46             energy = self.attn(torch.cat((hidden, encoder_output), 0))
     47             energy = self.v.dot(energy)
     48             return energy

RuntimeError: inconsistent tensor sizes at /opt/conda/conda-bld/pytorch_1512386481460/work/torch/lib/THC/generic/THCTensorMath.cu:157
@aevilorz
Copy link

aevilorz commented Jul 5, 2018

I think there could be a solution to attention calculation in matrix form in #107 .

@anantzoid
Copy link

anantzoid commented Oct 30, 2018

#119 addresses some of the mentioned issues.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants