-
Notifications
You must be signed in to change notification settings - Fork 87
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
train.py memory problem #1
Comments
What do you mean by "dies after a while"? |
it says. 'killed' after 20 minutes max |
the output of train are several files: |
I think "dies after a while" is because the seq_len is too long. |
interesting. It's a while ago so I don't remember if I used a sentence of a whole document as a sentence. but I guess i used sentences, so how would I chop them? |
@transfluxus I used Chinese corpus and it should be less than 300 words in each sentence; or crashed. I think it should be less than 1000 words for English corpus. I just split the sentence when encountered commas or full stops. |
i limited the sentence length to 100, still doesn't run through. actually already the train_word_embedding fails. |
is there a way to use a word embedding genereted with something else (like gensim for example).
This implementation dies after a while on my relatively large data set (with 32gb of memory)
The text was updated successfully, but these errors were encountered: