8000 Increasing max_phrase_words runs into memory issues · Issue #20 · thunlp/BERT-KPE · GitHub
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

Increasing max_phrase_words runs into memory issues #20

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
lumalik opened this issue Jul 19, 2022 · 1 comment
Open

Increasing max_phrase_words runs into memory issues #20

lumalik opened this issue Jul 19, 2022 · 1 comment

Comments

@lumalik
Copy link
lumalik commented Jul 19, 2022

Hi, first of all, thank you very much for the repository! :)

I want to retrain the model using a larger number of keyphrases and output longer keyphrases in general.

To achieve this I:

  • increase the number of max_phrase_words from 5 to 10 in scripts/config.py
  • increase max_gram from 5 to 10 parameter in my model (in bertkpe/networks/)

However, I see that every number bigger than 5 makes me run out of memory during the step
"start preparing (train) features for bert2joint (bert) ..."
I can also see that if I increase the numbers to 6, I run out of memory at a much later stage in the preparation step than if I increase it to something higher like 10, even though the operation is performed on batches. I suspect that there is a memory leak in one of the data loader functions.

@SunSiShining
Copy link
Collaborator

Which dataset are you using? In which line does the OOM you mentioned occur? I may need more information.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants
0