Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Can this model be trained on bert large? #74

Open
Umang9427 opened this issue Feb 13, 2020 · 4 comments
Open

Can this model be trained on bert large? #74

Umang9427 opened this issue Feb 13, 2020 · 4 comments

Comments

@Umang9427
Copy link

No description provided.

@ranjeetds
Copy link

Yes. While running run_ner.py script change --bert_model argument to --bert_model=bert-large-cased/uncased

@Umang9427
Copy link
Author

Yeah, but it will require higher specs of CUDA in machine, even google colab is not enough.

@ranjeetds
Copy link

You can always run it with a smaller batch size. You need to set --train_batch_size argument while running the script. For example, In my custom data case, some of the sentences are very long and hence I have to set max_seq_length to 512. I am using RTX 2080Ti GPU and with a batch size of 2 I can finish one epoch in ~4 hours (For ~60K training samples) for large bert model

@Swty13
Copy link

Swty13 commented Apr 21, 2020

@ranjeetds

Hi ,
I am training my custom NER model using BERT with 1 lakh dataset.Could you plz help me what hardware configuration is required to train this.

( I have VM having 32GB and 64GB RAM would it be possible to train on this or any GPU configuration is must to train this.)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants