Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[BERT/TF2] Global batch size not matching with the description #1378

Open
aandreas13 opened this issue Mar 5, 2024 · 0 comments
Open

[BERT/TF2] Global batch size not matching with the description #1378

aandreas13 opened this issue Mar 5, 2024 · 0 comments
Labels
bug Something isn't working

Comments

@aandreas13
Copy link

Hi, first of all this is such a great work and a very thorough documentation.

I just would like to ask a simple question. In the documentation of BERT/TF2, it is said that the global batch size is set to 61k (I assume it's rounded) for phase 1, and 30k for phase 2 training. However, if my understanding is correct
global_batch_size = batch_size * num_gpu * num_accumulation_steps
which if I use the described default parameter (60 * 64 * 8 = 30720 for phase 1, and 10 * 192 * 8 for phase 2 = 15360), which is exactly half of the set global batch size. Did I miss something here? or is there really a mistake?

Thanks in advance.

@aandreas13 aandreas13 added the bug Something isn't working label Mar 5, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

1 participant