Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Training Time of Reranker #782

Open
Impavidity opened this issue May 14, 2024 · 5 comments
Open

Training Time of Reranker #782

Impavidity opened this issue May 14, 2024 · 5 comments

Comments

@Impavidity
Copy link

Impavidity commented May 14, 2024

Thanks for the great work and the open-source models. BTW, I am quite interested in the following questions.

  1. Total time to train the LLMReranker, such as Gemma and MiniCPM, under what kind of hardware.
  2. Max length for the query/passage and batch size when training with LLM Reranker.

Many thanks!

@545999961
Copy link
Collaborator

We trained for 4 days on 8 * 40G A100 GPUs. During training, the total length of query plus passage was 1024, and the batch size was 128.

@Impavidity
Copy link
Author

Thank you for you quick follow up. Sorry I have another question: How many epoch you trained on all m3+fever+quora data? Do you do any downsampling?

@545999961
Copy link
Collaborator

Training for 1-2 epochs is enough.

@Impavidity
Copy link
Author

Impavidity commented May 17, 2024

Thanks for the reply. Sorry I have more questions here.

  1. for long context example (e.g. length > 1k), do we decrease the batch size during the training? If it is, is this done automatically?
  2. During the training, is left padding used or right padding (for padding_side of tokenizer)

@545999961
Copy link
Collaborator

  1. It will truncate long contexts, so there is no need to decrease the batch size.
  2. Follow the tokenizer's raw padding side.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants