Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[SDK] Use HuggingFace Data Collator for more Transformers in LLM Trainer #2032

Open
andreyvelich opened this issue Mar 15, 2024 · 3 comments
Assignees
Labels

Comments

@andreyvelich
Copy link
Member

andreyvelich commented Mar 15, 2024

More context: #2031 (comment).
Currently, we apply HuggingFace Data Collator only for AutoModelForCausalLM Transformer in HF LLM Trainer.

We need to investigate if we should apply it for other Transformers for language modelling models.

@live2awesome
Copy link

i am interested to contribute on this just i ping on this thread if any help is required
/assign

@live2awesome
Copy link

live2awesome commented Mar 20, 2024

what type of transformer we are looking . i have look for given below transformer model Data Collator can be used

  1. Masked Language Model - DataCollatorforLanguageModelling with mlm=True
  2. AutoModelForSeq2SeqLM - DataCollatorForSeq2Seq
  3. AutoModelForTokenClassification - DataCollatorForTokenClassification
  4. AutoModelForSequenceClassification - simple padding is sufficient
    There are also option for Permutation Language Modelling and Whole word mask . Kindly suggest @andreyvelich @johnugeorge

@andreyvelich
Copy link
Member Author

andreyvelich commented Mar 28, 2024

Thank you for your interest @live2awesome!
It would be nice if you could let us know what changes we need to make to our HF LLM Trainer to support Data Collators for other Transformers.
Also, we should discuss if we should add Data Collator by default to all supported transformers.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

2 participants