Issues: huggingface/transformers
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
OSError due to huggingface-hub FutureWarning about resume_download
bug
#31002
opened May 24, 2024 by
albertvillanova
Missing "config.json" when loading Llama-2-7b-chat-hf
#31001
opened May 24, 2024 by
stefanomezza
4 tasks
Trainer should throw a warning if max_sequence_length < number of tokens in dataset sample record.
Feature request
Request for a new feature
trainer
#30998
opened May 23, 2024 by
ronakk-google
loss calculation for PaliGemmaForConditionalGeneration potentially not cast to correct device
#30997
opened May 23, 2024 by
grahamannett
4 tasks
Add Nomic Embed Code to Transformers
New model
#30995
opened May 23, 2024 by
zanussbaum
2 tasks done
ViTLayer.forward() needs to be in "eager" mode when
output_attentions=True
#30978
opened May 23, 2024 by
alexsherstinsky
4 tasks
Autotokenizer."from_pretrained" read wrong config file. not "tokenizer_config.json", but "config.json"
#30977
opened May 23, 2024 by
daehuikim
2 of 4 tasks
RuntimeError: unique_by_key: failed to synchronize: cudaErrorIllegalAddress: an illegal memory access was encountered
Examples
Which is related to examples in general
#30976
opened May 23, 2024 by
HackXieHao
Training GPT2 with run_clm.py exceeds the described memory amount .
Examples
Which is related to examples in general
Good Second Issue
Issues that are more difficult to do than "Good First" issues - give it a try if you want!
#30969
opened May 22, 2024 by
CLL112
2 of 4 tasks
AttributeError: 'BertModel' object has no attribute 'attn_implementation'
#30965
opened May 22, 2024 by
pinnareet
2 of 4 tasks
Weird behaviour running AWQ code on RTX 4000 Ada that worked on Tesla T4
#30963
opened May 22, 2024 by
samrickman
2 of 4 tasks
google/siglip-so400m-patch14-384 inference output mismatch with pipeline output
Examples
Which is related to examples in general
Multimodal
#30951
opened May 22, 2024 by
aliencaocao
2 of 4 tasks
[LLaMA3] 'add_bos_token=True, add_eos_token=True' seems not taking effect
Core: Tokenization
Internals of the library; Tokenization.
Feature request
Request for a new feature
#30947
opened May 22, 2024 by
kiva12138
2 of 4 tasks
Llama Model throwing "RuntimeError: expected scalar type BFloat16 but found Float" when using torch.compile and AMP together
Compilation
Issues related to torchdynamo and torchinductor
#30945
opened May 21, 2024 by
JackCai1206
2 of 4 tasks
Trainer memory leak for evaluation with
compute_metrics
trainer
#30943
opened May 21, 2024 by
qubvel
2 of 4 tasks
Running out of memory while finetuning and inferencing VideoMAE due to which script is being killed.
#30939
opened May 21, 2024 by
pramodsarvi
1 of 4 tasks
Trained tokenizer has broken encoding for cyrillic
Core: Tokenization
Internals of the library; Tokenization.
#30937
opened May 21, 2024 by
Anna-Pinewood
2 of 4 tasks
CLIPTokenizerFast cause memory leak
Fast Tokenizers
Good Difficult Issue
#30930
opened May 21, 2024 by
janchen0611
2 of 4 tasks
Libraries import missing, unable to load image for inference and not able to load pipeline with the trained model
#30927
opened May 21, 2024 by
dejokz
2 of 4 tasks
Sink Cache Attention Scores are strange. CausalMask seems not working.
#30926
opened May 21, 2024 by
Tomorrowdawn
2 of 4 tasks
LLama3-70b LoRa results in OOM with torchrun but succeeds with python3 command
PEFT
#30923
opened May 20, 2024 by
premmotgi
4 tasks
[BUG] Offline loading of non-safe tensors fails
bug
#30920
opened May 20, 2024 by
pseudotensor
2 of 4 tasks
Previous Next
ProTip!
Type g p on any issue or pull request to go back to the pull request listing page.