-
Notifications
You must be signed in to change notification settings - Fork 27.4k
Issues: huggingface/transformers
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
When gradient checkpointing is enabled, flash_attn_kwargs cannot be passed into the decoder_layer
bug
#35509
opened Jan 4, 2025 by
efsotr
4 tasks
Reload Transformers imports
Feature request
Request for a new feature
#35508
opened Jan 4, 2025 by
KareemMusleh
Memory Access out of bounds in mra/cuda_kernel.cu::index_max_cuda_kernel()
bug
#35507
opened Jan 4, 2025 by
dingfen
4 tasks
Wav2Vec2BertForSequenceClassification. return_attention_mask work wrong
bug
#35495
opened Jan 3, 2025 by
HERIUN
2 of 4 tasks
Loss.. should be specified as either training loss or validation loss
bug
#35494
opened Jan 3, 2025 by
FlogramMatt
4 tasks
Instructions to raise PR for addition of shared library files(.so) and .cpp files
Feature request
Request for a new feature
#35492
opened Jan 2, 2025 by
vineel96
How to run the model on another machine and send the answer to another machine.
bug
#35485
opened Jan 2, 2025 by
ixn3rd3mxn
4 tasks
Possible bug when using cosine lr scheduler with gradient accumulation
bug
#35484
opened Jan 2, 2025 by
hanyin88
4 tasks
Request for a Vision Transformer Model for Digital Image Segmentation
New model
#35477
opened Jan 2, 2025 by
hanshengzhu0001
2 tasks done
LayerDrop broken in various Flax models (Whisper/BART/more...)
bug
#35468
opened Dec 31, 2024 by
sssshhhhhh
2 of 4 tasks
Support SDPA & Flash Attention 2 for LayoutLMv3
Feature request
Request for a new feature
#35467
opened Dec 31, 2024 by
stancld
"Is it possible for Hugging Face to implement a chat model for quick information retrieval similar to vLLM?"
Feature request
Request for a new feature
#35464
opened Dec 31, 2024 by
BeastyZ
Qwen2-VL used to work with
inputs_embeds
instead of input_ids
, but no more
bug
#35463
opened Dec 31, 2024 by
minostauros
2 of 4 tasks
How can I disable legacy processing in llava-next
bug
#35457
opened Dec 30, 2024 by
foreverpiano
1 of 4 tasks
Installation Error for transformers Package (🔥 maturin failed)
bug
#35454
opened Dec 29, 2024 by
SauceChord
2 of 4 tasks
[Feature Request] Add beam search text streaming visualization feature
Feature request
Request for a new feature
#35451
opened Dec 29, 2024 by
MosheOfer1
Support Constant Learning Rate with Cooldown
Feature request
Request for a new feature
#35449
opened Dec 29, 2024 by
LoserCheems
Tokenizer does not split text according to newly added input tokens
bug
Core: Tokenization
Internals of the library; Tokenization.
#35447
opened Dec 29, 2024 by
jiongjiongli
2 of 4 tasks
Previous Next
ProTip!
Mix and match filters to narrow down what you’re looking for.