-
Notifications
You must be signed in to change notification settings - Fork 26.4k
Pull requests: huggingface/transformers
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
Improve Error Messaging for Flash Attention 2 on CPU
#33655
opened Sep 22, 2024 by
sizhky
Loading 1�7
1 task done
Fix error string after refactoring into get_chat_template
#33652
opened Sep 22, 2024 by
tibor-reiss
Loading 1�7
feat(tokeniser): add lazy token conversion to ids
#33644
opened Sep 21, 2024 by
winstxnhdw
Loading 1�7
Fix typing in
load_balancing_loss_func function of modeling_mixtral.py.
#33641
opened Sep 21, 2024 by
PhilipMay
Loading 1�7
Fix the bug with loading DeepSpeed MoE models
#33640
opened Sep 21, 2024 by
foggy-frost-forest
Loading 1�7
1 of 5 tasks
Updated few changes to the code and updated the files which had few typos in it.
#33638
opened Sep 21, 2024 by
vignesh1507
Loading 1�7
Fix module initialization for root module under Zero3
#33632
opened Sep 20, 2024 by
Ben-Schneider-code
Loading 1�7
2 of 5 tasks
VLM Generate: tag
test_static_cache_matches_dynamic as flaky
#33630
opened Sep 20, 2024 by
gante
Loading 1�7
Uniformize kwargs for Udop processor and update docs
#33628
opened Sep 20, 2024 by
yonigozlan
Loading 1�7
5 tasks
Memory usage: new dynamic cache for models supporting sliding window attention
#33619
opened Sep 20, 2024 by
Cyrilvallez
Loading 1�7
Fix Llava conversion for LlavaQwen2ForCausalLM with Clip vision tower
#33613
opened Sep 20, 2024 by
Isotr0py
Loading 1�7
1 of 5 tasks
[feat] LlavaNext add feature size check to avoid CUDA Runtime Error
#33608
opened Sep 20, 2024 by
laurentd-lunit
Loading 1�7
1 of 5 tasks
Previous Next
ProTip!
Find all pull requests that aren't related to any open issues with -linked:issue.