Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
565 commits
Select commit Hold shift + click to select a range
1ed9877
🌐 [i18n-KO] Translated `gpt_neox_japanese.md` to Korean (#33894)
ahnjj Oct 9, 2024
2fe7778
🌐 [i18n-KO] Translated `rag.md` to Korean (#33989)
chhaewxn Oct 9, 2024
2e8de97
🌐 [i18n-KO] Translated `main_classes/quantization.md` to Korean (#33959)
fabxoe Oct 9, 2024
47da2c5
🌐 [i18n-KO] Translated `main_classes/configuration.md` to Korean (#3…
fabxoe Oct 9, 2024
1d45843
🌐 [i18n-KO] Translated `model_doc/mamba.md` to Korean (#33626)
fabxoe Oct 9, 2024
bb825dd
🌐 [i18n-KO] Translated `model_doc/autoformer.md` to Korean (#33574)
fabxoe Oct 9, 2024
acde6c7
🌐 [i18n-KO] Translated `model_doc/patchtsmixer.md` to Korean (#33587)
fabxoe Oct 9, 2024
dd43daf
🌐 [i18n-KO] Translated `model_doc/clip.md` to Korean (#33610)
fabxoe Oct 9, 2024
291e707
🌐 [i18n-KO] Translated `model_doc/paligemma.md` to Korean (#33612)
fabxoe Oct 9, 2024
adb14b9
🌐 [i18n-KO] Translated `model_doc/llama3.md` to Korean (#33635)
fabxoe Oct 9, 2024
48e8028
🌐 [i18n-KO] Translated `model_doc/mistral.md` to Korean (#33648)
fabxoe Oct 9, 2024
d6d07f9
🌐 [i18n-KO] Translated `model_doc/cohere.md` to Korean (#33885)
fabxoe Oct 9, 2024
cd9a3c4
🌐 [i18n-KO] Translated `model_doc/dbrx.md` to Korean (#33951)
fabxoe Oct 9, 2024
db5f117
🌐 [i18n-KO] Translated `model_doc/deberta-v2.md` to Korean (#33968)
fabxoe Oct 9, 2024
386401e
🌐 [i18n-KO] Translated `main_classes/onnx.md` to Korean (#33601)
fabxoe Oct 9, 2024
0d0ec1d
🌐 [i18n-KO] Translated `tokenization_utils.md` to Korean (#33813)
yijun-lee Oct 9, 2024
f0f8077
🌐 [i18n-KO] Translated `swin.md` to Korean (#33510)
mreraser Oct 9, 2024
c15d01f
🌐 [i18n-KO] Translated `file_utils.md` to Korean (#33803)
yijun-lee Oct 9, 2024
c674f2e
🌐 [i18n-KO] Translated `openai-gpt.md` to Korean (#33801)
yijun-lee Oct 9, 2024
5809b43
🌐 [i18n-KO] Translated `biogpt.md` to Korean (#33773)
yijun-lee Oct 9, 2024
109b1e7
🌐 [i18n-KO] Translated `blip.md` to Korean (#33515)
cjfghk5697 Oct 9, 2024
d31d076
🌐 [i18n-KO] Translated output.md to Korean (#33607)
4N3MONE Oct 9, 2024
6151bc4
🌐 [i18n-KO] Translated `image_processing_utils.md` to Korean (#33804)
yijun-lee Oct 9, 2024
698b36d
🌐 [i18n-KO] Translated `modular_transformers.md` to Korean (#33772)
yijun-lee Oct 9, 2024
e783f12
[`Patch helper`] update to not have to checkout main (#34006)
ArthurZucker Oct 9, 2024
faa0f63
Add gguf support for StableLM (#33793)
VladOS95-cyber Oct 9, 2024
cdee528
Fix Failed tests with mobile bert resize tokens embedding (#33950)
abuelnasr0 Oct 9, 2024
295a90c
Generate: remove most decoder-only LLMs `prepare_inputs_for_generatio…
gante Oct 9, 2024
5ee52ae
Mllama: fix tests (#34000)
zucchini-nlp Oct 9, 2024
4fb2870
Fix PIL dep for tests (#34028)
muellerzr Oct 9, 2024
48461c0
Make `pipeline` able to load `processor` (#32514)
qubvel Oct 9, 2024
36d410d
FEAT : Adding BitNet quantization method to HFQuantizer (#33410)
MekkCyber Oct 9, 2024
41794e6
🌐 [i18n-KO] Translated `model_doc/bart.md` to Korean (#33893)
fabxoe Oct 9, 2024
13929a0
🌐 [i18n-KO] Translated `model_doc/deberta.md` to Korean (#33967)
fabxoe Oct 9, 2024
03e6fa0
🌐 [i18n-KO] Translated `main_classes/keras_callbacks.md` to Korean (#…
fabxoe Oct 9, 2024
48b5420
🌐 [i18n-KO] Translated `model_doc/mamba2.md` to Korean (#33629)
fabxoe Oct 9, 2024
f0fbef1
🌐 [i18n-KO] Translated `main_classes/model.md` to Korean (#33606)
fabxoe Oct 9, 2024
32cc15c
🌐 [i18n-KO] Translated `model_doc/trajectory_transformer.md` to Korea…
fabxoe Oct 9, 2024
04f51c4
🌐 [i18n-KO] Translated `model_doc/time_series_transformer.md` to Kore…
fabxoe Oct 9, 2024
4ad9233
🌐 [i18n-KO] Translated `model_doc/informer.md` to Korean (#33585)
fabxoe Oct 9, 2024
8dca259
🌐 [i18n-KO] Translated `model_doc/graphormer.md` to Korean (#33569)
fabxoe Oct 9, 2024
6ac5f25
🌐 [i18n-KO] Translated `modeling_utils.md` to Korean (#33808)
yijun-lee Oct 9, 2024
61a6dce
🌐 [i18n-KO] Translated `main_classes/data_collator.md` to Korean (#33…
fabxoe Oct 9, 2024
973e606
🌐 [i18n-KO] Translated `model_doc/patchtst.md` to Korean (#33589)
fabxoe Oct 9, 2024
0354d44
🌐 [i18n-KO] Translated `text_generation.md` to Korean (#33777)
yijun-lee Oct 9, 2024
c02cf48
🌐 [i18n-KO] Translated `main_classes/callback.md` to Korean (#33572)
Jwaminju Oct 9, 2024
88d01d9
🌐 [i18n-KO] Translated `generation_utils.md` to Korean (#33818)
yijun-lee Oct 9, 2024
69b5ccb
Add Translate docs into Arabic - section files CONCEPTUAL GUIDES (#33…
AhmedAlmaghz Oct 9, 2024
a265600
add sdpa to OPT (#33298)
avishaiElmakies Oct 10, 2024
adea675
Phi3: fix attn for sliding window (#33586)
zucchini-nlp Oct 10, 2024
a84c413
HfArgumentParser: allow for hyhenated field names in long-options (#3…
djmarti Oct 10, 2024
66e08db
Fix pipelines tests (#34049)
qubvel Oct 10, 2024
c9afee5
Add gguf support for gpt2 (#34044)
VladOS95-cyber Oct 10, 2024
f8a260e
Sync QuestionAnsweringPipeline (#34039)
Rocketknight1 Oct 10, 2024
dda3f91
Specifying torch dtype in Qwen2VLForConditionalGeneration (#33953)
htahboub Oct 10, 2024
fb0c6b5
Universal Assisted Generation: Assisted generation with any assistant…
danielkorat Oct 10, 2024
4a3f1a6
check if eigenvalues of covariance matrix are complex. (#34037)
abuelnasr0 Oct 10, 2024
b2f09fb
[Docs] Update compressed_tensors.md (#33961)
mgoin Oct 10, 2024
a37a06a
Fix data_seed unused (#33731)
MekkCyber Oct 10, 2024
e7dfb91
[TESTS] ASR pipeline (#33925)
ylacombe Oct 10, 2024
8363fd8
Update Blip2 `is_pipeline_test_to_skip` method signature (#34067)
qubvel Oct 10, 2024
211f1d9
provide trust_remote_code for search feat extractor in model config (…
eaidova Oct 10, 2024
24b82f3
Small Fix to modular converter (#34051)
MekkCyber Oct 10, 2024
70b07d9
Default `synced_gpus` to `True` when using `FullyShardedDataParallel`…
Oct 10, 2024
7d97cca
Generate using exported model and enable gemma2-2b in ExecuTorch (#33…
guangy10 Oct 11, 2024
be9aeba
Idefics: fix position ids (#33907)
zucchini-nlp Oct 11, 2024
4b9bfd3
Update SSH workflow file (#34084)
ydshieh Oct 11, 2024
e878eaa
Tests: upcast `logits` to `float()` (#34042)
gante Oct 11, 2024
f052e94
Fix flax failures (#33912)
LysandreJik Oct 11, 2024
9dca0c9
Fix DAC slow tests (#34088)
ylacombe Oct 11, 2024
409dd2d
Fix failing conversion (#34010)
LysandreJik Oct 11, 2024
1c66be8
Fix PushToHubMixin when pusing to a PR revision (#34090)
Wauplin Oct 11, 2024
7b06473
avoid many failures for ImageGPT (#34071)
ydshieh Oct 11, 2024
3a24ba8
Fix NaNs in cost_matrix for mask2former (#34074)
ducha-aiki Oct 11, 2024
fd70464
Fix flaky tests (#34069)
zucchini-nlp Oct 11, 2024
37ac078
Generate: move `prepare_inputs_for_generation` in encoder-decoder llm…
gante Oct 11, 2024
80bee7b
Avoid many test failures for `LlavaNextVideoForConditionalGeneration`…
ydshieh Oct 11, 2024
144852f
refactor: benchmarks (#33896)
McPatate Oct 11, 2024
617b212
fix(ci): benchmarks dashboard was failing due to missing quotations (…
McPatate Oct 11, 2024
37ea040
Generate: Fix modern llm `generate` calls with `synced_gpus` (#34095)
gante Oct 12, 2024
7434c0e
Mistral-related models for QnA (#34045)
vasqu Oct 14, 2024
4c43917
Fix a typo (#34148)
PengWeixuan Oct 14, 2024
cb5ca32
Add GGUF for starcoder2 (#34094)
VladOS95-cyber Oct 14, 2024
013d3ac
Fixed error message in mllama (#34106)
dmgcsilva Oct 14, 2024
5114c9b
Specify that users should be careful with their own files (#34153)
LysandreJik Oct 14, 2024
fa3f2db
Add documentation for docker (#33156)
ArthurZucker Oct 14, 2024
dd4216b
Update README.md with Enterprise Hub (#34150)
gary149 Oct 15, 2024
23874f5
Idefics: enable generation tests (#34062)
zucchini-nlp Oct 15, 2024
293e627
Add sdpa for Vivit (#33757)
RUFFY-369 Oct 15, 2024
4de1bdb
Fix FSDP resume Initialization issue (#34032)
Itssshikhar Oct 15, 2024
5ee9e78
Fix default behaviour in TextClassificationPipeline for regression pr…
subhalingamd Oct 15, 2024
d314ce7
Generate: move `logits` to same device as `input_ids` (#34076)
gante Oct 15, 2024
6544271
Add support for inheritance from class with different suffix in modul…
yonigozlan Oct 15, 2024
d00f1ca
Fix optuna ddp hp search (#34073)
SunMarc Oct 15, 2024
0f49dea
[feat] LlavaNext add feature size check to avoid CUDA Runtime Error (…
laurentd-lunit Oct 15, 2024
67acb0b
🌐 [i18n-KO] Translated `vivit.md` to Korean (#33935)
mreraser Oct 15, 2024
8c33cf4
🌐 [i18n-KO] Translated `gemma2.md` to Korean (#33937)
yijun-lee Oct 15, 2024
554ed5d
🌐 [i18n-KO] Translated `trainer_utils.md` to Korean (#33817)
yijun-lee Oct 15, 2024
9d6998c
🌐 [i18n-KO] Translated `blip-2.md` to Korean (#33516)
cjfghk5697 Oct 15, 2024
d087165
IDEFICS: support inputs embeds (#34043)
zucchini-nlp Oct 16, 2024
9ba021e
Moshi integration (#33624)
ylacombe Oct 16, 2024
98bad9c
[fix] fix token healing tests and usage errors (#33931)
alpertunga-bile Oct 16, 2024
cc7d8b8
Revert `accelerate` error caused by `46d09af` (#34197)
steveepreston Oct 16, 2024
bd5dc10
Fix wrong name for llava onevision and qwen2_vl in tokenization auto …
yonigozlan Oct 16, 2024
3a10c61
Avoid using torch's Tensor or PIL's Image in chat template utils if n…
RezaRahemtola Oct 16, 2024
3f06f95
Revert "Fix FSDP resume Initialization issue" (#34193)
SunMarc Oct 16, 2024
6d2b203
Update `trainer._get_eval_sampler()` to support `group_by_length` arg…
larin92 Oct 17, 2024
aa3e35a
Fix warning message for fp32_cpu_offloading in bitsandbytes configs (…
amosyou Oct 17, 2024
fce1fcf
Ping team members for new failed tests in daily CI (#34171)
ydshieh Oct 17, 2024
b57c7bc
fix(Wav2Vec2ForCTC): torch export (#34023)
chrsmcgrr Oct 17, 2024
f2846ad
Fix for tokenizer.apply_chat_template with continue_final_message=Tru…
schoennenbeck Oct 17, 2024
7f50885
removes decord (#33987)
vrnvu Oct 17, 2024
9470c00
Llama3 and Llama2 are ExecuTorch compatible (#34101)
guangy10 Oct 17, 2024
1d2c29f
Fix bus error when using GPT2 on M1 macs (#34031)
chanind Oct 17, 2024
f51ac9e
Generate: visit non-llm `prepare_inputs_for_generation` (#34199)
gante Oct 17, 2024
c1c7e89
Fix Gradient Accumulation issue (#34191)
ArthurZucker Oct 17, 2024
7a06d07
Support Llama 3.2 conversion (text models) (#33778)
pcuenca Oct 17, 2024
6ba31a8
Enable users to use their own loss functions + deal with prefetching …
muellerzr Oct 17, 2024
b54109c
Fix-red-ci (#34230)
ArthurZucker Oct 17, 2024
5a5b590
BLIP: fix input expansion logic (#34225)
zucchini-nlp Oct 18, 2024
0437d6c
Fix broken test decorator `require_torch_up_to_2_accelerators` (#34201)
byi8220 Oct 18, 2024
e95ea47
Informative 2 (#34154)
LysandreJik Oct 18, 2024
6604764
add Glm (#33823)
Cyrilvallez Oct 18, 2024
e46e3bc
Fix UDOP dtype issue (#34180)
Rocketknight1 Oct 18, 2024
816f442
Only cast logits to float when computing loss (#34147)
Oct 18, 2024
ca541bd
Generation tests: don't rely on main input name (#34228)
zucchini-nlp Oct 21, 2024
24bdc94
Change Paligemma import logging to work with modular (#34211)
yonigozlan Oct 21, 2024
a412281
Add DetrImageProcessorFast (#34063)
yonigozlan Oct 21, 2024
f701b98
Add a doc section on writing generation prompts (#34248)
Rocketknight1 Oct 21, 2024
32590b5
Fix method name which changes in tutorial (#34252)
andimarafioti Oct 21, 2024
21d5025
Attn implementation for composite models (#32238)
zucchini-nlp Oct 22, 2024
5077bc0
VLM: add more modularity (#34175)
zucchini-nlp Oct 22, 2024
73d65e6
T5 compile compatibilty (#34089)
zucchini-nlp Oct 22, 2024
b644178
[docs] Fix GenerationConfig params (#34299)
stevhliu Oct 22, 2024
93352e8
Fix Korean doc _toctree.yml (#34293)
regisss Oct 22, 2024
681fc43
Sync video classification pipeline with huggingface_hub spec (#34288)
Rocketknight1 Oct 22, 2024
84b17e0
Update PR templates (#34065)
SunMarc Oct 22, 2024
eb6a734
[RT-DETR] Fix onnx inference bug for Optype (Where) (#33877)
YHallouard Oct 22, 2024
51e395d
Fix FA2 attention for models supporting sliding window (#34093)
Cyrilvallez Oct 22, 2024
104599d
Fix: tensor of examples of the same length triggers invalid stacking …
pbelcak Oct 22, 2024
c31a6ff
Add post_process_depth_estimation to image processors and support Zoe…
alex-bene Oct 22, 2024
7a08a77
Qwen2.5 is ExecuTorch Compatible (#34102)
guangy10 Oct 22, 2024
c14ccbc
Olmo is ExecuTorch Compatible (#34181)
guangy10 Oct 22, 2024
eef6b0b
Add option for running ffmpeg_microphone_live as a background process…
mikamerath Oct 22, 2024
96f67c0
Feature: Add `MLFLOW_MAX_LOG_PARAMS` to `MLflowCallback` (#34279)
cecheta Oct 22, 2024
e7c3fa7
Fix continue_final_message for image-text-to-text chat templates (#34…
yonigozlan Oct 22, 2024
a65a6ce
fix error in _get_eval_sampler when group_by_length enabled (#34237)
akakakakakaa Oct 22, 2024
688eeac
[docs] fix typo (#34235)
faaany Oct 22, 2024
4b14aa1
🌐 [i18n-KO] Translated `executorch.md` to Korean (#33888)
ahnjj Oct 22, 2024
b03dc0a
🌐 [i18n-KO] Translated `bert japanese.md` to Korean (#33890)
ahnjj Oct 22, 2024
644d528
🌐 [i18n-KO] Translated `model_doc/bartpho.md` to Korean (#33981)
Jwaminju Oct 22, 2024
049682a
Example doc for token classification of Llama and Dependent/Copied Mo…
h3110Fr13nd Oct 22, 2024
5ba85de
[docs] Fix Korean toctree (#34324)
stevhliu Oct 23, 2024
343c8cb
Added Deberta model type support (#34308)
FilipposVentirozos Oct 23, 2024
1fb575f
Support boolean tool args (#34208)
aymeric-roucher Oct 23, 2024
d9f7336
Enable Gradient Accumulation fix across all models + trainer fully in…
muellerzr Oct 23, 2024
c42b322
skip `test_pipeline_depth_estimation` temporarily (#34316)
ydshieh Oct 23, 2024
e50bf61
Fix red CI: benchmark script (#34351)
ArthurZucker Oct 23, 2024
b0f0c61
Add SynthID (watermerking by Google DeepMind) (#34350)
gante Oct 23, 2024
65753d6
Remove graph breaks for torch.compile() in flash_attention_forward wh…
Abhishek-TAMU Oct 24, 2024
0586381
Better defaults (#34026)
ArthurZucker Oct 24, 2024
f0e640a
Drop support for Python 3.8 (#34314)
ydshieh Oct 24, 2024
9643069
v4.47.0.dev0
ArthurZucker Oct 24, 2024
f0b3ef9
translated gguf.md into chinese (#34163)
blueingman Oct 24, 2024
b29c24f
CI: fix failures (#34371)
zucchini-nlp Oct 24, 2024
2112027
Zamba is an LM (#34342)
LysandreJik Oct 24, 2024
30c76d5
add code generation to natural language processing section (#34333)
furtnerthomas Oct 24, 2024
dd267fc
Add T5 GGUF loading support (#33389)
junejae Oct 24, 2024
6432ad8
Fix pil_torch_interpolation_mapping import in image_processing_detr_f…
yonigozlan Oct 24, 2024
450b9cb
Add code sample docstrings and checkpoint reference for GLM models (#…
h3110Fr13nd Oct 24, 2024
e288616
refactor: remove redundant if-condition and improve type correctness …
winstxnhdw Oct 24, 2024
fe35073
Ignore unsupported kwarg in ProcessorMixin call (#34285)
yonigozlan Oct 24, 2024
d9989e0
[PEFT] Add warning for missing key in LoRA adapter (#34068)
BenjaminBossan Oct 24, 2024
1c5918d
Fix `torch.fx` issue related to the new `loss_kwargs` keyword argumen…
michaelbenayoun Oct 24, 2024
4c6e0c9
Correct the new defaults (#34377)
Cyrilvallez Oct 24, 2024
a308d28
[auto. ping] Avoid sending empty info + add more team members (#34383)
ydshieh Oct 24, 2024
3d99f17
Fix glm (#34388)
Cyrilvallez Oct 24, 2024
940a6bd
Use non nested images and batched text Idefics2/3 (#34222)
yonigozlan Oct 25, 2024
5779bac
Fix onnx non-expotable inplace aten op (#34376)
IlyasMoutawwakil Oct 25, 2024
9f365fe
Fix right padding in LLaVA models (#34305)
zucchini-nlp Oct 25, 2024
2238553
no filter (#34391)
ydshieh Oct 25, 2024
8814043
SynthID: better example (#34372)
gante Oct 25, 2024
186b8dc
Tests: upgrade `test_eager_matches_sdpa_generate` (#34386)
gante Oct 25, 2024
e447185
Fix bnb training test failure (#34414)
matthewdouglas Oct 25, 2024
f73f5e6
Avoid check expected exception when it is on CUDA (#34408)
ydshieh Oct 25, 2024
6a62a6d
Fix typos in agents_advanced.md (#34405)
rudydel Oct 25, 2024
1d06379
[docs] Cache implementations (#34325)
stevhliu Oct 25, 2024
fddbd3c
Fix pix2struct (#34374)
IlyasMoutawwakil Oct 28, 2024
fc465bb
pin `tensorflow_probability<0.22` in docker files (#34381)
ydshieh Oct 28, 2024
9360f18
Tiny update after #34383 (#34404)
ydshieh Oct 28, 2024
92bcdff
Fix batch size handling in prediction_loop for DataLoaderShard (#34343)
zeus2611 Oct 28, 2024
8b3b9b4
exclude fsdp from delay_optimizer_creation (#34140)
eljandoubi Oct 28, 2024
c175343
New option called `"best"` for `args.save_strategy`. (#31817)
seanswyi Oct 28, 2024
fc1ae7f
[docs] update input documentation for MAMBA2 and MISTRAL models to in…
h3110Fr13nd Oct 28, 2024
1f7539c
🌐 [i18n-KO] Translated `model_doc/barthez.md` to Korean (#33980)
Jwaminju Oct 28, 2024
084e946
Apply linting to the important code blocks to make it readable (#34449)
ShubhamJagtap2000 Oct 28, 2024
a17f287
[i18n-ar] Translated file : `docs/source/ar/fast_tokenizers.md` into …
AhmedAlmaghz Oct 28, 2024
d21dbd1
enable average tokens across devices (#34373)
techkang Oct 28, 2024
6cc4a67
feat: run benchmarks on A100 (#34287)
McPatate Oct 28, 2024
a769ed4
Add `post_process_depth_estimation` for GLPN (#34413)
alex-bene Oct 28, 2024
fe76b60
LLaVA: latency issues (#34460)
zucchini-nlp Oct 29, 2024
808d6c5
Generation: fix test (#34369)
zucchini-nlp Oct 29, 2024
63ca6d9
Fix CI (#34458)
zucchini-nlp Oct 29, 2024
655bec2
use a tinymodel to test generation config which aviod timeout (#34482)
techkang Oct 29, 2024
a183519
🚨🚨🚨 [SuperPoint] Fix keypoint coordinate output and add post processi…
sbucaille Oct 29, 2024
439334c
Simplify running tests in a subprocess (#34213)
ydshieh Oct 29, 2024
626c610
Fix perplexity computation in perplexity.md (#34387)
Framartin Oct 29, 2024
9e3d704
Fixes for Modular Converter on Windows (#34266)
hlky Oct 29, 2024
004530a
Fix regression loading dtype (#34409)
SunMarc Oct 29, 2024
5392f12
Bert is ExecuTorch compatible (#34424)
guangy10 Oct 29, 2024
8755dd2
manual `head_dim` for `mixtral` model (#34281)
wavy-jung Oct 29, 2024
0ab0a42
fix-qwen2vl-no-position_ids (#33487)
simonJJJ Oct 29, 2024
56c45d5
Bug fix for drop path decay rate in swin transformer (#34291)
abhi-glitchhg Oct 29, 2024
34620e8
MobileBERT is ExecuTorch compatible (#34473)
guangy10 Oct 29, 2024
f339042
Albert is ExecuTorch compatible (#34476)
guangy10 Oct 29, 2024
e9ad460
Adding `optimizer_cls_and_kwargs` to `Trainer.__init__` (#34358)
apoorvkh Oct 29, 2024
4e2e880
Bump werkzeug from 3.0.3 to 3.0.6 in /examples/research_projects/deci…
dependabot[bot] Oct 29, 2024
f55595b
Fix performance in get_imports regexp (#34298)
AlekseyLobanov Oct 29, 2024
e4449bb
fix incorrect warning (#34416)
yonigozlan Oct 29, 2024
9bee9ff
Un-deprecate timeout arg in pipelines (#34382)
Rocketknight1 Oct 29, 2024
cd27761
Roberta is ExecuTorch compatible (#34425)
guangy10 Oct 30, 2024
25a9fc5
Fix format mistake in string repr of tokenizer objects (#34493)
gpetho Oct 30, 2024
0f764a5
Mllama: update docs (#34334)
zucchini-nlp Oct 30, 2024
913330c
VLMs: fix number of image tokens (#34332)
zucchini-nlp Oct 30, 2024
8a734ea
Tests: move `generate` tests to the right mixin and delete redundant …
gante Oct 30, 2024
241d790
fix pixtral processor (#34486)
molbap Oct 30, 2024
eab6c49
Use torch 2.5 in scheduled CI (#34465)
ydshieh Oct 30, 2024
5251fe6
Add GGUF for Mamba (#34200)
VladOS95-cyber Oct 30, 2024
9f06fb0
Fix super tiny extra space typo (#34440)
fzyzcjy Oct 30, 2024
48872fd
Add Image Processor Fast RT-DETR (#34354)
yonigozlan Oct 30, 2024
405b562
UPDATE Documentation for #TRANSLATING.md Documentation into Multiple …
anshumangahlot Oct 30, 2024
f385316
enable QA bf16 pipeline (#34483)
jiqing-feng Oct 31, 2024
1b86772
Fix: img size mismatch caused by incorrect unpadding in LLaVA-Next (#…
jp1924 Oct 31, 2024
dca93ca
Fix step shifting when accumulate gradient (#33673)
kibitzing Oct 31, 2024
ab98f0b
avoid calling `gc.collect` and `cuda.empty_cache` (#34514)
ydshieh Oct 31, 2024
4ca004e
Qwen2VL: skip base `input_ids`-`inputs_embeds` equivalence check (#34…
gante Oct 31, 2024
b5919e1
fix(DPT,Depth-Anything) Address expected_slice errors inside inferenc…
philkuz Oct 31, 2024
294c170
feat: add benchmarks pg indexes (#34536)
McPatate Oct 31, 2024
114dd81
make `test_eager_matches_sdpa_inference `less flaky (#34512)
ydshieh Oct 31, 2024
c443d8d
Bug Fix for issue #34294 (#34295)
fpgaminer Oct 31, 2024
203e270
Add image text to text pipeline (#34170)
yonigozlan Oct 31, 2024
df8640c
[CLIPSeg] Make interpolate_pos_encoding default to True (#34419)
NielsRogge Oct 31, 2024
2801d7b
update doc (#34478)
jiqing-feng Oct 31, 2024
b53e44e
[i18n-ar] Translated file : `docs/source/ar/multilingual.md` into Ara…
AhmedAlmaghz Oct 31, 2024
6beb3f1
Blip: get/set input embeddings correctly (#34152)
zucchini-nlp Nov 1, 2024
4cc0813
BLIP: enable generation tests (#34174)
zucchini-nlp Nov 1, 2024
86701f2
:red_circle: :red_circle: fix `query_pre_attn_scalar` different of `…
molbap Nov 1, 2024
e2ac16b
Large modular logic refactoring (#34487)
Cyrilvallez Nov 1, 2024
33868a0
[i18n-HI] Translated accelerate page to Hindi (#34443)
karthik-script Nov 1, 2024
b37a978
Merge remote-tracking branch 'zucchini-nlp/train_predict' into train_…
pdufour Nov 2, 2024
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
81 changes: 62 additions & 19 deletions .circleci/config.yml
Original file line number Diff line number Diff line change
Expand Up @@ -47,13 +47,13 @@ jobs:

- run:
name: "Retrieve Artifact Paths"
env:
CIRCLE_TOKEN: ${{ secrets.CI_ARTIFACT_TOKEN }}
# [reference] https://circleci.com/docs/api/v2/index.html#operation/getJobArtifacts
# `CIRCLE_TOKEN` is defined as an environment variables set within a context, see `https://circleci.com/docs/contexts/`
command: |
project_slug="gh/${CIRCLE_PROJECT_USERNAME}/${CIRCLE_PROJECT_REPONAME}"
job_number=${CIRCLE_BUILD_NUM}
url="https://circleci.com/api/v2/project/${project_slug}/${job_number}/artifacts"
curl -o test_preparation/artifacts.json ${url}
curl -o test_preparation/artifacts.json ${url} --header "Circle-Token: $CIRCLE_TOKEN"
- run:
name: "Prepare pipeline parameters"
command: |
Expand Down Expand Up @@ -82,22 +82,49 @@ jobs:
parallelism: 1
steps:
- checkout
- run: uv pip install -e .
- run: |
mkdir test_preparation
echo -n "tests" > test_preparation/test_list.txt
echo -n "all" > test_preparation/examples_test_list.txt
echo -n "tests/repo_utils" > test_preparation/test_repo_utils.txt
- run: uv pip install -U -e .
- run: echo 'export "GIT_COMMIT_MESSAGE=$(git show -s --format=%s)"' >> "$BASH_ENV" && source "$BASH_ENV"
- run: mkdir -p test_preparation
- run: python utils/tests_fetcher.py --fetch_all | tee tests_fetched_summary.txt
- run: python utils/tests_fetcher.py --filter_tests
- run: export "GIT_COMMIT_MESSAGE=$(git show -s --format=%s)" && echo $GIT_COMMIT_MESSAGE && python .circleci/create_circleci_config.py --fetcher_folder test_preparation
- run: |
echo -n "tests" > test_list.txt
python utils/tests_fetcher.py --filter_tests
mv test_list.txt test_preparation/filtered_test_list.txt
- run: python .circleci/create_circleci_config.py --fetcher_folder test_preparation
- run: cp test_preparation/generated_config.yml test_preparation/generated_config.txt
if [ ! -s test_preparation/generated_config.yml ]; then
echo "No tests to run, exiting early!"
circleci-agent step halt
fi

- store_artifacts:
path: test_preparation/generated_config.txt
path: test_preparation

- run:
name: "Retrieve Artifact Paths"
env:
CIRCLE_TOKEN: ${{ secrets.CI_ARTIFACT_TOKEN }}
command: |
project_slug="gh/${CIRCLE_PROJECT_USERNAME}/${CIRCLE_PROJECT_REPONAME}"
job_number=${CIRCLE_BUILD_NUM}
url="https://circleci.com/api/v2/project/${project_slug}/${job_number}/artifacts"
curl -o test_preparation/artifacts.json ${url}
- run:
name: "Prepare pipeline parameters"
command: |
python utils/process_test_artifacts.py

# To avoid too long generated_config.yaml on the continuation orb, we pass the links to the artifacts as parameters.
# Otherwise the list of tests was just too big. Explicit is good but for that it was a limitation.
# We used:

# https://circleci.com/docs/api/v2/index.html#operation/getJobArtifacts : to get the job artifacts
# We could not pass a nested dict, which is why we create the test_file_... parameters for every single job

- store_artifacts:
path: test_preparation/transformed_artifacts.json
- store_artifacts:
path: test_preparation/artifacts.json
- continuation/continue:
configuration_path: test_preparation/generated_config.yml
parameters: test_preparation/transformed_artifacts.json
configuration_path: test_preparation/generated_config.yml

check_code_quality:
working_directory: ~/transformers
Expand All @@ -110,7 +137,7 @@ jobs:
parallelism: 1
steps:
- checkout
- run: uv pip install -e .
- run: uv pip install -e ".[quality]"
- run:
name: Show installed libraries and their versions
command: pip freeze | tee installed.txt
Expand All @@ -135,13 +162,14 @@ jobs:
parallelism: 1
steps:
- checkout
- run: uv pip install -e .
- run: uv pip install -e ".[quality]"
- run:
name: Show installed libraries and their versions
command: pip freeze | tee installed.txt
- store_artifacts:
path: ~/transformers/installed.txt
- run: python utils/check_copies.py
- run: python utils/check_modular_conversion.py
- run: python utils/check_table.py
- run: python utils/check_dummies.py
- run: python utils/check_repo.py
Expand All @@ -158,13 +186,28 @@ workflows:
version: 2
setup_and_quality:
when:
not: <<pipeline.parameters.nightly>>
and:
- equal: [<<pipeline.project.git_url>>, https://github.com/huggingface/transformers]
- not: <<pipeline.parameters.nightly>>
jobs:
- check_circleci_user
- check_code_quality
- check_repository_consistency
- fetch_tests

setup_and_quality_2:
when:
not:
equal: [<<pipeline.project.git_url>>, https://github.com/huggingface/transformers]
jobs:
- check_circleci_user
- check_code_quality
- check_repository_consistency
- fetch_tests:
# [reference] https://circleci.com/docs/contexts/
context:
- TRANSFORMERS_CONTEXT

nightly:
when: <<pipeline.parameters.nightly>>
jobs:
Expand Down
11 changes: 10 additions & 1 deletion .circleci/create_circleci_config.py
Original file line number Diff line number Diff line change
Expand Up @@ -312,6 +312,15 @@ def job_name(self):
)


non_model_job = CircleCIJob(
"non_model",
docker_image=[{"image": "huggingface/transformers-torch-light"}],
marker="not generate",
parallelism=6,
pytest_num_workers=8,
)


# We also include a `dummy.py` file in the files to be doc-tested to prevent edge case failure. Otherwise, the pytest
# hangs forever during test collection while showing `collecting 0 items / 21 errors`. (To see this, we have to remove
# the bash output redirection.)
Expand All @@ -336,7 +345,7 @@ def job_name(self):
pytest_num_workers=1,
)

REGULAR_TESTS = [torch_and_tf_job, torch_and_flax_job, torch_job, tf_job, flax_job, hub_job, onnx_job, tokenization_job, processor_job, generate_job] # fmt: skip
REGULAR_TESTS = [torch_and_tf_job, torch_and_flax_job, torch_job, tf_job, flax_job, hub_job, onnx_job, tokenization_job, processor_job, generate_job, non_model_job] # fmt: skip
EXAMPLES_TESTS = [examples_torch_job, examples_tensorflow_job]
PIPELINE_TESTS = [pipelines_torch_job, pipelines_tf_job]
REPO_UTIL_TESTS = [repo_utils_job]
Expand Down
12 changes: 0 additions & 12 deletions .coveragerc

This file was deleted.

10 changes: 5 additions & 5 deletions .github/ISSUE_TEMPLATE/bug-report.yml
Original file line number Diff line number Diff line change
Expand Up @@ -37,25 +37,25 @@ body:
Models:

- text models: @ArthurZucker
- vision models: @amyeroberts
- speech models: @sanchit-gandhi
- vision models: @amyeroberts, @qubvel
- speech models: @ylacombe, @eustlb
- graph models: @clefourrier

Library:

- flax: @sanchit-gandhi
- generate: @zucchini-nlp (visual-language models) or @gante (all others)
- pipelines: @Narsil
- pipelines: @Rocketknight1
- tensorflow: @gante and @Rocketknight1
- tokenizers: @ArthurZucker
- tokenizers: @ArthurZucker and @itazap
- trainer: @muellerzr @SunMarc

Integrations:

- deepspeed: HF Trainer/Accelerate: @muellerzr
- ray/raytune: @richardliaw, @amogkam
- Big Model Inference: @SunMarc
- quantization (bitsandbytes, autogpt): @SunMarc
- quantization (bitsandbytes, autogpt): @SunMarc @MekkCyber

Documentation: @stevhliu

Expand Down
9 changes: 5 additions & 4 deletions .github/PULL_REQUEST_TEMPLATE.md
Original file line number Diff line number Diff line change
Expand Up @@ -40,25 +40,26 @@ members/contributors who may be interested in your PR.
Models:

- text models: @ArthurZucker
- vision models: @amyeroberts
- speech models: @sanchit-gandhi
- vision models: @amyeroberts, @qubvel
- speech models: @ylacombe, @eustlb
- graph models: @clefourrier

Library:

- flax: @sanchit-gandhi
- generate: @zucchini-nlp (visual-language models) or @gante (all others)
- pipelines: @Narsil
- pipelines: @Rocketknight1
- tensorflow: @gante and @Rocketknight1
- tokenizers: @ArthurZucker
- trainer: @muellerzr and @SunMarc
- chat templates: @Rocketknight1

Integrations:

- deepspeed: HF Trainer/Accelerate: @muellerzr
- ray/raytune: @richardliaw, @amogkam
- Big Model Inference: @SunMarc
- quantization (bitsandbytes, autogpt): @SunMarc
- quantization (bitsandbytes, autogpt): @SunMarc @MekkCyber

Documentation: @stevhliu

Expand Down
76 changes: 54 additions & 22 deletions .github/workflows/benchmark.yml
Original file line number Diff line number Diff line change
@@ -1,42 +1,74 @@
name: Self-hosted runner (benchmark)

on:
schedule:
- cron: "17 2 * * *"
workflow_call:
push:
branches: [main]
pull_request:
types: [ opened, labeled, reopened, synchronize ]

concurrency:
group: ${{ github.workflow }}-${{ github.head_ref || github.run_id }}
cancel-in-progress: true

env:
HF_HOME: /mnt/cache
TF_FORCE_GPU_ALLOW_GROWTH: true


jobs:
benchmark:
name: Benchmark
runs-on: [single-gpu, nvidia-gpu, a10, ci]
strategy:
matrix:
group: [aws-g5-4xlarge-cache, aws-p4d-24xlarge-plus]
runs-on:
group: ${{ matrix.group }}
if: |
(github.event_name == 'pull_request' && contains( github.event.pull_request.labels.*.name, 'run-benchmark') )||
(github.event_name == 'push' && github.ref == 'refs/heads/main')
container:
image: huggingface/transformers-all-latest-gpu
options: --gpus all --privileged --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/
image: huggingface/transformers-pytorch-gpu
options: --gpus all --privileged --ipc host
steps:
- name: Update clone
working-directory: /transformers
- name: Get repo
uses: actions/checkout@v4
with:
ref: ${{ github.event.pull_request.head.sha || github.sha }}

- name: Install libpq-dev & psql
run: |
git fetch && git checkout ${{ github.sha }}
apt update
apt install -y libpq-dev postgresql-client

- name: Install benchmark script dependencies
run: python3 -m pip install -r benchmark/requirements.txt

- name: Reinstall transformers in edit mode (remove the one installed during docker image build)
working-directory: /transformers
run: python3 -m pip uninstall -y transformers && python3 -m pip install -e .
run: python3 -m pip uninstall -y transformers && python3 -m pip install -e ".[torch]"

- name: Benchmark (daily)
if: github.event_name == 'schedule'
working-directory: /transformers
- name: Run database init script
run: |
python3 -m pip install optimum-benchmark>=0.3.0
HF_TOKEN=${{ secrets.TRANSFORMERS_BENCHMARK_TOKEN }} python3 benchmark/benchmark.py --repo_id hf-internal-testing/benchmark_results --path_in_repo $(date +'%Y-%m-%d') --config-dir benchmark/config --config-name generation --commit=${{ github.sha }} backend.model=google/gemma-2b backend.cache_implementation=null,static backend.torch_compile=false,true --multirun
psql -f benchmark/init_db.sql
env:
PGDATABASE: metrics
PGHOST: ${{ secrets.TRANSFORMERS_BENCHMARKS_PGHOST }}
PGUSER: transformers_benchmarks
PGPASSWORD: ${{ secrets.TRANSFORMERS_BENCHMARKS_PGPASSWORD }}

- name: Benchmark (merged to main event)
if: github.event_name == 'push' && github.ref_name == 'main'
working-directory: /transformers
- name: Run benchmark
run: |
python3 -m pip install optimum-benchmark>=0.3.0
HF_TOKEN=${{ secrets.TRANSFORMERS_BENCHMARK_TOKEN }} python3 benchmark/benchmark.py --repo_id hf-internal-testing/benchmark_results_merge_event --path_in_repo $(date +'%Y-%m-%d') --config-dir benchmark/config --config-name generation --commit=${{ github.sha }} backend.model=google/gemma-2b backend.cache_implementation=null,static backend.torch_compile=false,true --multirun
git config --global --add safe.directory /__w/transformers/transformers
if [ "$GITHUB_EVENT_NAME" = "pull_request" ]; then
commit_id=$(echo "${{ github.event.pull_request.head.sha }}")
elif [ "$GITHUB_EVENT_NAME" = "push" ]; then
commit_id=$GITHUB_SHA
fi
commit_msg=$(git show -s --format=%s | cut -c1-70)
python3 benchmark/llama.py "${{ github.head_ref || github.ref_name }}" "$commit_id" "$commit_msg"
env:
HF_TOKEN: ${{ secrets.HF_HUB_READ_TOKEN }}
# Enable this to see debug logs
# HF_HUB_VERBOSITY: debug
# TRANSFORMERS_VERBOSITY: debug
PGHOST: ${{ secrets.TRANSFORMERS_BENCHMARKS_PGHOST }}
PGUSER: transformers_benchmarks
PGPASSWORD: ${{ secrets.TRANSFORMERS_BENCHMARKS_PGPASSWORD }}
Loading