- 09 Apr, 2024 9 commits
-
-
Marc Sun authored
* revert back to torch 2.1.1 * run test * switch to torch 2.2.1 * udapte dockerfile * fix awq tests * fix test * run quanto tests * update tests * split quantization tests * fix * fix again * final fix * fix report artifact * build docker again * Revert "build docker again" This reverts commit 399a5f9d9308da071d79034f238c719de0f3532e. * debug * revert * style * new notification system * testing notfication * rebuild docker * fix_prev_ci_results * typo * remove warning * fix typo * fix artifact name * debug * issue fixed * debug again * fix * fix time * test notif with faling test * typo * issues again * final fix ? * run all quantization tests again * remove name to clear space * revert modfiication done on workflow * fix * build docker * build only quant docker * fix quantization ci * fix * fix report * better quantization_matrix * add print * revert to the basic one
-
Yih-Dar authored
Co-authored-by:
Wauplin <lucainp@gmail.com>
-
Yih-Dar authored
* fix mistral and mixtral * add pdb * fix mixtral tesst * fix * fix mistral ? * add fix gemma * fix mistral * fix * test * anoter test * fix * fix * fix mistral tests * fix them again * final fixes for mistral * fix padding right * fix whipser fa2 * fix * fix * fix gemma * test * fix llama * fix * fix * fix llama gemma * add class attribute * fix CI * clarify whisper * compute_capability * rename names in some comments * Add # fmt: skip * make style * Update tests/models/mistral/test_modeling_mistral.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * update * update --------- Co-authored-by:
Younes Belkada <younesbelkada@gmail.com> Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com> Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
NielsRogge authored
* Undo * Use tokenizer * Undo data collator
-
NielsRogge authored
* Fix data collator * Support feature extractors as well
-
Matt authored
* See if we can get tests to pass with the fixed weights * See if we can get tests to pass with the fixed weights * Replace the revisions now that we don't need them anymore
-
Raushan Turganbay authored
fix copies
-
Matthew Hoffman authored
* Add datasets.Dataset to Trainer's train_dataset and eval_dataset type hints * Add is_datasets_available check for importing datasets under TYPE_CHECKING guard https://github.com/huggingface/transformers/pull/30077/files#r1555939352
-
Sourab Mangrulkar authored
* fix sequence length errors * fix label column name error for vit * fix the lm_head embedding!=linear layer mismatches for Seq2Seq models
-
- 08 Apr, 2024 17 commits
-
-
Jonathan Tow authored
* init: add StableLm 2 support * add integration test for parallel residual and qk layernorm * update(modeling): match qk norm naming for consistency with phi/persimmon * fix(tests): run fwd/bwd on random init test model to jitter norm weights off identity * `use_parallel_residual`: add copy pointer to `GPTNeoXLayer.forward` * refactor: rename head states var in `StableLmLayerNormPerHead` * tests: update test model and add generate check
-
Felix Hirwa Nshuti authored
* adding env variable for mps and is_torch_mps_available for Pipeline * fix linting errors * Remove environment overide Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> --------- Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
DrAnaximandre authored
fix typo at ImportError
-
fxmarty authored
* remove controlflows * style * rename patch_ to padded_ following review comment * style
-
Younes Belkada authored
* Update trainer.py * fix copies
-
fxmarty authored
* fix falcon without attention_mask & alibi * add test * Update tests/models/falcon/test_modeling_falcon.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> --------- Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
Anton Vlasjuk authored
fix learning rate display issue in galore optimizer
-
Nick Doiron authored
* pass token to trainer.push_to_hub * fmt * Update src/transformers/trainer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * pass token to create_repo, update_folder --------- Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
Utkarsha Gupte authored
* ImportError: Trainer with PyTorch requires accelerate>=0.20.1 Fix Adding the evaluate and accelerate installs at the beginning of the cell to fix the issue * ImportError Fix: Trainer with PyTorch requires accelerate>=0.20.1 * Import Error Fix * Update installation.md * Update quicktour.md * rollback other lang changes * Update _config.py * updates for other languages * fixing error * Tutorial Update * Update tokenization_utils_base.py * Just use an optimizer string to pass the doctest? --------- Co-authored-by:
Matt <rocketknight1@gmail.com>
-
amyeroberts authored
* Patch fix - don't use safetensors for TF models * Skip test for TF for now * Update for another test
-
JINO ROHIT authored
-
Fanli Lin authored
* add bnb flag * move maker * add accelerator maker
-
Haz Sameen Shahgir authored
updated examples/pytorch/language-modeling scripts and requirements.txt to require datasets>=2.14.0 (#30120) updated requirements.txt and require_version() calls in examples/pytorch/language-modeling to require datasets>=2.14.0
-
Howard Liberty authored
* Make MLFlow version detection more robust and handles mlflow-skinny * Make function name more clear and refactor the logic * Further refactor
-
Xu Song authored
-
vaibhavagg303 authored
* add _torch_extract_fbank_features_batch function in feature_extractor_whisper * reformat feature_extraction_whisper.py file * handle batching in single function * add gpu test & doc * add batch test & device in each __call__ * add device arg in doc string --------- Co-authored-by:
vaibhav.aggarwal <vaibhav.aggarwal@sprinklr.com>
-
Cylis authored
-
- 05 Apr, 2024 13 commits
-
-
Raushan Turganbay authored
* clean-up whisper kwargs * failing test
-
Yih-Dar authored
* fix * fix --------- Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Kola authored
* Add docstrings and types for MambaCache * Update src/transformers/models/mamba/modeling_mamba.py * Update src/transformers/models/mamba/modeling_mamba.py * Update src/transformers/models/mamba/modeling_mamba.py * make fixup * import copy in generation_whisper * ruff * Revert "make fixup" This reverts commit c4fedd6f60e3b0f11974a11433bc130478829a5c.
-
Yih-Dar authored
* separate jobs * separate jobs * use channel name directly instead of ID * use channel name directly instead of ID * use channel name directly instead of ID --------- Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Michael Benayoun authored
* [WIP] fix fx * [WIP] fix fx * [WIP] fix fx * [WIP] fix fx * [WIP] fix fx * Apply changes to other models
-
Yih-Dar authored
* fix * fix --------- Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
miRx923 authored
Update quantizer_bnb_4bit.py: In the ValueError string there should be "....you need to set `llm_int8_enable_fp32_cpu_offload=True`...." instead of "`load_in_8bit_fp32_cpu_offload=True`". (#30013) * Update quantizer_bnb_4bit.py There is an mistake in ValueError on line 86 of quantizer_bnb_4bit.py. In the error string there should be "....you need to set `llm_int8_enable_fp32_cpu_offload=True`...." instead of "load_in_8bit_fp32_cpu_offload=True". I think you updated the BitsAndBytesConfig() arguments, but forgot to change the ValueError in quantizer_bnb_4bit.py. * Update quantizer_bnb_4bit.py Changed ValueError string "...you need to set load_in_8bit_fp32_cpu_offload=True..." to "....you need to set llm_int8_enable_fp32_cpu_offload=True...."
-
Marc Sun authored
fix bnb test
-
NielsRogge authored
* Add image processor to trainer * Replace tokenizer=image_processor everywhere
-
Adam Louly authored
* fix mixtral onnx export * fix qwen model
-
Wang, Yi authored
* if output is tuple like facebook/hf-seamless-m4t-medium, waveform is the first element Signed-off-by:
Wang, Yi <yi.a.wang@intel.com> * add test and fix batch issue Signed-off-by:
Wang, Yi <yi.a.wang@intel.com> * add dict output support for seamless_m4t Signed-off-by:
Wang, Yi <yi.a.wang@intel.com> --------- Signed-off-by:
Wang, Yi <yi.a.wang@intel.com>
-
Yih-Dar authored
skip test_encode_decode_fast_slow_all_tokens for now Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
Add whisper Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
- 04 Apr, 2024 1 commit
-
-
Saurabh Dash authored
* changes * addressing comments * smol fix
-