- 30 Oct, 2024 2 commits
- 28 Oct, 2024 4 commits
-
-
ydshieh authored
-
Yih-Dar authored
* update * update * update --------- Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
0.21 Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Ilyas Moutawwakil authored
* fix * fix and test use_cache test * style * remove atol
-
- 25 Oct, 2024 10 commits
-
-
Steven Liu authored
cache
-
Rudy Delouya authored
-
Yih-Dar authored
* update * update --------- Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Matthew Douglas authored
* Fix bnb training test: compatibility with OPTSdpaAttention
-
Joao Gante authored
-
Joao Gante authored
* better example * Update src/transformers/generation/configuration_utils.py * Update src/transformers/generation/logits_process.py * nits
-
Yih-Dar authored
* no filter * no filter * no filter --------- Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Raushan Turganbay authored
* fix right pad llavas * device mismatch
-
Ilyas Moutawwakil authored
* fix onnx non-expotable inplace op * mistral, qwen2, qwen2_vl, starcoder2 * fixup copies
-
Yoni Gozlan authored
* add support for non nested images and add tests * add tests error scenario * fix style * added single and no image to error tests
-
- 24 Oct, 2024 18 commits
-
-
Cyril Vallez authored
* Fix duplicated * fix import
-
Yih-Dar authored
* update * update --------- Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Cyril Vallez authored
* Correct the new defaults * CIs * add check * Update utils.py * Update utils.py * Add the max_length in generate test checking shape without passing length * style * CIs * fix fx CI issue
-
Michael Benayoun authored
* Fix FX * Unskip tests
-
Benjamin Bossan authored
When loading a LoRA adapter, so far, there was only a warning when there were unexpected keys in the checkpoint. Now, there is also a warning when there are missing keys. This change is consistent with https://github.com/huggingface/peft/pull/2118 in PEFT and the planned PR https://github.com/huggingface/diffusers/pull/9622 in diffusers. Apart from this change, the error message for unexpected keys was slightly altered for consistency (it should be more readable now). Also, besides adding a test for the missing keys warning, a test for unexpected keys warning was also added, as it was missing so far.
-
Yoni Gozlan authored
Fix accept any common kwargs
-
Winston H. authored
refactor: remove redundant if-condition and improve type correctness for `convert_tokens_to_ids` (#34030) * chore: remove redundant if-condition * fix: import `Iterable`
-
Vijay authored
* Add code sample docstrings and checkpoint reference for GLM models * Update modular_glm.py * Update modeling_glm.py
-
Yoni Gozlan authored
fix pil_torch_interpolation_mapping import
-
김준재 authored
* add: GGUFT5Converter * add: tensormapping for t5 * add: test code for t5 * fix: Remove whitespace from blank line * add: t5 fp16 tests * fix: whitespace formatting * fix: minor formatting * fix: testing every weights
-
Thomas Furtner authored
-
Lysandre Debut authored
* Zamba is an LM * Addition
-
Raushan Turganbay authored
fix
-
王一苇 authored
* translated gguf.md into chinese * Apply suggestions from code review I have updated the PR accordingly.Thank you very much for detailed guidance,and I 'll pay more attention to the details next time. Co-authored-by:
Isotr0py <2037008807@qq.com> * Apply suggestions from code review Co-authored-by:
Isotr0py <2037008807@qq.com> --------- Co-authored-by:
Isotr0py <2037008807@qq.com>
-
Arthur Zucker authored
-
Yih-Dar authored
* drop python 3.8 * update docker files --------- Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Arthur authored
* be nice to our usres * nit * fixup * default to -1 * oups * turbo nit * auto infer framework
-
Abhishek Maurya authored
Remove graph breaks for torch.compile() in flash_attention_forward when Lllama Model is padding free tuned (#33932) * fix: fixes for graph breaks Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * fix: formatting Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * fix: import error Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * fix: Add Fa2Kwargs Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * fix: PR Changes Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * PR changes Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * PR changes Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * PR changes Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * PR changes Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * Revert "PR changes" This reverts commit 39d2868e5c93cc5f3f3c7c6ff981b66614c0e0e4. * PR changes Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * fix: FlashAttentionKwarg Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * fix: FlashAttentionKwarg Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * PR Changes Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * PR Changes Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * PR Changes Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * PR Changes Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * PR Changes Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * addition of documentation Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * change in _flash_attention_forward Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * make fix-copies Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * revert make fix-copies Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> * fix copies * style * loss kwargs typing * style and pull latest changes --------- Signed-off-by:
Abhishek <maurya.abhishek@ibm.com> Co-authored-by:
Arthur Zucker <arthur.zucker@gmail.com>
-
- 23 Oct, 2024 6 commits
-
-
Joao Gante authored
* Add SynthIDTextWatermarkLogitsProcessor * esolving comments. * Resolving comments. * esolving commits, * Improving SynthIDWatermark tests. * switch to PT version * detector as pretrained model + style * update training + style * rebase * Update logits_process.py * Improving SynthIDWatermark tests. * Shift detector training to wikitext negatives and stabilize with lower learning rate. * Clean up. * in for 7B * cleanup * upport python 3.8. * README and final cleanup. * HF Hub upload and initiaze. * Update requirements for synthid_text. * Adding SynthIDTextWatermarkDetector. * Detector testing. * Documentation changes. * Copyrights fix. * Fix detector api. * ironing out errors * ironing out errors * training checks * make fixup and make fix-copies * docstrings and add to docs * copyright * BC * test docstrings * move import * protect type hints * top level imports * watermarking example * direct imports * tpr fpr meaning * process_kwargs * SynthIDTextWatermarkingConfig docstring * assert -> exception * example updates * no immutable dict (cant be serialized) * pack fn * einsum equivalent * import order * fix test on gpu * add detector example --------- Co-authored-by:
Sumedh Ghaisas <sumedhg@google.com> Co-authored-by:
Marc Sun <marc@huggingface.co> Co-authored-by:
sumedhghaisas2 <138781311+sumedhghaisas2@users.noreply.github.com> Co-authored-by:
raushan <raushan@huggingface.co>
-
Arthur authored
* dont'trigger always * fux * oups * update * ?? * ? * aie
-
Yih-Dar authored
skip Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Zach Mueller authored
* Enable grad accum fix across all models + trainer fully in forward() * handle peft case * Account for DDP: need to run scale tests * Use accelerator state * Quality * Guard * Experiment w/ only fairseq fix * Fairseq only * Revert multiply_grads fix * Mult by grad accum to fully bring back solution * Style * Good to go now * Skip fx tests for now * Bookmark * Working now
-
Aymeric Roucher authored
Support boolean tool arguments
-
Filippos Ventirozos authored
* Added Deberta model type for 'add_prefix_space' functionality * housekeeping --------- Co-authored-by:
Filippos Ventirozos <filippos.ventirozos@autotrader.co.uk>
-