- 14 Feb, 2025 11 commits
-
-
Kyle Sayers authored
* add is_composition flag to LlavaConfig Signed-off-by:
Kyle Sayers <kylesayrs@gmail.com> * WIP: pixtral text config Signed-off-by:
Kyle Sayers <kylesayrs@gmail.com> * fix style Signed-off-by:
Kyle Sayers <kylesayrs@gmail.com> * add test Signed-off-by:
Kyle Sayers <kylesayrs@gmail.com> * use is_composition for pixtral Signed-off-by:
Kyle Sayers <kylesayrs@gmail.com> * Revert "use is_composition for pixtral" This reverts commit a53d5f9fc5149c84419b0e9e03db6d99362add53. * Revert "Revert "use is_composition for pixtral"" This reverts commit 3ab1c99404e2c2963fba0bcf94b9786d6365db0f. --------- Signed-off-by:
Kyle Sayers <kylesayrs@gmail.com>
-
Raushan Turganbay authored
* llavas * add mroe models * fix `compile_forward` test for all models * fix copies * make style * also doesn't support cache class * fix some tests * not copied from * ci green? * fix tests * fix copies * fix tests * check with `numel` and remove `item` * fix copies * fix copies * Update src/transformers/models/cohere2/modeling_cohere2.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * opt remove cross attn * gemma2 * fixup * fixup * fix newly added test * maybe fixed? * green please? --------- Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
David LaPalomento authored
* archive_file may not be specified When loading a pre-trained model from a gguf file, resolved_archive_file may not be set. Guard against that case in the safetensors availability check. * Remap partial disk offload to cpu for GGUF files GGUF files don't support disk offload so attempt to remap them to the CPU when device_map is auto. If device_map is anything else but None, raise a NotImplementedError. * Don't remap auto device_map and raise RuntimeError If device_map=auto and modules are selected for disk offload, don't attempt to map them to any other device. Raise a runtime error when a GGUF model is configured to map any modules to disk. --------- Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com>
-
Arthur authored
* dito * add a test * upsate * test needs fa2 * update test and configuration * test requires fa2 * style
-
Mohamed Mekkouri authored
fix
-
Andrei Panferov authored
* new flute * new higgs working * small adjustments * progress and quallity * small updates * style --------- Co-authored-by:
Andrey Panferov <panferov.andrey3@wb.ru> Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com> Co-authored-by:
Mohamed Mekkouri <93391238+MekkCyber@users.noreply.github.com>
-
Raushan Turganbay authored
* allow processor to preprocess conversation + video metadata * allow callable * add test * fix test * nit: fix * add metadata frames_indices * Update src/transformers/processing_utils.py Co-authored-by:
Pablo Montalvo <39954772+molbap@users.noreply.github.com> * Update src/transformers/processing_utils.py Co-authored-by:
Pablo Montalvo <39954772+molbap@users.noreply.github.com> * port updates from Orr and add one more test * Update src/transformers/processing_utils.py Co-authored-by:
Pablo Montalvo <39954772+molbap@users.noreply.github.com> * typo * as dataclass * style * docstring + maek sure tests green --------- Co-authored-by:
Pablo Montalvo <39954772+molbap@users.noreply.github.com>
-
Isotr0py authored
* add qwen2 fast image processor to modular file Signed-off-by:
isotr0py <2037008807@qq.com> * fix modular Signed-off-by:
isotr0py <2037008807@qq.com> * fix circle import Signed-off-by:
isotr0py <2037008807@qq.com> * add docs Signed-off-by:
isotr0py <2037008807@qq.com> * fix typo Signed-off-by:
isotr0py <2037008807@qq.com> * add modular generated files Signed-off-by:
isotr0py <2037008807@qq.com> * revert qwen2vl fast image processor Signed-off-by:
isotr0py <2037008807@qq.com> * remove qwen2.5-vl image processor from modular Signed-off-by:
isotr0py <2037008807@qq.com> * re-generate qwen2.5-vl files Signed-off-by:
isotr0py <2037008807@qq.com> * remove unnecessary test Signed-off-by:
isotr0py <2037008807@qq.com> * fix auto map Signed-off-by:
isotr0py <2037008807@qq.com> * cleanup Signed-off-by:
isotr0py <2037008807@qq.com> * fix model_input_names Signed-off-by:
isotr0py <2037008807@qq.com> * remove import Signed-off-by:
isotr0py <2037008807@qq.com> * make fix-copies Signed-off-by:
isotr0py <2037008807@qq.com> --------- Signed-off-by:
isotr0py <2037008807@qq.com>
-
Raushan Turganbay authored
* decompose chat template docs * add docs * update model docs * qwen2-5 * pixtral * remove old chat template * also video as list frames supported * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * remove audio for now --------- Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com>
-
Raushan Turganbay authored
* fix * also the docstring
-
Amit Garg authored
* Added support for partial_rotary_factor * addressed comments * refactored
-
- 13 Feb, 2025 29 commits
-
-
Yoni Gozlan authored
* uniformize owlvit processor * uniformize owlv2 * nit * add positional arg test owlvit * run-slow: owlvit, owlv2 * run-slow: owlvit, owlv2 * remove one letter variable
-
Yoni Gozlan authored
* add support for initial shift in video processing and other fixes * revert modifications video loading functions
-
Yih-Dar authored
fix my bad Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Mohamed Mekkouri authored
fix
-
Wizyoung authored
fix load key name for _load_rng_state under torch.cuda Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com>
-
Yih-Dar authored
* speeddddd * speeddddd * speeddddd * speeddddd --------- Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Jiahao Li authored
* Optimize Qwen2VL vision model by precomputing cos/sin embeds before ViT blocks * Make rotary_pos_emb optional & fix type * Adapt pre-computed cos/sin to Qwen2.5VL * More concise
-
மனோஜ்குமார் பழனிச்சாமி authored
* Remove traces of the progressbar * Use tqdm auto
-
Joao Gante authored
* tmp commit * move tests to the right class * remove ALL all_generative_model_classes = ... * skip tf roberta * skip InstructBlipForConditionalGenerationDecoderOnlyTest * videollava * reduce diff * reduce diff * remove on vlms * fix a few more * manual rebase bits * more manual rebase * remove all manual generative model class test entries * fix up to ernie * a few more removals * handle remaining cases * recurrent gemma * it's better here * make fixup * tf idefics is broken * tf bert + generate is broken * don't touch tf :() * don't touch tf :( * make fixup * better comments for test skips * revert tf changes * remove empty line removal * one more * missing one
-
Arthur authored
* add disable compile code * fix
-
Arthur authored
* fix training issues * Update Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com> --------- Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com>
-
Elvir Crnčević authored
* Resolve vptq conflict * Rename spqr package to spqr_quant * Get rid of aqlm mention * Start working on tests * Resolve ruff code checks * Ruff format * Isort * Test updates * Add gpu tag * Rename to modules_to_not_convert * Config update * Docs and config update * Docs and config update * Update to update_torch_dtype * spqr config parameter validation * Ruff update * Apply ruff fixes * Test fixes * Ruff update * Mark tests as @slow again; Ruff; Docstring update * Ruff * Remove absolute path * Resolve typo * Remove redundandt log * Check accelerate/spqr availability * Ruff fix * Check if the config contains proper shapes * Ruff test * Documentation update * overview update * Ruff checks * Ruff code quality * Make style * Update docs/source/en/quantization/spqr.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update spqr.md * Enable gptqmodel (#35012) * gptqmodel Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * fix format Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * update readme Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * gptqmodel need use checkpoint_format (#1) * gptqmodel need use checkpoint_format * fix quantize * Update quantization_config.py * Update quantization_config.py * Update quantization_config.py --------- Co-authored-by:
ZX-ModelCloud <zx@modelcloud.ai> Co-authored-by:
Qubitium-ModelCloud <qubitium@modelcloud.ai> * Revert quantizer_gptq.py (#2) * revert quantizer_gptq.py change * pass **kwargs * limit gptqmodel and optimum version Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * fix format Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * fix warning Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * fix version check Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * revert unrelated changes Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * enable gptqmodel tests Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * fix requires gptq Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * Fix Transformer compat (#3) * revert quantizer_gptq.py change * pass **kwargs * add meta info * cleanup * cleanup * Update quantization_config.py * hf_select_quant_linear pass checkpoint_format and meta * fix GPTQTestCUDA * Update test_gptq.py * gptqmodel.hf_select_quant_linear() now does not select ExllamaV2 * cleanup * add backend * cleanup * cleanup * no need check exllama version * Update quantization_config.py * lower checkpoint_format and backend * check none * cleanup * Update quantization_config.py * fix self.use_exllama == False * spell * fix unittest * fix unittest --------- Co-authored-by:
LRL <lrl@lbx.dev> Co-authored-by:
Qubitium-ModelCloud <qubitium@modelcloud.ai> * fix format Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * fix format again Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * update gptqmodel version (#6) * update gptqmodel version * update gptqmodel version * fix unit test (#5) * update gptqmodel version * update gptqmodel version * "not self.use_exllama" is not equivalent to "self.use_exllama==False" * fix unittest * update gptqmodel version * backend is loading_attibutes (#7) * fix format and tests Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * fix memory check Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * fix device mismatch Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * fix result check Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * Update src/transformers/quantizers/quantizer_gptq.py Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com> * Update src/transformers/quantizers/quantizer_gptq.py Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com> * Update src/transformers/quantizers/quantizer_gptq.py Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com> * update tests Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * review: update docs (#10) * review: update docs (#12) * review: update docs * fix typo * update tests for gptqmodel Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * update document (#9) * update overview.md * cleanup * Update overview.md * Update overview.md * Update overview.md * update gptq.md * Update gptq.md * Update gptq.md * Update gptq.md * Update gptq.md * Update gptq.md * Update gptq.md --------- Co-authored-by:
Qubitium-ModelCloud <qubitium@modelcloud.ai> * typo * doc note for asymmetric quant * typo with apple silicon(e) * typo for marlin * column name revert: review * doc rocm support * Update docs/source/en/quantization/gptq.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/quantization/gptq.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/quantization/gptq.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/quantization/gptq.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/quantization/overview.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/quantization/overview.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> --------- Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> Co-authored-by:
LRL-ModelCloud <165116337+LRL-ModelCloud@users.noreply.github.com> Co-authored-by:
ZX-ModelCloud <zx@modelcloud.ai> Co-authored-by:
Qubitium-ModelCloud <qubitium@modelcloud.ai> Co-authored-by:
ZX-ModelCloud <165115237+ZX-ModelCloud@users.noreply.github.com> Co-authored-by:
LRL <lrl@lbx.dev> Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com> Co-authored-by:
Mohamed Mekkouri <93391238+MekkCyber@users.noreply.github.com> Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Fix : Nemotron Processor in GGUF conversion (#35708) * fixing nemotron processor * make style * Update docs/source/en/quantization/spqr.md Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Add missing TOC to doc --------- Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> Co-authored-by:
jiqing-feng <jiqing.feng@intel.com> Co-authored-by:
LRL-ModelCloud <165116337+LRL-ModelCloud@users.noreply.github.com> Co-authored-by:
ZX-ModelCloud <zx@modelcloud.ai> Co-authored-by:
Qubitium-ModelCloud <qubitium@modelcloud.ai> Co-authored-by:
ZX-ModelCloud <165115237+ZX-ModelCloud@users.noreply.github.com> Co-authored-by:
LRL <lrl@lbx.dev> Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com> Co-authored-by:
Mohamed Mekkouri <93391238+MekkCyber@users.noreply.github.com> Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
dependabot[bot] authored
Bump transformers in /examples/research_projects/adversarial Bumps [transformers](https://github.com/huggingface/transformers) from 4.38.0 to 4.48.0. - [Release notes](https://github.com/huggingface/transformers/releases) - [Commits](https://github.com/huggingface/transformers/compare/v4.38.0...v4.48.0 ) --- updated-dependencies: - dependency-name: transformers dependency-type: direct:production ... Signed-off-by:
dependabot[bot] <support@github.com> Co-authored-by:
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
-
dependabot[bot] authored
Bump transformers in /examples/tensorflow/language-modeling-tpu Bumps [transformers](https://github.com/huggingface/transformers) from 4.38.0 to 4.48.0. - [Release notes](https://github.com/huggingface/transformers/releases) - [Commits](https://github.com/huggingface/transformers/compare/v4.38.0...v4.48.0 ) --- updated-dependencies: - dependency-name: transformers dependency-type: direct:production ... Signed-off-by:
dependabot[bot] <support@github.com> Co-authored-by:
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
-
Joao Gante authored
* revert inputs_embeds len * Update test_utils.py * make fixup
-
Mohamed Mekkouri authored
* fix * fix
-
Arthur authored
test was weird
-
Joao Gante authored
skip modular checks based on diff
-
Pavel Iakubovskii authored
* Remove loading custom kernels * Remove config param * Fixup
-
Mohamed Mekkouri authored
* first commit * adding kernels * fix create_quantized_param * fix quantization logic * end2end * fix style * fix imports * fix consistency * update * fix style * update * udpate after review * make style * update * update * fix * update * fix docstring * update * update after review * update * fix scheme * update * update * fix * update * fix docstring * add source * fix test --------- Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com>
-
Lysandre Debut authored
* Helium documentation fixes * Update helium.md * Update helium.md * Update helium.md
-
Thomas Bauwens authored
* Add implementation for DataCollatorForMultipleChoice based on docs. * Add DataCollatorForMultipleChoice to import structure. * Remove custom DataCollatorForMultipleChoice implementations from example scripts. * Remove custom implementations of DataCollatorForMultipleChoice from docs in English, Spanish, Japanese and Korean. * Refactor torch version of DataCollatorForMultipleChoice to be more easily understandable. * Apply suggested changes and run make fixup. * fix copies, style and fixup * add missing documentation * nits * fix docstring * style * nits * isort --------- Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> Co-authored-by:
Arthur Zucker <arthur.zucker@gmail.com>
-
CL-ModelCloud authored
* Fix the bug in tokenizer.save_pretrained when saving tokenizer_class to tokenizer_config.json * Update tokenization_utils_base.py * Update tokenization_utils_base.py * Update tokenization_utils_base.py * add tokenizer class type test * code review * code opt * fix bug * Update test_tokenization_fast.py * ruff check * make style * code opt * Update test_tokenization_fast.py --------- Co-authored-by:
Qubitium-ModelCloud <qubitium@modelcloud.ai> Co-authored-by:
LRL-ModelCloud <165116337+LRL-ModelCloud@users.noreply.github.com>
-
Marco Edward Gorelli authored
-
gewenbin0992 authored
* qwen2.5vl: fix bugs when using flash2+bf16 or num_return_sequences>1 * fix * fix * fix * fix * add tests * fix test bugs * fix * fix failed tests * fix
-
Pavel Iakubovskii authored
* Trigger tests * [run-slow] beit, detr, dinov2, vit, textnet * Fix BEiT interpolate_pos_encoding * Fix DETR test * Update DINOv2 test * Fix textnet * Fix vit * Fix DPT * fix data2vec test * Fix textnet test * Update interpolation check * Fix ZoeDepth tests * Update interpolate embeddings for BEiT * Apply suggestions from code review
-
Lucain authored
-
Nerogar authored
fix gemma2 dtype issue when storing weights in float16 precision
-
Ben Schneider authored
* update env command to log deepspeed version * suppress deepspeed import logging * Add reminder to include configs to repro description in bug report. * make fixup * [WIP] update import utils for deepspeed * Change to using is_deepspeed_available() from integrations. * make fixup
-