- 18 Feb, 2025 1 commit
-
-
ydshieh authored
-
- 17 Feb, 2025 12 commits
-
-
Yih-Dar authored
* fix * fix * fix --------- Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Joao Gante authored
-
Marc Sun authored
add compressed_tensors in the dockerfile
-
dependabot[bot] authored
Bump transformers in /examples/research_projects/codeparrot/examples Bumps [transformers](https://github.com/huggingface/transformers) from 4.38.0 to 4.48.0. - [Release notes](https://github.com/huggingface/transformers/releases) - [Commits](https://github.com/huggingface/transformers/compare/v4.38.0...v4.48.0 ) --- updated-dependencies: - dependency-name: transformers dependency-type: direct:production ... Signed-off-by:
dependabot[bot] <support@github.com> Co-authored-by:
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
-
eustlb authored
-
Joao Gante authored
-
Arthur Zucker authored
-
ivarflakstad authored
-
Joao Gante authored
-
ivarflakstad authored
* Add hipsolver and hipblastlt as dependencies * Upgrade torch libs with rocm6.2.4 index
-
Yih-Dar authored
fix Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Fanli Lin authored
fix failed test
-
- 14 Feb, 2025 15 commits
-
-
Yih-Dar authored
* just skip * fix * fix * fix --------- Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
fix Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Mayank Mishra authored
* Modular GraniteMoE with shared Experts. Signed-off-by:
Shawn Tan <shawntan@ibm.com> * Modified * Import order. * Modified for style * Fix space. * Test * Remove extra granitemoe file. * New converted file and tests * Modified __init__ files. * Formatting. * Dummy PT objects * register granitemoe shared model Signed-off-by:
Sukriti-Sharma4 <sukriti.sharma4@ibm.com> * fix linting of a file Signed-off-by:
Sukriti-Sharma4 <sukriti.sharma4@ibm.com> * fix import in modeling file Signed-off-by:
Sukriti-Sharma4 <sukriti.sharma4@ibm.com> * update generated modeling file Signed-off-by:
Sukriti-Sharma4 <sukriti.sharma4@ibm.com> * add documentation Signed-off-by:
Sukriti-Sharma4 <sukriti.sharma4@ibm.com> * update docstrings Signed-off-by:
Sukriti-Sharma4 <sukriti.sharma4@ibm.com> * update generated modeling file Signed-off-by:
Sukriti-Sharma4 <sukriti.sharma4@ibm.com> * fix docstrings in config class Signed-off-by:
Sukriti-Sharma4 <sukriti.sharma4@ibm.com> * merge main Signed-off-by:
Sukriti-Sharma4 <sukriti.sharma4@ibm.com> --------- Signed-off-by:
Shawn Tan <shawntan@ibm.com> Signed-off-by:
Sukriti-Sharma4 <sukriti.sharma4@ibm.com> Co-authored-by:
Shawn Tan <shawntan@ibm.com> Co-authored-by:
Shawn Tan <shawn@wtf.sg> Co-authored-by:
Sukriti-Sharma4 <sukriti.sharma4@ibm.com> Co-authored-by:
Sukriti Sharma <Ssukriti@users.noreply.github.com>
-
ivarflakstad authored
-
Kyle Sayers authored
* add is_composition flag to LlavaConfig Signed-off-by:
Kyle Sayers <kylesayrs@gmail.com> * WIP: pixtral text config Signed-off-by:
Kyle Sayers <kylesayrs@gmail.com> * fix style Signed-off-by:
Kyle Sayers <kylesayrs@gmail.com> * add test Signed-off-by:
Kyle Sayers <kylesayrs@gmail.com> * use is_composition for pixtral Signed-off-by:
Kyle Sayers <kylesayrs@gmail.com> * Revert "use is_composition for pixtral" This reverts commit a53d5f9fc5149c84419b0e9e03db6d99362add53. * Revert "Revert "use is_composition for pixtral"" This reverts commit 3ab1c99404e2c2963fba0bcf94b9786d6365db0f. --------- Signed-off-by:
Kyle Sayers <kylesayrs@gmail.com>
-
Raushan Turganbay authored
* llavas * add mroe models * fix `compile_forward` test for all models * fix copies * make style * also doesn't support cache class * fix some tests * not copied from * ci green? * fix tests * fix copies * fix tests * check with `numel` and remove `item` * fix copies * fix copies * Update src/transformers/models/cohere2/modeling_cohere2.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * opt remove cross attn * gemma2 * fixup * fixup * fix newly added test * maybe fixed? * green please? --------- Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
David LaPalomento authored
* archive_file may not be specified When loading a pre-trained model from a gguf file, resolved_archive_file may not be set. Guard against that case in the safetensors availability check. * Remap partial disk offload to cpu for GGUF files GGUF files don't support disk offload so attempt to remap them to the CPU when device_map is auto. If device_map is anything else but None, raise a NotImplementedError. * Don't remap auto device_map and raise RuntimeError If device_map=auto and modules are selected for disk offload, don't attempt to map them to any other device. Raise a runtime error when a GGUF model is configured to map any modules to disk. --------- Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com>
-
Arthur authored
* dito * add a test * upsate * test needs fa2 * update test and configuration * test requires fa2 * style
-
Mohamed Mekkouri authored
fix
-
Andrei Panferov authored
* new flute * new higgs working * small adjustments * progress and quallity * small updates * style --------- Co-authored-by:
Andrey Panferov <panferov.andrey3@wb.ru> Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com> Co-authored-by:
Mohamed Mekkouri <93391238+MekkCyber@users.noreply.github.com>
-
Raushan Turganbay authored
* allow processor to preprocess conversation + video metadata * allow callable * add test * fix test * nit: fix * add metadata frames_indices * Update src/transformers/processing_utils.py Co-authored-by:
Pablo Montalvo <39954772+molbap@users.noreply.github.com> * Update src/transformers/processing_utils.py Co-authored-by:
Pablo Montalvo <39954772+molbap@users.noreply.github.com> * port updates from Orr and add one more test * Update src/transformers/processing_utils.py Co-authored-by:
Pablo Montalvo <39954772+molbap@users.noreply.github.com> * typo * as dataclass * style * docstring + maek sure tests green --------- Co-authored-by:
Pablo Montalvo <39954772+molbap@users.noreply.github.com>
-
Isotr0py authored
* add qwen2 fast image processor to modular file Signed-off-by:
isotr0py <2037008807@qq.com> * fix modular Signed-off-by:
isotr0py <2037008807@qq.com> * fix circle import Signed-off-by:
isotr0py <2037008807@qq.com> * add docs Signed-off-by:
isotr0py <2037008807@qq.com> * fix typo Signed-off-by:
isotr0py <2037008807@qq.com> * add modular generated files Signed-off-by:
isotr0py <2037008807@qq.com> * revert qwen2vl fast image processor Signed-off-by:
isotr0py <2037008807@qq.com> * remove qwen2.5-vl image processor from modular Signed-off-by:
isotr0py <2037008807@qq.com> * re-generate qwen2.5-vl files Signed-off-by:
isotr0py <2037008807@qq.com> * remove unnecessary test Signed-off-by:
isotr0py <2037008807@qq.com> * fix auto map Signed-off-by:
isotr0py <2037008807@qq.com> * cleanup Signed-off-by:
isotr0py <2037008807@qq.com> * fix model_input_names Signed-off-by:
isotr0py <2037008807@qq.com> * remove import Signed-off-by:
isotr0py <2037008807@qq.com> * make fix-copies Signed-off-by:
isotr0py <2037008807@qq.com> --------- Signed-off-by:
isotr0py <2037008807@qq.com>
-
Raushan Turganbay authored
* decompose chat template docs * add docs * update model docs * qwen2-5 * pixtral * remove old chat template * also video as list frames supported * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/chat_template_multimodal.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * remove audio for now --------- Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com>
-
Raushan Turganbay authored
* fix * also the docstring
-
Amit Garg authored
* Added support for partial_rotary_factor * addressed comments * refactored
-
- 13 Feb, 2025 12 commits
-
-
Yoni Gozlan authored
* uniformize owlvit processor * uniformize owlv2 * nit * add positional arg test owlvit * run-slow: owlvit, owlv2 * run-slow: owlvit, owlv2 * remove one letter variable
-
Yoni Gozlan authored
* add support for initial shift in video processing and other fixes * revert modifications video loading functions
-
Yih-Dar authored
fix my bad Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Mohamed Mekkouri authored
fix
-
Wizyoung authored
fix load key name for _load_rng_state under torch.cuda Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com>
-
Yih-Dar authored
* speeddddd * speeddddd * speeddddd * speeddddd --------- Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Jiahao Li authored
* Optimize Qwen2VL vision model by precomputing cos/sin embeds before ViT blocks * Make rotary_pos_emb optional & fix type * Adapt pre-computed cos/sin to Qwen2.5VL * More concise
-
மனோஜ்குமார் பழனிச்சாமி authored
* Remove traces of the progressbar * Use tqdm auto
-
Joao Gante authored
* tmp commit * move tests to the right class * remove ALL all_generative_model_classes = ... * skip tf roberta * skip InstructBlipForConditionalGenerationDecoderOnlyTest * videollava * reduce diff * reduce diff * remove on vlms * fix a few more * manual rebase bits * more manual rebase * remove all manual generative model class test entries * fix up to ernie * a few more removals * handle remaining cases * recurrent gemma * it's better here * make fixup * tf idefics is broken * tf bert + generate is broken * don't touch tf :() * don't touch tf :( * make fixup * better comments for test skips * revert tf changes * remove empty line removal * one more * missing one
-
Arthur authored
* add disable compile code * fix
-
Arthur authored
* fix training issues * Update Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com> --------- Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com>
-
Elvir Crnčević authored
* Resolve vptq conflict * Rename spqr package to spqr_quant * Get rid of aqlm mention * Start working on tests * Resolve ruff code checks * Ruff format * Isort * Test updates * Add gpu tag * Rename to modules_to_not_convert * Config update * Docs and config update * Docs and config update * Update to update_torch_dtype * spqr config parameter validation * Ruff update * Apply ruff fixes * Test fixes * Ruff update * Mark tests as @slow again; Ruff; Docstring update * Ruff * Remove absolute path * Resolve typo * Remove redundandt log * Check accelerate/spqr availability * Ruff fix * Check if the config contains proper shapes * Ruff test * Documentation update * overview update * Ruff checks * Ruff code quality * Make style * Update docs/source/en/quantization/spqr.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update spqr.md * Enable gptqmodel (#35012) * gptqmodel Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * fix format Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * update readme Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * gptqmodel need use checkpoint_format (#1) * gptqmodel need use checkpoint_format * fix quantize * Update quantization_config.py * Update quantization_config.py * Update quantization_config.py --------- Co-authored-by:
ZX-ModelCloud <zx@modelcloud.ai> Co-authored-by:
Qubitium-ModelCloud <qubitium@modelcloud.ai> * Revert quantizer_gptq.py (#2) * revert quantizer_gptq.py change * pass **kwargs * limit gptqmodel and optimum version Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * fix format Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * fix warning Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * fix version check Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * revert unrelated changes Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * enable gptqmodel tests Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * fix requires gptq Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * Fix Transformer compat (#3) * revert quantizer_gptq.py change * pass **kwargs * add meta info * cleanup * cleanup * Update quantization_config.py * hf_select_quant_linear pass checkpoint_format and meta * fix GPTQTestCUDA * Update test_gptq.py * gptqmodel.hf_select_quant_linear() now does not select ExllamaV2 * cleanup * add backend * cleanup * cleanup * no need check exllama version * Update quantization_config.py * lower checkpoint_format and backend * check none * cleanup * Update quantization_config.py * fix self.use_exllama == False * spell * fix unittest * fix unittest --------- Co-authored-by:
LRL <lrl@lbx.dev> Co-authored-by:
Qubitium-ModelCloud <qubitium@modelcloud.ai> * fix format Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * fix format again Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * update gptqmodel version (#6) * update gptqmodel version * update gptqmodel version * fix unit test (#5) * update gptqmodel version * update gptqmodel version * "not self.use_exllama" is not equivalent to "self.use_exllama==False" * fix unittest * update gptqmodel version * backend is loading_attibutes (#7) * fix format and tests Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * fix memory check Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * fix device mismatch Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * fix result check Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * Update src/transformers/quantizers/quantizer_gptq.py Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com> * Update src/transformers/quantizers/quantizer_gptq.py Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com> * Update src/transformers/quantizers/quantizer_gptq.py Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com> * update tests Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * review: update docs (#10) * review: update docs (#12) * review: update docs * fix typo * update tests for gptqmodel Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> * update document (#9) * update overview.md * cleanup * Update overview.md * Update overview.md * Update overview.md * update gptq.md * Update gptq.md * Update gptq.md * Update gptq.md * Update gptq.md * Update gptq.md * Update gptq.md --------- Co-authored-by:
Qubitium-ModelCloud <qubitium@modelcloud.ai> * typo * doc note for asymmetric quant * typo with apple silicon(e) * typo for marlin * column name revert: review * doc rocm support * Update docs/source/en/quantization/gptq.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/quantization/gptq.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/quantization/gptq.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/quantization/gptq.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/quantization/overview.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/quantization/overview.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> --------- Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> Co-authored-by:
LRL-ModelCloud <165116337+LRL-ModelCloud@users.noreply.github.com> Co-authored-by:
ZX-ModelCloud <zx@modelcloud.ai> Co-authored-by:
Qubitium-ModelCloud <qubitium@modelcloud.ai> Co-authored-by:
ZX-ModelCloud <165115237+ZX-ModelCloud@users.noreply.github.com> Co-authored-by:
LRL <lrl@lbx.dev> Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com> Co-authored-by:
Mohamed Mekkouri <93391238+MekkCyber@users.noreply.github.com> Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Fix : Nemotron Processor in GGUF conversion (#35708) * fixing nemotron processor * make style * Update docs/source/en/quantization/spqr.md Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Add missing TOC to doc --------- Signed-off-by:
jiqing-feng <jiqing.feng@intel.com> Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> Co-authored-by:
jiqing-feng <jiqing.feng@intel.com> Co-authored-by:
LRL-ModelCloud <165116337+LRL-ModelCloud@users.noreply.github.com> Co-authored-by:
ZX-ModelCloud <zx@modelcloud.ai> Co-authored-by:
Qubitium-ModelCloud <qubitium@modelcloud.ai> Co-authored-by:
ZX-ModelCloud <165115237+ZX-ModelCloud@users.noreply.github.com> Co-authored-by:
LRL <lrl@lbx.dev> Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com> Co-authored-by:
Mohamed Mekkouri <93391238+MekkCyber@users.noreply.github.com> Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-