- 01 Sep, 2023 7 commits
-
-
Sanchit Gandhi authored
* [VITS] Add to TTA pipeline * Update tests/pipelines/test_pipelines_text_to_audio.py Co-authored-by:
Yoach Lacombe <52246514+ylacombe@users.noreply.github.com> * remove extra spaces --------- Co-authored-by:
Yoach Lacombe <52246514+ylacombe@users.noreply.github.com>
-
Zach Mueller authored
* Revert frozen training arguments * TODO
-
Omar Sanseviero authored
Update musicgen.md
-
Yih-Dar authored
* update * update * update * update --------- Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Joao Gante authored
-
pkumc authored
* fix FSDP resume optimizer & scheduler * improve trainer code quality --------- Co-authored-by:
machi04 <machi04@meituan.com>
-
Matthijs Hollemans authored
* add VITS model * let's vits * finish TextEncoder (mostly) * rename VITS to Vits * add StochasticDurationPredictor * ads flow model * add generator * correctly set vocab size * add tokenizer * remove processor & feature extractor * add PosteriorEncoder * add missing weights to SDP * also convert LJSpeech and VCTK checkpoints * add training stuff in forward * add placeholder tests for tokenizer * add placeholder tests for model * starting cleanup * let the great renaming begin! * use config * global_conditioning * more cleaning * renaming variables * more renaming * more renaming * it never ends * reticulating the splines * more renaming * HiFi-GAN * doc strings for main model * fixup * fix-copies * don't make it a PreTrainedModel * fixup * rename config options * remove training logic from forward pass * simplify relative position * use actual checkpoint * style * PR review fixes * more review changes * fixup * more unit tests * fixup * fix doc test * add integration test * improve tokenizer tests * add tokenizer integration test * fix tests on GPU (gave OOM) * conversion script can handle repos from hub * add conversion script for all MMS-TTS checkpoints * automatically create a README for the converted checkpoint * small changes to config * push README to hub * only show uroman note for checkpoints that need it * remove conversion script because code formatting breaks the readme * make WaveNet layers configurable * rename variables * simplifying the math * output attentions and hidden states * remove VitsFlip in flow model * also got rid of the other flip * fix tests * rename more variables * rename tokenizer, add phonemization * raise error when phonemizer missing * re-order config docstrings to match method * change config naming * remove redundant str -> list * fix copyright: vits authors -> kakao enterprise * (mean, log_variances) -> (prior_mean, prior_log_variances) * if return dict -> if not return dict * speed -> speaking rate * Apply suggestions from code review Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * update fused tanh sigmoid * reduce dims in tester * audio -> output_values * audio -> output_values in tuple out * fix return type * fix return type * make _unconstrained_rational_quadratic_spline a function * all nn's to accept a config * add spectro to output * move {speaking rate, noise scale, noise scale duration} to config * path -> attn_path * idxs -> valid idxs -> padded idxs * output values -> waveform * use config for attention * make generation work * harden integration test * add spectrogram to dict output * tokenizer refactor * make style * remove 'fake' padding token * harden tokenizer tests * ron norm test * fprop / save tests deterministic * move uroman to tokenizer as much as possible * better logger message * fix vivit imports * add uroman integration test * make style * up * matthijs -> sanchit-gandhi * fix tokenizer test * make fix-copies * fix dict comprehension * fix config tests * fix model tests * make outputs consistent with reverse/not reverse * fix key concat * more model details * add author * return dict * speaker error * labels error * Apply suggestions from code review Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/models/vits/convert_original_checkpoint.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * remove uromanize * add docstrings * add docstrings for tokenizer * upper-case skip messages * fix return dict * style * finish tests * update checkpoints * make style * remove doctest file * revert * fix docstring * fix tokenizer * remove uroman integration test * add sampling rate * fix docs / docstrings * style * add sr to model output * fix outputs * style / copies * fix docstring * fix copies * remove sr from model outputs * Update utils/documentation_tests.txt Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * add sr as allowed attr --------- Co-authored-by:
sanchit-gandhi <sanchit@huggingface.co> Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
- 31 Aug, 2023 11 commits
-
-
Marc Sun authored
* remove torch_dtype override * style * Update src/transformers/modeling_utils.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> --------- Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
Sylvain Gugger authored
* Smarter check for * Use protected functions * Do others too * Apply suggestions from code review Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Address review comments --------- Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
Yih-Dar authored
update Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
David Reguera authored
* Add type hints to `TFBlipTextModel` * Add missing type hints to DPR family models * Add type hints to `TFLEDModel` * Add type hints to `TFLxmertForPreTraining` * Add missing type hints to `TFMarianMTModel` and `TFMarianModel` * Add missing type hints to `TFRagModel` & `TFRagTokenForGeneration` * Make type hints annotations consistent
-
Younes Belkada authored
fix instructblip test
-
raghavanone authored
* Save image_processor while saving pipeline (ImageSegmentationPipeline) * Fix black issues
-
Arthur authored
* Fix coellama * style
-
Arthur authored
[`TokenizerFast`] `can_save_slow_tokenizer` as a property for when `vocab_file`'s folder was removed (#25626) * pad token should be None by default * fix tests * nits * check if isfile vocabfile * add warning if sp model folder was deleted * save SPM when missing folder for sloz * update the ` can_save_slow_tokenizer` to be a property * first batch * second batch * missing one
-
Vibhor Kumar authored
* Modify single-GPU efficient training doc with now-available adamw_bnb_8bit optimizer * Apply suggestions from code review Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> --------- Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com>
-
Sourab Mangrulkar authored
* fix ds z3 checkpointing when `stage3_gather_16bit_weights_on_model_save=False` * refactoring
-
qihqi authored
* For xla tensors, use an alternative way to get a unique id Because xla tensors don't have storage. * add is_torch_tpu_available check
-
- 30 Aug, 2023 11 commits
-
-
NielsRogge authored
Fix docstrings
-
Yih-Dar authored
fix Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Joao Gante authored
-
Marc Sun authored
-
Yih-Dar authored
fix Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Lysandre Debut authored
-
Juan Pizarro authored
* Add Blip2 model in VQA pipeline * use require_torch_gpu for test_large_model_pt_blip2 * use can_generate in vqa pipeline * test Blip2ForConditionalGeneration using float16 * remove custom can_generate from Blip2ForConditionalGeneration
-
Yih-Dar authored
fix Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Aman Gupta Karmani authored
fix minor documentation typo
-
Nino Risteski authored
deleted unnecessary comma in the Adding a new model section.
-
- 29 Aug, 2023 11 commits
-
-
Joao Gante authored
-
Nino Risteski authored
_toctree.yml file. broken link, now fixed.
-
Haylee Schäfer authored
* support loading base64 images * add test * mention in docs * remove the logging * sort imports * update error message * Update tests/utils/test_image_utils.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * restructure to catch base64 exception * doesn't like the newline * download files * format * optimize imports * guess it needs a space? * support loading base64 images * add test * remove the logging * sort imports * restructure to catch base64 exception * doesn't like the newline * download files * optimize imports * guess it needs a space? --------- Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
amyeroberts authored
Allocate result array ahead of time
-
Sanchit Gandhi authored
-
Susnato Dhar authored
update checkpoints
-
Arthur authored
update warning to If you want to use the new behaviour, set `legacy=False`. instead of True -
Sohyun Sim authored
* docs: ko: community.md * feat: deepl draft * fix: manual edits * fix: resolve suggestions Co-authored-by:
Hyeonseo Yun <0525yhs@gmail.com> Co-authored-by:
SeongWooChoi <46990061+nuatmochoi@users.noreply.github.com> --------- Co-authored-by:
Hyeonseo Yun <0525yhs@gmail.com> Co-authored-by:
SeongWooChoi <46990061+nuatmochoi@users.noreply.github.com>
-
heuristicwave authored
* dos: ko: add_new_pipeline.mdx * feat: chatgpt draft * fix: manual edits * docs: ko: add_new_pipeline Update _toctree * Update docs/source/ko/add_new_pipeline.md Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/add_new_pipeline.md Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/add_new_pipeline.md Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/add_new_pipeline.md Co-authored-by:
SeongWooChoi <46990061+nuatmochoi@users.noreply.github.com> * Update docs/source/ko/add_new_pipeline.md Co-authored-by:
SeongWooChoi <46990061+nuatmochoi@users.noreply.github.com> * Update docs/source/ko/add_new_pipeline.md Co-authored-by:
SeongWooChoi <46990061+nuatmochoi@users.noreply.github.com> * Update docs/source/ko/add_new_pipeline.md Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/add_new_pipeline.md Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/add_new_pipeline.md Co-authored-by:
SeongWooChoi <46990061+nuatmochoi@users.noreply.github.com> * Update docs/source/ko/add_new_pipeline.md Co-authored-by:
SeongWooChoi <46990061+nuatmochoi@users.noreply.github.com> * Update docs/source/ko/add_new_pipeline.md Co-authored-by:
SeongWooChoi <46990061+nuatmochoi@users.noreply.github.com> --------- Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> Co-authored-by:
SeongWooChoi <46990061+nuatmochoi@users.noreply.github.com>
-
Joao Gante authored
Tests: detect lines removed from "utils/not_doctested.txt" and doctest ALL generation files (#25763)
-
Chau Nguyen authored
* Update trainer.py (error with checking steps in args.eval_accumulation_steps to gather tensors) While the deprecated code has the correct check (line 3772): "if args.eval_accumulation_steps is not None and (step + 1) % args.eval_accumulation_steps == 0:" The current code does not (line 3196): "if args.eval_accumulation_steps is not None and self.accelerator.sync_gradients:" We need to check "(step + 1) % args.eval_accumulation_steps == 0". Hence, the line 3196 should be modified to: "if args.eval_accumulation_steps is not None and (step + 1) % args.eval_accumulation_steps == 0 and self.accelerator.sync_gradients:" * Fix error with checking args.eval_accumulation_steps to gather tensors
-