- 17 Apr, 2023 15 commits
-
-
Sylvain Gugger authored
-
Sylvain Gugger authored
* Mark auto models as important * Annoying file with bad line endings
-
Zachary Mueller authored
* Use accelerate for device management * Add accelerate to setup Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Sylvain Gugger authored
Revert "Use code on the Hub from another repo (#22698)" This reverts commit ea7b0a53.
-
Sylvain Gugger authored
* Simplify update metadata job * Match more branch names * Install all what is necessary * Install all what is necessary * Forgot the dev * Install less stuff * This syntax?
-
Zachary Mueller authored
Remove accelerate from tf
-
Kunhao ZHENG authored
fix-squeeze-tuple
-
Yih-Dar authored
* fix --------- Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Sylvain Gugger authored
* initial work * Add other classes * Refactor code * Move warning and fix dynamic pipeline * Issue warning when necessary * Add test
-
Wonhyeong Seo authored
docs: ko: tasks/translation.mdx
-
Matt authored
-
fpgaminer authored
-
Jungnerd authored
fix: docs: ko: sagemaker anchors and `_toctree.yml` Co-authored-by:
Hyeonseo Yun <0525_hhgus@naver.com> Co-authored-by:
Gabriel Yang <gabrielwithhappy@gmail.com> Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> Co-authored-by:
Na Yeon Han <nayeon2.han@gmail.com> Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com>
-
Na Yeon Han authored
docs: ko: translated `custom_models.mdx` Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> Co-authored-by:
Gabriel Yang <gabrielwithhappy@gmail.com> Co-authored-by:
Jungnerd <46880056+jungnerd@users.noreply.github.com>
-
Yih-Dar authored
* fix --------- Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
- 15 Apr, 2023 1 commit
-
-
bcol authored
-
- 14 Apr, 2023 12 commits
-
-
oscar-garzon authored
-
amyeroberts authored
* Indexing fix - CLIP checkpoint conversion * Fix up
-
Joao Gante authored
-
Mayank Agarwal authored
* Fix word_ids hyperlink * Add suggested fix
-
Matt authored
* If EOS is None, don't add it to sequences * If EOS is None, don't add it to sequences
-
Sohyun Sim authored
* add ko preprocessing * translate preprocessing.mdx to korean * translate preprocessing.mdx * Update preprocessing.mdx Fixed the line 273 as below: 또한, 특징 추출기에 `sampling_rate` 인자를 추가하여 발생할 수 있는 조용한 오류(silent errors)를 더 잘 디버깅하는 것을 권장합니다. * translate Image part * translated preprocess.mdx * Update docs/source/ko/preprocessing.mdx Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/preprocessing.mdx Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/preprocessing.mdx Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/preprocessing.mdx Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/preprocessing.mdx Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/preprocessing.mdx Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/preprocessing.mdx Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/preprocessing.mdx Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * Update docs/source/ko/preprocessing.mdx * Update docs/source/ko/preprocessing.mdx * Update docs/source/ko/preprocessing.mdx * Update docs/source/ko/preprocessing.mdx * Update docs/source/ko/preprocessing.mdx * Update docs/source/ko/preprocessing.mdx * fixed translation --------- Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com>
-
Yih-Dar authored
* fix --------- Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Alexander Ljungberg authored
Fixed string format; better tokenizer message. Before: `Saving a {tokenizer_class} to {tokenizer_path}` After: `Saving a LlamaTokenizerFast to outdir.`
-
Joao Gante authored
-
Joao Gante authored
-
Sayak Paul authored
* add: tokenizer training script for TF TPU LM training. * add: script for preparing the TFRecord shards. * add: sequence of execution to readme. * remove limit from the tfrecord shard name. * Add initial train_model.py * Add basic training arguments and model init * Get up to the point of writing the data collator * Pushing progress so far! * Complete first draft of model training code * feat: grouping of texts efficiently. Co-authored-by:
Matt <rocketknight1@gmail.com> * Add proper masking collator and get training loop working * fix: things. * Read sample counts from filenames * Read sample counts from filenames * Draft README * Improve TPU warning * Use distribute instead of distribute.experimental * Apply suggestions from code review Co-authored-by:
Matt <Rocketknight1@users.noreply.github.com> * Modularize loading and add MLM probability as arg * minor refactoring to better use the cli args. * readme fillup. * include tpu and inference sections in the readme. * table of contents. * parallelize maps. * polish readme. * change script name to run_mlm.py * address PR feedback (round I). --------- Co-authored-by:
Matt <rocketknight1@gmail.com> Co-authored-by:
Matt <Rocketknight1@users.noreply.github.com>
-
Hyeonseo Yun authored
* docs: ko: init: tasks/sequence_classification.mdx * docs: ko: revised: change voca in tasks/sequence_classification.mdx * docs: ko: revised: [RE] change voca in tasks/sequence_classification.mdx * docs: ko: revised: spell check and sentence naturally in tasks/sequence_classification.mdx * docs: ko: revised: spell check and consistent vocabulary in tasks/sequence_classification.mdx * docs: ko: revised: Add full stop and change voca in tasks/sequence_classification.mdx * docs: ko: revised: sync first section templates in tasks/sequence_classification.mdx Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * fix: revert use of full-stops to colons * colons are used to emphasize the code block that follows * @0525hhgus @wonhyeongseo docs: ko: revised: sync second section templates in tasks/sequence_classification.mdx Co-Authored-By:
Wonhyeong Seo <wonhseo@kakao.com> * docs: ko: revised: change 'train', 'finetuning' in tasks/sequence_classification.mdx --------- Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com>
-
- 13 Apr, 2023 12 commits
-
-
Yih-Dar authored
* fix * style * fix --------- Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
fix Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Joao Gante authored
-
Ruiyang Sun authored
Bug in LlamaTokenizer when #22742
-
Stas Bekman authored
* [trainer] update url * style
-
Yih-Dar authored
fix Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
* fix --------- Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Gabriel Yang authored
translate training doc to Korean
-
Yih-Dar authored
* fix --------- Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
NielsRogge authored
* Add model to doc tests * Remove generate and replace by prepare_inputs_for_generation * More fixes * Remove print statements * Update integration tests * Fix generate * Remove model from auto mapping * Use auto processor * Fix integration tests * Fix test * Add inference code snippet * Remove is_encoder_decoder * Update docs * Remove notebook link
-
Rinat authored
* Update modeling_vilt.py Vilt compatible with model parallelism * Update modeling_switch_transformers.py switch_transformers compatible with model parallelism
-
Joel Lamy-Poirier authored
Fix indexing
-