1. 18 Jan, 2024 9 commits
    • Sanchit Gandhi's avatar
      [Whisper Tok] Move token ids to CPU when computing offsets (#28485) · 619ecfe2
      Sanchit Gandhi authored
      * move token ids to cpu
      
      * check for torch attr
      619ecfe2
    • Sanchit Gandhi's avatar
      [ASR Pipe] Update init to set model type and subsequently call parent init method (#28486) · 0eaa5ea3
      Sanchit Gandhi authored
      * add image processor arg
      
      * super
      
      * rm args
      0eaa5ea3
    • Jeremy Fowers's avatar
      Fix the documentation checkpoint for xlm-roberta-xl (#28567) · c662c78c
      Jeremy Fowers authored
      * Fix the documentation checkpoint for xlm-roberta-xl
      
      * Improve docstring consistency
      c662c78c
    • Yih-Dar's avatar
      Use `LoggingLevel` context manager in 3 tests (#28575) · 0754217c
      Yih-Dar authored
      
      * inside with LoggingLevel
      
      * remove is_flaky
      
      ---------
      
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      0754217c
    • Yoach Lacombe's avatar
      Add new meta w2v2-conformer BERT-like model (#28165) · d2cdefb9
      Yoach Lacombe authored
      
      * first commit
      
      * correct default value non causal
      
      * update config and modeling code
      
      * update converting checkpoint
      
      * clean modeling and fix tests
      
      * make style
      
      * add new config parameters to docstring
      
      * fix copied from statements
      
      * Apply suggestions from code review
      
      Co-authored-by: default avatarSanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
      
      * make position_embeddings_type docstrings clearer
      
      * clean converting script
      
      * remove function not used
      
      * clean modeling file
      
      * apply suggestion for test file + add convert script to not_doctested
      
      * modify tests according to review - cleaner logic and more tests
      
      * Apply nit suggestions from code review
      
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * add checker of valid position embeddings type
      
      * instantiate new layer norm layer with the right eps
      
      * fix freeze_feature_encoder since it can be None in some cases
      
      * add test same output in convert script
      
      * restore wav2vec2conformer and add new model
      
      * create processor and FE + clean
      
      * add new model code
      
      * fix convert script and set default config parameters
      
      * correct model id paths
      
      * make style
      
      * make fix-copies and cleaning files
      
      * fix copied from statements
      
      * complete .md and fixe copies
      
      * clean convert script argument defaults
      
      * fix config parameters docstrings
      
      * fix config docstring
      
      * add copied from and enrich FE tests
      
      * fix copied from and repo-consistency
      
      * add autotokenizer
      
      * make test input length shorter and change docstring code
      
      * fix docstrings and copied from
      
      * add add_adapter to ASR training example
      
      * make testing of adapters more robust
      
      * adapt to multi adapter layers
      
      * refactor input_values->input_features and remove w2v2-bert feature extractor
      
      * remove pretraining model
      
      * remove depreciated features and useless lines
      
      * add copied from and ignore statements to modeling tests
      
      * remove pretraining model #2
      
      * change import in convert script
      
      * change default in convert script
      
      * update readme and remove useless line
      
      * Update tests/models/wav2vec2_bert/test_processor_wav2vec2_bert.py
      
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * refactor BERT to Bert for consistency
      
      * remove useless ignore copy statement
      
      * add persistent to buffer in rotary
      
      * add eps in LayerNorm init and remove copied from
      
      * add adapter activation parameters and add copied from statements
      
      * Fix copied statements and add unitest.skip reasons
      
      * add copied statement in test_processor
      
      * refactor processor
      
      * make style
      
      * replace numpy random by torch rand
      
      * remove expected output CTC
      
      * improve converting script with processor class
      
      * Apply suggestions from code review
      
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * remove gumbel class
      
      * remove tests related to previously deleted class
      
      * Update src/transformers/models/wav2vec2_bert/configuration_wav2vec2_bert.py
      
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * correct typos
      
      * remove uused parameters
      
      * update processor to takes both text and audio
      
      * update checkpoints
      
      * update expected output and add ctc expected output
      
      * add label_attention_mask
      
      * replace pt with np in processor tests
      
      * fix typo
      
      * revert to behaviour with labels_attention_mask
      
      ---------
      
      Co-authored-by: default avatarSanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      d2cdefb9
    • hugo-syn's avatar
      chore: Fix multiple typos (#28574) · 5d8eb93e
      hugo-syn authored
      5d8eb93e
    • Arthur's avatar
      [`Core Tokenization`] Support a fix for spm fast models (#26678) · 81899778
      Arthur authored
      * fix
      
      * last attempt
      
      * current work
      
      * fix forward compatibility
      
      * save all special tokens
      
      * current state
      
      * revert additional changes
      
      * updates
      
      * remove tokenizer.model
      
      * add a test and the fix
      
      * nit
      
      * revert one more break
      
      * fix typefield issue
      
      * quality
      
      * more tests
      
      * fix fields for FC
      
      * more nits?
      
      * new additional changes
      
      * how
      
      * some updates
      
      * the fix
      
      * where do we stand
      
      * nits
      
      * nits
      
      * revert unrelated changes
      
      * nits nits nits
      
      * styling
      
      * don't break llama just yet
      
      * revert llama changes
      
      * safe arg check
      
      * fixup
      
      * Add a test for T5
      
      * Necessary changes
      
      * Tests passing, added tokens need to not be normalized. If the added tokens are normalized, it will the stripping which seems to be unwanted for a normal functioning
      
      * Add even more tests, when normalization is set to True (which does not work 😓 )
      
      * Add even more tests, when normalization is set to True (which does not work 😓 )
      
      * Update to main
      
      * nits
      
      * fmt
      
      * more and more test
      
      * comments
      
      * revert change as tests are failing
      
      * make the test more readble
      
      * nits
      
      * refactor the test
      
      * nit
      
      * updates
      
      * simplify
      
      * style
      
      * style
      
      * style convert slow
      
      * Update src/transformers/convert_slow_tokenizer.py
      81899778
    • Yih-Dar's avatar
      Use `weights_only` only if torch >= 1.13 (#28506) · a1668cc7
      Yih-Dar authored
      
      * fix
      
      * fix
      
      * fix
      
      ---------
      
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      a1668cc7
    • Yih-Dar's avatar
      Save `Processor` (#27761) · 3005f965
      Yih-Dar authored
      
      * save processor
      
      * Update tests/models/auto/test_processor_auto.py
      
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update tests/test_processing_common.py
      
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * fix
      
      ---------
      
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      3005f965
  2. 17 Jan, 2024 7 commits
  3. 16 Jan, 2024 10 commits
    • Joao Gante's avatar
    • inisis's avatar
      Add is_model_supported for fx (#28521) · 7142bdfa
      inisis authored
      
      * modify check_if_model_is_supported to return bool
      
      * add is_model_supported and have check_if_model_is_supported use that
      
      * Update src/transformers/utils/fx.py
      
      Fantastic
      
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      ---------
      
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      7142bdfa
    • fxmarty's avatar
      Clearer error for SDPA when explicitely requested (#28006) · 02f8738e
      fxmarty authored
      * clearer error for sdpa
      
      * better message
      02f8738e
    • Arthur's avatar
      [`SpeechT5Tokenization`] Add copied from and fix the... · fe23256b
      Arthur authored
      [`SpeechT5Tokenization`]  Add copied from and fix the `convert_tokens_to_string` to match the fast decoding scheme (#28522)
      
      * Add copied from and fix the `convert_tokens_to_string` to match the fast decoding scheme
      
      * fixup
      
      * add a small test
      
      * style test file
      
      * nites
      fe23256b
    • Arthur's avatar
      [`TokenizationRoformerFast`] Fix the save and loading (#28527) · 96d08831
      Arthur authored
      * cleanup
      
      * add a test
      
      * update the test
      
      * style
      
      * revert part that allows to pickle the tokenizer
      96d08831
    • Arthur's avatar
      [ `TokenizationUtils`] Fix `add_special_tokens` when the token is already there (#28520) · 716df5fb
      Arthur authored
      
      * fix adding special tokens when the token is already there.
      
      * add a test
      
      * add a test
      
      * nit
      
      * fix the test: make sure the order is preserved
      
      * Update tests/test_tokenization_common.py
      
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      ---------
      
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      716df5fb
    • Nima Yaqmuri's avatar
      Fix/speecht5 bug (#28481) · 07ae53e6
      Nima Yaqmuri authored
      * Fix bug in SpeechT5 speech decoder prenet's forward method
      
      - Removed redundant `repeat` operation on speaker_embeddings in the forward method. This line was erroneously duplicating the embeddings, leading to incorrect input size for concatenation and performance issues.
      - Maintained original functionality of the method, ensuring the integrity of the speech decoder prenet's forward pass remains intact.
      - This change resolves a critical bug affecting the model's performance in handling speaker embeddings.
      
      * Refactor SpeechT5 text to speech integration tests
      
      - Updated SpeechT5ForTextToSpeechIntegrationTests to accommodate the variability in sequence lengths due to dropout in the speech decoder pre-net. This change ensures that our tests are robust against random variations in generated speech, enhancing the reliability of our test suite.
      - Removed hardcoded dimensions in test assertions. Replaced with dynamic checks based on model configuration and seed settings, ensuring tests remain valid across different runs and configurations.
      - Added new test cases to thoroughly validate the shapes of generated spectrograms and waveforms. These tests leverage seed settings to ensure consistent and predictable behavior in testing, addressing potential issues in speech generation and vocoder processing.
      - Fixed existing test cases where incorrect assumptions about output shapes led to potential errors.
      
      * Fix bug in SpeechT5 speech decoder prenet's forward method
      
      - Removed redundant `repeat` operation on speaker_embeddings in the forward method. This line was erroneously duplicating the embeddings, leading to incorrect input size for concatenation and performance issues.
      - Maintained original functionality of the method, ensuring the integrity of the speech decoder prenet's forward pass remains intact.
      - This change resolves a critical bug affecting the model's performance in handling speaker embeddings.
      
      * Refactor SpeechT5 text to speech integration tests
      
      - Updated SpeechT5ForTextToSpeechIntegrationTests to accommodate the variability in sequence lengths due to dropout in the speech decoder pre-net. This change ensures that our tests are robust against random variations in generated speech, enhancing the reliability of our test suite.
      - Removed hardcoded dimensions in test assertions. Replaced with dynamic checks based on model configuration and seed settings, ensuring tests remain valid across different runs and configurations.
      - Added new test cases to thoroughly validate the shapes of generated spectrograms and waveforms. These tests leverage seed settings to ensure consistent and predictable behavior in testing, addressing potential issues in speech generation and vocoder processing.
      - Fixed existing test cases where incorrect assumptions about output shapes led to potential errors.
      
      * Enhance handling of speaker embeddings in SpeechT5
      
      - Refined the generate and generate_speech functions in the SpeechT5 class to robustly handle two scenarios for speaker embeddings: matching the batch size (one embedding per sample) and one-to-many (a single embedding for all samples in the batch).
      - The update includes logic to repeat the speaker embedding when a single embedding is provided for multiple samples, and a ValueError is raised for any mismatched dimensions.
      - Also added corresponding test cases to validate both scenarios, ensuring complete coverage and functionality for diverse speaker embedding situations.
      
      * Improve Test Robustness with Randomized Speaker Embeddings
      07ae53e6
    • fxmarty's avatar
      Fix mismatching loading in from_pretrained with/without accelerate (#28414) · 66db33dd
      fxmarty authored
      * fix mismatching behavior in from_pretrained with/without accelerate
      
      * meaningful refactor
      
      * remove added space
      
      * add test
      
      * fix model on the hub
      
      * comment
      
      * use tiny model
      
      * style
      66db33dd
    • Hamza FILALI's avatar
      Improving Training Performance and Scalability Documentation (#28497) · 002566f3
      Hamza FILALI authored
      
      * Improving Training Performance and Scaling documentation by adding PEFT techniques to suggestions to reduce memory requirements for training
      
      * Update docs/source/en/perf_train_gpu_one.md
      
      Co-authored-by: default avatarYounes Belkada <49240599+younesbelkada@users.noreply.github.com>
      
      ---------
      
      Co-authored-by: default avatarYounes Belkada <49240599+younesbelkada@users.noreply.github.com>
      002566f3
    • regisss's avatar
      Remove `task` arg in `load_dataset` in image-classification example (#28408) · 0cdcd7a2
      regisss authored
      * Remove `task` arg in `load_dataset` in image-classification example
      
      * Manage case where "train" is not in dataset
      
      * Add new args to manage image and label column names
      
      * Similar to audio-classification example
      
      * Fix README
      
      * Update tests
      0cdcd7a2
  4. 15 Jan, 2024 13 commits
  5. 13 Jan, 2024 1 commit