1. 18 Nov, 2024 3 commits
    • Ofek Lev's avatar
      Fix broken link (#34618) · eb0ab3ed
      Ofek Lev authored
      eb0ab3ed
    • Raushan Turganbay's avatar
      VLMs: `patch_size` -> `num_image_tokens` in processing (#33424) · 1646ffb4
      Raushan Turganbay authored
      * use num additional tokens
      
      * fix copies + docs
      
      * another fix copies :)
      
      * add docs
      
      * move order for BC
      1646ffb4
    • Shane A's avatar
      Add OLMo November 2024 (#34551) · 3ee24e22
      Shane A authored
      * Add model skeletion with transformers-cli add-new-model-like
      
      * Convert config to modular, add rms_norm_eps, delete clip_qkv
      
      * Convert model to modular, add RMSNorm
      
      * Add flash attention with qk norm and no qkv clipping
      
      * Add decoder layer with RMSNorm after attention/feedforward layers
      
      * Add base and causal model
      
      * Add converter improvements from OLMo repo
      
      * Update weight loading in OLMo to HF converter
      
      * Set correct default for rms_norm_eps
      
      * Set correct pipeline_model_mapping in test
      
      * Run make fixup
      
      * Fix model type
      
      * Re-run modular conversion
      
      * Manually set config docs to fix build errors
      
      * Convert olmo-1124 to olmo_1124 to fix flash attention docs errors
      
      * Start updating tests
      
      * Update tests
      
      * Copy upstream test_eager_matches_sdpa_inference_1_bfloat16 changes to olmo_1124
      
      * Rename input_layernorm and post_attention_layernorm to reflect their ops better
      
      * Use correct tokenizer
      
      * Remove test unsupported by GPT2 tokenizer
      
      * Create GenerationConfig outside of from_pretrained call
      
      * Use simpler init file structure
      
      * Add explicit __all__ to support simplified init
      
      * Make safetensor serialization the default
      
      * Update OLMo November 2024 docs
      3ee24e22
  2. 15 Nov, 2024 7 commits
  3. 13 Nov, 2024 4 commits
  4. 11 Nov, 2024 3 commits
  5. 10 Nov, 2024 2 commits
  6. 09 Nov, 2024 1 commit
  7. 08 Nov, 2024 1 commit
  8. 07 Nov, 2024 2 commits
  9. 05 Nov, 2024 12 commits
  10. 04 Nov, 2024 5 commits