1. 02 Feb, 2024 1 commit
  2. 01 Feb, 2024 1 commit
  3. 25 Jan, 2024 1 commit
  4. 24 Jan, 2024 1 commit
    • Steven Liu's avatar
      [docs] DeepSpeed (#28542) · 738ec75c
      Steven Liu authored
      * config
      
      * optim
      
      * pre deploy
      
      * deploy
      
      * save weights, memory, troubleshoot, non-Trainer
      
      * done
      738ec75c
  5. 12 Jan, 2024 1 commit
  6. 02 Jan, 2024 1 commit
  7. 20 Dec, 2023 1 commit
  8. 18 Dec, 2023 1 commit
  9. 15 Dec, 2023 2 commits
  10. 11 Dec, 2023 1 commit
  11. 28 Nov, 2023 1 commit
  12. 27 Nov, 2023 1 commit
  13. 24 Nov, 2023 2 commits
  14. 20 Nov, 2023 1 commit
  15. 13 Nov, 2023 1 commit
  16. 09 Nov, 2023 1 commit
  17. 06 Nov, 2023 2 commits
  18. 01 Nov, 2023 2 commits
  19. 31 Oct, 2023 2 commits
  20. 30 Oct, 2023 1 commit
  21. 27 Oct, 2023 1 commit
  22. 26 Oct, 2023 1 commit
    • Marc Sun's avatar
      add exllamav2 arg (#26437) · 8214d6e7
      Marc Sun authored
      * add_ xllamav2 arg
      
      * add test
      
      * style
      
      * add check
      
      * add doc
      
      * replace by use_exllama_v2
      
      * fix tests
      
      * fix doc
      
      * style
      
      * better condition
      
      * fix logic
      
      * add deprecate msg
      8214d6e7
  23. 25 Oct, 2023 1 commit
  24. 24 Oct, 2023 1 commit
  25. 16 Oct, 2023 1 commit
  26. 12 Oct, 2023 2 commits
  27. 11 Oct, 2023 1 commit
  28. 10 Oct, 2023 1 commit
  29. 22 Sep, 2023 1 commit
  30. 15 Sep, 2023 1 commit
    • Matt's avatar
      Tweaks to Chat Templates docs (#26168) · 2518e368
      Matt authored
      * Put tokenizer methods in the right alphabetical order in the docs
      
      * Quick tweak to ConversationalPipeline
      
      * Typo fixes in the developer doc
      
      * make fixup
      2518e368
  31. 14 Sep, 2023 1 commit
    • Matt's avatar
      Overhaul Conversation class and prompt templating (#25323) · 866df66f
      Matt authored
      
      
      * First commit while I figure this out
      
      * make fixup
      
      * Remove unused method
      
      * Store prompt attrib
      
      * Fix prompt argument for tests
      
      * Make same changes in fast tokenizer
      
      * Remove global prompts from fast tokenizer too
      
      * stash commit
      
      * stash commit
      
      * Migrate PromptConfig to its True Final Location
      
      * Replace Conversation entirely with the new class
      
      * Import/dependency fixes
      
      * Import/dependency fixes
      
      * Change format for lots of default prompts
      
      * More default prompt fixups
      
      * Revert llama old methods so we can compare
      
      * Fix some default configs
      
      * Fix some default configs
      
      * Fix misspelled kwarg
      
      * Fixes for Blenderbot
      
      * make fixup
      
      * little rebase cleanup
      
      * Add basic documentation
      
      * Quick doc fix
      
      * Truncate docstring for now
      
      * Add handling for the case when messages is a single string
      
      * Quick llama merges
      
      * Update conversational pipeline and tests
      
      * Add a couple of legacy properties for backward compatibility
      
      * More legacy handling
      
      * Add docstring for build_conversation_input_ids
      
      * Restructure PromptConfig
      
      * Let's start T E M P L A T I N G
      
      * Refactor all default configs to use templates instead
      
      * Revert changes to the special token properties since we don't need them anymore
      
      * More class templates
      
      * Make the sandbox even sandier
      
      * Everything replaced with pure templating
      
      * Remove docs for PromptConfig
      
      * Add testing and optional requirement boilerplate
      
      * Fix imports and make fixup
      
      * Fix LLaMA tests and add Conversation docstring
      
      * Finally get LLaMA working with the template system
      
      * Finally get LLaMA working with the template system
      
      * make fixup
      
      * make fixup
      
      * fmt-off for the long lists of test tokens
      
      * Rename method to apply_chat_template for now
      
      * Start on documentation
      
      * Make chat_template a property that reads through to the default if it's not set
      
      * Expand docs
      
      * Expand chat templating doc some more
      
      * trim/lstrip blocks by default and update doc
      
      * Few doc tweaks
      
      * rebase cleanup
      
      * Clarify docstring
      
      * rebase cleanup
      
      * rebase cleanup
      
      * make fixup
      
      * Quick doc edit
      
      * Reformat the standard template to match ChatML
      
      * Re-add PEFT check
      
      * Update docs/source/en/chat_templating.md
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * Add apply_chat_template to the tokenizer doc
      
      * make fixup
      
      * Add doc links
      
      * Fix chat links
      
      * Fix chat links
      
      * Explain system messages in the doc
      
      * Add chat template test
      
      * Proper save-loading for chat template attribute
      
      * Add test skips for layout models
      
      * Remove _build_conversation_input_ids, add default_chat_template to code_llama
      
      * Make sure all LLaMA models are using the latest template
      
      * Remove default_system_prompt block in code_llama because it has no default prompt
      
      * Update ConversationPipeline preprocess
      
      * Add correct #Copied from links to the default_chat_templates
      
      * Remove unneeded type checking line
      
      * Add a dummy mark_processsed method
      
      * Reorganize Conversation to have **deprecated_kwargs
      
      * Update chat_templating.md
      
      * Quick fix to LLAMA tests
      
      * Small doc tweaks
      
      * Add proper docstrings and "copied from" statements to all default chat templates
      
      * Merge use_default_system_prompt support for code_llama too
      
      * Improve clarity around self.chat_template
      
      * Docstring fix
      
      * Fix blenderbot default template
      
      * More doctest fix
      
      * Break out some tokenizer kwargs
      
      * Update doc to explain default templates
      
      * Quick tweaks to tokenizer args
      
      * Cleanups for tokenizer args
      
      * Add note about cacheing
      
      * Quick tweak to the chat-templating doc
      
      * Update the LLaMA template with error checking and correct system message embedding
      
      * make fixup
      
      * make fixup
      
      * add requires_jinja
      
      * Cleanup to expected output formatting
      
      * Add cacheing
      
      * Fix typo in llama default template
      
      * Update LLaMA tests
      
      * Update documentation
      
      * Improved legacy handling in the Conversation class
      
      * Update Jinja template with proper error handling
      
      * Quick bugfix
      
      * Proper exception raising
      
      * Change cacheing behaviour so it doesn't try to pickle an entire Jinja env
      
      * make fixup
      
      * rebase cleanup
      
      ---------
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      866df66f
  32. 13 Sep, 2023 1 commit
  33. 05 Sep, 2023 1 commit
  34. 29 Aug, 2023 1 commit