1. 16 Feb, 2024 1 commit
  2. 02 Feb, 2024 1 commit
    • Klaus Hipp's avatar
      [Docs] Fix spelling and grammar mistakes (#28825) · 721ee783
      Klaus Hipp authored
      * Fix typos and grammar mistakes in docs and examples
      
      * Fix typos in docstrings and comments
      
      * Fix spelling of `tokenizer` in model tests
      
      * Remove erroneous spaces in decorators
      
      * Remove extra spaces in Markdown link texts
      721ee783
  3. 30 Jan, 2024 1 commit
    • Matt's avatar
      Add tf_keras imports to prepare for Keras 3 (#28588) · 415e9a09
      Matt authored
      * Port core files + ESM (because ESM code is odd)
      
      * Search-replace in modelling code
      
      * Fix up transfo_xl as well
      
      * Fix other core files + tests (still need to add correct import to tests)
      
      * Fix cookiecutter
      
      * make fixup, fix imports in some more core files
      
      * Auto-add imports to tests
      
      * Cleanup, add imports to sagemaker tests
      
      * Use correct exception for importing tf_keras
      
      * Fixes in modeling_tf_utils
      
      * make fixup
      
      * Correct version parsing code
      
      * Ensure the pipeline tests correctly revert to float32 after each test
      
      * Ensure the pipeline tests correctly revert to float32 after each test
      
      * More tf.keras -> keras
      
      * Add dtype cast
      
      * Better imports of tf_keras
      
      * Add a cast for tf.assign, just in case
      
      * Fix callback imports
      415e9a09
  4. 02 Jun, 2023 1 commit
  5. 20 Apr, 2023 1 commit
  6. 14 Apr, 2023 1 commit
    • Sayak Paul's avatar
      [Examples] TPU-based training of a language model using TensorFlow (#21657) · 390e121f
      Sayak Paul authored
      
      
      * add: tokenizer training script for TF TPU LM training.
      
      * add: script for preparing the TFRecord shards.
      
      * add: sequence of execution to readme.
      
      * remove limit from the tfrecord shard name.
      
      * Add initial train_model.py
      
      * Add basic training arguments and model init
      
      * Get up to the point of writing the data collator
      
      * Pushing progress so far!
      
      * Complete first draft of model training code
      
      * feat: grouping of texts efficiently.
      Co-authored-by: default avatarMatt <rocketknight1@gmail.com>
      
      * Add proper masking collator and get training loop working
      
      * fix: things.
      
      * Read sample counts from filenames
      
      * Read sample counts from filenames
      
      * Draft README
      
      * Improve TPU warning
      
      * Use distribute instead of distribute.experimental
      
      * Apply suggestions from code review
      Co-authored-by: default avatarMatt <Rocketknight1@users.noreply.github.com>
      
      * Modularize loading and add MLM probability as arg
      
      * minor refactoring to better use the cli args.
      
      * readme fillup.
      
      * include tpu and inference sections in the readme.
      
      * table of contents.
      
      * parallelize maps.
      
      * polish readme.
      
      * change script name to run_mlm.py
      
      * address PR feedback (round I).
      
      ---------
      Co-authored-by: default avatarMatt <rocketknight1@gmail.com>
      Co-authored-by: default avatarMatt <Rocketknight1@users.noreply.github.com>
      390e121f