1. 28 Jun, 2021 3 commits
    • Bhadresh Savani's avatar
      [Examples] Added context manager to datasets map (#12367) · 04dbea31
      Bhadresh Savani authored
      * added cotext manager to datasets map
      
      * fixed style and spaces
      
      * fixed warning of deprecation
      
      * changed desc
      04dbea31
    • Sylvain Gugger's avatar
    • Taha ValizadehAslani's avatar
      Update run_mlm.py (#12344) · 9490d668
      Taha ValizadehAslani authored
      Before the code could not be used for validation only because of this line:
      extension = data_args.train_file.split(".")[-1]
      was assuming that extension must be extracted from the training dataset. This line would run regardless of the training or validation options of the user. This would lead to an error if the user only wants to run an evaluation only and does not want to do train (because the training file does not exist). I modified it to extract extension from the training file if the user wants to do train and extract it from the validation file if the user wants to run eval. This way the code can be used for both training and validation separately.
      9490d668
  2. 26 Jun, 2021 1 commit
  3. 25 Jun, 2021 6 commits
  4. 24 Jun, 2021 2 commits
  5. 23 Jun, 2021 4 commits
  6. 22 Jun, 2021 3 commits
  7. 21 Jun, 2021 2 commits
  8. 17 Jun, 2021 3 commits
  9. 15 Jun, 2021 3 commits
  10. 14 Jun, 2021 7 commits
  11. 11 Jun, 2021 1 commit
    • Suraj Patil's avatar
      Flax CLM script (#12023) · 15b498f3
      Suraj Patil authored
      * first draft
      
      * max_seq_length => block_size
      
      * fix arg names
      
      * fix typos
      
      * fix loss calculation
      
      * add max examples, fix  train eval steps, metrics
      
      * optimizer mask
      
      * fix perpelexity, metric logging
      
      * fix logging
      
      * data_collator = > data_loader
      
      * refactor loss_fn
      
      * support single GPU
      
      * pass distributed to write_metric
      
      * fix jitting
      
      * fix single device training
      
      * fix single device metrics
      
      * close inner progress bars once finished
      
      * add overwrite_cache arg
      
      * ifx dataset caching issue
      
      * add more logs
      
      * few small fixes,
      
      * address nicholas suggestions
      
      * fix docstr
      
      * address patricks suggestions
      
      * make flake happy
      
      * pass new new_dropout_rng to apply_gradients
      
      * reset train metrics after every epoc
      
      * remove distributed logis, small fixes
      15b498f3
  12. 10 Jun, 2021 5 commits