1. 23 Jun, 2021 1 commit
    • Suraj Patil's avatar
      Flax summarization script (#12230) · c0fe3c9a
      Suraj Patil authored
      * add summrization script
      
      * fix arguments, preprocessing, metrics
      
      * add generation and metrics
      
      * auto model, prediction loop
      
      * prettify
      
      * label smoothing
      
      * adress Sylvain and Patricks suggestions
      
      * dynamically import shift_tokens_right
      
      * fix shift_tokens_right_fn call
      c0fe3c9a
  2. 15 Jun, 2021 1 commit
  3. 14 Jun, 2021 3 commits
  4. 11 Jun, 2021 1 commit
    • Suraj Patil's avatar
      Flax CLM script (#12023) · 15b498f3
      Suraj Patil authored
      * first draft
      
      * max_seq_length => block_size
      
      * fix arg names
      
      * fix typos
      
      * fix loss calculation
      
      * add max examples, fix  train eval steps, metrics
      
      * optimizer mask
      
      * fix perpelexity, metric logging
      
      * fix logging
      
      * data_collator = > data_loader
      
      * refactor loss_fn
      
      * support single GPU
      
      * pass distributed to write_metric
      
      * fix jitting
      
      * fix single device training
      
      * fix single device metrics
      
      * close inner progress bars once finished
      
      * add overwrite_cache arg
      
      * ifx dataset caching issue
      
      * add more logs
      
      * few small fixes,
      
      * address nicholas suggestions
      
      * fix docstr
      
      * address patricks suggestions
      
      * make flake happy
      
      * pass new new_dropout_rng to apply_gradients
      
      * reset train metrics after every epoc
      
      * remove distributed logis, small fixes
      15b498f3
  5. 09 Jun, 2021 1 commit
  6. 03 Jun, 2021 2 commits
  7. 31 May, 2021 1 commit
    • Nicholas Vadivelu's avatar
      Remove redundant `nn.log_softmax` in `run_flax_glue.py` (#11920) · 1ab147d6
      Nicholas Vadivelu authored
      * Remove redundant `nn.log_softmax` in `run_flax_glue.py`
      
      `optax.softmax_cross_entropy` expects unnormalized logits, and so it already calls `nn.log_softmax`, so I believe it is not needed here. `nn.log_softmax` is idempotent so mathematically it shouldn't have made a difference.
      
      * Remove unused 'flax.linen' import
      1ab147d6
  8. 26 May, 2021 1 commit
  9. 24 May, 2021 1 commit
  10. 21 May, 2021 3 commits
  11. 19 May, 2021 1 commit
  12. 18 May, 2021 1 commit
  13. 17 May, 2021 1 commit
  14. 14 May, 2021 2 commits
  15. 12 May, 2021 1 commit
  16. 11 May, 2021 1 commit
  17. 04 May, 2021 1 commit
  18. 23 Apr, 2021 1 commit
  19. 21 Apr, 2021 1 commit