1. 04 Aug, 2022 2 commits
    • nlpcat's avatar
      change shape to support dynamic batch input in tf.function XLA generate for tf serving (#18372) · fc1d841b
      nlpcat authored
      
      
      * change shape to support dynamic batch input in tf.generate
      
      * add tests
      Co-authored-by: default avatarnlpcatcode <nlpcodecat@gmail.com>
      fc1d841b
    • Thomas Wang's avatar
      [BLOOM] Clean modeling code (#18344) · b69a62d5
      Thomas Wang authored
      
      
      * Cleanup some code
      
      * Improve signatures
      
      * Try to reduce the number of reshape/copies
      
      * I don't think we actually need the layer_num scaling trick
      
      * No need for duplication
      
      * Try to fix beam_search
      
      * Fix beam search
      
      * Removing layer num normalization seems to be breaking
      
      * Not sure self.layer_number normalization actually matters
      
      * Try and be backward compatible
      
      * Try to fix beam_search
      
      * Revert attempt to be backward compatible
      
      * Improve documentation on past_key_values format
      
      * Optimize the device allocation in case of hidden_states in multiple devices
      
      * No need to manually cast the values to a specific device
      
      * Rename with long version of variables
      
      * Improve type hinting
      
      * Add comment that explains that some methods return views
      
      * Actually i think the attention casting only makes sense when we use torch.float16
      
      * We don't actually need layer_number to be passed anymore
      
      * Fix FX test
      
      * Bypass torch.baddbmm
      
      * Apply suggestions from code review
      
      * Add comment about support for torchScript v1.11
      
      * fix ONNX support for bloom (#18456)
      Co-authored-by: default avatarNiklas Muennighoff <n.muennighoff@gmail.com>
      Co-authored-by: default avatarNouamane Tazi <nouamane98@gmail.com>
      b69a62d5
  2. 03 Aug, 2022 10 commits
  3. 02 Aug, 2022 10 commits
  4. 01 Aug, 2022 18 commits