"src/turbomind/models/llama/llama_utils.h" did not exist on "cc93136e6a166566fc6f0502c67aa99a94673db3"
  1. 29 Aug, 2023 1 commit
  2. 18 Aug, 2023 1 commit
  3. 06 Oct, 2022 1 commit
    • Antoine Adam's avatar
      remove numpy dependency · 4e38df05
      Antoine Adam authored
      According to the `setup.py` file, only dependencies are torch and einops. But the `bert_padding.py` file requires `numpy` only to multiply the elements of a `torch.Size` object. This change aims at allowing the use of FlashAttention without numpy.
      4e38df05
  4. 05 Aug, 2022 1 commit
  5. 02 Jun, 2022 1 commit
  6. 29 May, 2022 1 commit
  7. 20 May, 2022 1 commit