1. 20 Feb, 2023 1 commit
  2. 16 Feb, 2023 1 commit
  3. 15 Feb, 2023 1 commit
  4. 14 Feb, 2023 1 commit
    • charlie's avatar
      Somehow this verify test works · 996426be
      charlie authored
      * Changed the allocates to occur in the submodules
        * Incomplete, as the use_local_alloc variable in module does not work
        properly
      * added a hip::sync_stream before the return
      * not sure why the hip::sync_stream gets rid of the dangling reference
      error (code-wise it's because hip::sync_stream's output alias is -1)
      996426be
  5. 08 Feb, 2023 1 commit
  6. 17 Jan, 2023 1 commit
  7. 13 Jan, 2023 1 commit
  8. 11 Jan, 2023 1 commit
  9. 09 Jan, 2023 1 commit
  10. 02 Nov, 2022 1 commit
  11. 28 Oct, 2022 1 commit
  12. 27 Oct, 2022 1 commit
    • kahmed10's avatar
      Add JIT pad (#1411) · 0d841ded
      kahmed10 authored
      updated GPU pad to now use JIT version.
      added range functions for JIT kernels.
      0d841ded
  13. 26 Oct, 2022 1 commit
  14. 19 Oct, 2022 2 commits
  15. 13 Oct, 2022 2 commits
  16. 10 Oct, 2022 1 commit
  17. 04 Oct, 2022 1 commit
  18. 29 Sep, 2022 2 commits
  19. 27 Sep, 2022 1 commit
  20. 21 Sep, 2022 1 commit
  21. 19 Sep, 2022 1 commit
    • Paul Fultz II's avatar
      Improve layernorm and reductions performance (#1348) · 97a1ed2d
      Paul Fultz II authored
      Compute mean and variance in same reduction
      Set block size to numbers divisible by 32 instead powers of 2
      Global is also set exactly instead of being divisible by block size
      More exact matching of global/local can help get rid of branching/loops
      Reduce vectors first before doing dpp_reduce
      Explicitly vectorize array operators since the compiler doesnt always vectorize them
      Still uses old for loop when its computing at compile-time since the reinterpret_cast nor the all the vector types is supported
      97a1ed2d
  22. 16 Sep, 2022 1 commit
  23. 14 Sep, 2022 2 commits
  24. 13 Sep, 2022 1 commit
    • turneram's avatar
      Use rocblas_gemm_ex for batched gemms with broadcasted B (#1354) · a10a8ef1
      turneram authored
      Improves performance for 4/6 GEMMs used by huggingface BERT models with batch_size>1 by using a non-batched rocBLAS call for GEMMs where the B input has a broadcasted batch dimension.
      The four verify tests added reflect the actual configurations used by bert-base-cased, with varied batch sizes.
      
      Also adds a matcher to simplify_reshapes to move multibroadcasts after concats.
      a10a8ef1
  25. 07 Sep, 2022 1 commit
  26. 06 Sep, 2022 1 commit
  27. 31 Aug, 2022 1 commit
  28. 27 Aug, 2022 1 commit
  29. 17 Aug, 2022 1 commit
  30. 16 Aug, 2022 1 commit
  31. 25 Jul, 2022 1 commit
    • varunsh's avatar
      Add fpga target (#1304) · 8a30d698
      varunsh authored
      * Add is_supported to the target
      * Add get_target_assignments
      * Rename assignment to target_assignments
      * Add ref target header to test
      * Add fpga target
      * Make context const in compute
      8a30d698
  32. 06 Jul, 2022 1 commit
    • Paul Fultz II's avatar
      Verify load and save (#1265) · f2531606
      Paul Fultz II authored
      *In the verification tests, check that saving and reloading the program is the same program. This also fixes serialization to always load instructions in the same order. There is also fixes for deconv and quant_conv which didn't save the solution id, and was broken for serialization.
      f2531606
  33. 22 Jun, 2022 1 commit
  34. 07 Jun, 2022 1 commit
  35. 02 Jun, 2022 1 commit
  36. 26 May, 2022 1 commit