1. 11 Oct, 2022 2 commits
    • charlie's avatar
      Fix things · a88810da
      charlie authored
      convolution revert
      a88810da
    • charlie's avatar
      Redo design · d9d2215a
      charlie authored
      * doesn't make much sense to make broadcast use two inputs or handle
      dynamic shapes
      * compute the common shape for dynamic multibroadcast in the
      multibroadcast op
      * multibroadcast all combinations of the dynamic inputs
      d9d2215a
  2. 10 Oct, 2022 1 commit
  3. 07 Oct, 2022 1 commit
  4. 04 Oct, 2022 2 commits
  5. 03 Oct, 2022 2 commits
    • charlie's avatar
      More progress · b162c4ec
      charlie authored
      b162c4ec
    • Umang Yadav's avatar
      Add output_alias and runs_on_offload_target flags for the custom ops (#1309) · c9ffb38d
      Umang Yadav authored
      Adds two methods for the custom_ops virtual class.
      
      bool runs_on_offload_target(), if the custom op runs directly on the gpu then it should be set to true. in this case, custom op expects its parameters to reside in GPU memory and writes output to the GPU memory. If it is set to false then, custom op expects it's parameter to reside on the host and puts back the result into the host memory.
      
      output_alias, if output of the custom op is aliasing the input buffer. i.e. interpreting the same input buffer with differnet shape and strides.
      
      Update as_vector() in C++ API to handle non-standard shapes. It required exposing element_index to space_index conversion method for the shape class.
      c9ffb38d
  6. 02 Oct, 2022 1 commit
  7. 30 Sep, 2022 3 commits
  8. 29 Sep, 2022 6 commits
  9. 28 Sep, 2022 3 commits
  10. 27 Sep, 2022 4 commits
  11. 26 Sep, 2022 4 commits
  12. 23 Sep, 2022 2 commits
  13. 22 Sep, 2022 2 commits
  14. 21 Sep, 2022 2 commits
  15. 19 Sep, 2022 1 commit
    • Paul Fultz II's avatar
      Improve layernorm and reductions performance (#1348) · 97a1ed2d
      Paul Fultz II authored
      Compute mean and variance in same reduction
      Set block size to numbers divisible by 32 instead powers of 2
      Global is also set exactly instead of being divisible by block size
      More exact matching of global/local can help get rid of branching/loops
      Reduce vectors first before doing dpp_reduce
      Explicitly vectorize array operators since the compiler doesnt always vectorize them
      Still uses old for loop when its computing at compile-time since the reinterpret_cast nor the all the vector types is supported
      97a1ed2d
  16. 16 Sep, 2022 4 commits