1. 08 Dec, 2022 2 commits
    • Charlie Lin's avatar
      Dynamic ref flatten (#1482) · 4c32afcc
      Charlie Lin authored
      Changes flatten's compute_shape() to handle dynamic shapes
      Calculates the flattened shape with the min, max, and opt
      4c32afcc
    • shivadbhavsar's avatar
      fix issues with compiling lstm ops in fp16 mode (#1450) · 352c2465
      shivadbhavsar authored
      Currently, quantizing a program with rnn layers to fp16 results in segmentation faults due to a "convert" operation being applied to an "undefined" instruction.
      
      The following changes are implemented to fix this issue:
      
      Added is_undefined method to the instruction class that returns true if all inputs to the instruction are from an undefined op.
      Updated rewrite_rnn pass to use the new is_undefined method rather than checking ins->name()
      Updated the dead_code_elimination pass to also use this new method rather than only checking the instruction name
      352c2465
  2. 07 Dec, 2022 2 commits
  3. 06 Dec, 2022 3 commits
  4. 02 Dec, 2022 2 commits
    • Charlie Lin's avatar
      Refactor non-standard literal construction (#1443) · fdc3f00a
      Charlie Lin authored
      Fix problem with the contiguous operator constructing non-standard shape literals.  A non-standard literal will almost never be used, since a literal is known at compile time.  Added some comments on the intended behavior:
      
      - literal{shape, vector} constructor with a non-standard shape is intended to keep the same ordering as the given vector. The data buffer will be populated such that when the non-standard indexing is used the original order is as given.
      - literal{shape, argument} constructor directly copies the data buffer from the argument
      - Changed non-standard literal fill() to use tensor_view iterators as it handles non-standard shapes now
      - Changed the contiguous ref_ops_test to be more helpful
      fdc3f00a
    • Charlie Lin's avatar
      Dynamic ref pooling (#1449) · 0e40ebaa
      Charlie Lin authored
      Extends the pooling operators for dynamic shape inputs
      
      AveragePooling
      GlobalAveragePooling
      MaxPooling
      GlobalMaxPooling
      LpNormPooling
      GlobalLpNormPooling
      y.github.com>
      0e40ebaa
  5. 29 Nov, 2022 1 commit
  6. 28 Nov, 2022 1 commit
  7. 20 Nov, 2022 1 commit
  8. 18 Nov, 2022 1 commit
    • Umang Yadav's avatar
      Disable Find2.0 for now (#1462) · 493bb8d5
      Umang Yadav authored
      Disabling it untill int8 fix is in mainline from MIOpen and also so that QA tests could run migraphx-driver and unittests from MIGraphX.
      493bb8d5
  9. 17 Nov, 2022 2 commits
  10. 14 Nov, 2022 1 commit
  11. 13 Nov, 2022 1 commit
    • Charlie Lin's avatar
      Dyn ref multibroadcast; dyn binary (#1423) · d73c6d7c
      Charlie Lin authored
      Updated Multibroadcast op to have a two input version for dynamic shapes
      Current dynamic shape broadcasting logic
      dynamic_dimensions must be the same or one of them is {1, 1, 0} or {1, 1, 1}
      Works for dyn-dyn, dyn-static, and static-static shape combinations
      Changed common.cpp for multibroadcasting for binary ops with dynamic shapes
      Extended binary.hpp for dynamic shapes to test the new common.cpp stuff
      d73c6d7c
  12. 07 Nov, 2022 1 commit
  13. 06 Nov, 2022 1 commit
  14. 02 Nov, 2022 2 commits
  15. 01 Nov, 2022 2 commits
  16. 28 Oct, 2022 1 commit
  17. 27 Oct, 2022 2 commits
  18. 26 Oct, 2022 2 commits
  19. 24 Oct, 2022 1 commit
  20. 19 Oct, 2022 2 commits
  21. 18 Oct, 2022 1 commit
  22. 14 Oct, 2022 1 commit
  23. 13 Oct, 2022 2 commits
  24. 07 Oct, 2022 1 commit
  25. 04 Oct, 2022 2 commits
  26. 03 Oct, 2022 1 commit
    • Umang Yadav's avatar
      Add output_alias and runs_on_offload_target flags for the custom ops (#1309) · c9ffb38d
      Umang Yadav authored
      Adds two methods for the custom_ops virtual class.
      
      bool runs_on_offload_target(), if the custom op runs directly on the gpu then it should be set to true. in this case, custom op expects its parameters to reside in GPU memory and writes output to the GPU memory. If it is set to false then, custom op expects it's parameter to reside on the host and puts back the result into the host memory.
      
      output_alias, if output of the custom op is aliasing the input buffer. i.e. interpreting the same input buffer with differnet shape and strides.
      
      Update as_vector() in C++ API to handle non-standard shapes. It required exposing element_index to space_index conversion method for the shape class.
      c9ffb38d
  27. 29 Sep, 2022 1 commit