1. 24 Jan, 2023 1 commit
  2. 21 Jan, 2023 7 commits
  3. 17 Jan, 2023 4 commits
  4. 16 Jan, 2023 1 commit
  5. 13 Jan, 2023 2 commits
  6. 11 Jan, 2023 3 commits
  7. 09 Jan, 2023 2 commits
  8. 06 Jan, 2023 2 commits
  9. 04 Jan, 2023 1 commit
  10. 29 Dec, 2022 1 commit
  11. 14 Dec, 2022 2 commits
  12. 13 Dec, 2022 2 commits
  13. 11 Dec, 2022 1 commit
    • Umang Yadav's avatar
      change target flag (#1488) · b41c1f01
      Umang Yadav authored
      HIP had change in previous rocm releases to use --offload-arch instead of --cuda-gpu-arch.
      
      This should be backwards compatbile. hipRTC also supports --offload-arch.
      b41c1f01
  14. 08 Dec, 2022 4 commits
    • Charlie Lin's avatar
      Dynamic ref dot operator (#1457) · d411aa69
      Charlie Lin authored
      Extends dot MIGX operator to handle dynamic input shapes
      Only allow dot between two dynamic shapes that have exactly matching outer dimensions
      Inner dimensions must also match correspondingly
      Updates dot related tests
      Change check_shapes to use shape.ndim()
      ONNX parsers for GEMM and MatMult will be updated in a separate PR
      d411aa69
    • Charlie Lin's avatar
      Dynamic reference Softmax (#1475) · 8e7d2efe
      Charlie Lin authored
      No major changes required, use dyn_output and pass dynamic shape when calling compute_shape()
      Adds dynamic shape tests
      8e7d2efe
    • Charlie Lin's avatar
      Dynamic ref flatten (#1482) · 4c32afcc
      Charlie Lin authored
      Changes flatten's compute_shape() to handle dynamic shapes
      Calculates the flattened shape with the min, max, and opt
      4c32afcc
    • shivadbhavsar's avatar
      fix issues with compiling lstm ops in fp16 mode (#1450) · 352c2465
      shivadbhavsar authored
      Currently, quantizing a program with rnn layers to fp16 results in segmentation faults due to a "convert" operation being applied to an "undefined" instruction.
      
      The following changes are implemented to fix this issue:
      
      Added is_undefined method to the instruction class that returns true if all inputs to the instruction are from an undefined op.
      Updated rewrite_rnn pass to use the new is_undefined method rather than checking ins->name()
      Updated the dead_code_elimination pass to also use this new method rather than only checking the instruction name
      352c2465
  15. 07 Dec, 2022 2 commits
  16. 06 Dec, 2022 3 commits
  17. 02 Dec, 2022 2 commits
    • Charlie Lin's avatar
      Refactor non-standard literal construction (#1443) · fdc3f00a
      Charlie Lin authored
      Fix problem with the contiguous operator constructing non-standard shape literals.  A non-standard literal will almost never be used, since a literal is known at compile time.  Added some comments on the intended behavior:
      
      - literal{shape, vector} constructor with a non-standard shape is intended to keep the same ordering as the given vector. The data buffer will be populated such that when the non-standard indexing is used the original order is as given.
      - literal{shape, argument} constructor directly copies the data buffer from the argument
      - Changed non-standard literal fill() to use tensor_view iterators as it handles non-standard shapes now
      - Changed the contiguous ref_ops_test to be more helpful
      fdc3f00a
    • Charlie Lin's avatar
      Dynamic ref pooling (#1449) · 0e40ebaa
      Charlie Lin authored
      Extends the pooling operators for dynamic shape inputs
      
      AveragePooling
      GlobalAveragePooling
      MaxPooling
      GlobalMaxPooling
      LpNormPooling
      GlobalLpNormPooling
      y.github.com>
      0e40ebaa