1. 20 Aug, 2023 2 commits
  2. 18 Aug, 2023 2 commits
  3. 16 Aug, 2023 1 commit
  4. 13 Aug, 2023 1 commit
  5. 12 Aug, 2023 3 commits
  6. 11 Aug, 2023 2 commits
  7. 10 Aug, 2023 1 commit
  8. 09 Aug, 2023 2 commits
  9. 08 Aug, 2023 2 commits
  10. 07 Aug, 2023 1 commit
  11. 06 Aug, 2023 2 commits
  12. 04 Aug, 2023 1 commit
  13. 03 Aug, 2023 2 commits
  14. 31 Jul, 2023 2 commits
    • Lakhinder Walia's avatar
      Lw/fix half shape (#2000) · e4dc75ea
      Lakhinder Walia authored
      * Use shape of Instruction (instead of a default) in add_return()
      
      * Instruction validation fix: not to use a default shape value for comparison
      
      * Fix instruction::replace() to recompute shape for "@return"
      
      * handle the case of missing shape in an Instruction related Test
      
      * use compute_shape() to get op shapes + test case for tuple_type
      
      * add test case shape_test/return_shape_tuple
      
      * Add test for @return to check for half type
      
      * Move @return unit-tests around..; Address review comments
      
      * Broken comparison fix: comparison to a (default) shape of tuple_type
      
      * Test cases: (add) return_shape_empty & (modify) return_shape_tuple
      
      * modify the assert() statement
      e4dc75ea
    • Artur Wojcik's avatar
  15. 30 Jul, 2023 1 commit
    • Paul Fultz II's avatar
      Enable tuning for MLIR (#1965) · be6ecff6
      Paul Fultz II authored
      * Add initial tuning support
      
      * Format
      
      * Add extra param
      
      * Format
      
      * Use exauhstive flag
      
      * Format
      
      * Set expected shapes
      
      * Format
      
      * Format
      
      * Fix missing symbol
      
      * Format
      
      * Add missing license header
      
      * Format
      
      * Update src/targets/gpu/include/migraphx/gpu/mlir.hpp
      be6ecff6
  16. 29 Jul, 2023 1 commit
  17. 28 Jul, 2023 3 commits
  18. 27 Jul, 2023 1 commit
  19. 26 Jul, 2023 1 commit
  20. 25 Jul, 2023 3 commits
  21. 23 Jul, 2023 1 commit
  22. 22 Jul, 2023 3 commits
  23. 21 Jul, 2023 2 commits
    • Umang Yadav's avatar
      Add back clamping and add tests (#1969) · 6957243c
      Umang Yadav authored
      Fixes #1957
      
      Clamping was removed in #1853.
      
      Turns out clamping as necessary to handle overflow/underflow cases. during downcasting, if it overflowed then without clamping it returned infinity.
      6957243c
    • Umang Yadav's avatar
      Use `optimize_module` pass for the quantization to fp16 (#1974) · 6f1f4b59
      Umang Yadav authored
      Fixes #1746
      
      BatchNorm only has x as the runtime input parameter for the following equation. All the other parameters are compile-time constants and related operations can be const-folded before quantizing to fp16 to preserve precision.
      6f1f4b59