1. 06 Aug, 2023 1 commit
  2. 30 Jul, 2023 1 commit
    • Paul Fultz II's avatar
      Enable tuning for MLIR (#1965) · be6ecff6
      Paul Fultz II authored
      * Add initial tuning support
      
      * Format
      
      * Add extra param
      
      * Format
      
      * Use exauhstive flag
      
      * Format
      
      * Set expected shapes
      
      * Format
      
      * Format
      
      * Fix missing symbol
      
      * Format
      
      * Add missing license header
      
      * Format
      
      * Update src/targets/gpu/include/migraphx/gpu/mlir.hpp
      be6ecff6
  3. 29 Jul, 2023 1 commit
  4. 28 Jul, 2023 2 commits
  5. 26 Jul, 2023 1 commit
  6. 25 Jul, 2023 1 commit
  7. 22 Jul, 2023 1 commit
  8. 21 Jul, 2023 1 commit
    • Umang Yadav's avatar
      Make global workitems multiple of local workitems (#1976) · 3216fe52
      Umang Yadav authored
      HIP requires global work items in multiple of local work items. If it is not it is not guaranteed to generate correct results all the time.
      Fixes #1977
      Fixes #1644
      MIGraphX CI has moved to rocm-5.6 which doesn't require hipRTC workarounds
      3216fe52
  9. 19 Jul, 2023 1 commit
  10. 18 Jul, 2023 1 commit
  11. 17 Jul, 2023 2 commits
  12. 13 Jul, 2023 2 commits
    • Krzysztof Drewniak's avatar
      [NFC] Update MLIR usage to account for upstream merge (#1924) · c4765a6d
      Krzysztof Drewniak authored
      Allows the rocMLIR CI (which builds rocMLIR tip against MIGraphX tip) to pass.
      c4765a6d
    • Charlie Lin's avatar
      Update deconvolution -> convolution_backwards and Dynamic Shape Support (#1801) · 4edf1195
      Charlie Lin authored
      Renames deconvolution -> convolution_backwards to be more consistent with the literature
      Note: this is not the cross-correlation operator (which is the adjoint of convolution). This is technically a standard convolution operator combined with an upsampling operator rather than a downsampling operator.
      Adds unit tests for the padding, strides, dilations, and other op attributes.
      Throws on auto_pad attribute since it has not been implemented
      Previously it read the attribute and set it but then did nothing with it
      Extended for dynamic shapes
      Does not support using asymmetric padding (padding_L != padding_R) and output_shape with dynamic shapes.
      4edf1195
  13. 11 Jul, 2023 1 commit
  14. 08 Jul, 2023 2 commits
  15. 06 Jul, 2023 1 commit
  16. 05 Jul, 2023 1 commit
  17. 02 Jul, 2023 1 commit
    • Paul Fultz II's avatar
      Improvement to ck integration (#1859) · 3c9df3b4
      Paul Fultz II authored
      Add a CI job to test CK
      Add MIGRAPHX_TUNE_CK env variable to only do tuning for CK
      Continue tuning even when there is invalid configs
      Fix a bug with parallel compilation not using all available threads
      Add additional test for gemms using half types
      Removed int32 as supported type since it doesnt pass our test suite
      3c9df3b4
  18. 29 Jun, 2023 1 commit
  19. 28 Jun, 2023 2 commits
  20. 22 Jun, 2023 1 commit
  21. 21 Jun, 2023 1 commit
  22. 17 Jun, 2023 1 commit
    • turneram's avatar
      Update CK commit hash and add gfx940 to supported archs (#1842) · b8898d7e
      turneram authored
      * Add initial ck_gemm code
      
      * Format
      
      * Add additional src files
      
      * Format
      
      * Add include
      
      * Simplify fuse_ck
      
      * Format
      
      * Rename var
      
      * Enable pass
      
      * Update ck version
      
      * Fix include
      
      * Add group stride
      
      * Disable warnings for ck headers
      
      * Format
      
      * Add unpack array
      
      * Add interface to enable tuning
      
      * Format
      
      * Update compile_ops to handle tuning config
      
      * Format
      
      * Add some comments
      
      * Move time_op to migraphx_gpu
      
      * Add banchmarking
      
      * Refactor
      
      * Format
      
      * Add lift class macro
      
      * Use device name
      
      * Format
      
      * Generate configs
      
      * Format
      
      * Pass tuning parameter
      
      * Move data type to is_ck_gemm matcher
      
      * Format
      
      * Add problem_cache to avoid retuning same configs
      
      * Format
      
      * Format
      
      * Mark the problems
      
      * Format
      
      * Use is_null
      
      * Format
      
      * Resize vector
      
      * Only tune with exaustive tuning
      
      * Format
      
      * Use assert
      
      * FOrmat
      
      * Tidy fixes
      
      * More tidy fixes
      
      * Format
      
      * Add license to missing files
      
      * Format
      
      * Use transform
      
      * Format
      
      * Fix tidy
      
      * Format
      
      * Fix cppcheck issues
      
      * Format
      
      * Add static_assert
      
      * Add ops header
      
      * Add assertion in batcher
      
      * Format
      
      * Improve the batch fold check
      
      * Format
      
      * Add where op workaround for CK
      
      * Skip if any input is not a supported ck type
      
      * Format
      
      * Check batch is standard
      
      * Format
      
      * Remove redundant static keyword
      
      * Update commit hash
      
      * Fix error when running without --exhaustive-tune
      
      * Formatting
      
      * Formatting
      
      * Remove fuse_ck_gemm_softmax_gemm
      
      * Update ck hash
      
      * Correct spelling mistake
      
      * Remove commented out logic from fuse_ck
      
      * Remove unused include and add comment
      
      * Formatting
      
      * Remove redundant get_shape and remove ck_gemm from names
      
      * Formatting
      
      * Allow for mixed types with int8 gemms
      
      * Formatting
      
      * Add back find_package from merge
      
      * Update CK commit hash and add gfx940 to fuse_ops supported archs
      
      * Formatting
      
      * Update CK hash
      b8898d7e
  23. 15 Jun, 2023 1 commit
  24. 14 Jun, 2023 1 commit
  25. 09 Jun, 2023 2 commits
  26. 08 Jun, 2023 2 commits
  27. 06 Jun, 2023 2 commits
  28. 31 May, 2023 1 commit
  29. 24 May, 2023 2 commits
  30. 23 May, 2023 1 commit
  31. 20 May, 2023 1 commit