- 24 May, 2023 1 commit
-
-
umangyadav authored
-
- 23 May, 2023 1 commit
-
-
umangyadav authored
-
- 20 May, 2023 1 commit
-
-
Umang Yadav authored
* use half hip functions to compute max and min * add verify test for min and max
-
- 19 May, 2023 1 commit
-
-
Zhuoran Yin authored
Co-authored-by:Paul Fultz II <pfultz2@yahoo.com>
-
- 17 May, 2023 2 commits
-
-
Chris Austen authored
Move CI to support the rocm5.5 release
-
shivadbhavsar authored
Adding support for broadcasted scalars to unsqueeze op. Specifying steps other than 1 is disallowed in this implementation since we want the output the always be a tensor. We can support varying step sizes if we allow a broadcasted scalar output from this op.
-
- 08 May, 2023 1 commit
-
-
Umang Yadav authored
-
- 06 May, 2023 1 commit
-
-
Chris Austen authored
Remove various file not required for what we use Github runners for
-
- 05 May, 2023 3 commits
-
-
Charlie Lin authored
Python API with documentation updates
-
Manupa Karunaratne authored
Adds support for slice,transpose,contigous and reshape fusions into input tensors for a fused mlir kernel.
-
kahmed10 authored
add option to print tf supported ops sort both onnx and tf ops alphabetically
-
- 04 May, 2023 2 commits
-
-
Paul Fultz II authored
When multiplying either the input or output across the K dimensions then the multiple can be applied to the constant which can then be folded with propagate_const.
-
Zhuoran Yin authored
Exposed the mlir_enabled() call the decide for lowering pipeline's enablement Disabled the rewrite quantization pipeline in mlir compilation Added quant convolution as anchor ops Fixed the return type expectations Added the fall back hip implementation for quantizelinear and dequantizelinear Will need advises to improve the implementation for quantizelinear
-
- 03 May, 2023 1 commit
-
-
Charlie Lin authored
Relies on Removed split_single_dyn_dim compile flag #1711 Exposes dynamic_dimension as a opaque object with dynamic_dimensions and optimals Exposes ONNX dyn_input_dims and default_dyn_dim to run with dynamic batch Updates api.py to be able to create objects from aggregate initialization (used for dynamic_dimension) Uses offload copy for now
-
- 02 May, 2023 1 commit
-
-
Paul Fultz II authored
Improves the constant propagation for bert models. Larger batch size no longer use as large of constants. Also improves the speed of model compilation
-
- 28 Apr, 2023 1 commit
-
-
Charlie Lin authored
-
- 25 Apr, 2023 2 commits
-
-
kahmed10 authored
update rocBLAS version check to support 3.0 and above with simplified logic
-
Chris Austen authored
-
- 24 Apr, 2023 3 commits
-
-
Charlie Lin authored
Updates the hip::copy_to_gpu and hip::copy_from_gpu operators to work with dynamic shapes Allows for offload_copy to be used with dynamic batch Changed assert in select_module because the argument might now be smaller with how offload_copy will work with dynamic batch. (maximum buffer size will be used)
-
Paul Fultz II authored
This fixes #1700
-
Paul Fultz II authored
-
- 21 Apr, 2023 1 commit
-
-
Umang Yadav authored
-
- 20 Apr, 2023 1 commit
-
-
Umang Yadav authored
Solves #1311
-
- 19 Apr, 2023 1 commit
-
-
shivadbhavsar authored
Expose get_shape and get_operator methods for instruction_ref object in the python API.
-
- 18 Apr, 2023 2 commits
-
-
Paul Fultz II authored
This will show whats being replaced with a constant. This is useful for debugging where a literal comes from.
-
Ted Themistokleous authored
Ensure that we don't have empty inputs when computing shape for pointwise function
-
- 17 Apr, 2023 2 commits
-
-
Charlie Lin authored
Fixes the above behavior This needs to be changed to allow for setting static shapes with map_dyn_input_dims since you cannot also use map_input_dims
-
shivadbhavsar authored
Expose the shape::type_t values to be used by the python api and is required by torch_migraphx to support torchbench models.
-
- 13 Apr, 2023 1 commit
-
-
Zhuoran Yin authored
-
- 12 Apr, 2023 1 commit
-
-
Paul Fultz II authored
-
- 11 Apr, 2023 1 commit
-
-
Paul Fultz II authored
-
- 10 Apr, 2023 2 commits
-
-
Umang Yadav authored
-
Charlie Lin authored
Adds a matcher to split_single_dyn_dim to find all broadcast or multibroadcast with two static shape inputs and replaces the instruction with the one input version. Sorts the get_output_parameters() list to ensure the correct ordering. (Was getting an error for some models.)
-
- 09 Apr, 2023 1 commit
-
-
Paul Fultz II authored
* Enable hiprtc by default
-
- 07 Apr, 2023 1 commit
-
-
Paul Fultz II authored
Converts can be inserted when the scales and input differ in the onnx file(we are already doing this implicit conversion in the ref implementation). This will also improve the compile-time of quantizelinear.hpp since we can remove the nested visit method.
-
- 06 Apr, 2023 2 commits
-
-
Charlie Lin authored
Examples.. bin/driver verify /codes/onnx_models/resnet50-v1-7/resnet50-v1-7.onnx --split-single-dyn-dim --batch 3 --dyn-input-dim @data "[{min:1, max:4}, 3, 224, 224]" bin/driver compile /codes/onnx_models/resnet50-v1-7/resnet50-v1-7.onnx --split-single-dyn-dim --default-dyn-dim "{min:1, max:10}" --output resnet50_batch1-10.mxr bin/driver perf resnet50_batch1-10.mxr --batch 4 -
Paul Fultz II authored
Automatically fuse multiple reductions and pointwise operations.
-
- 05 Apr, 2023 3 commits
-
-
Paul Fultz II authored
* Add MIGRAPHX_VALIDATE_MATCHES env variable to validate each matcher
-
Paul Fultz II authored
This will replace conv(x+a, w) with conv(x, w) + conv(a, w) where a is a constant so conv(a, w) can be replaced with a constant.
-
Paul Fultz II authored
-