- 20 Oct, 2023 1 commit
-
-
Zakor Gyula authored
-
- 19 Oct, 2023 1 commit
-
-
Paul Fultz II authored
-
- 17 Oct, 2023 1 commit
-
-
Charlie Lin authored
-
- 11 Oct, 2023 3 commits
-
-
Ted Themistokleous authored
* Fix scatter operator for nonstandard shapes remove standard() shape check for scatter inputs. * Add nostandard input tests for scatter --------- Co-authored-by:Chris Austen <causten@users.noreply.github.com>
-
Artur Wojcik authored
-
Artur Wojcik authored
-
- 06 Oct, 2023 1 commit
-
-
Charlie Lin authored
-
- 04 Oct, 2023 1 commit
-
-
Artur Wojcik authored
-
- 28 Sep, 2023 1 commit
-
-
Charlie Lin authored
-
- 27 Sep, 2023 1 commit
-
-
Ted Themistokleous authored
Modify reshapes to use reshape_lazy for aliasing and then reshape for a reshape copy operation to eliminate contiguous
-
- 21 Sep, 2023 1 commit
-
-
Chris Austen authored
* Rectify flipped coordinate_transformation_mode logic in ROIAlign * Handle both opset 10 and 16 versions * Fix version check and clang tidy warning Co-authored-by:Dino Musić <dino.music@htecgroup.com>
-
- 16 Sep, 2023 1 commit
-
-
Charlie Lin authored
Implements a fill operator that sets the values in an output buffer to a given value Will be used when parsing ONNX ConstantOfShape Can also be used when a buffer needs to be filled with a value that is determined at runtime
-
- 14 Sep, 2023 1 commit
-
-
Brian Pickrell authored
New op that populates a shape with random numbers with a uniform distribution. The rand_uniform op. can implement the Onnx RandomUniform instruction, and can also create the random number sequence necessary to implement Multinomial. (At this time, our Onnx Multinomial parsing generates a random sequence of numbers when parsing as a workaround, so that the resulting program uses the same "random" set every time.) Arguments: shape, seed. Shape is required; can be static or dynamic. Seed is still optional in this version. If it's not given at inference time, use the value in the creation attribute seed. Update: deleted A boolean use_auto_seed causes any given seed to be ignored.
-
- 10 Sep, 2023 2 commits
-
-
Charlie Lin authored
Makes a version of allocate that takes in dimensions and allocates a buffer Going to create a simplify_dynamic_ops compiler pass that will use the use_shape_attr flag The ONNX op ConstantOfShape needs the buffer to be filled with a specific value, so going to make another PR for that and a fill operator
-
Lakhinder Walia authored
Reduce memory footprint by std::move of temporary (potentially very large) containers. Minor cleanup for performance optimization: e.g. of Index() calculation -- which can get repeated millions of times in large tensors/vectors in a single Visit.
-
- 29 Aug, 2023 1 commit
-
-
Brian Pickrell authored
Adds support for dynamic input shape in pooling operator along with auto-padding. This combination requires that the padding (and therefore the output shape) can't be computed until runtime.
-
- 18 Aug, 2023 1 commit
-
-
Charlie Lin authored
Allows slice to work with variable starts, ends, and axes input Outputs a dynamic shape even with a static shape data input when the starts and ends are variable
-
- 08 Aug, 2023 1 commit
-
-
Paul Fultz II authored
-
- 25 Jul, 2023 1 commit
-
-
Brian Pickrell authored
* Add dynamic input to prefix_scan_op * Added a shape test. The op should return the same dynamic shape as the input. * add 2d shape test for prefix_scan
-
- 23 Jul, 2023 1 commit
-
-
Charlie Lin authored
-
- 22 Jul, 2023 1 commit
-
-
Charlie Lin authored
Throwing on these calls catches dynamic shape errors earlier rather than having to backpedal from a bad call
-
- 21 Jul, 2023 1 commit
-
-
Umang Yadav authored
Fixes #1957 Clamping was removed in #1853. Turns out clamping as necessary to handle overflow/underflow cases. during downcasting, if it overflowed then without clamping it returned infinity.
-
- 13 Jul, 2023 1 commit
-
-
Charlie Lin authored
Renames deconvolution -> convolution_backwards to be more consistent with the literature Note: this is not the cross-correlation operator (which is the adjoint of convolution). This is technically a standard convolution operator combined with an upsampling operator rather than a downsampling operator. Adds unit tests for the padding, strides, dilations, and other op attributes. Throws on auto_pad attribute since it has not been implemented Previously it read the attribute and set it but then did nothing with it Extended for dynamic shapes Does not support using asymmetric padding (padding_L != padding_R) and output_shape with dynamic shapes.
-
- 10 Jul, 2023 1 commit
-
-
Brian Pickrell authored
Changes to the way Pooling operation calculates pooling when there's padding. Old code would clip off any padding values before computing; for instance if an Average pooling window contained 0 1 2 where the 0 is padding, the result was 1.5 instead of 1.0. See Issue 1766
-
- 08 Jul, 2023 1 commit
-
-
Artur Wojcik authored
Export API symbols for migraphx, migraphx_ref, migraphx_cpu, migrphx_gpu, migraphx_device, migraphx_tf, and migraphx_onnx. There is a separate PR for migrahx_c. API symbol exporting affects only Windows. It is transparent on Linux.
-
- 02 Jul, 2023 1 commit
-
-
Charlie Lin authored
Updates ref version of clip to work with dynamic shapes Encountered in agentmodel
-
- 23 Jun, 2023 1 commit
-
-
Umang Yadav authored
Fixes #1852 Fixes #1847
-
- 17 Jun, 2023 1 commit
-
-
Umang Yadav authored
* Fix convert for the NaNs * NaNs can't be compared, use std::isnan() * formatting * formatting * formatting * add extra tests
-
- 16 Jun, 2023 1 commit
-
-
Charlie Lin authored
* initial * Added tests and new functionality * Update optimals handling * Simplify conditionals * Ref test, update docs * Remove comment, suggestion unclear --------- Co-authored-by:Umang Yadav <29876643+umangyadav@users.noreply.github.com>
-
- 15 Jun, 2023 1 commit
-
-
Brian Pickrell authored
* fix parse_instancenorm to create broadcast and multibroadcast instructions with two dynamic shape arguments instead of 1. Their make_op() functions don't support dynamic shapes when called with one input. This caused an error when parsing an ONNX 3duunet model * Use add_common_op() to create multibroadcast op. * add verification and parsing test for instance_norm with dynamic input. Parse test doesn't pass. * fix for test; still doesn't pass * another fix for test; still doesn't pass * work in progress, instance_norm_dyn_batch_test works but instance_norm_test doesn't * fix onnx instancenorm tests to match parser changes. Passes all check tests * Updated comments explaining usage of add_common_op() * hand-merged conflicts with develop * fix instance_norm_half_test after merge * add Onnx test instance_norm_dyn_batch_half_test * add shape test cases broadcast_1in_dyn_error and multibroadcast_1in_dyn_error_0
-
- 12 Jun, 2023 1 commit
-
-
Paul Fultz II authored
-
- 31 May, 2023 1 commit
-
-
Umang Yadav authored
partially solves #1656 This PR only handles compilation part of multitarget.
-
- 17 May, 2023 1 commit
-
-
shivadbhavsar authored
Adding support for broadcasted scalars to unsqueeze op. Specifying steps other than 1 is disallowed in this implementation since we want the output the always be a tensor. We can support varying step sizes if we allow a broadcasted scalar output from this op.
-
- 04 May, 2023 1 commit
-
-
Zhuoran Yin authored
Exposed the mlir_enabled() call the decide for lowering pipeline's enablement Disabled the rewrite quantization pipeline in mlir compilation Added quant convolution as anchor ops Fixed the return type expectations Added the fall back hip implementation for quantizelinear and dequantizelinear Will need advises to improve the implementation for quantizelinear
-
- 24 Apr, 2023 1 commit
-
-
Charlie Lin authored
Updates the hip::copy_to_gpu and hip::copy_from_gpu operators to work with dynamic shapes Allows for offload_copy to be used with dynamic batch Changed assert in select_module because the argument might now be smaller with how offload_copy will work with dynamic batch. (maximum buffer size will be used)
-
- 18 Apr, 2023 1 commit
-
-
Ted Themistokleous authored
Ensure that we don't have empty inputs when computing shape for pointwise function
-
- 10 Apr, 2023 1 commit
-
-
Charlie Lin authored
Adds a matcher to split_single_dyn_dim to find all broadcast or multibroadcast with two static shape inputs and replaces the instruction with the one input version. Sorts the get_output_parameters() list to ensure the correct ordering. (Was getting an error for some models.)
-
- 07 Apr, 2023 1 commit
-
-
Paul Fultz II authored
Converts can be inserted when the scales and input differ in the onnx file(we are already doing this implicit conversion in the ref implementation). This will also improve the compile-time of quantizelinear.hpp since we can remove the nested visit method.
-
- 06 Apr, 2023 1 commit
-
-
Paul Fultz II authored
Automatically fuse multiple reductions and pointwise operations.
-
- 05 Apr, 2023 1 commit
-
-
Paul Fultz II authored
This will replace conv(x+a, w) with conv(x, w) + conv(a, w) where a is a constant so conv(a, w) can be replaced with a constant.
-