Commit f265f8d
authored
[5693592][ONNX-customOp][Autocast] Fix QuantizeLinear node output type (#671)
## What does this PR do?
**Type of change:** Bug fix
**Overview:** The output type of Q nodes was being set incorrectly. This
PR fixes that.
## Usage
```python
$ python -m modelopt.onnx.autocast --onnx_path=$MODEL_NAME.onnx
```
## Testing
See bug 5693592 for more details:
```python
$ python -m modelopt.onnx.autocast --onnx_path=$MODEL_NAME.onnx --low_precision_type=fp16 --data_max=inf --init_max=inf --keep_io_types
```
## Before your PR is "*Ready for review*"
<!-- If you haven't finished some of the above items you can still open
`Draft` PR. -->
- **Make sure you read and follow [Contributor
guidelines](https://github.com/NVIDIA/Model-Optimizer/blob/main/CONTRIBUTING.md)**
and your commits are signed.
- **Is this change backward compatible?**: Yes
- **Did you write any new necessary tests?**: No
- **Did you add or update any necessary documentation?**: No
- **Did you update
[Changelog](https://github.com/NVIDIA/Model-Optimizer/blob/main/CHANGELOG.rst)?**:
No
Signed-off-by: gcunhase <[email protected]>1 parent ee8a1f4 commit f265f8d
1 file changed
+2
-0
lines changed| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
296 | 296 | | |
297 | 297 | | |
298 | 298 | | |
| 299 | + | |
| 300 | + | |
299 | 301 | | |
300 | 302 | | |
301 | 303 | | |
| |||
0 commit comments