-
Notifications
You must be signed in to change notification settings - Fork 7
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ONNX converting issues #10
Comments
They're just some normal warnings, it's OK to ignore them. |
Thank you for your answer. So it doesn't affect inference quality? |
Yes, none of these warnings have effects on quality. The code reported by TracerWarnings is used to check whether variables meet the requirements, rather than the inference part. Therefore, it can be ignored. Constant folding is a method used in ONNX to optimize the Slice operation. However, this optimization is not applicable at the location where the warning occurred, and not performing constant folding meets our expectations. As for the last warning, I think it is some kind of bug in PyTorch that causes ONNX to be unable to recognize the type. However, it won't have any effect if your model infers correctly. |
Hello! Model converted into onnx with your scripts has very poor performance in NVIDIA Triton Inference Server Have you faced such problem? Thanks. |
Very sorry! When writing the Onnx runtime, I specified the CPU inference, please wait Line 157 in c114751
|
Thanks, I'll give it a try I use the converted model in NVIDIA Triton Inference server I'll test pure pytorch inference, pure onnx inference, and triton inference with pytorch model and onnx model |
ok |
pls wait a while for svc branch |
Have done 50 test inferences for each model with same input text for some reason onnxruntime in triton makes execution slower, trying to find bottleneck |
The server will convert pth to onnx before loading. Instead of using onnx. |
The ONNX model will perform some initialization operations during the first reasoning after the Session is loaded, and this factor should also be considered |
Hello, I faced these errors while converting to onnx
TracerWarning: Converting a tensor to a Python boolean might cause the trace to be incorrect. We can't record the data flow of Python values, so this value will be treated as a constant in the future. This means that the trace might not generalize to other inputs!
assert (discriminant >= 0).all()
Warning: Constant folding - Only steps=1 can be constant folded for opset >= 10 onnx::Slice op. Constant folding not applied.
WARNING: The shape inference of prim::Constant type is missing, so it may result in wrong shape inference for the exported graph. Please consider adding it in symbolic function.
What may be wrong? Thanks
The text was updated successfully, but these errors were encountered: