-
Notifications
You must be signed in to change notification settings - Fork 2.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
failure of SigLIP2 FP32 to FP16 #4373
Comments
Are the FP32 results good? Can you try exporting a model without simplifying it (i.e. remove the |
Yes, the FP32 results are good. ![]() and the following is the onnx model's diference between opset_version 16(left) and 17(right). ![]() ![]() I think dustynv/l4t-pytorch:2.2-r35.4.1 and nvcr.io/nvidia/pytorch:25.01-py3 have the same problem in Layer Normalization op with opset_version 16. |
Use |
The log of nx with opset_version 16: The log of agx with opset_version 16: The log of RTX3090 with opset_version 16: |
The logs has no problem, you can export onnx weith ops=17, and use |
The log of agx with opset_version 17 and the cosine similarity is 0.6441. polygraphy run model/img_en_ori.onnx --trt --onnxrt --onnx-outputs mark all --trt-outputs mark all > comparison_results.txt |
I am trying to convert an SigLIP2 model to TensorRT and use fp16, but the cosine similarity between onnx and trt is 0.6463.
I used the following code convert to onnx.
and use the command to fp16 trt engine.
Environment
AGX with dustynv/l4t-pytorch:r36.4.0
NX with dustynv/l4t-pytorch:2.2-r35.4.1
ubuntu 22.04, RTX 3090 with nvcr.io/nvidia/pytorch:25.01-py3
The text was updated successfully, but these errors were encountered: