-
Notifications
You must be signed in to change notification settings - Fork 470
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Update: ONNX conversion doc #2686
Conversation
Actually, @b8zhong you know what would be helpful? A snippet of code that compares the encoded output of the hgf model vs. the ONNX model - and take the two vectors and compute l1 norm or something - to confirm that the conversion worked. would you might adding this? |
@lintool any idea why such a large L1 diff?
Maybe |
Ty for the hint.. I don't think I would have ever thought of that 😛 There we go: Acceptable I think? |
Yup, that's good. The page as written is specific to SPLADE++, can you make it generic, using SPLADE++ as an example? |
Thoughts? |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
minor nits, I'll fix and then I'll merge.
Update onnx + onnxruntime
Don't think
quantize_onnx_model.py
supports model name flag? Did not work w/ SPLADE++ Ensemble Distil earlier.https://github.com/castorini/anserini/blob/master/src/main/python/onnx/quantize_onnx_model.py#L19-L24