-
Notifications
You must be signed in to change notification settings - Fork 216
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Failed analyse for node #678 "/Expand" MultiBroadcastTo #1200
Comments
Hey, thanks for your interest in tract. Performance without the call to into_optimized() is often ridiculously bad. It is meant to be used for debugging only. Your problem look like a typical case of ONNX problematic symbol usage in dimensions. Its semantics and tract's are different. It's usually pretty easy to fix. You can try tract main branch: I have relaxed how tract interprets symbols coming from ONNX (ignoring the output shapes altogether by default). If it does not fix, I'll need a link to the ONNX model to have a look. |
Hey, Thanks for the fast response! Here is the link to the model: Is there a link to any troubleshooting guide or any tools you use, so I can troubleshoot models myself in the future? Thanks :) |
A good starting point is the command line interface, there are some recipes in https://github.com/sonos/tract/blob/main/doc/cli-recipe.md . It does help, but the documentation is a bit lacking so it takes a bit of trial and error. There are some examples of usage for debugging in some discussions too. |
I'm surprised by the fact you're using a lot of .with_output_fact() calls in fixing the ONNX interface. Most of the time, people end up overwritting the inputs ( |
I tried using onnx-simplifier and setting num_points to 1 and now I'm getting a panic from unwrapping a none in the
|
Hello, would you give a shot to the branch https://github.com/sonos/tract/tree/resize-op ? |
Trying to run a quantized model of the decoder of SAM.
If I run it without into_optimized(), it does run but the inference time is around 1min 50sec on my M1 Pro Macbook.
Whereas I expect this to run under a second(which is the performance I see in python)
I have an ONNX file and here is the code I use to load and run inference:
When I try running with into_optimized() it gives me the following error:
Is there a way to fix this, or alternatively is there a way to make the inference faster?
Thank you in advance :)
The text was updated successfully, but these errors were encountered: