Skip to content

Instantly share code, notes, and snippets.

@lovettchris
Last active August 15, 2022 21:09
Show Gist options
  • Star 0 You must be signed in to star a gist
  • Fork 0 You must be signed in to fork a gist
  • Save lovettchris/3c588ee252fa7bbee1e2cd1763235d7f to your computer and use it in GitHub Desktop.
Save lovettchris/3c588ee252fa7bbee1e2cd1763235d7f to your computer and use it in GitHub Desktop.
SNPE Performance Comparison between 1.61 and 1.64
Display the source blob
Display the rendered blob
Raw
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Display the source blob
Display the rendered blob
Raw
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Display the source blob
Display the rendered blob
Raw
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.

SNPE Performance Comparison between 1.61 and 1.64

I upgraded my Qualcomm 888 testing pipeline to a new version of the SNPE SDK moving from snpe-1.61.0 to snpe-1.64.0. Then re-ran all our 8000 models we have been testing. The results are mixed. On the one hand snpe-1.64.0 seems to have fixed some quantization bugs such that a lot more models are getting an interesting non-zero F1 accuracy score and all scores are the same or better than before, but on the other hand the mean inference times have slowed down some. The slow down is more evident in the smaller faster models. So my question is for smaller models the perf hit is pretty bad, is this a known issue and something that will likely improve in future versions of the SDK ?

Here's the results on the smaller faster models:

fast

Here's the result on medium sized models:

medium

Here's the results on larger models:

large

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment