-
Notifications
You must be signed in to change notification settings - Fork 72
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Added QNN EP for ARM64 builds. #41
Conversation
Not ready for review since the exported projects are not great yet... I'm not sure what to do there... |
AIDevGallery/Samples/ModelsDefinitions/embeddings.modelgroup.json
Outdated
Show resolved
Hide resolved
@azchohfi, I made few changes to address my initial comments, please take a look. In addition, the exported projects still need work to include ORT.QNN instead of ORT.DML. Also, the execution provider in the exported project defaults to CPU even if NPU is selected in the app. Once this is added, I would be good to merge this in |
Not sure why the build is not showing up as executed here. It is passing. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Approved with one minor comment.
else if (hardwareAccelerator == HardwareAccelerator.QNN) | ||
{ | ||
Dictionary<string, string> options = new() | ||
{ | ||
{ "backend_path", "QnnHtp.dll" }, | ||
{ "htp_performance_mode", "high_performance" }, | ||
{ "htp_graph_finalization_optimization_mode", "3" } | ||
}; | ||
_sessionOptions.AppendExecutionProvider("QNN", options); | ||
_chunkSize = 8; | ||
} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Should be removed if we're not doing NPU for all-Mini
No description provided.