Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Enable QNN HTP spill fill buffer setting to save RAM usage. #22853

Open
wants to merge 3 commits into
base: main
Choose a base branch
from

Conversation

HectorSVC
Copy link
Contributor

Description

Enable QNN HTP spill fill buffer setting to save RAM usage.
This feature is available after QNN 2.28. Need to re-generate QNN context binary.
https://docs.qualcomm.com/bundle/publicresource/topics/80-63442-50/htp_backend.html#qnn-htp-backend-api

Copy link
Contributor

@github-actions github-actions bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

You can commit the suggested changes from lintrunner.

Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
@HectorSVC HectorSVC closed this Nov 15, 2024
@HectorSVC HectorSVC reopened this Nov 15, 2024
@HectorSVC HectorSVC added the ep:QNN issues related to QNN exeution provider label Nov 15, 2024
@HectorSVC
Copy link
Contributor Author

@chiwwang, could you help to take a look?

@chiwwang
Copy link

Hi Hector,
This looks good for me but let me ping others and see if they can also take a look.

@HectorSVC
Copy link
Contributor Author

HectorSVC commented Nov 22, 2024

Comments from QC: The approach has the limitation that it always gets the max spill fill buffer size form the 1st QNN context. The max spill file buffer size should be across all QNN contexts. To fill the gap, we need to go through all QNN context to:

  1. Load the QNN context binary buffer and extract the max spill fille buffer size for each QNN context
  2. Compare the max spill fille buffer size across all QNN context and track the index of the QNN context
  3. Load and deserialize the QNN context (to get the graph info for future execute) which has the max spill fille buffer size first, also set the max spill fill buffer, set the group handle to 0.
  4. Load and deserialize other QNN contexts, set the max spill buffer size, and set the group handle to the context in step3.

Considering this feature is mostly target for large models which has large context binary size, so there will be big overhead for step 1 & 2. Another approach is we dump the max spill fill buffer size for each Qnn context in EPContext node when we generate the model to make this information ready ahead of time instead of get it during normal session creation time. We can get the information from all EPContext nodes to get the max size and load that one first.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ep:QNN issues related to QNN exeution provider
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants