Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

What is execution context memory? #2698

Open
wxsms opened this issue Jan 16, 2025 · 1 comment
Open

What is execution context memory? #2698

wxsms opened this issue Jan 16, 2025 · 1 comment
Labels
triaged Issue has been triaged by maintainers

Comments

@wxsms
Copy link

wxsms commented Jan 16, 2025

We found that with same model & same inference image & same configuration, model running on different devices will result in different execution context memory, and the difference could be very large. for example, on 4090, it takes 2304 MiB, and on A10 it takes 6384 MiB. Is this expected or configurable?

@nv-guomingz
Copy link
Collaborator

I'd like to say it's a expected but not configurable(depends on your device hw).

The value refers to a memory which mainly used for temporary storage required by layer implementations.
Different hw will has their own optimization, so the value difference is expected.

@nv-guomingz nv-guomingz added the triaged Issue has been triaged by maintainers label Jan 21, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
triaged Issue has been triaged by maintainers
Projects
None yet
Development

No branches or pull requests

2 participants