Skip to content

[Performance] Onnx session utilizes more GPU and CPU ram on Nvidia H100 than on Nvidia A100 #24543

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
razor182 opened this issue Apr 25, 2025 · 0 comments
Labels
performance issues related to performance regressions

Comments

@razor182
Copy link

Describe the issue

Hi everyone.

I have a model which consist of 4 onnx files.

  1. I load it on Nvidia A100 and get next RAM utilization results: CPU - 4,3 Gb, GPU - 800 Mb.
Image
  1. I load it on Nvidia H100 and get next RAM utilization results: CPU - 6,7 Gb, GPU - 1854 Mb.
Image

I use TensorrtExecutionProvider. Cuda, onnxruntime and tenorrt versions are same.

What's the reason of such differ? Is there any way to reduce memory utilization for H100?

To reproduce

Urgency

No response

Platform

Linux

OS Version

RHEL8

ONNX Runtime Installation

Released Package

ONNX Runtime Version or Commit ID

1.18.1

ONNX Runtime API

Python

Architecture

X64

Execution Provider

TensorRT

Execution Provider Library Version

CUDA 11.8 / CUDA 12.4

Model File

No response

Is this a quantized model?

No

@razor182 razor182 added the performance issues related to performance regressions label Apr 25, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
performance issues related to performance regressions
Projects
None yet
Development

No branches or pull requests

1 participant