You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
For the example when model loading the RAM required close to 5 GB and VRAM use 2.1 GB. How can i reduce RAM uses for loading the model infernce fime. basically i try to figure out which is the issue for taking more RAM. Here i found when i initialize the GPT block then this model used closed to 5 GB RAM. this RAM is not GPU memory.
saiful9379
changed the title
why xtts v2 inferrence time used RAM double(or more 3x) then GPU or VRAM
why xtts v2 inference time used RAM double(or more 3x) then GPU or VRAM
Aug 20, 2024
Describe the bug
For the example when model loading the RAM required close to 5 GB and VRAM use 2.1 GB. How can i reduce RAM uses for loading the model infernce fime. basically i try to figure out which is the issue for taking more RAM. Here i found when i initialize the GPT block then this model used closed to 5 GB RAM. this RAM is not GPU memory.
To Reproduce
Inference used RAM : 4634.7890625
Expected behavior
Expected low RAM use when inference
Logs
No response
Environment
Additional context
No response
Tasks
The text was updated successfully, but these errors were encountered: