CPU Memory Leak During Inference with trf Model #13582
Unanswered
Amriteshwork
asked this question in
Help: Coding & Implementations
Replies: 1 comment
-
This is most likely caused by the lexeme cache and string store: |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I am trying to get the entity from the text in an infinite loop. The text is being sent in a batch of 16 or less depending upon how many sentences I get from the backend. The model is loaded on the GPU, so I think the use of CPU is used to load and unload from gpu, and then cup memory should return to its original state. But the CPU memory is continuously increasing while GPU memory is as it is. following is the reference code:
I am using spacy 3.7.4.
memory at each batch are (in bytes):
335206469 --> 335366716 --> 335411483 --> 335700046 --> 335890747 ...
Beta Was this translation helpful? Give feedback.
All reactions