My server configuration is as follows:
- apache 2.4.23.
- Mod_wsgi 4.5.9
By using the Django framework and apache server, we call the Keras deep learning model. And after the successful calling of the model, the model has been always running in the GPU memory, which causes the GPU memory can not be released except by shutting down the apache server.
So, is there any way to control the release of GPU memory when calling a Keras model by Apache+Mod_wsgi+Django?
Thanks!