What is the required GPU memory for inference? I encountered an OOM error when using the 4090
What is the required GPU memory for inference? I encountered an OOM error when using the 4090