thanks for pointing out. I was trying to offload to two running instance so I can get a faster inference speed. But after I shut one off, it seem to be fine.
most likely due to I was running larger model size ( now is size small )
This is my nvidia-smi
+---------------------------------------------------------------------------------------+
| NVIDIA-SMI 535.161.07 Driver Version: 535.161.07 CUDA Version: 12.2 |
|-----------------------------------------+----------------------+----------------------+
|...
For those who have trouble running CodeProjectAI YOLO with GPU (CUDA), this is something you can try
I installed my CodeProjectAI in LXC in Proxmox, with 4GPU Sharing among each other LXC containers.
( # is a prompt, )
1. enter
# cd /usr/bin/codeproject.ai-server-2.9.5
For some reason, running...