[HW Accel Support]: Issue with Nvidia Tesla P4 and any GPU features #19796
Replies: 3 comments 8 replies
-
|
Your setup and troubleshooting steps are thorough. Based on Frigate documentation and the details you provided:
Key points:
For more details on hardware acceleration and troubleshooting, see: 📚 Sources: Have another question? Just tag @inkeep. |
Beta Was this translation helpful? Give feedback.
-
|
is this AI response the only response I am going to get? it might be worth adding that I am running Ollama without any issue using nvidia-container-toolkit. I have also compiled my own ffmpeg binary (using CUDA 12.2 & v525 drivers) and left it running encoding one of the RTSP streams for hours and it works fine without causing a kernel panic. I also more recently tried a v570+ driver that was made available for my aging card, and it worked seemingly fine for around 10-15 minutes before it crashed the PVE server again. so I dont think its a low level issue with my system because it works for other VMs passing through the vGPU - I think its something specific that frigate does, and I just want some help or pointers trying to work out what is causing it. it also seems I am not alone - but this other thread has too many variables - #3652 |
Beta Was this translation helpful? Give feedback.
-
|
hi, I also want to using P106-100 for object detection. Does it work? |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Describe the problem you are having
I have had a stable Frigate setup running on Proxmox with a USB Coral for years. Its great, thank you for all your hard work.
Over time I added an Nvidia P106 GPU (full passthrough) to test tensor object identification and more recently face recognition, etc. It all works, but my main issue was this GPU having no HW decoder and the CPU being under pressure adding timestamps to recorded video (6 streams @ 1080p). I needed something else reasonably cheap, so went with a tesla P4.
I recently added the Tesla P4 (vGPU) and tried to get HW acceleration working. I realise the built-in ffmpeg requires drivers higher than what this card supports, and compiled my own ffmpeg binaries that work. The problem is that by enabling the GPU in the docker compose file, I get maybe 10 minutes of the system working before it crashes my entire proxmox server with a kernel panic.
Ive checked the kernel logs from journalctl in both the VM and host, there is no error at all before it crashes.
I have tried running this setup with just the GPU enabled in the docker config and all other HW acceleration / GPU features off in the frigate config, but still something runs on the GPU (I can see the process in nvidia-smi) - pushes it only up to around 4% utilisation. and it crashes very quickly. if I remove the GPU from the docker-compose file it runs fine again.
I have run other VMs with CUDA on this setup and done benchmarking tests, and they are all stable. I really dont think its a driver problem in itself, but more of a compatibility problem with Frigate. I want to try and rebuild any GPU / CUDA related components with CUDA 12.2.2 (the highest supported by this GPU / driver) but there doesnt appear to be a documented way to run frigate outside of docker.
Do you have any suggestions?
Version
0.16.0-c2f8de9
Frigate config file
docker-compose file or Docker CLI command
Relevant Frigate log output
Relevant go2rtc log output
FFprobe output from your camera
Install method
Proxmox via Docker
Object Detector
Coral
Network connection
Wired
Camera make and model
Multiple - but no camera issue
Screenshots of the Frigate UI's System metrics pages
Nothing noteworthy
Any other information that may be helpful
highest supported driver version for the Tesla P4 is 535.261.03 - CUDA 12.2.2 - Nvidia container toolkit 1.14.3
Beta Was this translation helpful? Give feedback.
All reactions