Closed 503Dev closed 1 year ago
it's having an issue with the labelmap, not sure why as other users have not seen this issue
it's having an issue with the labelmap, not sure why as other users have not seen this issue
Thank you. I did suspect this as I have worked with custom models in the past and defined my labelmaps and recognized the error. The confusing thing to me here is that this is the stock models and not custom therefore I do not have a custom labelmap to define and as a result I am not sure how to proceed or fix this.
I attempted to grab the labelmap.txt from the repo, drop it in my _modelcache and specify it via the config but the same error happens.
I don't know why it would happen unless maybe something went wrong while the model was generated, like I said others have not seen this issue. Does running the 320 model work for you?
I don't know why it would happen unless maybe something went wrong while the model was generated, like I said others have not seen this issue. Does running the 320 model work for you?
It does not. I am going to backup and then purge all of it and do a fresh setup from scratch. I am coming from a very weathered 0.12 install w/ previous use of extensive custom models, etc. Maybe something in the database or in my files is irregular. I will attempt a full fresh run and update shortly.
it won't be the database file itself but yeah could be something weird
it won't be the database file itself but yeah could be something weird
Well, I am perplexed. Just wiped everything and did a fresh pull and started with a minimum viable config. Everything starts smooth, detect process launches, the tensorrt engine loads ok and then same exact thing:
frigate | 2023-10-19 18:17:35.384802509 Process camera_processor:sala:
frigate | 2023-10-19 18:17:35.386412649 Traceback (most recent call last):
frigate | 2023-10-19 18:17:35.386459460 File "/usr/lib/python3.9/multiprocessing/process.py", line 315, in _bootstrap
frigate | 2023-10-19 18:17:35.386462508 self.run()
frigate | 2023-10-19 18:17:35.386465280 File "/usr/lib/python3.9/multiprocessing/process.py", line 108, in run
frigate | 2023-10-19 18:17:35.386467735 self._target(*self._args, **self._kwargs)
frigate | 2023-10-19 18:17:35.386482408 File "/opt/frigate/frigate/video.py", line 506, in track_camera
frigate | 2023-10-19 18:17:35.386484884 process_frames(
frigate | 2023-10-19 18:17:35.386487348 File "/opt/frigate/frigate/video.py", line 881, in process_frames
frigate | 2023-10-19 18:17:35.386489451 detect(
frigate | 2023-10-19 18:17:35.386491886 File "/opt/frigate/frigate/video.py", line 580, in detect
frigate | 2023-10-19 18:17:35.386494419 region_detections = object_detector.detect(tensor_input)
frigate | 2023-10-19 18:17:35.386496948 File "/opt/frigate/frigate/object_detection.py", line 225, in detect
frigate | 2023-10-19 18:17:35.386499442 (self.labels[int(d[0])], float(d[1]), (d[2], d[3], d[4], d[5]))
frigate | 2023-10-19 18:17:35.386511808 KeyError: -14
In frigate web UI the image is blank (no frames recvd) but on the system page I can see the inference speed updating properly and the detector process showing activity and FPS. Very unusual. I am stumped as to what is going on, I followed the 0.13-beta2 docs and started fresh with all minimal configs.
I tested both 0.13-beta1 and 0.13-beta2 and the result was the same.
Maybe @NateMeyer has a better idea.
What driver version are you running?
Maybe @NateMeyer has a better idea.
What driver version are you running?
NVIDIA-SMI 525.125.06 Driver Version: 525.125.06 CUDA Version: 12.0
0.13 updated the TensorRT library and now needs driver 530 or later.
@NateMeyer Thank you. This can be closed as absolute user error. I actually read that and assumed mine were up-to-date because I updated everything on my fresh Debian 12 install. Of course, thats a dumb assumption and my drivers were up to date but the repo drivers for Nvidia are not 5.30+ because they aren't flagged as stable.
Anyway, I had to purge those + Nouveau and then manually use the .run installer from the Nvidia site. Rebooted and tested, works like a charm. Performance is significantly better on 0.13.0-beta2 vs 0.12.x.
Describe the problem you are having
Upon switching to 0.13.0-beta2-tensorrt, the tensorrt detector no longer works correctly. I have forced the models to rebuild and ensured they are in the _modelcache folder as expected. The engine loads and I can see the processes in nvidia-smi as expected, but the camera streams are unavailable and the log produces errors.
Version
0.13.0-beta2-tensorrt
Frigate config file
docker-compose file or Docker CLI command
Relevant log output
Operating system
Debian
Install method
Docker Compose
Coral version
Other
Any other information that may be helpful
TensorRT via PCI-E Nvidia GPU. Configuration works fine with 0.12 using TensorRT. I tested with CPU detectors on 0.13.0-beta2 to validate, and it functions as expected.