Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

running error #2

Open
laybebe opened this issue Jul 6, 2023 · 4 comments
Open

running error #2

laybebe opened this issue Jul 6, 2023 · 4 comments

Comments

@laybebe
Copy link

laybebe commented Jul 6, 2023

Thanks for your wonderful jobs! I followed the “GET_STARTED.md” to test the TAO model on the Open-vocabulary MOT benchmark, however, the code threw an error.
1688635366852

@siyuanliii
Copy link
Collaborator

Have you run the conversion scripts for the TAO annotations file?

@laybebe
Copy link
Author

laybebe commented Jul 14, 2023

Thanks for your reply. I ensure that I have not omitted any steps. I will double-check and run it again.

@AndyBarcia
Copy link

This problem is apparently caused because the trackers doesn't actually detect anything with the default configuration (at least for me), because this line sets the score threshold too high:

rcnn_test_cfg.score_thr = 1 / len(text_features) + 0.01

https://github.com/SysCV/ovtrack/blob/e4f20e0974753539a47cf96fae46efad8adfd3b6/ovtrack/models/roi_heads/ovtrack_roi_head.py#L526C9-L526C64

Changing it to the following gives me the best results, although I'm not sure if this is the correct way of doing it.

rcnn_test_cfg.score_thr = (1 / len(text_features)) * 1.001

@siyuanliii
Copy link
Collaborator

Thanks for pointing out this problem, and sorry for the inconvenience. I accidentally added this line. It was intended for dynamic adjusting the inference threshold on any video with any class prompt. If you check in detail, you will find the confidence score varies depending on the number of classes to track. This is due to the nature of the softmax function. I added that line to dynamically adjust the proper threshold for any number of class prompts. However, this is not needed for inference on standard benchmarks. I update the code accordingly.

I also noticed that adding the line you suggested to dynamically adjust the rcnn confidence threshold works very well! It can give better overall but only slightly worse classification performance on TAO validation benchmark.
I will add this line as default with a hyperparameter "dynamic_rcnn_thre" to control it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants