This post has been de-listed
It is no longer included in search results and normal feeds (front page, hot posts, subreddit posts, etc). It remains visible only via the author's post history.
Hey, I'm currently developing an object detector and tracker for an autonomous driving application as part of my master thesis. I built my own dataset with 4,5k images containing the objects I want to detect and I'm already quite satisfied with my results compared to the dataset size (~90 % AP for the classes i care most about). Nevertheless I still have a good amount of False Positives when evaluating my Yolo v4 based detector on unseen (video)data. My object tracker can already mitigate some of these FPs (e.g. when they only occurred in one frame) but I was wondering if there are other worthy strategies to further mitigate this problem. Surely I could simply increase the dataset size or increase the network input size, but I'm looking for ideas/strategies beside that. I'd be grateful for some tips, ideas or papers that are worth reading :)
Subreddit
Post Details
- Posted
- 4 years ago
- Reddit URL
- View post on reddit.com
- External URL
- reddit.com/r/computervis...