Sentence Directed Video Object Codetection
- 總結
- Researchers at Purdue University have developed a new method of video object codetection using the audio and visual cues in the video to help interpret and detect objects. This technology allows for video object detection without the need for object pre-learning. Using an algorithm that first generates sentences that describe an object's appearance/movement in the video by comparing subtle differences against a static background, it produces and displays a bounding box around an object while it is present in the video field. Using this technology, object detection is more robust, allowing for the detection of more objects, faster, and more accurately than ever before.
- 技術優勢
- Objects do not need to be pre-learned Increased accuracy Detects nearly any sized object Detects multiple objects simultaneously Works with fast object movement and motion blur
- 技術應用
- Surveillance Security Autonomous vehicles Facial recognition Computer vision Medical imaging Robotics
- 詳細技術說明
- Jeffrey SiskindPurdue Electrical and Computer Engineering
- *Abstract
-
- *Background
- Video object detection is a broad field with many applications ranging from self-driving cars to security surveillance. Video object codetection normally requires the pre-learning of objects before they can be detected. Pre-learning is a problem with many applications given it is not possible to pre-learn all potential scenarios. In addition, most of the work in object detection focuses on still images, which is simpler than the changing and moving viewpoints found in videos.
- *IP Issue Date
- None
- *IP Type
- Provisional
- *Stage of Development
- Prototype testing
- *Web Links
- Purdue Office of Technology CommercializationPurdueInnovation and EntrepreneurshipJeffrey SiskindPurdue Electrical and Computer Engineering
- 國家
- United States
- 申請號碼
- None
- 國家/地區
- 美國

欲了解更多信息,請點擊 這裡