黃正能教授→: Open Long-Tailed Recognition (OLTR) in Deep Learning Tasks

12月21日 9:30,線上

發布者:韋鈺 發布時間:2020-12-08瀏覽次數💁🏻:5770

報告內容🗼:Open Long-Tailed Recognition (OLTR) in Deep Learning Tasks

報告人➞:黃正能 教授

報告時間:12月21日 9:30

報告方式🧝:線上(騰訊會議💪🏼:329 808 289)


報告人簡介

Real world data often have a long-tailed and open-ended distribution. A practical recognition system must classify among majority and minority classes, generalize from a few known instances, and acknowledge novelty upon a never seen instance. The Open Long-Tailed Recognition (OLTR) of a deep learning architecture should be optimized for the classification accuracy over a balanced test set which include head, tail, and open classes, i.e., the OLTR must handle imbalanced classification, few-shot learning, and open-set recognition in one integrated algorithm, whereas existing classification approaches focus only on one aspect and deliver poorly over the entire class spectrum. We propose an OLTR deep learning platform, which uses re-balanced sampling strategy to improve the recognition accuracy of tail classes, without degrading the accuracy of head classes. To simultaneously determine the open classes, the embedding features are metric learned based on an auto-encoder architecture, and the dimensionality reduced features are tested by an innovative adaptive outlier factor (AOF) algorithm, which is an unsupervised anomaly detection method, to predict the open classes.



  報告內容簡介:

 Multiple object tracking (MOT) and video object segmentation (VOS)  are crucial tasks in computer vision society. Further improvement and significance  can be achieved by effectively combining these two tasks together, i.e.,  multiple object tracking and segmentation (MOTS). However, most tracking-by-detection MOT methods, with available detected bounding boxes, cannot effectively handle static, slow-moving and fast moving camera scenarios simultaneously due to ego-motion and frequent occlusion. In this work, we propose a novel tracking framework, called “instance-aware MOT” (IAMOT), that can track multiple objects in either static or moving cameras by jointly considering the instance-level features and object motions. Overall, when evaluated on the MOTS20 and KITTI-MOTS dataset, our proposed method won the first place in Track3 of the BMTT Challenge in IEEE CVPR 2020 workshop. When Lidar information is available, we further propose a multi-stage framework called “Lidar and monocular Image Fusion based multi-object Tracking and Segmentation (LIFTS)” for MOTS. This proposed framework is also evaluated on BMTT Challenge 2020 Track2: KITTI-MOTS dataset and achieves the 2nd place ranking in the competition.





 





  






























恒达平台专业提供:恒达平台🫅🏿、恒达🫄🏽、恒达娱乐等服务,提供最新官网平台、地址、注册、登陆、登录、入口、全站、网站、网页、网址、娱乐、手机版、app、下载、欧洲杯、欧冠、nba、世界杯、英超等,界面美观优质完美,安全稳定,服务一流🥱,恒达平台欢迎您。 恒达平台官網xml地圖