蘇黎世聯邦理工學院計算機視覺實驗室由計算機視覺領域著名學者Luc Van Gool, 以及醫療影像教授Ender Konukoglu和計算機視覺及系統教授Fisher Yu的研究組組成,是歐洲乃至世界最頂尖的CV/ML研究機構之一。ETHZ CVL關注訊號從採集、分析到處理的全流程,旨在開發通用的概念和方法,研究領域包括視覺場景理解、醫學影像分析、機器人、具身智慧、高效神經網路與計算等。
CVL對標國際研究前沿,以實際應用驅動研究,並將與工業界的密切合作視為重要優勢。研究成果發表於計算機視覺(CVPR、 ECCV、ICCV)、機器學習(NeurIPS、ICML、ICLR)、人工智慧(AAAI)、機器人(ICRA)、醫學影像(MICCAI)等領域的頂級會議。其中,Luc Van Gool教授提出的SURF運算元是計算機視覺領域的經典演算法,谷歌學術引用超過20000次。CVL組織的「PASCAL VOC挑戰賽」對學術界和工業界產生深遠影響。
CVL由影像通訊與理解(Image Communication and Understanding,ICU),生物醫學影像計算(Biomedical Image Computing,BMIC),視覺智慧和系統(Visual Intelligence and Systems,VIS),計算機輔助醫學應用(Computer-assisted Applications in Medicine,CAiM)共4個研究小組構成。其中:
影像通訊與理解(ICU)小組由Luc Van Gool教授領導,Dengxin Dai、Radu Timofte、Martin Danelljan等三位講師,分別在自動駕駛、影像處理、目標跟蹤等領域展開研究。此外,ICU小組的成員對高效網路設計與計算、大規模3D場景理解、影像和影片解析以及3D重構等方向有著廣泛的興趣。相關主頁:https://icu.ee.ethz.ch/
生物醫學影像計算(BMIC)小組由Ender Konukoglu教授領導。面向生物醫學領域的前沿挑戰,研發理論合理且實際可行的前沿技術解決方案。相關主頁:https://bmic.ee.ethz.ch/
視覺智慧和系統(VIS)小組由Fisher Yu教授領導。藉助影像處理、機器學習、統計學、深度學習和控制理論,研究可以在實際環境中執行復雜任務的感知機器人系統。相關主頁:https://cv.ethz.ch/
計算機輔助醫學應用(CAiM)小組由Orçun Göksel教授領導。CAiM關注醫療資料分析和資訊提取,研究基礎涉及多個交叉學科,包括工程學、電腦科學和醫學。小組成員具備跨學科背景,力求研發前沿的醫學成像和影像分析技術,並應用於臨床實踐。相關主頁:https://caim.ee.ethz.ch/group.html
直播地址:https://jmq.h5.xeknow.com/s/30jrSR(點選閱讀原文直達)
6月7日 20:00-21:00
分享主題:New Opportunities in Monocular 2D and 3D Object Tracking
分享摘要:Object tracking is foundational for video analysis and a key component for perceptionin autonomous systems, such as self-driving cars and drones. Due to its importance, tracking has been studied extensively in the literature. However, the availability of large-scale driving video data brings new research opportunities in the field.
In this talk, I will discuss our recent findings in multiple object tracking (MOT), after briefly reviewing the current works and trends on the topic. Then, I will introduce our new tracking method based on Quasi-Dense Similarity Learning. Our method isconceptually more straight forward yet more effective than the previous works. It boosts almost ten percent of accuracy on the large-scale BDD100K and WaymoMOT datasets.
I will also talk about how to use the 2D tracking method for monocular 3D object tracking. Our quasi-dense 3D tracking pipeline achieves impressive improvements on the nuScenes 3D tracking benchmark with five times tracking accuracy of the popular tracking methods. Our works point to some interesting directions in MOT research and technology in the era of ubiquitous videos.
6月9日 19:00-20:00
分享摘要:The world we live inis highly structured. Things that are semantically related are typically presented in a similar way: both trucks and buses have wheels and cabins, for example. Things also undergo continuous variations over time; in a video clip, content among frames are highly correlated. Our humans also interact with the environment constantly and communicate with each other frequently. Overall, there exist rich structures between human and environment and over both spaceand time.
Therefore, it is highly needed to understand this visual world from a structured view. In this talk, I will advocate the value of structured information in intelligent visual perception. As examples, I will present a line of my recent work on semantic segmentation, human semantic parsing and fake video detection.
6月10日 19:00-20:00
分享摘要:In recent year, deep neural networks has boosted the performances of computer vision algorithms for the task of visual recognition, object detection, semantic segmentation etc. Yet, it often comes with the increase of the model complexity in terms of number of parameters and computational cost. This increases the energy consumption, latency, andtransmission of pre-trained models and becomes a huge challenge for the deployment of deep neural networks on edge devices. Thus, network compression and model acceleration becomes a feasible solution to this problem.
In this talk, we will report the recent development and thought of model acceleration in the Computer Vision Lab. We introduce our work on model acceleration and network architecture optimization from the perspective of learning filter basis, group sparsity, differentiable network pruning, and the advantage of heterogenous network architectures.
直播地址:https://jmq.h5.xeknow.com/s/30jrSR(點選閱讀原文直達)
掃碼入群:針對本次分享主題,歡迎大家進群一起交流。
如群已超出人數限制,請新增其他小助手:syncedai2、syncedai3、syncedai4 或 syncedai5,備註「cv lab」即可加入。
地址:https://jmq.h5.xeknow.com/s/30jrSR