UvA-Bosch Delta Lab Deep Learning Seminar
How Do Neural Networks See?
Christopher Olah – January 29
Abstract: Neural networks greatly exceed anything humans can design directly at computer vision by building up their own internal hierarchy of internal visual concepts. So, what are they detecting? How do they implement these detectors? How do they fit together to create the behavior of the network as a whole? At a more practical level, can we use these techniques to audit neural networks? Or find cases where the right decision is made for bad reasons? To allow human feedback on the decision process, rather than just the final decision? Or to improve our ability to design models?
Vacancy – PhD position in Machine Learning and Deep Learning. [closed]
PhD position in Machine Learning and Deep Learning. This position is funded by Bosch Research, is within UvA-Bosch Delta Lab whose research focuses on deep learning and applications to intelligent vehicles. As PhD candidate, you will perform cutting edge research in AMLAB in the field of machine learning, will be supervised by Prof dr. Max Welling (promotor). The research topic is ‘Methods for Robust Feature Learning’ where the goals are to learn features that are robust or invariant to changing conditions, to learn classifiers that perform well on multiple domains, to learn representations that can be transferred between domains and to learn classifiers that do not suffer from adversarial examples. Keywords: Bayesian Regularization, Model Uncertainty, Domain Transfer, Representation Learning.
Vacancy – PhD position in Machine Learning and Deep Learning. [closed]
PhD position in Machine Learning and Deep Learning. This position is funded by Bosch Research, is within UvA-Bosch Delta Lab whose research focuses on deep learning and applications to intelligent vehicles. As PhD candidate, you will perform cutting edge research in AMLAB in the field of machine learning, will be supervised by Prof dr. Arnold Smeulders (promotor). The research topic is ‘Learning to Follow Objects over Multiple Cameras‘ where the goals are to follow objects over multiple camera’s regardless variations in pose, illumination, occlusion, scale and other sources of variance in spite of the fact that most subjects are seen for the first time. This is a hard problem which requires learning generics on many examples and subsequently tailoring it to the current specific case. Keywords: Object Tracking, Generic to specific, Representation learning.
Colloquium at the ICAI opening and the 1st UvA-Bosch Delta Lab Deep Learning Seminar
Adaptive Deep Learning for Perception, Action, and Explanation
Prof. Dr. Trevor Darrell – October 19
Abstract: Learning of layered or “deep” representations has provided significant advances in computer vision in recent years, but has traditionally been limited to fully supervised settings with very large amounts of training data, where the model lacked interpretability. New results in adversarial adaptive representation learning show how such methods can also excel when learning across modalities and domains, and further can be trained or constrained to provide natural language explanations or multimodal visualizations to their users. I’ll present recent long-term recurrent network models that learn cross-modal description and explanation, using implicit and explicit approaches, which can be applied to domains including fine-grained recognition and visuomotor policies.
Prof. Darrell is on the faculty of the CS and EE Divisions of the EECS Department at UC Berkeley. He leads Berkeley’s DeepDrive (BDD) Industrial Consortia, is co-Director of the Berkeley Artificial Intelligence Research (BAIR) lab, and is Faculty Director of PATH at UC Berkeley. Darrell’s group develops algorithms for large-scale perceptual learning, including object and activity recognition and detection, for a variety of applications including autonomous vehicles, media search, and multimodal interaction with robots and mobile devices. His areas of interest include computer vision, machine learning, natural language processing, and perception-based human computer interfaces. Prof. Darrell previously led the vision group at the International Computer Science Institute in Berkeley, and was on the faculty of the MIT EECS department from 1999-2008, where he directed the Vision Interface Group. He was a member of the research staff at Interval Research Corporation from 1996-1999, and received the S.M., and PhD. degrees from MIT in 1992 and 1996, respectively. He obtained the B.S.E. degree from the University of Pennsylvania in 1988.
Vacancy – PhD position in Machine Learning and Deep Learning. [filled]
Colloquium – Challenges of multiple instance learning in medical image analysis
Veronika Cheplygina – Friday, January 26
Abstract: Data is often only weakly annotated: for example, for a medical image, we might know the patient’s diagnosis, but not where the abnormalities are located. Multiple instance learning (MIL), is aimed at learning classifiers from such data. In this talk, I will share a number of lessons I have learnt about MIL so far: (1) researchers do not agree on what MIL is, (2) there is no “one size fits all” approach (3) we need more thorough evaluation methods. I will give examples from several applications, including computer-aided diagnosis in chest CT images. I will also briefly discuss my work on crowdsourcing medical image annotations, and why MIL might be useful in this case.
Veronika Cheplygina is an assistant professor at the Medical Image Analysis group, Eindhoven University of Technology since February 2017. She received her Ph.D. from the Delft University of Technology for her thesis “Dissimilarity-Based Multiple Instance Learning“ in 2015. As part of her PhD, she was a visiting researcher at the Max Planck Institute for Intelligent Systems in Tuebingen, Germany. From 2015 to 2016 she was a postdoc at the Biomedical Imaging Group Rotterdam, Erasmus MC. Her research interests are centered around learning scenarios where few labels are available, such as multiple instance learning, transfer learning, and crowdsourcing. Next to research, Veronika blogs about academic life at http://www.veronikach.com.
Delta Lab Opening – Towards Affordable Self-driving Cars
Raquel Urtasun – April 6, 2017
The revolution of self-driving cars will happen in the near future. Most solutions rely on expensive 3D sensors such as LIDAR as well as hand-annotated maps. Unfortunately, this is neither cost effective nor scalable, as one needs to have a very detailed up-to-date map of the world. In this talk, I’ll review our current efforts in the domain of autonomous driving. In particular, I’ll present our work on stereo, optical flow, appearance-less localization, 3D object detection as well as creating HD maps from visual information alone. This results in a much more scalable and cost-effective solution to self-driving cars.
Raquel Urtasun is an Associate Professor in the Department of Computer Science at the University of Toronto and a Canada Research Chair in Machine Learning and Computer Vision. Prior to this, she was an AssistantProfessor at the Toyota Technological Institute at Chicago (TTIC), anacademic computer science institute affiliated with the University ofChicago. She received her Ph.D. degree from the Computer Science department at Ecole Polytechnique Federal de Lausanne (EPFL) in 2006 and did her postdoc at MIT and UC Berkeley. Her research interests includemachine learning, computer vision, robotics and remote sensing. Her recent work involves perception algorithms for self-driving cars, deep structured models and exploring problems at the intersection of vision and language. Her lab was selected as an NVIDIA NVAIL lab. She is a recipient of an NSERC EWR Steacie Award (awarded to the top 6 scientistsin Canada), an NVIDIA Pioneers of AI Award, a Ministry of Education and Innovation Early Researcher Award, three Google Faculty Research Awards,an Amazon Faculty Research Award, a Connaught New Researcher Award and a Best Paper Runner up Prize awarded at the Conference on Computer Vision and Pattern Recognition (CVPR). She is also Program Chair of CVPR 2018, an Editor of the International Journal in Computer Vision (IJCV) and has served as Area Chair of multiple machine learning and vision conferences (i.e., NIPS, UAI, ICML, ICLR, CVPR, ECCV, ICCV).
|15.30-15.35||Welcome Prof. Max Welling, Faculty of Science|
|15.35-15.40||Prof. Peter van Tienderen, Dean Faculty of Science|
|15.40-16.25||Key-note speech Prof. Raquel Urtasun, University of Toronto
Title: Towards Affordable Self-driving Cars
|16.25-16.35||Dr. Michael Bolle, President Corporate Research and Advanced Development, Robert Bosch GmbH|
|16.35-16.40||Prof. dr. ir. K.I.J. (Karen) Maex
Rector Magnificus University of Amsterdam
|16.40-16.45||Ms. Simone Kukenheim, Alderman City of Amsterdam|
|16.45-16.50||Opening UvA-Bosch DELTA Lab|
|16.50-18.00||Drinks and Closure|