Menu

[ELLIS Meetup] Efficient Hyperparameter Optimization in the Age of Deep Learning

Speaker: Josif Grabocka
Date: 2022-09-23 

Josif Grabocka from the Representation Learning Research Group at Uni Freiburg will give a talk titled “Efficient Hyperparameter Optimization in the Age of Deep Learning“.

Abstract

Hyperparameter Optimization (HPO) is essential in designing state-of-the-art Machine Learning systems. Unfortunately, existing HPO techniques are not efficient enough for Deep Learning, due to the high cost of evaluating multiple hyperparameter configurations in complex Deep Learning systems. In this talk, I will present three approaches for scaling HPO for Deep Learning along the lines of transfer learning, gray-box HPO, and pipeline optimization with pre-trained deep networks. The talk will end with a demonstration of the capability of HPO to achieve state-of-the-art predictive results on tabular datasets.

[ELLIS Meetup] Prior-data Fitted Networks (PFNs)

Speaker: Samuel Müller
Date: 2022-09-09

Location: Building 101, Room 00.036

Samuel Müller from the Machine Learning Lab will give a talk titled “Prior-data Fitted Networks (PFNs): Use neural networks for 100x faster Bayesian predictions“.

Abstract

Bayesian methods can be expensive and complicated to approximate with e.g. Markov Chain Monte Carlo (MCMC) or Variational Inference (VI) approaches. Prior-data Fitted Networks (PFNs) are a new, cheap and simple method to accurately approximate Bayesian predictions. I will explain how to build a PFN out of a Transformer by learning to model artificial data. I present the results from our paper that introduces PFNs, in which PFNs beat VI and MCMC for some standard tasks. As well as some more recent results with our new TabPFN, where we show that a simple PFN can replace a full AutoML tool for small datasets.

[ELLIS Meetup] Learning to solve arbitrary mobile manipulation tasks

Speaker: Daniel Honerkamp
Date: 2022-08-26

Abstract

Mobile manipulation remains a critical challenge across both service and industrial settings and is a key component to visions such as household assistants. But for this it requires the combination of a wide range of capabilities, such as perception and exploration in unknown environments while controlling large, continuous action spaces for simultaneous navigation and manipulation.
In this talk I will first provide an overview of the main challenges and current benchmarks. I will then summarize and walk through the pipelines of current state-of-the-art approaches with a focus on both the high-level task learning and the low-level motion execution on robotic agents. Lastly, I will discuss potential paths forward and ways to integrate these low- and high-level components.

[ELLIS Meetup] Introduction to Vision-and-Language Navigation

Speaker: Chenguang Huang
Date: 2022-08-12

Abstract

Robot navigation has been an actively researched challenge for decades. With different focuses, navigation goals can be defined in many different ways such as goal images, relative positions, semantic objects, and sound sources. Recently, with the advances in both computer vision and natural language processing methods, researchers propose to define the navigation goal with natural language instructions and define the task as Vision-and-Language Navigation. It is the very task that requires an agent to combine natural language instructions with visual perception data and navigate in an unseen environment. In this talk, I will give an overview of the Vision-and-Language Navigation tasks and introduce the current state of the research. In the end, I will discuss some future research directions.

[ELLIS Meetup] Robust Deep Neural Networks for Computer Vision

Speaker: Adam Kortylewski
Date: 2022-07-29

Abstract

Deep learning sparked a tremendous increase in the performance of computer vision systems over the past decade. However, Deep Neural Networks (DNNs) are still far from reaching human-level performance at visual recognition tasks. The most important limitation of DNNs is that they fail to give reliable predictions in unseen or adverse viewing conditions, which would not fool a human observer, such as when objects are partially occluded, seen in an unusual pose or context, or in bad weather. This lack of robustness in DNNs is generally acknowledged, but the problem largely remains unsolved.
In this talk, I will give an overview of the principles underlying my work on building robust deep neural networks for computer vision. My working hypothesis is that vision systems need a causal 3D understanding images by following an analysis-by-synthesis approach. I will discuss a new type of neural network architecture that implements such an approach, and I will show that these generative neural network models are vastly superior to traditional models in terms of robustness, learning efficiency and because they can solve many vision tasks at once. Finally, I will give a brief outlook on current projects of mine and future research directions.

[ELLIS Meetup] Robots Learning (Through) Interactions

Speaker: Jens Kober
Date: 2022-06-24

Abstract

The acquisition and self-improvement of novel motor skills is among the most important problems in robotics. I will discuss various learning techniques we developed that enable robots to have complex interactions with their environment and humans. Complexity arises from dealing with high-dimensional input data, non-linear dynamics in general and contacts in particular, multiple reference frames, and variability in objects, environments, and tasks. A human teacher is always involved in the learning process, either directly (providing data) or indirectly (designing the optimization criterion), which raises the question: How to best make use of the interactions with the human teacher to render the learning process efficient and effective? I will discuss various methods we have developed in the fields of supervised learning, imitation learning, reinforcement learning, and interactive learning. All these concepts will be illustrated with benchmark tasks and real robot experiments ranging from fun (ball-in-a-cup) to more applied (sorting products).

[ELLIS Meetup] Deep Learning 2.0

Speaker: Frank Hutter
Date: 2022-05-05

Abstract

Deep Learning (DL) has been able to revolutionize learning from raw data (images, text, speech, etc) by replacing domain-specific hand-crafted features with features that are jointly learned for the particular task at hand. I propose to take deep learning to the next level, by also jointly (meta-)learning other, currently hand-crafted, elements of the deep learning pipeline: neural architectures and their initializations, training pipelines and their hyperparameters, self-supervised learning pipelines, etc.

I dub this new research direction Deep Learning 2.0 (DL 2.0), since

  • Like deep learning, it replaces previous hand-crafted solutions with learned solutions (but now on the meta-level),
  • It allows for a qualitatively new level of automated customization to the task at hand, and
  • It allows for automatically finding meta-level tradeoffs with the objectives specified by the user, such as algorithmic fairness, interpretability, calibration of uncertainty estimates, robustness, energy consumption, etc. It can thus achieve Trustworthy AI by design, breaking qualitatively new ground for deep learning.

[ELLIS Meetup] Neural Ensemble Search

Speaker: Arbër Zela
Date: 2022-04-15

Abstract

Ensembles of neural networks achieve superior performance compared to standalone networks not only in terms of accuracy on in-distribution data but also on data with distributional shift alongside improved uncertainty calibration. Diversity among networks in an ensemble is believed to be key for building strong ensembles, but typical approaches only ensemble different weight vectors of a fixed architecture. Instead, we investigate neural architecture search (NAS) for explicitly constructing ensembles to exploit diversity among networks of varying architectures and to achieve robustness against distributional shift. By directly optimizing ensemble performance, our methods implicitly encourage diversity among networks, without the need to explicitly define diversity. We find that the resulting ensembles are more diverse compared to ensembles composed of a fixed architecture and are therefore also more powerful. We show significant improvements in ensemble performance on image classification tasks both for in-distribution data and during distributional shift with better uncertainty calibration.

[ELLIS Meetup] SAT Competitions + SAT Solving

Speaker: Armin Biere
Date: 2022-04-01

Abstract

Armin Bieres research interests are applied formal methods, more specifically formal verification of hardware and software, using model checking and related techniques with the focus on developing efficient SAT and SMT solvers. He is the author and co-author of more than 220 papers and served on the program committee of more than 160 international conferences and workshops. His most influential work is his contribution to Bounded Model Checking. Decision procedures for SAT, QBF and SMT, developed by him or under his guidance rank at the top of many international competitions and were awarded 98 medals including 55 gold medals. He is a recipient of an IBM faculty award in 2012, received the TACAS most influential paper in the first 20 years of TACAS in 2014 award, the HVC’15 award on the most influential work in the last five years in formal verification, simulation, and testing, the ETAPS 2017 Test of Time Award, the CAV Award in 2018, and the IJCAI-JAIR 2019 Award.

[ELLIS Meetup] Research in the Robot Learning Lab

Speaker: Abhinav Valada
Date: 2022-03-18

Abstract

The research of the Robot Learning Lab lies at the intersection of robotics, machine learning and computer vision with a focus on tackling fundamental robot perception, state estimation and planning problems using learning approaches to enable robots to reliably operate in more complex domains and diverse environments. The overall goal of this research is to develop scalable lifelong robot learning systems that continuously learn multiple tasks from what they perceive and experience by interacting with the real-world. The groups approach is to design deep learning algorithms that facilitate transfer of information through self-supervised multimodal and multitask learning by exploiting complementary features and cross-modal interdependencies. These techniques in turn enable robots to perceive more robustly and reason about the environment more effectively.