• White Facebook Icon
  • White Instagram Icon
  • White Twitter Icon
  • White LinkedIn Icon

3rd ICML 2021 Workshop on Human in the Loop Learning

Recent years have witnessed the rising need for machine learning systems that have humans in the learning loop. Such systems can be applied to computer vision, natural language processing, robotics, and human-computer interaction. Creating and running such systems call for interdisciplinary research of artificial intelligence, machine learning, and cognitive science, which we abstract as Human in the Loop Learning (HILL). The HILL workshop aims to bring together researchers and practitioners working on the broad areas of HILL, ranging from the interactive/active learning algorithms for real-world decision-making systems (e.g., autonomous driving vehicles, robotic systems, etc.), lifelong learning systems that retain knowledge from different tasks and selectively transfer knowledge to learn new tasks over a lifetime, models with strong explainability, as well as human-inspired learning. The HILL workshop continues the previous effort to provide a platform for researchers from interdisciplinary areas to share their recent research. In this year’s workshop, a special feature is to encourage the exploration of human-inspired learning.

July 24

Virtual Conference


Call for Papers

We welcome high-quality submissions on algorithms and system designs in the broad area of human in the loop learning. A few (non-exhaustive) topics of interest include:

  • The topics of HILL include but are not limited to:

  • Interactive/Active machine learning algorithms for autonomous decision-making systems,

  • Lifelong learning systems that learn a sequence of tasks and leverage their shared structure to enable knowledge transfer over a lifetime,

  • Online learning and active learning,

  • Comparison of human in the loop learning and label-efficient learning,

  • Psychology driven human concept learning,

  • Explainable AI,

  • Human-inspired learning,

  • Design, testing, and assessment of interactive systems for data analytics,

  • Model understanding tools (debugging, visualization, introspection, etc.).


These topics span a variety of scientific disciplines and application domains like machine learning, human-computer interaction, cognitive science, and robotics. It is an opportunity for scientists in these disciplines to share their perspectives, discuss solutions to common problems and highlight the challenges in the field to help guide future research. The target audience for the workshop includes people who are interested in using machines to solve problems by having a human be an integral part of the learning process. 

We invite submissions of full papers, as well as works-in-progress, position papers, and papers describing open problems and challenges. While original contributions are preferred, we also invite submissions of high-quality work that has recently been published in other venues or is concurrently submitted. We encourage creative ML approaches, as well as interdisciplinarity and perspectives from outside traditional ML. Papers should be 4-8 pages in length (excluding references) formatted using the ICML template. All the submissions should be anonymous. The accepted papers are allowed to get submitted to other conference venues. 

Papers can be submitted through CMT:


Important Dates

Submission deadline: 27th June 2021 (23:59 AoE)

Acceptance notification: 7th, July 2021 (23:59 AoE)

Workshop Date: 24th July

Speakers of 2nd HILL Workshop at ICML 2020


Professor at Princeton University


Professor at University of Toronto


Assistant professor at Stanford University

Professor at University of Toronto

Professor at Tsinghua University

Assistant professor at UC Berkeley

Sergey Levine .jpeg

Postdoc Researcher at Facebook AI Research (FAIR)

Assistant professor at UC Berkeley

Research Scientist at Carnegie Mellon University

Associate professor at Carnegie Mellon University

Professor at

University of Tübingen

Agenda of 2nd HILL Workshop
at ICML 2020


Accepted Papers on 2nd HILL Workshop at ICML 2020



A Linear Bandit for Seasonal Environments

Giuseppe Di Benedetto (Oxford Univerisity); Vito Bellini (Amazon); Giovanni Zappella (Amazon)

Online Ride-Sharing Pricing with Fairness

Yupeng Li (University of Toronto/Shenzhen Research Institute of Big Data)*; Mengjia Xia (Cornell University); Dacheng Wen (The University of Hong Kong); Cheng Zhang (Didi Chuxing); Meng Ai (Didi Chuxing); Qun (Tracy) Li (DiDi)

Deep Active Learning: Unified and Principled Method for Query and Training

Changjian Shui (Université Laval)*; Fan Zhou (Laval University); Christian Gagné (Université Laval); Boyu Wang (University of Western Ontario)

GLAD: Localized Anomaly Detection via Human-in-the-Loop Learning

Md Rakibul Islam (Washington State university)*; Shubhomoy Das (School of EECS, Washington State University, Pullman); Janardhan Rao Doppa (Washington State University); Sriraam Natarajan (UT Dallas)

Human-Centric Efficiency Improvements in Image Annotation for Autonomous Driving

Frédéric Ratle (Samasource)*; Martine Bertrand (Samasource)

Online Learning for Distributed and Personal Recommendations - a Fair approach

Martin Tegnér (IKEA Retail, Oxford-Man Institute, University of Oxford)*

Yet Another Study on Active Learning and Human Pose Estimation

Sinan Kaplan (Lappeenranta University of Technology)*; Lasse Lensu (Lappeenranta University of Technology)

Program Synthesis with Pragmatic Communication

Yewen Pu (MIT)*; Marta Kryven (Massachusetts Institute of Technology); Kevin M Ellis (MIT); Joshua Tenenbaum (MIT); Armando Solar-Lezama (MIT)

Preference learning along multiple criteria: A game-theoretic perspective

Kush Bhatia (UC Berkeley)*; Ashwin Pananjady (UC Berkeley); Peter Bartlett (); Anca Dragan (EECS Department, University of California, Berkeley); Martin Wainwright (UC Berkeley)

Bridging the Gap: Providing Post-Hoc Symbolic Explanations for Sequential Decision-Making Problems with Inscrutable Representations

Sarath Sreedharan (Arizona State University)*; Utkarsh Soni (Arizona State University); Mudit Verma (Arizona State University); Siddharth Srivastava (Arizona State University); Subbarao Kambhampati (Arizona State University)

Interactive Segmentation of RGB-D Indoor Scenes using Deep Learning

Maximilian Ruethlein (Friedrich-Alexander University Erlangen-Nuernberg)*; Franz Koeferl (Friedrich-Alexander University Erlangen-Nuernberg); Wolfgang Mehringer (Friedrich-Alexander University Erlangen-Nuernberg); Bjoern Eskofier (Friedrich-Alexander University Erlangen-Nuernberg)

Interactive learning of cognitive programs

Sunayana Rane (MIT)*; Miguel Lázaro-Gredilla (Vicarious AI); Dileep George (Vicarious )

The Need for Standardised Explainability

Othman Benchekroun (Dathena)*; Adel Rahimi (Dathena); Qini Zhang (Dathena); Tetiana Kodliuk (Dathena)

Quick Question: Interrupting Users for Microtasks with Reinforcement Learning

Bo-Jhang Ho (UCLA); Bharathan Balaji (Amazon)*; Mehmet Koseoglu (UCLA); Sandeep Singh Sandha (University of California - Los Angeles); Siyou Pei (UCLA); Mani Srivastava (UC Los Angeles)

Adwait Sahasrabhojanee (USRA/NASA Ames); David Iverson (NASA Ames); shawn r wolfe (); Kevin Bradner (NASA Ames); Nikunj Oza (NASA Ames)*

Active Learning Strategies to Reduce Anomaly Detection False Alarm Rates

SCRAM: Simple Checks for Realtime Analysis of Model Training for Non-Expert ML Programmers

Eldon Schoop (University of California, Berkeley)*; Forrest Huang (University of California, Berkeley); Bjoern Hartmann (University of California, Berkeley)

AVID: Learning Multi-Stage Tasks via Pixel-Level Translation of Human Videos

Laura M Smith (UC Berkeley)*; Nikita Dhawan (UC Berkeley); Marvin Zhang (UC Berkeley); Pieter Abbeel (UC Berkeley); Sergey Levine (UC Berkeley)

Personalized Stress Detection with Self-supervised Learned Features

Stefan Matthes (Fortiss GmbH); Zhiwei Han (fortiss GmbH)*; Tianming Qiu (fortiss GmbH); Bruno Michel (IBM Zurich Research Lab); Sören Klinger (fortiss GmbH); Hao Shen (fortiss GmbH); Yuanting Liu (fortiss GmbH); Bashar Altakrouri (IBM Deutschland GmbH)

Metric-Free Individual Fairness in Online Learning

Yahav Bechavod (Hebrew University of Jerusalem)*; Steven Wu (University of Minnesota); Christopher Jung (University of Pennsylvania)

Bias in Multimodal AI: Testbed for Fair Automatic

Alejandro Peña (Universidad Autonoma de Madrid); Ignacio Serna (Universidad Autonoma de Madrid); Aythami Morales (Universidad Autonoma de Madrid)*; Julian Fierrez (Universidad Autonoma de Madrid)

Explanation Augmented Feedback in Human-in-the-Loop Reinforcement Learning

Lin Guan (Arizona State University)*; Mudit Verma (Arizona State University); Subbarao Kambhampati (Arizona State University)

Not all Failure Modes are Created Equal: Training Deep Neural Networks for Explicable (Mis)Classification

Alberto Olmo (Arizona State University)*; Sailik Sengupta (Arizona State University); Subbarao Kambhampati (Arizona State University)

Battlesnake Challenge: A Multi-agent Reinforcement Learning Playground with Human-in-the-loop

Jonathan Chung (Amazon Web Services)*; Runfei Luo (Amazon Web Services); Xavier Raffin (Amazon Web Services); Scott Perry (Amazon Web Services)

Faster Human-Machine Collaboration Bounding Box Annotation Framework Based on Active Learning

Minzhe Liu (Nanjing University)*; LI DU (Nanjing University); Yuan Du (Nanjing University); Ruofan Guo (Nanjing University); Xiaoliang Chen (University of California, Irvine)

Combining Human and Machine Intelligence to Assess Stroke Rehabilitation Exercises

Min Hun Lee (Carnegie Mellon University)*; Daniel Siewiorek (Carnegie Mellon University); Asim Smailagic (Carnegie Mellon University); Alexandre Bernardino (Instituto Superior Técnico); Sergi Bermudez (University of Madeira)

Personalized Size Recommendations with Human in the Loop

Leonidas Lefakis (Zalando)*; Evgenii Koriagin (Zalando SE); Julia Lasserre (Zalando Research); Reza Shirvany (Zalando SE)

A Prospective Human-in-the-Loop Experiment using Reinforcement Learning with Planning for Optimizing Energy Demand Response

Lucas Spangher (U.C. Berkeley)*; Manan Khattar (University of California at Berkeley); Akash Gokul (University of California at Berkeley); Akaash Tawade (University of California at Berkeley); Adam Bouyamourn (University of California at Berkeley); Alex R Devonport (U.C. Berkeley); Costas J. Spanos (University of California at Berkeley)

Learning Interpretable Models for Black-Box Agents

Pulkit Verma (Arizona State University)*; Siddharth Srivastava (Arizona State University)

Assisted Robust Reward Design

Jerry Zhi-Yang He (EECS Department, University of California, Berkeley)*; Anca Dragan (EECS Department, University of California, Berkeley)

Better Transferability with Attribute Attention for Generalized Zero-Shot Learning

Ruofan Guo (Nanjing University)*; LI DU (Nanjing University); Yuan Du (Nanjing University); Minzhe Liu (Nanjing University); Xiaoliang Chen (University of California, Irvine)

Human Explanation-based Learning for Machine Comprehension

Qinyuan Ye (University of Southern California)*; Xiao Huang (University of Southern California); Elizabeth Boschee (University of Southern California); Xiang Ren (University of Southern California)

Soliciting Stakeholders’ Fairness Notions in Child Maltreatment Predictive Systems

Hao-Fei Cheng (University of Minnesota)*; Paige Bullock (Kenyon College); Alexandra Chouldechova (CMU); Steven Wu (University of Minnesota); Haiyi Zhu (Carnegie Mellon University)

Improve black-box sequential anomaly detector relevancy with limited user feedback

Chris Kong (Amazon research); Lifan Chen (Amazon research); Ming Chen (Amazon research); Laurent Callot (Amazon research)*; Parminder Bhatia (Amazon)

Feature Expansive Reward Learning: Rethinking Human Input

Andreea Bobu (UC Berkeley)*; Marius Wiggert (UC Berkeley); Claire Tomlin (UC Berkeley); Anca Dragan (EECS Department, University of California, Berkeley)



Shanghang Zhang, UC Berkeley
Xin Wang, PhD at UC Berkeley,
Shiji Zhou, PhD at Tsinghua University,
Fisher Yu, Assistant professor at ETH Zurich,
Li Erran Li, Senior applied scientist at Amazon,
Kalesha Bullard, Postdoc. Researcher at F,
Wenwu Zhu, Professor at Tsinghua University, 
Trevor Darrell, Professor at UC Berkeley