Winter 2025 Colloquium

Organizers: Jerry Savage, Abhishek Gupta, Maya Cakmak, Josh Smith

Learning-from-observation2.0
Katsu Ikeuchi (Principal Researcher/Research Manager, Microsoft) 01/31/2025

Abstract: We are developing a Learning-from-Observation (LfO) system that acquires robotic behaviors through the observation of human demonstrations. Unlike the bottom-up approach known as "Learning-from-Demonstration" or "Imitation Learning," which replicates human movements as they are, we are employing a top-down approach (top-down learning-from-observation). This method entails observing only the critical components of human actions through a task model representation (akin to Minsky's frame), generating an abstract representation based on these observations, which is subsequently mapped onto the robot's behavior. The advantages of this top-down approach include the ability to generalize and correct observational errors by utilizing an intermediate task model representation, thereby enhancing the affinity with large language models. Furthermore, by tailoring the mapping to each individual robot, the system can be applied to different robotic platforms without necessitating significant modifications to the recognition system. The initial step of the system involves the utilization of a large language model (LLM) to comprehend the "what-to-do" from human demonstrations and subsequently retrieve the corresponding task model. This task model directs the CNN-based observation module to focus on specific aspects of human behavior and fills in the requisite parameters for "how-to-do," thereby completing the intermediate representation. Based on this finalized task model, the system activates the appropriate agents from a pre-trained group of agents—trained through reinforcement learning on the "how-to-do" aspect—to execute the robot's actions. This presentation will provide a comprehensive overview of the system architecture, the design methodologies for the pre-trained skill sets, and other pertinent details. Furthermore, it will discuss a comparison between this hybrid approach, which integrates traditional robotic techniques with LLMs, and end-to-end (E2E) methodologies, including foundation models.

Biography: Dr. Ikeuchi joined Microsoft in 2015, following distinguished tenures at MIT's Artificial Intelligence Laboratory, Japan's National Institute of Advanced Industrial Science and Technology (AIST), Carnegie Mellon University's Robotics Institute (CMU-RI), and the University of Tokyo. His research interests span computer vision, robotics, and Intelligent Transportation Systems (ITS). He has served as the Editor-in-Chief of the International Journal of Computer Vision (IJCV) and the International Journal of Intelligent Transportation Systems (IJITS), as well as the Encyclopedia of Computer Vision. Dr. Ikeuchi has also chaired numerous international conferences, including IROS95, CVPR96, ICCV03, ITSW07, ICRA09, ICPR12, and ICCV17. He has been the recipient of several prestigious awards, such as the IEEE PAMI Distinguished Researcher Award, the Okawa Award, the Funai Award, the IEICE Outstanding Achievements and Contributions Award, as well as the Medal of Honor with Purple Ribbon from the Emperor of Japan. Dr. Ikeuchi is a Fellow of IEEE, IAPR, IEICE, IPSJ, and RSJ. He earned his Ph.D. in Information Engineering from the University of Tokyo and his Bachelor's degree in Mechanical Engineering from Kyoto University.

Peter Stone (Professor of Computer Science, University of Texas at Austin) 02/07/2025
Towards Task Planning with Learned Manipulation Skills
Sebastian Castro (Roboticist/Applied Scientist) 02/21/2025

Abstract: Robotics as a field has a constantly growing repository of fundamental techniques for perception, motion planning, navigation, and control. Lately, this has been accelerated by robots becoming more ubiquitous in industry, as well as a surge of research in machine learning and optimization based approaches. As we become equipped with the ability to program robots with a variety of skills, it naturally follows to think about composition of these skills to achieve high-level goal specifications. This talk will introduce the landscape of tools that enable planning at the task level, as well as common behavior abstractions to ground task plans to robust, executable skills in the real world. Then, we describe some promising research directions in composition of learned manipulation skills, with a goal of operationalizing robot task planning for real-world tasks while reducing dependency on domain-specific, hand-engineered solutions.

Biography: Sebastian Castro is a roboticist and applied scientist at The AI Institute, working on task planning and composition of learned manipulation skills. He holds Bachelor’s and Master’s degrees from Cornell University in mechanical engineering, with a concentration on dynamics, systems, and control, applied to high-level planning and control of modular robots. His prior professional experience includes technical content development and marketing for robotics competitions with MathWorks, and robotics software engineering at MIT Computer Science and Artificial Intelligence Laboratory (CSAIL), Boston Dynamics, and PickNik Robotics. Sebastian also devotes personal time to robotics education through blog posts, open-source software, and education-focused talks and workshops.

Russ Tedrake (Toyota Professor of EECS, MIT) 02/28/2025
Jeannette Bohg (Assistant Professor of Computer Science, Stanford University) 03/07/2025
Xiaolong Wang (Assistant Professor of Computer Science and Engineering, UC San Diego) 03/14/2025