Schedule: August 28th, 2023 (KST)

Time

Title

Speaker

09:00-09:05am Session: Opening Remarks
09:05-09:35am Keynote talk: "Embodied AI: Machine Learning to Learning Machines"

Abstract: Machine learning (including deep learning) has changed the paradigm of AI from rule-based “manual” programming to data-driven “automatic” programming. However, the current paradigm of machine learning requires some external system that provides them with data, making their scalability limited. Here we argue that the learner can feed itself the data autonomously if it is embodied, i.e. equipped with sensors and actuators. With the perception-action cycle the embodied AI can continually learn to solve problems in a self-teaching way by doing new actions, observing their outcomes, and correcting their own predictions like the humans and animals do. In this talk, I will show some of our studies in this direction of “(embodied) learning machine” research and discuss its implications for achieving truly human-level general AI.

09:35-10:05am Invited Talk 1: "Following Instructions and Asking Questions"

Abstract: As we move towards the creation of embodied agents that understand natural language, several new challenges and complexities arise for grounding (e.g. complex state-spaces), planning (e.g. long horizons), and social interaction (e.g. asking for help or clarifications). In this talk, I'll discuss several recent results both on improvements to embodied instruction following within ALFRED and initial steps towards building agents that ask questions or model theory-of-mind.

10:05-10:35am Invited Talk 2: "Scaling Robot Learning by Understanding Videos"

Abstract: True gains of machine learning in AI sub-fields such as computer vision and natural language processing have come about from the use of large-scale diverse datasets for learning. In this talk, I will discuss if and how we can leverage large-scale diverse data in the form of egocentric videos (first-person videos of humans conducting different tasks) to similarly scale up policy learning for robots. I will discuss the challenges this presents, and some of our initial efforts towards tackling them. In particular, I will describe techniques to acquire low-level visuomotor subroutines, high-level value functions, and an interactive understanding of objects from in-the-wild egocentric videos.

10:35-10:45am 10-min Coffee Break/Social Networking
10:45-11:55am

3-min teaser talks for poster presentations / Poster Session

Paper 1: ECLAIR: Event-Cognizant Language Interaction Embodied Robots
Jinyeon Kim, Byeonghwi Kim, Cheolhong Min, Yuyeong Kim, Taewoong Kim and Jonghyun Choi

Paper 2: GVCCI: Lifelong Learning of Visual Grounding for Language-Guided Robotic Manipulation
Junghyun Kim, Gi-Cheon Kang, Jaein Kim, Suyeon Shin and Byoung-Tak Zhang

Paper 3: Continual Fine-tuning using Linearized Deep Neural Networks
Hyounguk Shon and Junmo Kim

Paper 4: The Dialog Must Go On: Improving Visual Dialog via Generative Self-Training
Gi-Cheon Kang, Sungdong Kim, Jin-Hwa Kim, Donghyun Kwak and Byoung-Tak Zhang

Paper 5: Towards Robust Robot Perception: Enhancing Offline Reinforcement Learning with Adversarial Attacks and Defenses
Thanh Nguyen, Tung Luu and Chang Dong Yoo

Paper 6: CLIP is Also a Good Teacher in the Presence of Noisy Labels
Yeonguk Yu, Minhwan Ko and Kyoobin Lee

Paper 7: Leveraging Class-agnostic Object Proposals for Open-world Object Detection
Assefa Seyoum Wahd, Minsu Jang and Seung-Ik Lee

Paper 8: A Dataset Design for Question Generation Based on Human Cognition
Minjung Shin, Minsu Chang, Miyoung Cho and Jeh-Kwang Ryu

Paper 9: Reducing Object Hallucination for Image Captioning using Large Vision-Language Models with Reinforcement Learning
Hee Suk Yoon, Eunseop Yoon and Chang D. Yoo

Paper 10: A Data Set for Clarifying Ambiguous Questions with Intermediate Questions in Visual Question Answering
Gyu-Min Park and Seong-Bae Park

Paper 11: Spatio-Temporal Graph Random Walks to Understand Long Video Semantic Structure
Eun-Sol Kim, Hoyoung Yoon, Minseo Kim and Kyuyoung Lee

11:55-12:00pm Closing remarks