| CARVIEW |
Select Language
HTTP/2 200
server: GitHub.com
content-type: text/html; charset=utf-8
last-modified: Wed, 28 Jul 2021 08:21:08 GMT
access-control-allow-origin: *
strict-transport-security: max-age=31556952
etag: W/"61011374-5eae"
expires: Mon, 29 Dec 2025 22:03:14 GMT
cache-control: max-age=600
content-encoding: gzip
x-proxy-cache: MISS
x-github-request-id: 38CF:3827E5:961139:A86619:6952F84A
accept-ranges: bytes
age: 0
date: Mon, 29 Dec 2025 21:53:14 GMT
via: 1.1 varnish
x-served-by: cache-bom-vanm7210040-BOM
x-cache: MISS
x-cache-hits: 0
x-timer: S1767045194.445811,VS0,VE206
vary: Accept-Encoding
x-fastly-request-id: 47be94e0fe1c563a2f0fd08f7d1319784c7fd7bd
content-length: 5119
RSS 2021 VLRR Workshop
Visual perception is essential for achieving robot autonomy in the real
world. To perform complex robot tasks in unknown environments, a robot
needs to actively acquire knowledge through physical interactions and
conduct sophisticated reasoning of the observed objects. This invites a
series of research challenges in developing computational tools to close
the perception-action loop. Given the recent advances in computer vision
and deep learning, we look for new potential solutions for performing
real-world robotic tasks in an effective and computationally efficient manner.
We focus on the two parallel themes in this workshop:
We're inviting submissions! If you're interested in (remotely) presenting a spotlight talk, please submit a short paper (or extended abstract) to CMT. We suggest extended abstracts of 2 pages in the RSS format. A maximum of 4 pages will be considered. References will not count towards the page limit. The review process is double-blind. Significant overlap with work submitted to other venues is acceptable, but it must be explicitly stated at the time of submission.
Important Dates:
For further information, please contact us at rssvlrr [AT] gmail [DOT] com
Visual Learning and Reasoning for Robotics
Full-day workshop at RSS 2021
Virtual Conference
July 13, 2021, Pacific Time (PT)
Welcome! This workshop includes three live events:
For the panel discussion, you can also post questions at this link.
- Invited Talks (25 min talk + 5 min Q&A)
- Spotlight Talks (4 min talk + 2 min Q&A)
- Panel Discussion (60 min)
For the panel discussion, you can also post questions at this link.
Schedule
| Time (PT) | Invited Speaker | Title |
|---|---|---|
| 10:15 - 10:30 | - |
| Video | |
| 10:30 - 11:00 |
Andrew Davison Imperial College London |
| Video | |
| 11:00 - 11:30 |
Raquel Urtasun University of Toronto / Waabi |
| Video | |
| 11:30 - 12:00 |
Spotlight Talks + Q&A |
ZePHyR: Zero-shot Pose Hypothesis Rating
Brian Okorn (Carnegie Mellon University); Qiao Gu (Carnegie Mellon University)*; Martial Hebert (Carnegie Mellon University); David Held (Carnegie Mellon University) ST-DETR: Spatio-Temporal Object Traces Attention Detection Transformer Eslam Bakr (Valeo)*; Ahmad ElSallab (Valeo Deep Learning Research) Lifelong Interactive 3D Object Recognition for Real-Time Robotic Manipulation Hamed Ayoobi (University of Groningen)*; S. Hamidreza Kasaei (University of Groningen); Ming Cao (University of Groningen); Rineke Verbrugge (University of Groningen); Bart Verheij (University of Groningen) Predicting Diverse and Plausible State Foresight For Robotic Pushing Tasks Lingzhi Zhang (University of Pennsylvania)*; Shenghao Zhou (University of Pennsylvania); Jianbo Shi (University of Pennsylvania) Learning by Watching: Physical Imitation of Manipulation Skills from Human Videos Haoyu Xiong (University of Toronto, Vector Institute)*; Quanzhou Li (University of Toronto, Vector Institute); Yun-Chun Chen (University of Toronto, Vector Institute); Homanga Bharadhwaj (University of Toronto, Vector Institute); Samarth Sinha (University of Toronto, Vector Institute); Animesh Garg (University of Toronto, Vector Institute, NVIDIA) |
| 12:00 - 12:30 |
Abhinav Gupta CMU / Facebook AI Research |
| Video | |
| 12:30 - 1:00 |
Shuran Song Columbia University |
| Video | |
| 1:00 - 2:30 | - | Break |
| 2:30 - 3:00 |
Saurabh Gupta UIUC |
| Video | |
| 3:00 - 3:30 |
Sergey Levine UC Berkeley / Google |
| Video | |
| 3:30 - 4:00 |
Spotlight Talks + Q&A |
3D Neural Scene Representations for Visuomotor Control
Yunzhu Li (MIT)*; Shuang Li (MIT); Vincent Sitzmann (MIT); Pulkit Agrawal (MIT); Antonio Torralba (MIT) Stabilizing Deep Q-Learning with ConvNets and Vision Transformers under Data Augmentation Nicklas A Hansen (UC San Diego)*; Hao Su (UC San Diego); Xiaolong Wang (UC San Diego) Learning Vision-Guided Quadrupedal Locomotion End-to-End with Cross-Modal Transformers Ruihan Yang (UC San Diego)*; Minghao Zhang (Tsinghua University); Nicklas A Hansen (UC San Diego); Huazhe Xu (UC Berkeley); Xiaolong Wang (UC San Diego) Interaction Prediction and Monte-Carlo Tree Search for Robot Manipulation in Clutter Baichuan Huang (Rutgers University)*; Abdeslam Boularias (Rutgers University); Jingjin Yu (Rutgers University) A Simple Method for Complex In-Hand Manipulation Tao Chen (MIT)*; Jie Xu (MIT); Pulkit Agrawal (MIT) |
| 4:00 - 5:00 | Invited Speakers |
| Video | |
Introduction
We focus on the two parallel themes in this workshop:
- How could a robot’s interaction with the physical world facilitate the development of its visual perception?
- How a deep understanding of the physical world through visual learning and reasoning could give rise to effective and robust robotic control?
Call for Papers
Important Dates:
- Paper Submission: June 20, 2021 (11:59 pm PST)
- Review Due: June 29, 2021 (11:59 pm PST)
- Author Notification: July 1, 2021 (11:59 pm PST)
- Camera-Ready Version: July 8, 2021 (11:59 pm PST)
- Conference Date: July 13, 2021
Organizers
Kuan Fang Stanford University |
David Held CMU |
Yuke Zhu UT Austin / NVIDIA |
Dinesh Jayaraman Univ. of Pennsylvania |
Animesh Garg Univ. of Toronto / NVIDIA |
Lin Sun Magic Leap |
Yu Xiang NVIDIA |
Greg Dudek McGill / Samsung |













