vecId stringlengths 12 23 | id stringlengths 2 13 | conference stringclasses 11
values | year float64 2.02k 2.03k | title stringlengths 6 189 | abstract stringlengths 10 4.74k | author stringlengths 0 7.45k | aff stringlengths 0 7.16k | status stringclasses 11
values | track stringclasses 4
values | keywords stringlengths 0 804 | github stringlengths 0 141 | site stringlengths 0 193 | gsCitation float64 -1 11.1k | arxiv stringlengths 0 12 | text stringlengths 58 4.82k | vector list |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
corl_2024_0gDbaEtVrd | 0gDbaEtVrd | corl | 2,024 | One Model to Drift Them All: Physics-Informed Conditional Diffusion Model for Driving at the Limits | Enabling autonomous vehicles to reliably operate at the limits of handling— where tire forces are saturated — would improve their safety, particularly in scenarios like emergency obstacle avoidance or adverse weather conditions.
However, unlocking this capability is challenging due to the task's dynamic nature and the ... | Franck Djeumou;Thomas Jonathan Lew;NAN DING;Michael Thompson;Makoto Suminaka;Marcus Greiff;John Subosits | ;Toyota Research Institute;Toyota Research Institute;Toyota Research Institute;;Mitsubishi Electric Research Labs;Toyota Motor Corporation | Poster | main | Diffusion Models;Learning for Control;Autonomous Drifting;Model Predictive Control | https://openreview.net/forum?id=0gDbaEtVrd | 8 | One Model to Drift Them All: Physics-Informed Conditional Diffusion Model for Driving at the Limits
Enabling autonomous vehicles to reliably operate at the limits of handling— where tire forces are saturated — would improve their safety, particularly in scenarios like emergency obstacle avoidance or adverse weather con... | [
-0.08241964876651764,
-0.032408084720373154,
-0.0076278685592114925,
-0.008603793568909168,
-0.056566838175058365,
-0.014160123653709888,
-0.012650201097130775,
0.038742393255233765,
0.013929951936006546,
0.025631848722696304,
-0.03769281134009361,
0.015881802886724472,
0.014012813568115234,... | ||
corl_2024_1IzW0aniyg | 1IzW0aniyg | corl | 2,024 | EscIRL: Evolving Self-Contrastive IRL for Trajectory Prediction in Autonomous Driving | While deep neural networks (DNN) and inverse reinforcement learning (IRL) have both been commonly used in autonomous driving to predict trajectories through learning from expert demonstrations, DNN-based methods suffer from data-scarcity, while IRL-based approaches often struggle with generalizability, making both hard... | Siyue Wang;Zhaorun Chen;Zhuokai Zhao;Chaoli Mao;Yiyang Zhou;Jiayu He;Albert Sibo Hu | CIDI Intelligent Driving(Chengdu) Technology Co., Ltd.;University of Chicago;University of Chicago;CIDI;Xi'an Jiaotong University;cidi-lab;CiDi | Poster | main | Reinforcement Learning;Trajectory Prediction;Autonomous Driving | https://github.com/SiyueWang-CiDi/EscIRL | https://openreview.net/forum?id=1IzW0aniyg | 2 | EscIRL: Evolving Self-Contrastive IRL for Trajectory Prediction in Autonomous Driving
While deep neural networks (DNN) and inverse reinforcement learning (IRL) have both been commonly used in autonomous driving to predict trajectories through learning from expert demonstrations, DNN-based methods suffer from data-scarc... | [
-0.10358572751283646,
-0.02554372139275074,
0.03456570580601692,
-0.00735588651150465,
-0.022332191467285156,
-0.00653908122330904,
-0.01786760985851288,
0.002717269118875265,
0.00391231058165431,
0.025933559983968735,
-0.022406447678804398,
-0.013597947545349598,
-0.05583976209163666,
0.0... | |
corl_2024_1TEZ1hiY5m | 1TEZ1hiY5m | corl | 2,024 | Learning Robotic Locomotion Affordances and Photorealistic Simulators from Human-Captured Data | Learning reliable affordance models which satisfy human preferences is often hindered by a lack of high-quality training data. Similarly, learning visuomotor policies in simulation can be challenging due to the high cost of photo-realistic rendering. We present PAWS: a comprehensive robot learning framework that uses a... | Alejandro Escontrela;Justin Kerr;Kyle Stachowicz;Pieter Abbeel | University of California, Berkeley;University of California, Berkeley;University of California, Berkeley;Covariant | Poster | main | Navigation;Dataset;Real2Sim | https://openreview.net/forum?id=1TEZ1hiY5m | 1 | Learning Robotic Locomotion Affordances and Photorealistic Simulators from Human-Captured Data
Learning reliable affordance models which satisfy human preferences is often hindered by a lack of high-quality training data. Similarly, learning visuomotor policies in simulation can be challenging due to the high cost of p... | [
-0.07279437780380249,
-0.03138977289199829,
-0.02232784405350685,
0.017638063058257103,
-0.03934932127594948,
-0.006511509418487549,
0.03912510722875595,
0.014443032443523407,
-0.003031542059034109,
0.05956583097577095,
-0.01861899346113205,
0.013807762414216995,
0.010808918625116348,
0.02... | ||
corl_2024_1jc2zA5Z6J | 1jc2zA5Z6J | corl | 2,024 | Get a Grip: Multi-Finger Grasp Evaluation at Scale Enables Robust Sim-to-Real Transfer | This work explores conditions under which multi-finger grasping algorithms can attain robust sim-to-real transfer. While numerous large datasets facilitate learning *generative* models for multi-finger grasping at scale, reliable real-world dexterous grasping remains challenging, with most methods degrading when deploy... | Tyler Ga Wei Lum;Albert H. Li;Preston Culbertson;Krishnan Srinivasan;Aaron Ames;Mac Schwager;Jeannette Bohg | Stanford University;;California Institute of Technology;Stanford University;California Institute of Technology;Stanford University;Stanford University | Poster | main | Multi-Fingered Grasping;Large-Scale Grasp Dataset;Sim-to-Real | https://github.com/tylerlum/get_a_grip | https://openreview.net/forum?id=1jc2zA5Z6J | 2 | Get a Grip: Multi-Finger Grasp Evaluation at Scale Enables Robust Sim-to-Real Transfer
This work explores conditions under which multi-finger grasping algorithms can attain robust sim-to-real transfer. While numerous large datasets facilitate learning *generative* models for multi-finger grasping at scale, reliable rea... | [
-0.03876669332385063,
-0.003993328660726547,
-0.028224874287843704,
-0.027979275211691856,
-0.03162546083331108,
-0.01708795130252838,
-0.016218911856412888,
0.02588224783539772,
0.012364912778139114,
0.01879769004881382,
-0.02739361859858036,
0.03145543113350868,
-0.012100422754883766,
0.... | |
corl_2024_1tCteNSbFH | 1tCteNSbFH | corl | 2,024 | Trajectory Improvement and Reward Learning from Comparative Language Feedback | Learning from human feedback has gained traction in fields like robotics and natural language processing in recent years. While prior works mostly rely on human feedback in the form of comparisons, language is a preferable modality that provides more informative insights into user preferences. In this work, we aim to i... | Zhaojing Yang;Miru Jun;Jeremy Tien;Stuart Russell;Anca Dragan;Erdem Biyik | University of Southern California;University of Southern California;University of California, Berkeley;University of California, Berkeley;University of California, Berkeley;University of Southern California | Poster | main | Learning from Human Language Feedback;Reward Learning;Human-Robot Interaction | https://github.com/USC-Lira/language-preference-learning | https://openreview.net/forum?id=1tCteNSbFH | 8 | Trajectory Improvement and Reward Learning from Comparative Language Feedback
Learning from human feedback has gained traction in fields like robotics and natural language processing in recent years. While prior works mostly rely on human feedback in the form of comparisons, language is a preferable modality that provi... | [
-0.07589001953601837,
-0.026844175532460213,
0.0035684595350176096,
-0.01956217736005783,
-0.017877396196126938,
0.01708180643618107,
0.01022100355476141,
-0.015986697748303413,
0.005335139576345682,
0.004511468578130007,
-0.023455893620848656,
-0.049532558768987656,
-0.041557926684617996,
... | |
corl_2024_2AZfKk9tRI | 2AZfKk9tRI | corl | 2,024 | Multi-agent Reinforcement Learning with Hybrid Action Space for Free Gait Motion Planning of Hexapod Robots | Legged robots are able to overcome challenging terrains through diverse gaits formed by contact sequences. However, environments characterized by discrete footholds present significant challenges. In this paper, we tackle the problem of free gait motion planning for hexapod robots walking in randomly generated plum blo... | Huiqiao Fu;Kaiqiang Tang;Peng Li;Guizhou Deng;Chunlin Chen | Nanjing University;Nanjing University;Institute of Software, Chinese Academy of Sciences;;Nanjing University | Poster | main | Free Gait;Hexapod Robot;Hybrid Action Space;Multi-agent Reinforcement Learning | https://openreview.net/forum?id=2AZfKk9tRI | 0 | Multi-agent Reinforcement Learning with Hybrid Action Space for Free Gait Motion Planning of Hexapod Robots
Legged robots are able to overcome challenging terrains through diverse gaits formed by contact sequences. However, environments characterized by discrete footholds present significant challenges. In this paper, ... | [
-0.06565209478139877,
-0.04211918264627457,
-0.027843035757541656,
0.016576699912548065,
-0.008565688505768776,
-0.013175884261727333,
0.01770424284040928,
-0.0015992468688637018,
0.023060070350766182,
0.014030634425580502,
-0.02924337051808834,
-0.018358943983912468,
-0.01213927287608385,
... | ||
corl_2024_2CScZqkUPZ | 2CScZqkUPZ | corl | 2,024 | Genetic Algorithm for Curriculum Design in Multi-Agent Reinforcement Learning | As the deployment of autonomous agents in real-world scenarios grows, so does the interest in their application to competitive environments with other robots. Self-play in Reinforcement Learning (RL) enables agents to develop competitive strategies. However, the complexity arising from multi-agent interactions and the ... | Yeeho Song;Jeff Schneider | Carnegie Mellon University;Carnegie Mellon University | Poster | main | Reinforcement Learning;Multiagent Learning;Curricular Learning | https://github.com/yeehos/GEnetic-Multiagent-Selfplay | https://openreview.net/forum?id=2CScZqkUPZ | 1 | Genetic Algorithm for Curriculum Design in Multi-Agent Reinforcement Learning
As the deployment of autonomous agents in real-world scenarios grows, so does the interest in their application to competitive environments with other robots. Self-play in Reinforcement Learning (RL) enables agents to develop competitive stra... | [
-0.062280625104904175,
-0.008072194643318653,
-0.02093593031167984,
-0.01941092126071453,
-0.023214031010866165,
-0.006650735158473253,
-0.013423847034573555,
0.016822170466184616,
0.014393451623618603,
0.01168232411146164,
-0.048197824507951736,
-0.017697637900710106,
-0.04812251403927803,
... | |
corl_2024_2LLu3gavF1 | 2LLu3gavF1 | corl | 2,024 | Robot See Robot Do: Imitating Articulated Object Manipulation with Monocular 4D Reconstruction | Humans can learn to manipulate new objects by simply watching others; providing robots with the ability to learn from such demonstrations would enable a natural interface specifying new behaviors. This work develops Robot See Robot Do (RSRD), a method for imitating articulated object manipulation from a single monocula... | Justin Kerr;Chung Min Kim;Mingxuan Wu;Brent Yi;Qianqian Wang;Ken Goldberg;Angjoo Kanazawa | University of California, Berkeley;University of California, Berkeley;Xi'an Jiaotong University;University of California, Berkeley;University of California, Berkeley;University of California, Berkeley;University of California, Berkeley | Poster | main | Feature Fields;Visual Imitation;Grasping;Articulated Objects | https://github.com/kerrj/rsrd | https://openreview.net/forum?id=2LLu3gavF1 | 15 | Robot See Robot Do: Imitating Articulated Object Manipulation with Monocular 4D Reconstruction
Humans can learn to manipulate new objects by simply watching others; providing robots with the ability to learn from such demonstrations would enable a natural interface specifying new behaviors. This work develops Robot See... | [
-0.03082842193543911,
-0.006290764082223177,
0.0055412063375115395,
-0.0056561692617833614,
-0.020251845940947533,
0.036438606679439545,
-0.003494868753477931,
-0.013547215610742569,
0.00957410130649805,
0.021888915449380875,
-0.03914253041148186,
-0.02190730907022953,
0.02856595441699028,
... | |
corl_2024_2SYFDG4WRA | 2SYFDG4WRA | corl | 2,024 | Manipulate-Anything: Automating Real-World Robots using Vision-Language Models | Large-scale endeavors like RT-1 and widespread community efforts such as Open-X-Embodiment have contributed to growing the scale of robot demonstration data. However, there is still an opportunity to improve the quality, quantity, and diversity of robot demonstration data. Although vision-language models have been show... | Jiafei Duan;Wentao Yuan;Wilbert Pumacay;Yi Ru Wang;Kiana Ehsani;Dieter Fox;Ranjay Krishna | NVIDIA;University of Washington, Seattle;Universidad Nacional de Ingeniería;University of Washington;Allen Institute for Artificial Intelligence;Department of Computer Science;University of Washington | Poster | main | Robot Learning; Multimodal Large Language Model; Data Generation; Imitation Learning; Behavior Cloning | https://github.com/Robot-MA/manipulate-anything/tree/main | https://openreview.net/forum?id=2SYFDG4WRA | 39 | Manipulate-Anything: Automating Real-World Robots using Vision-Language Models
Large-scale endeavors like RT-1 and widespread community efforts such as Open-X-Embodiment have contributed to growing the scale of robot demonstration data. However, there is still an opportunity to improve the quality, quantity, and divers... | [
-0.02514158934354782,
-0.008075951598584652,
0.00026492003235034645,
0.05216602981090546,
-0.018597763031721115,
0.007120681460946798,
0.0010187235893681645,
0.029239557683467865,
0.022501908242702484,
0.0035764926578849554,
-0.030531710013747215,
-0.016179591417312622,
0.011389398947358131,... | |
corl_2024_2sg4PY1W9d | 2sg4PY1W9d | corl | 2,024 | Learning Transparent Reward Models via Unsupervised Feature Selection | In complex real-world tasks such as robotic manipulation and autonomous driving, collecting expert demonstrations is often more straightforward than specifying precise learning objectives and task descriptions. Learning from expert data can be achieved through behavioral cloning or by learning a reward function, i.e., ... | Daulet Baimukashev;Gokhan Alcan;Kevin Sebastian Luck;Ville Kyrki | Aalto University;Aalto University;Vrije Universiteit Amsterdam;Aalto University | Poster | main | Inverse reinforcement learning;Reinforcement learning;Imitation learning;Robots;Reward learning;Robot learning | https://github.com/baimukashev/reward-learning | https://openreview.net/forum?id=2sg4PY1W9d | 0 | Learning Transparent Reward Models via Unsupervised Feature Selection
In complex real-world tasks such as robotic manipulation and autonomous driving, collecting expert demonstrations is often more straightforward than specifying precise learning objectives and task descriptions. Learning from expert data can be achiev... | [
-0.07196776568889618,
-0.015986589714884758,
-0.00032329271198250353,
0.008860094472765923,
-0.002367299050092697,
-0.027737576514482498,
0.006587674375623465,
0.020503317937254906,
0.009253667667508125,
0.016895558685064316,
-0.026144539937376976,
-0.011020064353942871,
-0.02031590044498443... | |
corl_2024_3NI5SxsJqf | 3NI5SxsJqf | corl | 2,024 | Accelerating Visual Sparse-Reward Learning with Latent Nearest-Demonstration-Guided Explorations | Recent progress in deep reinforcement learning (RL) and computer vision enables artificial agents to solve complex tasks, including locomotion, manipulation, and video games from high-dimensional pixel observations. However, RL usually relies on domain-specific reward functions for sufficient learning signals, requirin... | Ruihan Zhao;ufuk topcu;Sandeep P. Chinchali;Mariano Phielipp | University of Texas at Austin;University of Texas, Austin;Intel Labs;University of Texas at Austin | Poster | main | Computer Vision;Sparse Reward;RL from Demonstrations | https://openreview.net/forum?id=3NI5SxsJqf | 0 | Accelerating Visual Sparse-Reward Learning with Latent Nearest-Demonstration-Guided Explorations
Recent progress in deep reinforcement learning (RL) and computer vision enables artificial agents to solve complex tasks, including locomotion, manipulation, and video games from high-dimensional pixel observations. However... | [
-0.0834430605173111,
0.0011691030813381076,
0.000011807649570982903,
0.019890496507287025,
-0.040191490203142166,
0.024238044396042824,
0.04127371311187744,
0.014488715678453445,
0.0060081989504396915,
0.052879612892866135,
-0.026122603565454483,
-0.012697450816631317,
-0.021140649914741516,... | ||
corl_2024_3ZAgXBRvla | 3ZAgXBRvla | corl | 2,024 | FlowBotHD: History-Aware Diffuser Handling Ambiguities in Articulated Objects Manipulation | We introduce a novel approach to manipulate articulated objects with ambiguities, such as opening a door, in which multi-modality and occlusions create ambiguities about the opening side and direction. Multi-modality occurs when the method to open a fully closed door (push, pull, slide) is uncertain, or the side from w... | Yishu Li;Wen Hui Leng;Yiming Fang;Ben Eisner;David Held | Tsinghua University;Carnegie Mellon University;Carnegie Mellon University;Carnegie Mellon University;Carnegie Mellon University | Poster | main | Ambiguity;Multi-modality;Occlusion;Articulated Objects;Diffusion | https://openreview.net/forum?id=3ZAgXBRvla | 0 | FlowBotHD: History-Aware Diffuser Handling Ambiguities in Articulated Objects Manipulation
We introduce a novel approach to manipulate articulated objects with ambiguities, such as opening a door, in which multi-modality and occlusions create ambiguities about the opening side and direction. Multi-modality occurs when ... | [
-0.026064615696668625,
-0.057717181742191315,
-0.03862813487648964,
0.04605373740196228,
-0.02265183813869953,
-0.02737722173333168,
-0.018610885366797447,
0.016004424542188644,
0.03281516209244728,
0.024864517152309418,
-0.03626544401049614,
-0.010491476394236088,
0.004495677072554827,
0.... | ||
corl_2024_3bcujpPikC | 3bcujpPikC | corl | 2,024 | FREA: Feasibility-Guided Generation of Safety-Critical Scenarios with Reasonable Adversariality | Generating safety-critical scenarios, which are essential yet difficult to collect at scale, offers an effective method to evaluate the robustness of autonomous vehicles (AVs). Existing methods focus on optimizing adversariality while preserving the naturalness of scenarios, aiming to achieve a balance through data-dri... | Keyu Chen;Yuheng Lei;Hao Cheng;Haoran Wu;Wenchao Sun;Sifa Zheng | Tsinghua University;The University of Hong Kong;Tsinghua University;Tsinghua University;Tsinghua University;Tsinghua University | Poster | main | Feasibility;Scenario Generation;Autonomous Driving | https://github.com/CurryChen77/FREA | https://openreview.net/forum?id=3bcujpPikC | 3 | FREA: Feasibility-Guided Generation of Safety-Critical Scenarios with Reasonable Adversariality
Generating safety-critical scenarios, which are essential yet difficult to collect at scale, offers an effective method to evaluate the robustness of autonomous vehicles (AVs). Existing methods focus on optimizing adversaria... | [
-0.05088040232658386,
0.0014859603252261877,
0.009876783937215805,
0.00214184052310884,
-0.025515025481581688,
0.026768330484628677,
0.0055697206407785416,
0.030098002403974533,
0.028882110491394997,
0.03786100447177887,
-0.019660038873553276,
0.00851124431937933,
0.012776218354701996,
0.0... | |
corl_2024_3i7j8ZPnbm | 3i7j8ZPnbm | corl | 2,024 | UMI-on-Legs: Making Manipulation Policies Mobile with Manipulation-Centric Whole-body Controllers | We introduce UMI-on-Legs, a new framework that combines real-world and simulation data for quadruped manipulation systems. We scale task-centric data collection in the real world using a handheld gripper (UMI), providing a cheap way to demonstrate task-relevant manipulation skills without a robot. Simultaneously, we s... | Huy Ha;Yihuai Gao;Zipeng Fu;Jie Tan;Shuran Song | Columbia University;Stanford University;Stanford University;Google;Stanford University | Poster | main | Manipulation;Visuo-motor Policy;Whole-body Controller | https://github.com/real-stanford/umi-on-legs | https://openreview.net/forum?id=3i7j8ZPnbm | 45 | UMI-on-Legs: Making Manipulation Policies Mobile with Manipulation-Centric Whole-body Controllers
We introduce UMI-on-Legs, a new framework that combines real-world and simulation data for quadruped manipulation systems. We scale task-centric data collection in the real world using a handheld gripper (UMI), providing ... | [
-0.0619102381169796,
-0.04926946386694908,
-0.051376257091760635,
0.015542242676019669,
-0.02860807441174984,
-0.015015543438494205,
-0.004687159787863493,
0.02539243921637535,
0.007161259185522795,
0.016346151009202003,
-0.04269034415483475,
0.014331758953630924,
0.04878896474838257,
0.01... | |
corl_2024_3jNEz3kUSl | 3jNEz3kUSl | corl | 2,024 | PointPatchRL - Masked Reconstruction Improves Reinforcement Learning on Point Clouds | Perceiving the environment via cameras is crucial for Reinforcement Learning (RL) in robotics. While images are a convenient form of representation, they often complicate extracting important geometric details, especially with varying geometries or deformable objects. In contrast, point clouds naturally represent this ... | Balazs Gyenes;Nikolai Franke;Philipp Becker;Gerhard Neumann | Karlsruher Institut für Technologie;Karlsruher Institut für Technologie;FZI Forschungszentrum Informatik ;Karlsruhe Institute of Technology | Poster | main | Point Clouds;Self-Supervised Learning;Reinforcement Learning | https://github.com/balazsgyenes/pprl | https://openreview.net/forum?id=3jNEz3kUSl | 0 | PointPatchRL - Masked Reconstruction Improves Reinforcement Learning on Point Clouds
Perceiving the environment via cameras is crucial for Reinforcement Learning (RL) in robotics. While images are a convenient form of representation, they often complicate extracting important geometric details, especially with varying ... | [
-0.0005533811636269093,
0.003310643369331956,
-0.002264579990878701,
-0.02245519310235977,
-0.01668255589902401,
0.006612210068851709,
0.018842756748199463,
0.02692081779241562,
-0.008500116877257824,
0.01292489655315876,
-0.044982995837926865,
-0.029879746958613396,
-0.00106081273406744,
... | |
corl_2024_3wBqoPfoeJ | 3wBqoPfoeJ | corl | 2,024 | Twisting Lids Off with Two Hands | Manipulating objects with two multi-fingered hands has been a long-standing challenge in robotics, due to the contact-rich nature of many manipulation tasks and the complexity inherent in coordinating a high-dimensional bimanual system. In this work, we share novel insights into physical modeling, real-time perception,... | Toru Lin;Zhao-Heng Yin;Haozhi Qi;Pieter Abbeel;Jitendra Malik | ;;University of California, Berkeley;Covariant;University of California, Berkeley | Poster | main | Bimanual Manipulation;Sim-to-Real;Reinforcement Learning | https://openreview.net/forum?id=3wBqoPfoeJ | 28 | Twisting Lids Off with Two Hands
Manipulating objects with two multi-fingered hands has been a long-standing challenge in robotics, due to the contact-rich nature of many manipulation tasks and the complexity inherent in coordinating a high-dimensional bimanual system. In this work, we share novel insights into physica... | [
-0.07024779915809631,
-0.0027922471053898335,
0.011767160147428513,
0.0257587730884552,
-0.03630154952406883,
-0.006841589696705341,
0.003927838988602161,
0.026057858020067215,
-0.013225204311311245,
0.004425536375492811,
-0.03054414689540863,
-0.005902273114770651,
-0.021253790706396103,
... | ||
corl_2024_46SluHKoE9 | 46SluHKoE9 | corl | 2,024 | Continuously Improving Mobile Manipulation with Autonomous Real-World RL | We present a fully autonomous real-world RL framework for mobile manipulation that can learn policies without extensive instrumentation or human supervision. This is enabled by 1) task-relevant autonomy, which guides exploration towards object interactions and prevents stagnation near goal states, 2) efficient policy l... | Russell Mendonca;Emmanuel Panov;Bernadette Bucher;Jiuguang Wang;Deepak Pathak | Carnegie Mellon University;;Boston Dynamics AI Institute;;Carnegie Mellon University | Poster | main | Continual Learning;Mobile Manipulation;Reinforcement Learning | https://openreview.net/forum?id=46SluHKoE9 | 3 | Continuously Improving Mobile Manipulation with Autonomous Real-World RL
We present a fully autonomous real-world RL framework for mobile manipulation that can learn policies without extensive instrumentation or human supervision. This is enabled by 1) task-relevant autonomy, which guides exploration towards object int... | [
-0.07026921212673187,
-0.01020097080618143,
0.0038463440723717213,
0.005133121274411678,
-0.023292528465390205,
-0.022677112370729446,
0.009576231241226196,
-0.011739508248865604,
0.01568375900387764,
0.025232017040252686,
-0.05314948409795761,
-0.02355361357331276,
0.011170716024935246,
0... | ||
corl_2024_4Of4UWyBXE | 4Of4UWyBXE | corl | 2,024 | RP1M: A Large-Scale Motion Dataset for Piano Playing with Bi-Manual Dexterous Robot Hands | Endowing robot hands with human-level dexterity is a long-lasting research objective. Bi-manual robot piano playing constitutes a task that combines challenges from dynamic tasks, such as generating fast while precise motions, with slower but contact-rich manipulation problems. Although reinforcement learning based app... | Yi Zhao;Le Chen;Jan Schneider;Quankai Gao;Juho Kannala;Bernhard Schölkopf;Joni Pajarinen;Dieter Büchler | Max Planck Institute for Intelligent Systems;;Max Planck Institute for Intelligent Systems;University of Southern California;Aalto University;;Aalto University;Max Planck Institute for Intelligent Systems, Max-Planck Institute | Poster | main | Bi-manual dexterous robot hands;dataset for robot piano playing;imitation learning;robot learning at scale | https://openreview.net/forum?id=4Of4UWyBXE | 2 | RP1M: A Large-Scale Motion Dataset for Piano Playing with Bi-Manual Dexterous Robot Hands
Endowing robot hands with human-level dexterity is a long-lasting research objective. Bi-manual robot piano playing constitutes a task that combines challenges from dynamic tasks, such as generating fast while precise motions, wit... | [
-0.06796083599328995,
-0.019620096310973167,
-0.05255991220474243,
0.03242897242307663,
-0.01860787346959114,
0.0005445427959784865,
0.02771787904202938,
0.026185356080532074,
-0.0025683974381536245,
0.034075018018484116,
-0.02713135816156864,
0.014388701878488064,
0.00401105172932148,
0.0... | ||
corl_2024_55tYfHvanf | 55tYfHvanf | corl | 2,024 | Bimanual Dexterity for Complex Tasks | To train generalist robot policies, machine learning methods often require a substantial amount of expert human teleoperation data. An ideal robot for humans collecting data is one that closely mimics them: bimanual arms and dexterous hands. However, creating such a bimanual teleoperation system with over 50 DoF is a s... | Kenneth Shaw;Yulong Li;Jiahui Yang;Mohan Kumar Srirama;Ray Liu;Haoyu Xiong;Russell Mendonca;Deepak Pathak | Carnegie Mellon University;Carnegie Mellon University;;Carnegie Mellon University;;Carnegie Mellon University;Carnegie Mellon University;Carnegie Mellon University | Poster | main | Dexterous Manipulation;Bimanual;Behavior Cloning | https://openreview.net/forum?id=55tYfHvanf | 21 | Bimanual Dexterity for Complex Tasks
To train generalist robot policies, machine learning methods often require a substantial amount of expert human teleoperation data. An ideal robot for humans collecting data is one that closely mimics them: bimanual arms and dexterous hands. However, creating such a bimanual teleope... | [
-0.023834049701690674,
-0.04077768698334694,
-0.038669146597385406,
-0.0038899767678231,
0.007083381991833448,
0.02219616435468197,
0.032249391078948975,
0.015512841753661633,
0.02033236436545849,
0.03689947724342346,
-0.026827426627278328,
-0.012284137308597565,
0.013140732422471046,
-0.0... | ||
corl_2024_56IzghzjfZ | 56IzghzjfZ | corl | 2,024 | IMAGINATION POLICY: Using Generative Point Cloud Models for Learning Manipulation Policies | Humans can imagine goal states during planning and perform actions to match those goals. In this work, we propose IMAGINATION POLICY, a novel multi-task key-frame policy network for solving high-precision pick and place tasks. Instead of learning actions directly, IMAGINATION POLICY generates point clouds to imagine de... | Haojie Huang;Karl Schmeckpeper;Dian Wang;Ondrej Biza;Yaoyao Qian;Haotian Liu;Mingxi Jia;Robert Platt;Robin Walters | Northeastern University;The Robotics and AI Institute;Northeastern University;Northeastern University;Northeastern University;Northeastern University;Brown University;Northeastern University;Northeastern University | Poster | main | Manipulation policy learning;Generative model;Geometric learning | https://openreview.net/forum?id=56IzghzjfZ | 7 | IMAGINATION POLICY: Using Generative Point Cloud Models for Learning Manipulation Policies
Humans can imagine goal states during planning and perform actions to match those goals. In this work, we propose IMAGINATION POLICY, a novel multi-task key-frame policy network for solving high-precision pick and place tasks. In... | [
-0.011953826993703842,
-0.01901201345026493,
0.0062805903144180775,
0.01804577000439167,
-0.032465819269418716,
-0.008374120108783245,
0.013748281635344028,
0.025195980444550514,
0.0065704635344445705,
0.02639228291809559,
-0.017134739086031914,
-0.016831062734127045,
-0.018100982531905174,
... | ||
corl_2024_5Awumz1VKU | 5Awumz1VKU | corl | 2,024 | Learning Differentiable Tensegrity Dynamics using Graph Neural Networks | Tensegrity robots are composed of rigid struts and flexible cables. They constitute an emerging class of hybrid rigid-soft robotic systems and are promising systems for a wide array of applications, ranging from locomotion to assembly. They are difficult to control and model accurately, however, due to their compliance... | Nelson Chen;Kun Wang;William R. Johson III;Rebecca Kramer-Bottiglio;Kostas Bekris;Mridul Aanjaneya | Rutgers University;;;Yale University;Rutgers University;Rutgers University, New Brunswick | Poster | main | graph neural networks;differentiable simulation;tensegrity robots | https://github.com/nchen9191/tensegrity_gnn_simulator_public | https://openreview.net/forum?id=5Awumz1VKU | 0 | Learning Differentiable Tensegrity Dynamics using Graph Neural Networks
Tensegrity robots are composed of rigid struts and flexible cables. They constitute an emerging class of hybrid rigid-soft robotic systems and are promising systems for a wide array of applications, ranging from locomotion to assembly. They are dif... | [
-0.03626123070716858,
-0.021227890625596046,
-0.006162190344184637,
-0.008034423924982548,
-0.0035757345613092184,
-0.07000765949487686,
0.00605124281719327,
-0.008986719883978367,
0.02315097488462925,
0.02500009350478649,
-0.04323241114616394,
-0.010845084674656391,
-0.0029077401850372553,
... | |
corl_2024_5W0iZR9J7h | 5W0iZR9J7h | corl | 2,024 | DexGraspNet 2.0: Learning Generative Dexterous Grasping in Large-scale Synthetic Cluttered Scenes | Grasping in cluttered scenes remains highly challenging for dexterous hands due to the scarcity of data. To address this problem, we present a large-scale synthetic dataset, encompassing 1319 objects, 8270 scenes, and 426 million grasps. Beyond benchmarking, we also explore data-efficient learning strategies from grasp... | Jialiang Zhang;Haoran Liu;Danshi Li;XinQiang Yu;Haoran Geng;Yufei Ding;Jiayi Chen;He Wang | Peking University;Peking University;New York University;Institute of Computing Technology, Chinese Academy of Sciences;Peking University;Peking University;Peking University;Peking University | Poster | main | Dexterous Grasping;Synthetic Data;Generative Models | https://openreview.net/forum?id=5W0iZR9J7h | 10 | DexGraspNet 2.0: Learning Generative Dexterous Grasping in Large-scale Synthetic Cluttered Scenes
Grasping in cluttered scenes remains highly challenging for dexterous hands due to the scarcity of data. To address this problem, we present a large-scale synthetic dataset, encompassing 1319 objects, 8270 scenes, and 426 ... | [
-0.045684266835451126,
-0.009716851636767387,
-0.033613529056310654,
-0.017691826447844505,
-0.027606403455138206,
-0.058263447135686874,
-0.022032396867871284,
0.025290178135037422,
-0.014048006385564804,
0.004453557077795267,
-0.03088301606476307,
0.001333477790467441,
0.000898008467629551... | ||
corl_2024_5iXG6EgByK | 5iXG6EgByK | corl | 2,024 | Promptable Closed-loop Traffic Simulation | Simulation stands as a cornerstone for safe and efficient autonomous driving development. At its core a simulation system ought to produce realistic, reactive, and controllable traffic patterns. In this paper, we propose ProSim, a multimodal promptable closed-loop traffic simulation framework. ProSim allows the user to... | Shuhan Tan;Boris Ivanovic;Yuxiao Chen;Boyi Li;Xinshuo Weng;Yulong Cao;Philipp Kraehenbuehl;Marco Pavone | NVIDIA;NVIDIA;University of California, Berkeley;NVIDIA;NVIDIA;Apple;Stanford University;California Institute of Technology | Poster | main | Autonomous Driving;Scenario Generation;Traffic Simulation | https://openreview.net/forum?id=5iXG6EgByK | 5 | Promptable Closed-loop Traffic Simulation
Simulation stands as a cornerstone for safe and efficient autonomous driving development. At its core a simulation system ought to produce realistic, reactive, and controllable traffic patterns. In this paper, we propose ProSim, a multimodal promptable closed-loop traffic simul... | [
-0.06938907504081726,
-0.04614853858947754,
0.013532347977161407,
-0.028265517204999924,
-0.030242256820201874,
-0.0001450512936571613,
0.0027780570089817047,
-0.00005455660721054301,
0.004826383199542761,
0.041419144719839096,
-0.03353066369891167,
-0.010539527982473373,
0.00786076951771974... | ||
corl_2024_5lSkn5v4LK | 5lSkn5v4LK | corl | 2,024 | EquiGraspFlow: SE(3)-Equivariant 6-DoF Grasp Pose Generative Flows | Traditional methods for synthesizing 6-DoF grasp poses from 3D observations often rely on geometric heuristics, resulting in poor generalizability, limited grasp options, and higher failure rates. Recently, data-driven methods have been proposed that use generative models to learn the distribution of grasp poses and ge... | Byeongdo Lim;Jongmin Kim;Jihwan Kim;Yonghyeon Lee;Frank C. Park | Seoul National University;Seoul National University;Seoul National University;Korea Institute for Advanced Study;Seoul National University | Poster | main | 6-DoF grasp pose generation;equivariance;generative models;continuous normalizing flows | https://github.com/bdlim99/EquiGraspFlow | https://openreview.net/forum?id=5lSkn5v4LK | 7 | EquiGraspFlow: SE(3)-Equivariant 6-DoF Grasp Pose Generative Flows
Traditional methods for synthesizing 6-DoF grasp poses from 3D observations often rely on geometric heuristics, resulting in poor generalizability, limited grasp options, and higher failure rates. Recently, data-driven methods have been proposed that us... | [
-0.06999378651380539,
-0.02757769636809826,
-0.03995508700609207,
-0.03074442595243454,
-0.012594538740813732,
-0.035304516553878784,
-0.0001327482023043558,
0.05714590847492218,
-0.01826750859618187,
0.0082154031842947,
-0.01804131455719471,
0.003558047581464052,
-0.006953234784305096,
0.... | |
corl_2024_5u9l6U61S7 | 5u9l6U61S7 | corl | 2,024 | GenSim2: Scaling Robot Data Generation with Multi-modal and Reasoning LLMs | Robotic simulation today remains challenging to scale up due to the human efforts required to create diverse simulation tasks and scenes. Simulation-trained policies also face scalability issues as many sim-to-real methods focus on a single task. To address these challenges, this work proposes GenSim2, a scalable frame... | Pu Hua;Minghuan Liu;Annabella Macaluso;Yunfeng Lin;Weinan Zhang;Huazhe Xu;Lirui Wang | Electronic Engineering, Tsinghua University;Shanghai Jiaotong University;University of California, San Diego;;Shanghai Jiaotong University;Tsinghua University;Massachusetts Institute of Technology | Poster | main | Generative Simulation; Robotics; Learning | https://github.com/GenSim2/GenSim2 | https://openreview.net/forum?id=5u9l6U61S7 | 11 | GenSim2: Scaling Robot Data Generation with Multi-modal and Reasoning LLMs
Robotic simulation today remains challenging to scale up due to the human efforts required to create diverse simulation tasks and scenes. Simulation-trained policies also face scalability issues as many sim-to-real methods focus on a single task... | [
-0.069534070789814,
-0.012155462056398392,
0.0076470766216516495,
0.014059105888009071,
-0.047136079519987106,
-0.012573335319757462,
-0.04464741051197052,
0.023215169087052345,
0.007154915016144514,
0.017114222049713135,
-0.04624461382627487,
0.012025456875562668,
-0.002484023105353117,
0... | |
corl_2024_67tTQeO4HQ | 67tTQeO4HQ | corl | 2,024 | In-Flight Attitude Control of a Quadruped using Deep Reinforcement Learning | We present the development and real world demonstration of an in-flight attitude control law for a small low-cost quadruped with a five-bar-linkage leg design using only its legs as reaction masses. The control law is trained using deep reinforcement learning (DRL) and specifically through Proximal Policy Optimization ... | Tarek El-Agroudi;Finn Gross Maurer;Jørgen Anker Olsen;Kostas Alexis | Norwegian University of Science and Technology;Norwegian University of Science and Technology;Norwegian University of Science and Technology;Norwegian University of Science and Technology | Poster | main | Deep Reinforcement Learning;Legged Robotics | https://github.com/ntnu-arl/Eurepus-RL and https://github.com/ntnu-arl/Eurepus-design | https://openreview.net/forum?id=67tTQeO4HQ | 1 | In-Flight Attitude Control of a Quadruped using Deep Reinforcement Learning
We present the development and real world demonstration of an in-flight attitude control law for a small low-cost quadruped with a five-bar-linkage leg design using only its legs as reaction masses. The control law is trained using deep reinfor... | [
-0.05853114277124405,
-0.021699439734220505,
-0.004046677611768246,
-0.014392297714948654,
-0.04184495657682419,
0.0183048564940691,
0.018526844680309296,
0.030745498836040497,
0.004358849488198757,
0.03198494017124176,
-0.03962506353855133,
-0.012819874100387096,
0.0187210850417614,
0.018... | |
corl_2024_6FGlpzC9Po | 6FGlpzC9Po | corl | 2,024 | Steering Your Generalists: Improving Robotic Foundation Models via Value Guidance | Large, general-purpose robotic policies trained on diverse demonstration datasets have been shown to be remarkably effective both for controlling a variety of robots in a range of different scenes, and for acquiring broad repertoires of manipulation skills. However, the data that such policies are trained on is general... | Mitsuhiko Nakamoto;Oier Mees;Aviral Kumar;Sergey Levine | University of California, Berkeley;Electrical Engineering & Computer Science Department, University of California, Berkeley;Google DeepMind;Google | Poster | main | generalist policies;value functions;robot reinforcement learning | https://openreview.net/forum?id=6FGlpzC9Po | 7 | Steering Your Generalists: Improving Robotic Foundation Models via Value Guidance
Large, general-purpose robotic policies trained on diverse demonstration datasets have been shown to be remarkably effective both for controlling a variety of robots in a range of different scenes, and for acquiring broad repertoires of m... | [
-0.045459598302841187,
-0.0627002939581871,
-0.0339532271027565,
0.012251456268131733,
-0.033915501087903976,
-0.026332613080739975,
0.005192014388740063,
-0.005833352915942669,
0.021069863811135292,
0.02355976589024067,
-0.042252905666828156,
0.006276631262153387,
0.015533600933849812,
0.... | ||
corl_2024_6X3ybeVpDi | 6X3ybeVpDi | corl | 2,024 | Online Transfer and Adaptation of Tactile Skill: A Teleoperation Framework | This paper presents a teleoperation framework designed for online learning and adaptation of tactile skills, which provides an intuitive interface without need for physical access to execution robot. The proposed tele-teaching approach utilizes periodical Dynamical Movement Primitives (DMP) and Recursive Least Square (... | Xiao Chen;Tianle Ni;Kübra Karacan;Hamid Sadeghian;Sami Haddadin | Technische Universität München;Technische Universität München;Technische Universität München;; | Poster | main | Learning from Demonstration;Online Adaptation;Tactile Skill;Teleoperation;Autonomy Allocation | https://openreview.net/forum?id=6X3ybeVpDi | 1 | Online Transfer and Adaptation of Tactile Skill: A Teleoperation Framework
This paper presents a teleoperation framework designed for online learning and adaptation of tactile skills, which provides an intuitive interface without need for physical access to execution robot. The proposed tele-teaching approach utilizes ... | [
-0.07193850725889206,
0.006770683452486992,
-0.0006140531622804701,
0.032565515488386154,
-0.0016294256784021854,
0.013458573259413242,
0.05030176043510437,
-0.010855171829462051,
0.03252871707081795,
0.023586999624967575,
-0.02794746682047844,
0.012768625281751156,
0.013237789273262024,
0... | ||
corl_2024_6oESa4g05O | 6oESa4g05O | corl | 2,024 | Distribution Discrepancy and Feature Heterogeneity for Active 3D Object Detection | LiDAR-based 3D object detection is a critical technology for the development of autonomous driving and robotics. However, the high cost of data annotation limits its advancement. We propose a novel and effective active learning (AL) method called Distribution Discrepancy and Feature Heterogeneity (DDFH), which simultan... | Huang-Yu Chen;Jia-Fong Yeh;Jiawei;Pin-Hsuan Peng;Winston H. Hsu | National Taiwan University;Sony Group Corporation;National Taiwan University;National Taiwan University;National Taiwan University | Poster | main | Active Learning;LiDAR 3D Object Detection;Autonomous Driving | https://github.com/Coolshanlan/DDFH-active-3Ddet | https://openreview.net/forum?id=6oESa4g05O | 0 | Distribution Discrepancy and Feature Heterogeneity for Active 3D Object Detection
LiDAR-based 3D object detection is a critical technology for the development of autonomous driving and robotics. However, the high cost of data annotation limits its advancement. We propose a novel and effective active learning (AL) metho... | [
-0.0291119571775198,
-0.07518510520458221,
-0.021860692650079727,
0.013068311847746372,
-0.01922386884689331,
-0.0023606696631759405,
0.05726183205842972,
0.008413783274590969,
0.023838309571146965,
0.01931295171380043,
-0.05665607377886772,
-0.05131116136908531,
0.0058393171057105064,
0.0... | |
corl_2024_7E3JAys1xO | 7E3JAys1xO | corl | 2,024 | D$^3$RoMa: Disparity Diffusion-based Depth Sensing for Material-Agnostic Robotic Manipulation | Depth sensing is an important problem for 3D vision-based robotics. Yet, a real-world active stereo or ToF depth camera often produces noisy and incomplete depth which bottlenecks robot performances. In this work, we propose D3RoMa, a learning-based depth estimation framework on stereo image pairs that predicts clean a... | Songlin Wei;Haoran Geng;Jiayi Chen;Congyue Deng;Cui Wenbo;Chengyang Zhao;Xiaomeng Fang;Leonidas Guibas;He Wang | Peking University;Peking University;Stanford University;University of Chinese Academy of Sciences;Peking University;;Stanford University;Peking University;Peking University | Poster | main | Depth Estimation;Diffusion Model;Stereo Vision | https://openreview.net/forum?id=7E3JAys1xO | 4 | D$^3$RoMa: Disparity Diffusion-based Depth Sensing for Material-Agnostic Robotic Manipulation
Depth sensing is an important problem for 3D vision-based robotics. Yet, a real-world active stereo or ToF depth camera often produces noisy and incomplete depth which bottlenecks robot performances. In this work, we propose D... | [
-0.06711392849683762,
-0.006707737687975168,
-0.03746829181909561,
-0.0024857011158019304,
-0.015581325627863407,
0.008837033063173294,
-0.00551514932885766,
0.0184599868953228,
0.029024217277765274,
0.03712102398276329,
-0.0352201908826828,
0.012547307647764683,
0.007027589250355959,
-0.0... | ||
corl_2024_7c5rAY8oU3 | 7c5rAY8oU3 | corl | 2,024 | Automated Creation of Digital Cousins for Robust Policy Learning | Training robot policies in the real world can be unsafe, costly, and difficult to scale. Simulation serves as an inexpensive and potentially limitless source of training data, but suffers from the semantics and physics disparity between simulated and real-world environments. These discrepancies can be minimized by trai... | Tianyuan Dai;Josiah Wong;Yunfan Jiang;Chen Wang;Cem Gokmen;Ruohan Zhang;Jiajun Wu;Li Fei-Fei | ;Stanford University;Stanford University;Computer Science Department, Stanford University;Stanford University;Stanford University;Stanford University;Stanford University | Poster | main | Real-to-Sim; Digital Twin; Sim-to-Real Transfer | https://github.com/cremebrule/digital-cousins | https://openreview.net/forum?id=7c5rAY8oU3 | 11 | Automated Creation of Digital Cousins for Robust Policy Learning
Training robot policies in the real world can be unsafe, costly, and difficult to scale. Simulation serves as an inexpensive and potentially limitless source of training data, but suffers from the semantics and physics disparity between simulated and real... | [
-0.07176607847213745,
-0.018415160477161407,
-0.022469528019428253,
0.024761950597167015,
-0.026542840525507927,
0.006422571372240782,
-0.005394770763814449,
0.049561791121959686,
0.028854208067059517,
0.059148285537958145,
-0.015933282673358917,
0.014057664200663567,
0.0009384011500515044,
... | |
corl_2024_7ddT4eklmQ | 7ddT4eklmQ | corl | 2,024 | ACE: A Cross-platform and visual-Exoskeletons System for Low-Cost Dexterous Teleoperation | Bimanual robotic manipulation with dexterous hands has a large potential workability and a wide workspace as it follows the most natural human workflow.
Learning from human demonstrations has proven highly effective for learning a dexterous manipulation policy. To collect such data, teleoperation serves as a straightfo... | Shiqi Yang;Minghuan Liu;Yuzhe Qin;Runyu Ding;Jialong Li;Xuxin Cheng;Ruihan Yang;Sha Yi;Xiaolong Wang | University of California, San Diego;Shanghai Jiaotong University;University of California, San Diego;Electrical and Electronic Engineering, University of Hong Kong;University of California, San Diego;University of California, San Diego;University of California, San Diego;Carnegie Mellon University;University of Califor... | Poster | main | Teleopration System; Hardware; Imitation Learning; Robot Learning; Exoskeletons | https://github.com/ACETeleop/ACETeleop | https://openreview.net/forum?id=7ddT4eklmQ | 37 | ACE: A Cross-platform and visual-Exoskeletons System for Low-Cost Dexterous Teleoperation
Bimanual robotic manipulation with dexterous hands has a large potential workability and a wide workspace as it follows the most natural human workflow.
Learning from human demonstrations has proven highly effective for learning a... | [
-0.02170124277472496,
-0.04018063098192215,
-0.0462910458445549,
0.025571174919605255,
0.0192570760846138,
0.010989494621753693,
0.004795752465724945,
0.018525676801800728,
-0.0032033405732363462,
0.04580961912870407,
-0.0028561577200889587,
-0.002186094643548131,
0.0290707778185606,
0.005... | |
corl_2024_7vzDBvviRO | 7vzDBvviRO | corl | 2,024 | UBSoft: A Simulation Platform for Robotic Skill Learning in Unbounded Soft Environments | It is desired to equip robots with the capability of interacting with various soft materials as they are ubiquitous in the real world. While physics simulations are one of the predominant methods for data collection and robot training, simulating soft materials presents considerable challenges. Specifically, it is sign... | Chunru Lin;Jugang Fan;Yian Wang;Zeyuan Yang;Zhehuan Chen;Lixing Fang;Tsun-Hsuan Wang;Zhou Xian;Chuang Gan | University of Massachusetts at Amherst;South China University of Technology;, Tsinghua University;University of Massachusetts at Amherst;Tsinghua University;Liquid AI;Carnegie Mellon University;University of Massachusetts at Amherst;NVIDIA | Poster | main | Soft-Body Manipulation;Locomotion;Physics Simulation | https://openreview.net/forum?id=7vzDBvviRO | 1 | UBSoft: A Simulation Platform for Robotic Skill Learning in Unbounded Soft Environments
It is desired to equip robots with the capability of interacting with various soft materials as they are ubiquitous in the real world. While physics simulations are one of the predominant methods for data collection and robot traini... | [
-0.05729326978325844,
-0.009101850911974907,
-0.04483179748058319,
-0.03314456343650818,
-0.01432331744581461,
0.0058159674517810345,
-0.015245024114847183,
0.010120335966348648,
0.043652016669511795,
0.024941375479102135,
-0.007889806292951107,
0.024554258212447166,
0.03386349231004715,
0... | ||
corl_2024_7wMlwhCvjS | 7wMlwhCvjS | corl | 2,024 | GenDP: 3D Semantic Fields for Category-Level Generalizable Diffusion Policy | Diffusion-based policies have shown remarkable capability in executing complex robotic manipulation tasks but lack explicit characterization of geometry and semantics, which often limits their ability to generalize to unseen objects and layouts. To enhance the generalization capabilities of Diffusion Policy, we introdu... | Yixuan Wang;Guang Yin;Binghao Huang;Tarik Kelestemur;Jiuguang Wang;Yunzhu Li | University of Illinois, Urbana Champaign;University of Illinois, Urbana Champaign;University of Illinois Urbana-Champaign;Boston Dynamics AI Institute;;University of Illinois Urbana-Champaign | Poster | main | Semantic Fields;Category-Level Generalization;Imitation Learning;Diffusion Models | https://github.com/WangYixuan12/gendp | https://openreview.net/forum?id=7wMlwhCvjS | 14 | GenDP: 3D Semantic Fields for Category-Level Generalizable Diffusion Policy
Diffusion-based policies have shown remarkable capability in executing complex robotic manipulation tasks but lack explicit characterization of geometry and semantics, which often limits their ability to generalize to unseen objects and layouts... | [
-0.07962880283594131,
-0.012606070376932621,
0.0007941459771245718,
-0.009352008812129498,
-0.024683469906449318,
-0.015304116532206535,
-0.017336765304207802,
0.0007799037848599255,
0.023042766377329826,
0.0060478150844573975,
-0.027983106672763824,
0.026925764977931976,
0.00261601037345826... | |
corl_2024_7yMZAUkXa4 | 7yMZAUkXa4 | corl | 2,024 | MimicTouch: Leveraging Multi-modal Human Tactile Demonstrations for Contact-rich Manipulation | Tactile sensing is critical to fine-grained, contact-rich manipulation tasks, such as insertion and assembly. Prior research has shown the possibility of learning tactile-guided policy from teleoperated demonstration data. However, to provide the demonstration, human users often rely on visual feedback to control the r... | Kelin Yu;Yunhai Han;Qixian Wang;Vaibhav Saxena;Danfei Xu;Ye Zhao | Georgia Institute of Technology;Georgia Institute of Technology;;Georgia Institute of Technology;NVIDIA;Georgia Institute of Technology | Poster | main | Tactile Sensing;Learning from Human;Data Collection;Imitation Learning;Reinforcement Learning | https://openreview.net/forum?id=7yMZAUkXa4 | 16 | MimicTouch: Leveraging Multi-modal Human Tactile Demonstrations for Contact-rich Manipulation
Tactile sensing is critical to fine-grained, contact-rich manipulation tasks, such as insertion and assembly. Prior research has shown the possibility of learning tactile-guided policy from teleoperated demonstration data. How... | [
-0.041096411645412445,
-0.03896057605743408,
-0.010946154594421387,
0.016828907653689384,
-0.02021678350865841,
-0.03437589481472969,
0.008999045938253403,
0.002058733254671097,
-0.011664236895740032,
0.044263340532779694,
-0.0010356958955526352,
-0.013431824743747711,
0.018891092389822006,
... | ||
corl_2024_82bpTugrMt | 82bpTugrMt | corl | 2,024 | Monocular Event-Based Vision for Obstacle Avoidance with a Quadrotor | We present the first static-obstacle avoidance method for quadrotors using just an onboard, monocular event camera. Quadrotors are capable of fast and agile flight in cluttered environments when piloted manually, but vision-based autonomous flight in unknown environments is difficult in part due to the sensor limitatio... | Anish Bhattacharya;Marco Cannici;Nishanth Rao;Yuezhan Tao;Vijay Kumar;Nikolai Matni;Davide Scaramuzza | University of Pennsylvania;University of Zurich;;University of Pennsylvania;University of Pennsylvania;School of Engineering and Applied Science, University of Pennsylvania; | Poster | main | event-based vision;learning for control;simulation-to-real transfer;aerial robotics | https://openreview.net/forum?id=82bpTugrMt | 5 | Monocular Event-Based Vision for Obstacle Avoidance with a Quadrotor
We present the first static-obstacle avoidance method for quadrotors using just an onboard, monocular event camera. Quadrotors are capable of fast and agile flight in cluttered environments when piloted manually, but vision-based autonomous flight in ... | [
-0.003234506817534566,
-0.03464537113904953,
-0.016393175348639488,
-0.025631941854953766,
0.0035584268625825644,
0.01089122798293829,
-0.005356418434530497,
0.038194410502910614,
0.0021899817511439323,
0.03554671257734299,
-0.028166968375444412,
-0.017942357808351517,
0.010177664458751678,
... | ||
corl_2024_8Ar8b00GJC | 8Ar8b00GJC | corl | 2,024 | Autonomous Improvement of Instruction Following Skills via Foundation Models | Intelligent robots capable of improving from autonomously collected experience have the potential to transform robot learning: instead of collecting costly teleoperated demonstration data, large-scale deployment of fleets of robots can quickly collect larger quantities of autonomous data useful for training better robo... | Zhiyuan Zhou;Pranav Atreya;Abraham Lee;Homer Rich Walke;Oier Mees;Sergey Levine | University of California, Berkeley;University of California, Berkeley;University of California, Berkeley;University of California, Berkeley;Electrical Engineering & Computer Science Department, University of California, Berkeley;Google | Poster | main | Autonomous Improvement;Instruction Following Skills;Scaled Data Collection | https://github.com/rail-berkeley/soar | https://openreview.net/forum?id=8Ar8b00GJC | 12 | Autonomous Improvement of Instruction Following Skills via Foundation Models
Intelligent robots capable of improving from autonomously collected experience have the potential to transform robot learning: instead of collecting costly teleoperated demonstration data, large-scale deployment of fleets of robots can quickly... | [
-0.029551200568675995,
-0.03472913056612015,
0.006278242915868759,
0.011557884514331818,
-0.02209867537021637,
-0.032454539090394974,
0.006698950193822384,
0.0031067593954503536,
0.009819579310715199,
0.01484032440930605,
-0.007692927960306406,
-0.03596813604235649,
0.004114606883376837,
0... | |
corl_2024_8JLmTZsxGh | 8JLmTZsxGh | corl | 2,024 | Learning Performance-oriented Control Barrier Functions Under Complex Safety Constraints and Limited Actuation | Control Barrier Functions (CBFs) offer an elegant framework for constraining nonlinear control system dynamics to an invariant subset of a pre-specified safe set. However, finding a CBF that simultaneously promotes performance by maximizing the resulting control invariant set while accommodating complex safety constrai... | Lakshmideepakreddy Manda;Shaoru Chen;Mahyar Fazlyab | Whiting School of Engineering;Microsoft Research;Johns Hopkins University | Poster | main | Control Barrier Functions;Safety;Hamilton-Jacobi Partial Differential Equation | https://github.com/o4lc/PINN-CBF | https://openreview.net/forum?id=8JLmTZsxGh | 8 | Learning Performance-oriented Control Barrier Functions Under Complex Safety Constraints and Limited Actuation
Control Barrier Functions (CBFs) offer an elegant framework for constraining nonlinear control system dynamics to an invariant subset of a pre-specified safe set. However, finding a CBF that simultaneously pro... | [
-0.01760326512157917,
-0.01763986237347126,
-0.08710139989852905,
-0.0031725221779197454,
0.03304729610681534,
-0.00009006348409457132,
0.025233788415789604,
0.029332676902413368,
0.012662640772759914,
0.026880662888288498,
-0.04464861750602722,
-0.014153976924717426,
0.020878717303276062,
... | |
corl_2024_8LPXeGhhbH | 8LPXeGhhbH | corl | 2,024 | RAM: Retrieval-Based Affordance Transfer for Generalizable Zero-Shot Robotic Manipulation | This work proposes a retrieve-and-transfer framework for zero-shot robotic manipulation, dubbed RAM, featuring generalizability across various objects, environments, and embodiments. Unlike existing approaches that learn manipulation from expensive in-domain demonstrations, RAM capitalizes on a retrieval-based affordan... | Yuxuan Kuang;Junjie Ye;Haoran Geng;Jiageng Mao;Congyue Deng;Leonidas Guibas;He Wang;Yue Wang | Peking University;University of Southern California;Peking University;;Stanford University;Stanford University;Peking University;NVIDIA | Poster | main | Hierarchical Retrieval;Affordance Transfer;Zero-Shot Robotic Manipulation;Visual Foundation Models | https://github.com/yxKryptonite/RAM_code | https://openreview.net/forum?id=8LPXeGhhbH | 26 | RAM: Retrieval-Based Affordance Transfer for Generalizable Zero-Shot Robotic Manipulation
This work proposes a retrieve-and-transfer framework for zero-shot robotic manipulation, dubbed RAM, featuring generalizability across various objects, environments, and embodiments. Unlike existing approaches that learn manipulat... | [
-0.007961239665746689,
0.00681333988904953,
-0.009581259451806545,
0.025864772498607635,
-0.015552189201116562,
-0.00617458950728178,
0.008947137743234634,
0.035140540450811386,
0.012571352533996105,
0.029882419854402542,
-0.04487917572259903,
0.011775229126214981,
0.012238091789186,
0.045... | |
corl_2024_8PcRynpd1m | 8PcRynpd1m | corl | 2,024 | Safe Bayesian Optimization for the Control of High-Dimensional Embodied Systems | Learning to move is a primary goal for animals and robots, where ensuring safety is often important when optimizing control policies on the embodied systems. For complex tasks such as the control of human or humanoid control, the high-dimensional parameter space adds complexity to the safe optimization effort. Current ... | Yunyue Wei;Zeji Yi;Hongda Li;Saraswati Soedarmadji;Yanan Sui | Tsinghua University;Tsinghua University;Tsinghua University;California Institute of Technology;Tsinghua University | Poster | main | Safe Bayesian Optimization;High-dimensional Embodied System | https://openreview.net/forum?id=8PcRynpd1m | 0 | Safe Bayesian Optimization for the Control of High-Dimensional Embodied Systems
Learning to move is a primary goal for animals and robots, where ensuring safety is often important when optimizing control policies on the embodied systems. For complex tasks such as the control of human or humanoid control, the high-dimen... | [
-0.020048297941684723,
-0.027521364390850067,
-0.015562563203275204,
0.028223149478435516,
-0.01579965278506279,
-0.014054672792553902,
0.02321581542491913,
-0.014064156450331211,
0.012025185860693455,
0.05344948545098305,
-0.06259166449308395,
0.005159070249646902,
0.05618075653910637,
0.... | ||
corl_2024_8XFT1PatHy | 8XFT1PatHy | corl | 2,024 | Splat-MOVER: Multi-Stage, Open-Vocabulary Robotic Manipulation via Editable Gaussian Splatting | We present Splat-MOVER, a modular robotics stack for open-vocabulary
robotic manipulation, which leverages the editability of Gaussian Splatting (GSplat)
scene representations to enable multi-stage manipulation tasks. Splat-MOVER
consists of: (i) ASK-Splat, a GSplat representation that distills semantic and grasp
affor... | Olaolu Shorinwa;Johnathan Tucker;Aliyah Smith;Aiden Swann;Timothy Chen;Roya Firoozi;Monroe David Kennedy;Mac Schwager | Stanford University;;;Stanford University;Stanford University;;Stanford University;Stanford University | Poster | main | Gaussian Splatting;Robotic Grasping;Robotic Manipulation;Scene Editing | https://splatmover.github.io | https://openreview.net/forum?id=8XFT1PatHy | 23 | Splat-MOVER: Multi-Stage, Open-Vocabulary Robotic Manipulation via Editable Gaussian Splatting
We present Splat-MOVER, a modular robotics stack for open-vocabulary
robotic manipulation, which leverages the editability of Gaussian Splatting (GSplat)
scene representations to enable multi-stage manipulation tasks. Splat-M... | [
-0.04607889801263809,
-0.023553885519504547,
-0.005222458858042955,
0.01498299092054367,
-0.00491012167185545,
0.021147053688764572,
-0.01934652216732502,
0.0019934452138841152,
0.014523671939969063,
0.04982694238424301,
-0.016259897500276566,
0.039868902415037155,
0.022690366953611374,
0.... | |
corl_2024_8Yu0TNJNGK | 8Yu0TNJNGK | corl | 2,024 | AnyRotate: Gravity-Invariant In-Hand Object Rotation with Sim-to-Real Touch | Human hands are capable of in-hand manipulation in the presence of different hand motions. For a robot hand, harnessing rich tactile information to achieve this level of dexterity still remains a significant challenge. In this paper, we present AnyRotate, a system for gravity-invariant multi-axis in-hand object rotatio... | Max Yang;chenghua lu;Alex Church;Yijiong Lin;Christopher J. Ford;Haoran Li;Efi Psomopoulou;David A.W. Barton;Nathan F. Lepora | University of Bristol;University of Bristol;;University of Bristol;University of Bristol;University of Bristol;University of Bristol;University of Bristol;University of Bristol | Poster | main | Tactile Sensing;In-hand Object Rotation;Reinforcement Learning | https://openreview.net/forum?id=8Yu0TNJNGK | 19 | AnyRotate: Gravity-Invariant In-Hand Object Rotation with Sim-to-Real Touch
Human hands are capable of in-hand manipulation in the presence of different hand motions. For a robot hand, harnessing rich tactile information to achieve this level of dexterity still remains a significant challenge. In this paper, we present... | [
-0.06469301134347916,
-0.003942988812923431,
-0.005412860307842493,
0.031039955094456673,
-0.03313043713569641,
-0.046998560428619385,
0.026858987286686897,
0.024600517004728317,
0.022136732935905457,
0.0015596970915794373,
-0.02908012643456459,
0.03830065205693245,
-0.0013998777139931917,
... | ||
corl_2024_928V4Umlys | 928V4Umlys | corl | 2,024 | DriveVLM: The Convergence of Autonomous Driving and Large Vision-Language Models | A primary hurdle of autonomous driving in urban environments is understanding complex and long-tail scenarios, such as challenging road conditions and delicate human behaviors. We introduce DriveVLM, an autonomous driving system leveraging Vision-Language Models (VLMs) for enhanced scene understanding and planning capa... | Xiaoyu Tian;Junru Gu;Bailin Li;Yicheng Liu;Yang Wang;Zhiyong Zhao;Kun Zhan;Peng Jia;XianPeng Lang;Hang Zhao | IIIS, Tsinghua University;Tsinghua University;Li Auto;Tsinghua University;LiAuto;LiAuto;Li Auto;LiAuto;Tsinghua University;Li Auto | Poster | main | Autonomous Driving;Vision Language Model;Dual System | https://openreview.net/forum?id=928V4Umlys | 190 | DriveVLM: The Convergence of Autonomous Driving and Large Vision-Language Models
A primary hurdle of autonomous driving in urban environments is understanding complex and long-tail scenarios, such as challenging road conditions and delicate human behaviors. We introduce DriveVLM, an autonomous driving system leveraging... | [
-0.06513936072587967,
-0.01409372128546238,
0.026219148188829422,
-0.024306561797857285,
-0.05422090366482735,
0.048353161662817,
0.01169834565371275,
0.021669792011380196,
0.024380836635828018,
0.022876763716340065,
-0.014316546730697155,
-0.03386949375271797,
0.030601385980844498,
0.0106... | ||
corl_2024_97QXO0uBEO | 97QXO0uBEO | corl | 2,024 | Handling Long-Term Safety and Uncertainty in Safe Reinforcement Learning | Safety is one of the key issues preventing the deployment of reinforcement learning techniques in real-world robots. While most approaches in the Safe Reinforcement Learning area do not require prior knowledge of constraints and robot kinematics and rely solely on data, it is often difficult to deploy them in complex r... | Jonas Günster;Puze Liu;Jan Peters;Davide Tateo | Technische Universität Darmstadt;TU Darmstadt;TU Darmstadt;Technische Universität Darmstadt | Poster | main | Safe Reinforcement Learning;Chance Constraint;Distributional RL | https://github.com/cube1324/d-atacom | https://openreview.net/forum?id=97QXO0uBEO | 2 | Handling Long-Term Safety and Uncertainty in Safe Reinforcement Learning
Safety is one of the key issues preventing the deployment of reinforcement learning techniques in real-world robots. While most approaches in the Safe Reinforcement Learning area do not require prior knowledge of constraints and robot kinematics a... | [
-0.042707160115242004,
-0.015853989869356155,
-0.015133354812860489,
-0.00014289747923612595,
-0.048699818551540375,
0.010515594854950905,
0.0022626544814556837,
0.021884575486183167,
-0.024956759065389633,
0.017314225435256958,
-0.013910169713199139,
0.0016012814594432712,
-0.00760934641584... | |
corl_2024_9HkElMlPbU | 9HkElMlPbU | corl | 2,024 | Contrastive Imitation Learning for Language-guided Multi-Task Robotic Manipulation | Developing robots capable of executing various manipulation tasks, guided by natural language instructions and visual observations of intricate real-world environments, remains a significant challenge in robotics. Such robot agents need to understand linguistic commands and distinguish between the requirements of diff... | Teli Ma;Jiaming Zhou;Zifan Wang;Ronghe Qiu;Junwei Liang | Hong Kong University of Science and Technology;Hong Kong University of Science and Technology (Guangzhou);the Hong Kong University of Science and Technology(Guangzhou);Hong Kong University of Science and Technology;Hong Kong University of Science and Technology | Poster | main | Contrastive imitation learning;Multi-task learning;Robotic manipulation | https://openreview.net/forum?id=9HkElMlPbU | 11 | Contrastive Imitation Learning for Language-guided Multi-Task Robotic Manipulation
Developing robots capable of executing various manipulation tasks, guided by natural language instructions and visual observations of intricate real-world environments, remains a significant challenge in robotics. Such robot agents need ... | [
-0.02258148230612278,
-0.01576460525393486,
0.0237437654286623,
0.0025344209279865026,
0.005428595934063196,
-0.01434403844177723,
0.002077810000628233,
0.0013306286418810487,
-0.00007184990681707859,
0.004206354729831219,
-0.03955449163913727,
-0.04475708678364754,
-0.02876187115907669,
0... | ||
corl_2024_9XV3dBqcfe | 9XV3dBqcfe | corl | 2,024 | Generalized Animal Imitator: Agile Locomotion with Versatile Motion Prior | The agility of animals, particularly in complex activities such as running, turning, jumping, and backflipping, stands as an exemplar for robotic system design. Transferring this suite of behaviors to legged robotic systems introduces essential inquiries: How can a robot be trained to learn multiple locomotion behavior... | Ruihan Yang;Zhuoqun Chen;Jianhan Ma;Chongyi Zheng;Yiyu Chen;Quan Nguyen;Xiaolong Wang | University of California, San Diego;University of California, San Diego;University of California, San Diego;Princeton University;University of Southern California;University of Southern California;University of California, San Diego | Poster | main | Legged Robots;Imitation Learning;Agile Locomotion | https://openreview.net/forum?id=9XV3dBqcfe | 19 | Generalized Animal Imitator: Agile Locomotion with Versatile Motion Prior
The agility of animals, particularly in complex activities such as running, turning, jumping, and backflipping, stands as an exemplar for robotic system design. Transferring this suite of behaviors to legged robotic systems introduces essential i... | [
-0.08329824358224869,
-0.0342290885746479,
0.0021859263069927692,
0.039374638348817825,
-0.014923958107829094,
-0.009880946949124336,
-0.007275545969605446,
0.01154952123761177,
-0.01415026094764471,
-0.005989158526062965,
-0.025765033438801765,
-0.02830052189528942,
-0.008380161598324776,
... | ||
corl_2024_9aZ4ehSTRc | 9aZ4ehSTRc | corl | 2,024 | Guided Reinforcement Learning for Robust Multi-Contact Loco-Manipulation | Reinforcement learning (RL) has shown remarkable proficiency in developing robust control policies for contact-rich applications. However, it typically requires meticulous Markov Decision Process (MDP) designing tailored to each task and robotic platform. This work addresses this challenge by creating a systematic appr... | Jean Pierre Sleiman;Mayank Mittal;Marco Hutter | ETHZ - ETH Zurich;NVIDIA;ETHZ - ETH Zurich | Poster | main | Whole-body Loco-Manipulation;Reinforcement Learning;Legged Mobile Manipulators | https://openreview.net/forum?id=9aZ4ehSTRc | 5 | Guided Reinforcement Learning for Robust Multi-Contact Loco-Manipulation
Reinforcement learning (RL) has shown remarkable proficiency in developing robust control policies for contact-rich applications. However, it typically requires meticulous Markov Decision Process (MDP) designing tailored to each task and robotic p... | [
-0.03437739983201027,
-0.03488831967115402,
-0.0015270481817424297,
0.004949506372213364,
-0.035070788115262985,
-0.012462720274925232,
0.016020890325307846,
0.01626722514629364,
0.015938779339194298,
-0.010327817872166634,
-0.05287988856434822,
-0.0206373892724514,
-0.007189329247921705,
... | ||
corl_2024_9dsBQhoqVr | 9dsBQhoqVr | corl | 2,024 | Fleet Supervisor Allocation: A Submodular Maximization Approach | In real-world scenarios, the data collected by robots in diverse and unpredictable environments is crucial for enhancing their perception and decision-making models. This data is predominantly collected under human supervision, particularly through imitation learning (IL), where robots learn complex tasks by observing ... | Oguzhan Akcin;Ahmet Ege Tanriverdi;Kaan Kale;Sandeep P. Chinchali | The University of Texas at Austin;Bogazici University;Bogazici University;University of Texas at Austin | Poster | main | Imitation Learning;Submodular Maximization;Fleet Learning | https://github.com/UTAustin-SwarmLab/Fleet-Supervisor-Allocation | https://openreview.net/forum?id=9dsBQhoqVr | 0 | Fleet Supervisor Allocation: A Submodular Maximization Approach
In real-world scenarios, the data collected by robots in diverse and unpredictable environments is crucial for enhancing their perception and decision-making models. This data is predominantly collected under human supervision, particularly through imitati... | [
-0.04397568851709366,
-0.05126776546239853,
-0.03571629524230957,
0.032553914934396744,
-0.03560468181967735,
-0.025224633514881134,
0.006180593278259039,
0.01975557580590248,
0.050970129668712616,
0.0551370307803154,
-0.01766282320022583,
0.0009952196851372719,
0.026452381163835526,
0.047... | |
corl_2024_9iG3SEbMnL | 9iG3SEbMnL | corl | 2,024 | ReKep: Spatio-Temporal Reasoning of Relational Keypoint Constraints for Robotic Manipulation | Representing robotic manipulation tasks as constraints that associate the robot and the environment is a promising way to encode desired robot behaviors. However, it remains unclear how to formulate the constraints such that they are 1) versatile to diverse tasks, 2) free of manual labeling, and 3) optimizable by off-t... | Wenlong Huang;Chen Wang;Yunzhu Li;Ruohan Zhang;Li Fei-Fei | NVIDIA;Computer Science Department, Stanford University;University of Illinois Urbana-Champaign;Stanford University;Stanford University | Poster | main | Structural Representation;Model-Based Planning;Foundation Models | https://github.com/huangwl18/ReKep | https://openreview.net/forum?id=9iG3SEbMnL | 97 | ReKep: Spatio-Temporal Reasoning of Relational Keypoint Constraints for Robotic Manipulation
Representing robotic manipulation tasks as constraints that associate the robot and the environment is a promising way to encode desired robot behaviors. However, it remains unclear how to formulate the constraints such that th... | [
-0.03746436536312103,
0.0018312942702323198,
-0.005376403219997883,
-0.0023323087953031063,
-0.013194532133638859,
0.046070296317338943,
0.022021600976586342,
0.02087905816733837,
0.009969611652195454,
0.031014522537589073,
-0.04146326705813408,
0.006260952912271023,
0.0026835945900529623,
... | |
corl_2024_9jJP2J1oBP | 9jJP2J1oBP | corl | 2,024 | Leveraging Mutual Information for Asymmetric Learning under Partial Observability | Even though partial observability is prevalent in robotics, most reinforcement learning studies avoid it due to the difficulty of learning a policy that can efficiently memorize past events and seek information. Fortunately, in many cases, learning can be done in an asymmetric setting where states are available during ... | Hai Huu Nguyen;Long Dinh Van The;Christopher Amato;Robert Platt | Northeastern University;Hanoi University of Science and Technology;Northeastern University;Northeastern University | Poster | main | Partial Observability;Mutual Information;Reinforcement Learning | https://sites.google.com/view/mi-asym-pomdp | https://openreview.net/forum?id=9jJP2J1oBP | 0 | Leveraging Mutual Information for Asymmetric Learning under Partial Observability
Even though partial observability is prevalent in robotics, most reinforcement learning studies avoid it due to the difficulty of learning a policy that can efficiently memorize past events and seek information. Fortunately, in many cases... | [
-0.03441100940108299,
0.00043985957745462656,
-0.007730146870017052,
0.005041194148361683,
-0.02213762328028679,
-0.040405429899692535,
0.029232285916805267,
0.04825888201594353,
0.02433810755610466,
-0.002541937865316868,
-0.053949784487485886,
-0.0424920953810215,
-0.004052399192005396,
... | |
corl_2024_A1hpY5RNiH | A1hpY5RNiH | corl | 2,024 | What Makes Pre-Trained Visual Representations Successful for Robust Manipulation? | Inspired by the success of transfer learning in computer vision, roboticists have investigated visual pre-training as a means to improve the learning efficiency and generalization ability of policies learned from pixels. To that end, past work has favored large object interaction datasets, such as first-person videos o... | Kaylee Burns;Zach Witzel;Jubayer Ibn Hamid;Tianhe Yu;Chelsea Finn;Karol Hausman | Stanford University;Stanford University;;Google Brain;Google; | Poster | main | representation learning;manipulation;visual features | https://openreview.net/forum?id=A1hpY5RNiH | 19 | What Makes Pre-Trained Visual Representations Successful for Robust Manipulation?
Inspired by the success of transfer learning in computer vision, roboticists have investigated visual pre-training as a means to improve the learning efficiency and generalization ability of policies learned from pixels. To that end, past... | [
-0.0523376390337944,
-0.029005952179431915,
-0.01887066848576069,
0.04072779044508934,
-0.021446485072374344,
-0.007251488044857979,
-0.012739100493490696,
0.021502481773495674,
0.03466155380010605,
0.026094157248735428,
-0.016238853335380554,
-0.024115631356835365,
-0.002008858136832714,
... | ||
corl_2024_A6ikGJRaKL | A6ikGJRaKL | corl | 2,024 | KOROL: Learning Visualizable Object Feature with Koopman Operator Rollout for Manipulation | Learning dexterous manipulation skills presents significant challenges due to complex nonlinear dynamics that underlie the interactions between objects and multi-fingered hands. Koopman operators have emerged as a robust method for modeling such nonlinear dynamics within a linear framework.
However, current methods rel... | Hongyi Chen;ABULIKEMU ABUDUWEILI;Aviral Agrawal;Yunhai Han;Harish Ravichandar;Changliu Liu;Jeffrey Ichnowski | Carnegie Mellon University;Carnegie Mellon University;Carnegie Mellon University;Georgia Institute of Technology;Georgia Institute of Technology;Carnegie Mellon University;Carnegie Mellon University | Poster | main | Manipulation;Koopman Operator;Visual Representation Learning | https://github.com/hychen-naza/KOROL | https://openreview.net/forum?id=A6ikGJRaKL | 5 | KOROL: Learning Visualizable Object Feature with Koopman Operator Rollout for Manipulation
Learning dexterous manipulation skills presents significant challenges due to complex nonlinear dynamics that underlie the interactions between objects and multi-fingered hands. Koopman operators have emerged as a robust method f... | [
-0.03883194550871849,
-0.029328925535082817,
0.0050123766995966434,
0.00584156159311533,
-0.02217371203005314,
0.01530731376260519,
0.019639572128653526,
0.017757602035999298,
0.017273135483264923,
0.00316068553365767,
-0.033987268805503845,
-0.03868287801742554,
0.012549575418233871,
0.03... | |
corl_2024_AEq0onGrN2 | AEq0onGrN2 | corl | 2,024 | Physically Embodied Gaussian Splatting: A Visually Learnt and Physically Grounded 3D Representation for Robotics | For robots to robustly understand and interact with the physical world, it is highly beneficial to have a comprehensive representation -- modelling geometry, physics, and visual observations -- that informs perception, planning, and control algorithms. We propose a novel dual "Gaussian-Particle" representation that mo... | Jad Abou-Chakra;Krishan Rana;Feras Dayoub;Niko Suenderhauf | Queensland University of Technology;Queensland University of Technology;University of Adelaide;Queensland University of Technology | Poster | main | 3D Representation;Gaussian Splatting;Robotics;Tracking;Physics | https://openreview.net/forum?id=AEq0onGrN2 | 9 | Physically Embodied Gaussian Splatting: A Visually Learnt and Physically Grounded 3D Representation for Robotics
For robots to robustly understand and interact with the physical world, it is highly beneficial to have a comprehensive representation -- modelling geometry, physics, and visual observations -- that informs... | [
-0.06009404733777046,
-0.03272244334220886,
0.0038845038507133722,
0.0036623317282646894,
-0.023236863315105438,
-0.052647773176431656,
0.019869202747941017,
-0.01624896563589573,
0.02566906437277794,
0.056950896978378296,
-0.01827891729772091,
0.017193783074617386,
0.05032783001661301,
0.... | ||
corl_2024_AGG1zlrrMw | AGG1zlrrMw | corl | 2,024 | Neural Attention Field: Emerging Point Relevance in 3D Scenes for One-Shot Dexterous Grasping | One-shot transfer of dexterous grasps to novel scenes with object and context variations has been a challenging problem. While distilled feature fields from large vision models have enabled semantic correspondences across 3D scenes, their features are point-based and restricted to object surfaces, limiting their capabi... | Qianxu Wang;Congyue Deng;Tyler Ga Wei Lum;Yuanpei Chen;Yaodong Yang;Jeannette Bohg;Yixin Zhu;Leonidas Guibas | Peking University;Stanford University;Stanford University;PsiRobot;Peking University;Stanford University;Peking University;Stanford University | Poster | main | Desterous Grasping;One-Shot Manipulation;Distilled Feature Field;Neural Implicit Field;Self-Supervised Learning | https://openreview.net/forum?id=AGG1zlrrMw | 2 | Neural Attention Field: Emerging Point Relevance in 3D Scenes for One-Shot Dexterous Grasping
One-shot transfer of dexterous grasps to novel scenes with object and context variations has been a challenging problem. While distilled feature fields from large vision models have enabled semantic correspondences across 3D s... | [
-0.0241929329931736,
-0.03498559072613716,
-0.04251872003078461,
-0.019466619938611984,
-0.009497899562120438,
-0.032939329743385315,
0.0005743784713558853,
0.0004122509271837771,
0.021911265328526497,
-0.030567120760679245,
-0.018162809312343597,
-0.0026913732290267944,
0.011870110407471657... | ||
corl_2024_AhEE5wrcLU | AhEE5wrcLU | corl | 2,024 | Velociraptor: Leveraging Visual Foundation Models for Label-Free, Risk-Aware Off-Road Navigation | Traversability analysis in off-road regimes is a challenging task that requires understanding of multi-modal inputs such as camera and LiDAR. These measurements are often sparse, noisy, and difficult to interpret, particularly in the off-road setting. Existing systems are very engineering-intensive, often requiring han... | Samuel Triest;Matthew Sivaprakasam;Shubhra Aich;David Fan;Wenshan Wang;Sebastian Scherer | Carnegie Mellon University;Carnegie Mellon University;Carnegie Mellon University;Jet Propulsion Laboratory;School of Computer Science, Carnegie Mellon University;Near Earth Autonomy Inc. | Poster | main | Field Robotics;Self-Supervised Learning;Visual Foundation Models | https://openreview.net/forum?id=AhEE5wrcLU | 2 | Velociraptor: Leveraging Visual Foundation Models for Label-Free, Risk-Aware Off-Road Navigation
Traversability analysis in off-road regimes is a challenging task that requires understanding of multi-modal inputs such as camera and LiDAR. These measurements are often sparse, noisy, and difficult to interpret, particula... | [
-0.07319066673517227,
-0.051326584070920944,
-0.01769240014255047,
0.019610628485679626,
-0.013436912558972836,
0.0037386836484074593,
0.03866254910826683,
-0.008152471855282784,
0.01418185606598854,
0.020746666938066483,
-0.030617164447903633,
-0.017841389402747154,
0.04261074960231781,
0... | ||
corl_2024_AsbyZRdqPv | AsbyZRdqPv | corl | 2,024 | Simple Masked Training Strategies Yield Control Policies That Are Robust to Sensor Failure | Sensor failure is common when robots are deployed in the real world, as sensors naturally wear out over time. Such failures can lead to catastrophic outcomes, including damage to the robot from unexpected robot behaviors such as falling during walking. Previous work has tried to address this problem by recovering missi... | Skand Skand;Bikram Pandit;Chanho Kim;Li Fuxin;Stefan Lee | Oregon State University;Oregon State University;Oregon State University;; | Poster | main | Reinforcement Learning;Robustness;Sensorimotor Learning | https://openreview.net/forum?id=AsbyZRdqPv | 1 | Simple Masked Training Strategies Yield Control Policies That Are Robust to Sensor Failure
Sensor failure is common when robots are deployed in the real world, as sensors naturally wear out over time. Such failures can lead to catastrophic outcomes, including damage to the robot from unexpected robot behaviors such as ... | [
-0.07531550526618958,
0.0050316136330366135,
-0.01373191736638546,
-0.007295372895896435,
-0.022814959287643433,
-0.025410115718841553,
-0.004301142878830433,
0.08125261962413788,
-0.01194891519844532,
0.009941871277987957,
-0.021713418886065483,
0.003117920132353902,
0.011911574751138687,
... | ||
corl_2024_AuJnXGq3AL | AuJnXGq3AL | corl | 2,024 | Scaling Cross-Embodied Learning: One Policy for Manipulation, Navigation, Locomotion and Aviation | Modern machine learning systems rely on large datasets to attain broad generalization, and this often poses a challenge in robotic learning, where each robotic platform and task might have only a small dataset. By training a single policy across many different kinds of robots, a robotic learning method can leverage muc... | Ria Doshi;Homer Rich Walke;Oier Mees;Sudeep Dasari;Sergey Levine | University of California, Berkeley;University of California, Berkeley;Electrical Engineering & Computer Science Department, University of California, Berkeley;Google;Carnegie Mellon University | Poster | main | Imitation Learning;Cross-Embodiment | https://github.com/rail-berkeley/crossformer | https://openreview.net/forum?id=AuJnXGq3AL | 53 | Scaling Cross-Embodied Learning: One Policy for Manipulation, Navigation, Locomotion and Aviation
Modern machine learning systems rely on large datasets to attain broad generalization, and this often poses a challenge in robotic learning, where each robotic platform and task might have only a small dataset. By training... | [
-0.07338128238916397,
-0.039050642400979996,
-0.06327752768993378,
0.0031712513882666826,
-0.04041501507163048,
-0.01965438388288021,
-0.020244384184479713,
0.020631570369005203,
0.01913813315331936,
0.006324064917862415,
-0.013855786994099617,
-0.008753206580877304,
0.04174251854419708,
-... | |
corl_2024_AzP6kSEffm | AzP6kSEffm | corl | 2,024 | Dynamics-Guided Diffusion Model for Sensor-less Robot Manipulator Design | We present Dynamics-Guided Diffusion Model (DGDM), a data-driven framework for generating task-specific manipulator designs without task-specific training. Given object shapes and task specifications, DGDM generates sensor-less manipulator designs that can blindly manipulate objects towards desired motions and poses u... | Xiaomeng Xu;Huy Ha;Shuran Song | Stanford University;Columbia University;Stanford University | Poster | main | manipulator design;hardware optimization;diffusion model | https://github.com/real-stanford/dgdm | https://openreview.net/forum?id=AzP6kSEffm | 1 | Dynamics-Guided Diffusion Model for Sensor-less Robot Manipulator Design
We present Dynamics-Guided Diffusion Model (DGDM), a data-driven framework for generating task-specific manipulator designs without task-specific training. Given object shapes and task specifications, DGDM generates sensor-less manipulator design... | [
-0.0826910212635994,
-0.04431409016251564,
-0.009603174403309822,
-0.01783829927444458,
-0.0064468239434063435,
-0.0005485053407028317,
-0.01731969229876995,
-0.011570305563509464,
0.024499719962477684,
0.03556929901242256,
-0.03499704599380493,
-0.018634093925356865,
0.016380833461880684,
... | |
corl_2024_B2X57y37kC | B2X57y37kC | corl | 2,024 | Learning to Look: Seeking Information for Decision Making via Policy Factorization | Many robot manipulation tasks require active or interactive exploration behavior in order to be performed successfully. Such tasks are ubiquitous in embodied domains, where agents must actively search for the information necessary for each stage of a task, e.g., moving the head of the robot to find information relevant... | Shivin Dass;Jiaheng Hu;Ben Abbatematteo;Peter Stone;Roberto Martín-Martín | University of Texas at Austin;University of Texas, Austin;University of Texas at Austin;Brown University;University of Texas at Austin | Poster | main | Active Vision;Manipulation;Interactive Perception | https://openreview.net/forum?id=B2X57y37kC | 0 | Learning to Look: Seeking Information for Decision Making via Policy Factorization
Many robot manipulation tasks require active or interactive exploration behavior in order to be performed successfully. Such tasks are ubiquitous in embodied domains, where agents must actively search for the information necessary for ea... | [
-0.0359133705496788,
-0.04787207394838333,
-0.00766512006521225,
0.02892814390361309,
-0.04649365693330765,
-0.0027195808943361044,
-0.01887873373925686,
0.002312109339982271,
0.014370935969054699,
0.05986803397536278,
0.008186683990061283,
-0.021924292668700218,
-0.013160163536667824,
0.0... | ||
corl_2024_B45HRM4Wb4 | B45HRM4Wb4 | corl | 2,024 | ResPilot: Teleoperated Finger Gaiting via Gaussian Process Residual Learning | Dexterous robot hand teleoperation allows for long-range transfer of human manipulation expertise, and could simultaneously provide a way for humans to teach these skills to robots. However, current methods struggle to reproduce the functional workspace of the human hand, often limiting them to simple grasping tasks. W... | Patrick Naughton;Jinda Cui;Karankumar Patel;Soshi Iba | Honda Research Institution US;Honda Research Institution US;Honda Research Institution US;Honda R&D | Poster | main | Teleoperation;Dexterous Manipulation;Gaussian Process | https://openreview.net/forum?id=B45HRM4Wb4 | 2 | ResPilot: Teleoperated Finger Gaiting via Gaussian Process Residual Learning
Dexterous robot hand teleoperation allows for long-range transfer of human manipulation expertise, and could simultaneously provide a way for humans to teach these skills to robots. However, current methods struggle to reproduce the functional... | [
-0.07066745311021805,
-0.002491601277142763,
-0.003368362784385681,
0.009068488143384457,
-0.01652918942272663,
-0.02181326411664486,
-0.012683660723268986,
0.017930127680301666,
-0.004294486250728369,
0.04565741866827011,
-0.055849477648735046,
0.005857613403350115,
0.024727966636419296,
... | ||
corl_2024_B7Lf6xEv7l | B7Lf6xEv7l | corl | 2,024 | DiffusionSeeder: Seeding Motion Optimization with Diffusion for Rapid Motion Planning | Running optimization across many parallel seeds leveraging GPU compute [2] have relaxed the need for a good initialization, but this can fail if the problem is highly non-convex as all seeds could get stuck in local minima. One such setting is collision-free motion optimization for robot manipulation, where optimizatio... | Huang Huang;Balakumar Sundaralingam;Arsalan Mousavian;Adithyavairavan Murali;Ken Goldberg;Dieter Fox | University of California, Berkeley;NVIDIA;NVIDIA;;University of California, Berkeley;Department of Computer Science | Poster | main | Robot Motion Planning;Diffusion Model | https://openreview.net/forum?id=B7Lf6xEv7l | 4 | DiffusionSeeder: Seeding Motion Optimization with Diffusion for Rapid Motion Planning
Running optimization across many parallel seeds leveraging GPU compute [2] have relaxed the need for a good initialization, but this can fail if the problem is highly non-convex as all seeds could get stuck in local minima. One such s... | [
-0.06357908993959427,
-0.03325091674923897,
-0.04167751967906952,
0.0271587111055851,
-0.0345035195350647,
0.04065266251564026,
-0.01577141135931015,
0.020022669807076454,
0.039893507957458496,
0.05382397025823593,
-0.026646282523870468,
0.004675909876823425,
-0.007064870558679104,
0.01338... | ||
corl_2024_BmvUg1FIWC | BmvUg1FIWC | corl | 2,024 | Neural Inverse Source Problem | Reconstructing unknown external source functions is an important perception capability for a large range of robotics domains including manipulation, aerial, and underwater robotics. In this work, we propose a Physics-Informed Neural Network (PINN) based approach for solving the inverse source problems in robotics, join... | Youngsun Wi;Jayjun Lee;Miquel Oller;Nima Fazeli | University of Michigan;;University of Michigan - Ann Arbor;University of Michigan | Poster | main | Inverse source problem;Physics informed neural network | https://openreview.net/forum?id=BmvUg1FIWC | 2 | Neural Inverse Source Problem
Reconstructing unknown external source functions is an important perception capability for a large range of robotics domains including manipulation, aerial, and underwater robotics. In this work, we propose a Physics-Informed Neural Network (PINN) based approach for solving the inverse sou... | [
-0.04130327329039574,
-0.016444474458694458,
-0.01121597457677126,
0.021213840693235397,
-0.0028719266410917044,
-0.004839641507714987,
0.028691157698631287,
0.009590267203748226,
0.030583912506699562,
0.052959639579057693,
-0.056820109486579895,
0.0342007577419281,
-0.01624770276248455,
0... | ||
corl_2024_Bq4XOaU4sV | Bq4XOaU4sV | corl | 2,024 | Bridging the Sim-to-Real Gap from the Information Bottleneck Perspective | Reinforcement Learning (RL) has recently achieved remarkable success in robotic control. However, most works in RL operate in simulated environments where privileged knowledge (e.g., dynamics, surroundings, terrains) is readily available. Conversely, in real-world scenarios, robot agents usually rely solely on local st... | Haoran He;Peilin Wu;Chenjia Bai;Hang Lai;Lingxiao Wang;Ling Pan;Xiaolin Hu;Weinan Zhang | Hong Kong University of Science and Technology;Shanghai Jiaotong University;Shanghai AI Laboratory;Shanghai Jiaotong University;Northwestern University;Montreal Institute for Learning Algorithms (MILA);Tsinghua University;Shanghai Jiaotong University | Poster | main | Sim-to-Real;Information Bottleneck;Reinforcement Learning;Locomotion | https://github.com/tinnerhrhe/HIB_Policy | https://openreview.net/forum?id=Bq4XOaU4sV | 9 | Bridging the Sim-to-Real Gap from the Information Bottleneck Perspective
Reinforcement Learning (RL) has recently achieved remarkable success in robotic control. However, most works in RL operate in simulated environments where privileged knowledge (e.g., dynamics, surroundings, terrains) is readily available. Converse... | [
-0.06423180550336838,
-0.04795466363430023,
0.03560385853052139,
0.004638698883354664,
-0.006142047233879566,
-0.0549686998128891,
0.02008911594748497,
0.03802446648478508,
-0.0005741789354942739,
0.021251767873764038,
-0.033793170005083084,
-0.022757498547434807,
-0.04383772984147072,
0.0... | |
corl_2024_CPQW5kc0pe | CPQW5kc0pe | corl | 2,024 | VoxAct-B: Voxel-Based Acting and Stabilizing Policy for Bimanual Manipulation | Bimanual manipulation is critical to many robotics applications. In contrast to single-arm manipulation, bimanual manipulation tasks are challenging due to higher-dimensional action spaces. Prior works leverage large amounts of data and primitive actions to address this problem, but may suffer from sample inefficiency ... | I-Chun Arthur Liu;Sicheng He;Daniel Seita;Gaurav S. Sukhatme | University of Southern California;;;University of Southern California | Poster | main | bimanual manipulation;voxel representation;vision language models | https://github.com/VoxAct-B/voxactb | https://openreview.net/forum?id=CPQW5kc0pe | 14 | VoxAct-B: Voxel-Based Acting and Stabilizing Policy for Bimanual Manipulation
Bimanual manipulation is critical to many robotics applications. In contrast to single-arm manipulation, bimanual manipulation tasks are challenging due to higher-dimensional action spaces. Prior works leverage large amounts of data and primi... | [
-0.035452429205179214,
-0.004802964627742767,
-0.017716987058520317,
0.012235794216394424,
-0.02127883955836296,
0.01930413395166397,
0.029399126768112183,
0.048168059438467026,
0.00801878236234188,
0.019931610673666,
-0.024028664454817772,
-0.034880317747592926,
0.02116810902953148,
-0.01... | |
corl_2024_CpXiqz6qf4 | CpXiqz6qf4 | corl | 2,024 | SonicSense: Object Perception from In-Hand Acoustic Vibration | We introduce SonicSense, a holistic design of hardware and software to enable rich robot object perception through in-hand acoustic vibration sensing. While previous studies have shown promising results with acoustic sensing for object perception, current solutions are constrained to a handful of objects with simple ge... | Jiaxun Liu;Boyuan Chen | Duke University;Duke University | Poster | main | Tactile Perception;Object State Estimation;Audio;Acoustic Vibration Sensing | https://github.com/generalroboticslab/SonicSense?tab=readme-ov-file | https://openreview.net/forum?id=CpXiqz6qf4 | 5 | SonicSense: Object Perception from In-Hand Acoustic Vibration
We introduce SonicSense, a holistic design of hardware and software to enable rich robot object perception through in-hand acoustic vibration sensing. While previous studies have shown promising results with acoustic sensing for object perception, current so... | [
-0.038686275482177734,
0.005439676810055971,
-0.005802012048661709,
0.015626877546310425,
0.015599004924297333,
-0.03998696804046631,
0.007678723894059658,
0.0013750165235251188,
0.04058156907558441,
0.03344634920358658,
0.016908986493945122,
0.009606534615159035,
0.016072828322649002,
-0.... | |
corl_2024_CskuWHDBAr | CskuWHDBAr | corl | 2,024 | Enhancing Visual Domain Robustness in Behaviour Cloning via Saliency-Guided Augmentation | In vision-based behaviour cloning (BC), traditional image-level augmentation methods such as pixel shifting enhance in-domain performance but often struggle with visual domain shifts, including distractors, occlusion, and changes in lighting and backgrounds. Conversely, superimposition-based augmentation, proven effect... | Zheyu Zhuang;RUIYU WANG;Nils Ingelhag;Ville Kyrki;Danica Kragic | KTH Royal Institute of Technology;KTH Royal Institute of Technology;;Aalto University;KTH | Poster | main | Behaviour Cloning;Visuomotor Policy;Data Augmentation | https://github.com/Zheyu-Zhuang/RoboSaGA | https://openreview.net/forum?id=CskuWHDBAr | 2 | Enhancing Visual Domain Robustness in Behaviour Cloning via Saliency-Guided Augmentation
In vision-based behaviour cloning (BC), traditional image-level augmentation methods such as pixel shifting enhance in-domain performance but often struggle with visual domain shifts, including distractors, occlusion, and changes i... | [
-0.06226715072989464,
-0.014366479590535164,
-0.01053293701261282,
0.0041824690997600555,
-0.01064459327608347,
-0.017353292554616928,
0.042690031230449677,
0.04544422775506973,
0.03260372206568718,
0.006908750627189875,
-0.021084483712911606,
0.011268009431660175,
-0.021903298795223236,
0... | |
corl_2024_Czs2xH9114 | Czs2xH9114 | corl | 2,024 | WoCoCo: Learning Whole-Body Humanoid Control with Sequential Contacts | Humanoid activities involving sequential contacts are crucial for complex robotic interactions and operations in the real world and are traditionally solved by model-based motion planning, which is time-consuming and often relies on simplified dynamics models.
Although model-free reinforcement learning (RL) has become... | Chong Zhang;Wenli Xiao;Tairan He;Guanya Shi | ETHZ - ETH Zurich;Carnegie Mellon University;Carnegie Mellon University;Carnegie Mellon University | Poster | main | Whole-Body Humanoid Control;Multi-Contact Control;Reinforcement Learning | https://openreview.net/forum?id=Czs2xH9114 | 45 | WoCoCo: Learning Whole-Body Humanoid Control with Sequential Contacts
Humanoid activities involving sequential contacts are crucial for complex robotic interactions and operations in the real world and are traditionally solved by model-based motion planning, which is time-consuming and often relies on simplified dynami... | [
-0.046264369040727615,
0.004479791969060898,
-0.018066272139549255,
-0.008559643290936947,
-0.0087389275431633,
-0.007690807804465294,
-0.0030822991393506527,
-0.020557856187224388,
-0.006987464148551226,
-0.0026547766756266356,
-0.009331942535936832,
-0.028354395180940628,
0.018627109006047... | ||
corl_2024_DDIoRSh8ID | DDIoRSh8ID | corl | 2,024 | Multi-Task Interactive Robot Fleet Learning with Visual World Models | Recent advancements in large-scale multi-task robot learning offer the potential for deploying robot fleets in household and industrial settings, enabling them to perform diverse tasks across various environments. However, AI-enabled robots often face challenges with generalization and robustness when exposed to real-w... | Huihan Liu;Yu Zhang;Vaarij Betala;Evan Zhang;James Liu;Crystal Ding;Yuke Zhu | ;ShanghaiTech University;;;;;Computer Science Department, University of Texas, Austin | Poster | main | Robot Manipulation;Interactive Imitation Learning;Fleet Learning | https://openreview.net/forum?id=DDIoRSh8ID | 4 | Multi-Task Interactive Robot Fleet Learning with Visual World Models
Recent advancements in large-scale multi-task robot learning offer the potential for deploying robot fleets in household and industrial settings, enabling them to perform diverse tasks across various environments. However, AI-enabled robots often face... | [
-0.0532115176320076,
-0.04185471311211586,
-0.01719502918422222,
0.038808684796094894,
-0.002481950679793954,
-0.020062435418367386,
-0.011046561412513256,
0.016762569546699524,
0.031231213361024857,
0.04632974788546562,
0.010802126489579678,
-0.003903901670128107,
-0.008019332773983479,
0... | ||
corl_2024_DSdAEsEGhE | DSdAEsEGhE | corl | 2,024 | SoloParkour: Constrained Reinforcement Learning for Visual Locomotion from Privileged Experience | Parkour poses a significant challenge for legged robots, requiring navigation through complex environments with agility and precision based on limited sensory inputs.
In this work, we introduce a novel method for training end-to-end visual policies, from depth pixels to robot control commands, to achieve agile and saf... | Elliot Chane-Sane;Joseph Amigo;Thomas Flayols;Ludovic Righetti;Nicolas Mansard | LAAS / CNRS;New York University;;Max-Planck Institute;LAAS / CNRS | Poster | main | Reinforcement Learning;Agile Locomotion;Visuomotor Control | https://github.com/Gepetto/SoloParkour | https://openreview.net/forum?id=DSdAEsEGhE | 7 | SoloParkour: Constrained Reinforcement Learning for Visual Locomotion from Privileged Experience
Parkour poses a significant challenge for legged robots, requiring navigation through complex environments with agility and precision based on limited sensory inputs.
In this work, we introduce a novel method for training ... | [
-0.0784868374466896,
-0.03239520266652107,
0.00666824309155345,
0.027374407276511192,
-0.051869239658117294,
0.019141780212521553,
-0.00876793172210455,
0.004503948148339987,
0.011896699666976929,
0.003883271012455225,
-0.012644280679523945,
-0.01934482716023922,
0.01447170227766037,
0.015... | |
corl_2024_Dftu4r5jHe | Dftu4r5jHe | corl | 2,024 | Context-Aware Replanning with Pre-Explored Semantic Map for Object Navigation | Pre-explored Semantic Map, constructed through prior exploration using visual language models (VLMs), has proven effective as a foundational element for training-free robotic applications. However, existing approaches assume the map's accuracy and do not provide effective mechanisms for revising decisions based on inco... | Po-Chen Ko;Hung-Ting Su;CY Chen;Jia-Fong Yeh;Min Sun;Winston H. Hsu | National Taiwan University;National Taiwan University;Sony Group Corporation;National Tsing Hua University;National Taiwan University;National Taiwan University | Poster | main | VLMs;map;navigation;uncertainty;multi-view consistency;robotics | https://github.com/CARe-maps/CARe_experiments | https://openreview.net/forum?id=Dftu4r5jHe | 0 | Context-Aware Replanning with Pre-Explored Semantic Map for Object Navigation
Pre-explored Semantic Map, constructed through prior exploration using visual language models (VLMs), has proven effective as a foundational element for training-free robotic applications. However, existing approaches assume the map's accurac... | [
-0.06104801222681999,
-0.011282220482826233,
-0.008098394609987736,
0.018075617030262947,
0.004113984294235706,
-0.0036072563380002975,
0.05012675002217293,
0.02167593128979206,
0.010967540554702282,
0.016798384487628937,
0.015697002410888672,
-0.012448389083147049,
-0.004294462967664003,
... | |
corl_2024_DsFQg0G4Xu | DsFQg0G4Xu | corl | 2,024 | Learning Long-Horizon Action Dependencies in Sampling-Based Bilevel Planning | Autonomous robots will need the ability to make task and motion plans that involve long sequences of actions, e.g. to prepare a meal. One challenge is that the feasibility of actions late in the plan may depend on much earlier actions. This issue is exacerbated if these dependencies exist at a purely geometric level, m... | Bartłomiej Cieślar;Leslie Pack Kaelbling;Tomás Lozano-Pérez;Jorge Mendez-Mendez | Imperial College London;Massachusetts Institute of Technology;Massachusetts Institute of Technology;Massachusetts Institute of Technology | Poster | main | task and motion planning;long-horizon;learning for planning | https://openreview.net/forum?id=DsFQg0G4Xu | 0 | Learning Long-Horizon Action Dependencies in Sampling-Based Bilevel Planning
Autonomous robots will need the ability to make task and motion plans that involve long sequences of actions, e.g. to prepare a meal. One challenge is that the feasibility of actions late in the plan may depend on much earlier actions. This is... | [
-0.02016174979507923,
-0.020947271957993507,
0.0025319079868495464,
0.022742751985788345,
-0.0075933863408863544,
0.012241062708199024,
0.006648888811469078,
0.00020748554379679263,
-0.037218812853097916,
0.008430342189967632,
-0.06119595840573311,
-0.01009022630751133,
0.00980968214571476,
... | ||
corl_2024_E4K3yLQQ7s | E4K3yLQQ7s | corl | 2,024 | Visual Manipulation with Legs | Animals have the ability to use their arms and legs for both locomotion and manipulation. We envision quadruped robots to have the same versatility. This work presents a system that empowers a quadruped robot to perform object interactions with its legs, drawing inspiration from non-prehensile manipulation techniques. ... | Xialin He;Chengjing Yuan;Wenxuan Zhou;Ruihan Yang;David Held;Xiaolong Wang | Shanghai Jiaotong University;University of California, San Diego;Carnegie Mellon University;University of California, San Diego;Carnegie Mellon University;University of California, San Diego | Poster | main | Legged robots;Non-prehensile manipulation;Reinforcement Learning | https://openreview.net/forum?id=E4K3yLQQ7s | 2 | Visual Manipulation with Legs
Animals have the ability to use their arms and legs for both locomotion and manipulation. We envision quadruped robots to have the same versatility. This work presents a system that empowers a quadruped robot to perform object interactions with its legs, drawing inspiration from non-prehen... | [
-0.052956998348236084,
-0.032059844583272934,
-0.02877117320895195,
0.06423243880271912,
-0.027061063796281815,
-0.005647117737680674,
0.0040474142879247665,
-0.004432658664882183,
0.0012449967907741666,
-0.00848477054387331,
-0.030049055814743042,
-0.008465978316962719,
0.025275785475969315... | ||
corl_2024_EM0wndCeoD | EM0wndCeoD | corl | 2,024 | BiGym: A Demo-Driven Mobile Bi-Manual Manipulation Benchmark | We introduce BiGym, a new benchmark and learning environment for mobile bi-manual demo-driven robotic manipulation. BiGym features 40 diverse tasks set in home environments, ranging from simple target reaching to complex kitchen cleaning. To capture the real-world performance accurately, we provide human-collected demo... | Nikita Chernyadev;Nicholas Backshall;Xiao Ma;Yunfan Lu;Younggyo Seo;Stephen James | Dyson;;Dyson Robot Learning Lab;;Dyson;Dyson | Poster | main | Bi-Manual Manipulation;Mobile Manipulation;Benchmark | https://github.com/chernyadev/bigym | https://openreview.net/forum?id=EM0wndCeoD | 10 | BiGym: A Demo-Driven Mobile Bi-Manual Manipulation Benchmark
We introduce BiGym, a new benchmark and learning environment for mobile bi-manual demo-driven robotic manipulation. BiGym features 40 diverse tasks set in home environments, ranging from simple target reaching to complex kitchen cleaning. To capture the real-... | [
-0.0676727220416069,
-0.05718345195055008,
-0.02511409856379032,
-0.028328552842140198,
-0.023234302178025246,
-0.0053057293407619,
-0.006104643456637859,
-0.010367085225880146,
-0.004715942777693272,
0.04680696502327919,
-0.01775469072163105,
0.018478412181138992,
0.011100206524133682,
0.... | |
corl_2024_EPujQZWemk | EPujQZWemk | corl | 2,024 | ViPER: Visibility-based Pursuit-Evasion via Reinforcement Learning | In visibility-based pursuit-evasion tasks, a team of mobile pursuer robots with limited sensing capabilities is tasked with detecting all evaders in a multiply-connected planar environment, whose map may or may not be known to pursuers beforehand. This requires tight coordination among multiple agents to ensure that th... | Yizhuo Wang;Yuhong Cao;Jimmy Chiun;Subhadeep Koley;Mandy Pham;Guillaume Adrien Sartoretti | National University of Singapore;National University of Singapore;;Indian Institute of Engineering Science and Technology, Shibpur;University of California, Berkeley;National University of Singapore | Poster | main | MARL;pursuit-evasion;graph attention;path planning | https://github.com/marmotlab/ViPER | https://openreview.net/forum?id=EPujQZWemk | 1 | ViPER: Visibility-based Pursuit-Evasion via Reinforcement Learning
In visibility-based pursuit-evasion tasks, a team of mobile pursuer robots with limited sensing capabilities is tasked with detecting all evaders in a multiply-connected planar environment, whose map may or may not be known to pursuers beforehand. This ... | [
-0.10091979801654816,
-0.039164017885923386,
-0.0034324252046644688,
0.023836543783545494,
-0.04815613478422165,
0.010859282687306404,
0.020102214068174362,
0.01737113669514656,
0.042916927486658096,
0.005870886612683535,
-0.02246171608567238,
-0.026381833478808403,
-0.016191385686397552,
... | |
corl_2024_EdVNB2kHv1 | EdVNB2kHv1 | corl | 2,024 | Scaling Robot Policy Learning via Zero-Shot Labeling with Foundation Models | A central challenge towards developing robots that can relate human language to their perception and actions is the scarcity of natural language annotations in diverse robot datasets. Moreover, robot policies that follow natural language instructions are typically trained on either templated language or expensive human... | Nils Blank;Moritz Reuss;Marcel Rühle;Ömer Erdinç Yağmurlu;Fabian Wenzel;Oier Mees;Rudolf Lioutikov | Karlsruher Institut für Technologie;Karlsruher Institut für Technologie;Karlsruher Institut für Technologie;Karlsruher Institut für Technologie;Karlsruher Institut für Technologie;Electrical Engineering & Computer Science Department, University of California, Berkeley;Karlsruher Institut für Technologie | Poster | main | Foundation Models;Language-conditioned Imitation Learning;Data Labeling | https://openreview.net/forum?id=EdVNB2kHv1 | 7 | Scaling Robot Policy Learning via Zero-Shot Labeling with Foundation Models
A central challenge towards developing robots that can relate human language to their perception and actions is the scarcity of natural language annotations in diverse robot datasets. Moreover, robot policies that follow natural language instru... | [
-0.09158093482255936,
-0.01840776763856411,
-0.007440950721502304,
0.04674290865659714,
-0.008439183235168457,
-0.025129808112978935,
0.023298189043998718,
0.0006330531905405223,
0.01294038537889719,
0.030404869467020035,
-0.033390406519174576,
-0.007463845890015364,
0.016411302611231804,
... | ||
corl_2024_EifoVoIyd5 | EifoVoIyd5 | corl | 2,024 | What Matters in Range View 3D Object Detection | Lidar-based perception pipelines rely on 3D object detection models to interpret complex scenes. While multiple representations for lidar exist, the range view is enticing since it losslessly encodes the entire lidar sensor output. In this work, we achieve state-of-the-art amongst range view 3D object detection models ... | Benjamin Wilson;Nicholas Autio Mitchell;Jhony Kaesemodel Pontes;James Hays | Georgia Institute of Technology;NVIDIA;Latitude AI;Georgia Institute of Technology | Poster | main | 3D Object Detection;3D Perception;Autonomous Driving | https://github.com/benjaminrwilson/range-view-3d-detection | https://openreview.net/forum?id=EifoVoIyd5 | 2 | What Matters in Range View 3D Object Detection
Lidar-based perception pipelines rely on 3D object detection models to interpret complex scenes. While multiple representations for lidar exist, the range view is enticing since it losslessly encodes the entire lidar sensor output. In this work, we achieve state-of-the-art... | [
-0.04061802849173546,
-0.022916005924344063,
-0.048785384744405746,
-0.0012248759157955647,
0.01111161895096302,
-0.036826036870479584,
0.003281528130173683,
0.021767470985651016,
0.021147625520825386,
0.010901966132223606,
-0.03926895558834076,
-0.03606034815311432,
0.004354861564934254,
... | |
corl_2024_EiqQEsOMZt | EiqQEsOMZt | corl | 2,024 | TaMMa: Target-driven Multi-subscene Mobile Manipulation | For everyday service robotics, the ability to navigate back and forth based on tasks in multi-subscene environments and perform delicate manipulations is crucial and highly practical.
While existing robotics primarily focus on complex tasks within a single scene or simple tasks across scalable scenes individually, robo... | Jiawei Hou;Tianyu Wang;Tongying Pan;Shouyan Wang;Xiangyang Xue;Yanwei Fu | Fudan University;Fudan University;Fudan University;Fudan University;Fudan University;Fudan University, | Poster | main | Multi-subscene;3D Gaussians;Scene Inpainting;Target-driven Mobile Manipulation | https://openreview.net/forum?id=EiqQEsOMZt | 2 | TaMMa: Target-driven Multi-subscene Mobile Manipulation
For everyday service robotics, the ability to navigate back and forth based on tasks in multi-subscene environments and perform delicate manipulations is crucial and highly practical.
While existing robotics primarily focus on complex tasks within a single scene o... | [
-0.015687253326177597,
-0.0005512507050298154,
-0.029403137043118477,
-0.012293152511119843,
-0.03461052477359772,
0.02092253603041172,
-0.0013739131391048431,
-0.009754551574587822,
0.008452704176306725,
0.07167595624923706,
-0.0415475070476532,
0.025125641375780106,
0.021685045212507248,
... | ||
corl_2024_EyEE7547vy | EyEE7547vy | corl | 2,024 | Event3DGS: Event-Based 3D Gaussian Splatting for High-Speed Robot Egomotion | By combining differentiable rendering with explicit point-based scene representations, 3D Gaussian Splatting (3DGS) has demonstrated breakthrough 3D reconstruction capabilities.
However, to date 3DGS has had limited impact on robotics, where high-speed egomotion is pervasive: Egomotion introduces motion blur and l... | Tianyi Xiong;Jiayi Wu;Botao He;Cornelia Fermuller;Yiannis Aloimonos;Heng Huang;Christopher Metzler | University of Maryland, College Park;University of Maryland, College Park;University of Maryland, College Park;University of Maryland, College Park;University of Maryland, College Park;Department of Computer Science, University of Maryland, College Park;University of Maryland, College Park | Poster | main | Event-based 3D Reconstruction;Gaussian Splatting;High-speed Robot Egomotion | https://openreview.net/forum?id=EyEE7547vy | 11 | Event3DGS: Event-Based 3D Gaussian Splatting for High-Speed Robot Egomotion
By combining differentiable rendering with explicit point-based scene representations, 3D Gaussian Splatting (3DGS) has demonstrated breakthrough 3D reconstruction capabilities.
However, to date 3DGS has had limited impact on robotics, whe... | [
-0.05147513747215271,
-0.04223791882395744,
-0.010484986007213593,
-0.022999925538897514,
-0.03179017826914787,
-0.009186002425849438,
0.014507644809782505,
-0.01998293027281761,
-0.005251804832369089,
0.028531081974506378,
-0.007472647819668055,
0.036613646894693375,
0.02519748918712139,
... | ||
corl_2024_F0rWEID2gb | F0rWEID2gb | corl | 2,024 | Environment Curriculum Generation via Large Language Models | Recent work has demonstrated that a promising strategy for teaching robots a wide range of complex skills is by training them on a curriculum of progressively more challenging environments. However, developing an effective curriculum of environment distributions currently requires significant expertise, which must be r... | William Liang;Sam Wang;Hung-Ju Wang;Osbert Bastani;Dinesh Jayaraman;Yecheng Jason Ma | University of Pennsylvania;University of Pennsylvania;;University of Pennsylvania;University of Pennsylvania; | Poster | main | Large Language Models;Environment Curriculum;Quadrupeds;Sim-To-Real Reinforcement Learning | https://openreview.net/forum?id=F0rWEID2gb | 4 | Environment Curriculum Generation via Large Language Models
Recent work has demonstrated that a promising strategy for teaching robots a wide range of complex skills is by training them on a curriculum of progressively more challenging environments. However, developing an effective curriculum of environment distributio... | [
-0.06761855632066727,
0.010745727457106113,
0.0013582432875409722,
-0.02639269083738327,
-0.027114003896713257,
0.011217355728149414,
-0.030406154692173004,
-0.027927793562412262,
-0.0013836742145940661,
0.0003022813762072474,
0.00015417506801895797,
0.01618332415819168,
-0.01379743963479995... | ||
corl_2024_FHnVRmeqxf | FHnVRmeqxf | corl | 2,024 | FlowRetrieval: Flow-Guided Data Retrieval for Few-Shot Imitation Learning | Imitation learning policies in robotics tend to require an extensive amount of demonstrations. It is critical to develop few-shot adaptation strategies that rely only on a small amount of task-specific human demonstrations. Prior works focus on learning general policies from large scale dataset with diverse behaviors. ... | Li-Heng Lin;Yuchen Cui;Amber Xie;Tianyu Hua;Dorsa Sadigh | Stanford University;Stanford University;;;Stanford University | Poster | main | Data Retrieval;Few-shot Learning;Imitation Learning | https://openreview.net/forum?id=FHnVRmeqxf | 9 | FlowRetrieval: Flow-Guided Data Retrieval for Few-Shot Imitation Learning
Imitation learning policies in robotics tend to require an extensive amount of demonstrations. It is critical to develop few-shot adaptation strategies that rely only on a small amount of task-specific human demonstrations. Prior works focus on l... | [
-0.06562439352273941,
-0.005118375178426504,
-0.03705248609185219,
0.018107673153281212,
-0.03352194279432297,
-0.0009906899649649858,
0.013758191838860512,
0.022457154467701912,
0.003000505268573761,
0.011556153185665607,
-0.03890874981880188,
-0.03417709842324257,
-0.006515122950077057,
... | ||
corl_2024_FO6tePGRZj | FO6tePGRZj | corl | 2,024 | Mobile ALOHA: Learning Bimanual Mobile Manipulation using Low-Cost Whole-Body Teleoperation | Imitation learning from human demonstrations has shown impressive performance in robotics. However, most results focus on table-top manipulation, lacking the mobility and dexterity necessary for generally useful tasks. In this work, we develop a system for imitating mobile manipulation tasks that are bimanual and requi... | Zipeng Fu;Tony Z. Zhao;Chelsea Finn | Stanford University;Stanford University;Google | Poster | main | Mobile Manipulation;Imitation Learning | https://github.com/MarkFzp/mobile-aloha | https://openreview.net/forum?id=FO6tePGRZj | 15 | Mobile ALOHA: Learning Bimanual Mobile Manipulation using Low-Cost Whole-Body Teleoperation
Imitation learning from human demonstrations has shown impressive performance in robotics. However, most results focus on table-top manipulation, lacking the mobility and dexterity necessary for generally useful tasks. In this w... | [
-0.04515749216079712,
-0.02227894403040409,
-0.047368522733449936,
0.043358687311410904,
-0.02389037422835827,
0.002148962579667568,
0.00419018417596817,
0.041035231202840805,
-0.023628048598766327,
0.023478148505091667,
-0.03290313482284546,
-0.019262200221419334,
0.009120501577854156,
0.... | |
corl_2024_G0jqGG8Tta | G0jqGG8Tta | corl | 2,024 | Not All Errors Are Made Equal: A Regret Metric for Detecting System-level Trajectory Prediction Failures | Robot decision-making increasingly relies on data-driven human prediction models when operating around people. While these models are known to mispredict in out-of-distribution interactions, only a subset of prediction errors impact downstream robot performance.
We propose characterizing such ``system-level'' predict... | Kensuke Nakamura;Thomas Tian;Andrea Bajcsy | Carnegie Mellon University;University of California, Berkeley;Carnegie Mellon University | Poster | main | Human-Robot Interaction;Trajectory Prediction;Failure Detection | https://openreview.net/forum?id=G0jqGG8Tta | 1 | Not All Errors Are Made Equal: A Regret Metric for Detecting System-level Trajectory Prediction Failures
Robot decision-making increasingly relies on data-driven human prediction models when operating around people. While these models are known to mispredict in out-of-distribution interactions, only a subset of predict... | [
-0.043198224157094955,
-0.052999190986156464,
-0.03446981683373451,
0.013767584227025509,
-0.022153884172439575,
-0.0220614206045866,
0.013462459668517113,
-0.005737264174968004,
-0.022893579676747322,
0.02664753422141075,
-0.05336903780698776,
-0.019028669223189354,
-0.011261865496635437,
... | ||
corl_2024_G8UcwxNAoD | G8UcwxNAoD | corl | 2,024 | Teaching Robots with Show and Tell: Using Foundation Models to Synthesize Robot Policies from Language and Visual Demonstration | We introduce a modular, neuro-symbolic framework for teaching robots new skills through language and visual demonstration. Our approach, ShowTell, composes a mixture of foundation models to synthesize robot manipulation programs that are easy to interpret and generalize across a wide range of tasks and environments. Sh... | Michael Murray;Abhishek Gupta;Maya Cakmak | University of Washington;University of Washington;University of Washington, Seattle | Poster | main | learning from demonstration;language model planning;neuro-symbolic reasoning | https://openreview.net/forum?id=G8UcwxNAoD | 2 | Teaching Robots with Show and Tell: Using Foundation Models to Synthesize Robot Policies from Language and Visual Demonstration
We introduce a modular, neuro-symbolic framework for teaching robots new skills through language and visual demonstration. Our approach, ShowTell, composes a mixture of foundation models to sy... | [
-0.02947467379271984,
-0.02433808520436287,
0.010600049048662186,
0.004034555982798338,
-0.019519032910466194,
-0.013485876843333244,
0.0031309835612773895,
0.027662858366966248,
0.004429139196872711,
0.0071725440211594105,
-0.04127948358654976,
0.007279945537447929,
0.006247958168387413,
... | ||
corl_2024_GGuNkjQSrk | GGuNkjQSrk | corl | 2,024 | Action Space Design in Reinforcement Learning for Robot Motor Skills | Practitioners often rely on intuition to select action spaces for learning. The choice can substantially impact final performance even when choosing among configuration-space representations such as joint position, velocity, and torque commands. We examine action space selection considering a wheeled-legged robot, a qu... | Julian Eßer;Gabriel B. Margolis;Oliver Urbann;Sören Kerner;Pulkit Agrawal | Fraunhofer IML;;Fraunhofer IML;;Massachusetts Institute of Technology | Poster | main | Reinforcement Learning;Action Spaces;Sim-to-Real | https://openreview.net/forum?id=GGuNkjQSrk | 1 | Action Space Design in Reinforcement Learning for Robot Motor Skills
Practitioners often rely on intuition to select action spaces for learning. The choice can substantially impact final performance even when choosing among configuration-space representations such as joint position, velocity, and torque commands. We ex... | [
-0.0536654070019722,
-0.02753252536058426,
-0.017016759142279625,
0.014650250785052776,
-0.032615453004837036,
-0.005593982990831137,
0.004940200597047806,
0.003611917607486248,
0.05138177052140236,
0.010303978808224201,
-0.00026588686159811914,
-0.015396115370094776,
-0.020534291863441467,
... | ||
corl_2024_GVX6jpZOhU | GVX6jpZOhU | corl | 2,024 | RoboPoint: A Vision-Language Model for Spatial Affordance Prediction in Robotics | From rearranging objects on a table to putting groceries into shelves, robots must plan precise action points to perform tasks accurately and reliably. In spite of the recent adoption of vision language models (VLMs) to control robot behavior, VLMs struggle to precisely articulate robot actions using language. We intro... | Wentao Yuan;Jiafei Duan;Valts Blukis;Wilbert Pumacay;Ranjay Krishna;Adithyavairavan Murali;Arsalan Mousavian;Dieter Fox | University of Washington, Seattle;NVIDIA;NVIDIA;Universidad Nacional de Ingeniería;University of Washington;;NVIDIA;Department of Computer Science | Poster | main | Foundation Model;Affordance Prediction;Open-world Manipulation | https://github.com/wentaoyuan/RoboPoint | https://openreview.net/forum?id=GVX6jpZOhU | 53 | RoboPoint: A Vision-Language Model for Spatial Affordance Prediction in Robotics
From rearranging objects on a table to putting groceries into shelves, robots must plan precise action points to perform tasks accurately and reliably. In spite of the recent adoption of vision language models (VLMs) to control robot behav... | [
-0.03870873153209686,
-0.026715250685811043,
0.010382143780589104,
0.018127551302313805,
-0.029828062281012535,
0.0010414187563583255,
0.00782322883605957,
0.0023689414374530315,
0.001032263389788568,
0.0096130957826972,
-0.03420431166887283,
-0.0045547764748334885,
0.034405726939439774,
0... | |
corl_2024_HlxRd529nG | HlxRd529nG | corl | 2,024 | Detect Everything with Few Examples | Few-shot object detection aims at detecting novel categories given only a few example images. It is a basic skill for a robot to perform tasks in open environments. Recent methods focus on finetuning strategies, with complicated procedures that prohibit a wider application. In this paper, we introduce DE-ViT, a few-sho... | Xinyu Zhang;Yuhan Liu;Yuting Wang;Abdeslam Boularias | Rutgers University;Rutgers University;Amazon;, Rutgers University | Poster | main | Robot Vision;Object Detection and Recognition;Few-shot Learning | http://github.com/mlzxy/devit | https://openreview.net/forum?id=HlxRd529nG | 25 | Detect Everything with Few Examples
Few-shot object detection aims at detecting novel categories given only a few example images. It is a basic skill for a robot to perform tasks in open environments. Recent methods focus on finetuning strategies, with complicated procedures that prohibit a wider application. In this p... | [
-0.03763836249709129,
-0.0371696874499321,
-0.019035493955016136,
0.033276062458753586,
0.0031838486902415752,
-0.017782684415578842,
-0.019558249041438103,
0.026372089982032776,
0.008607432246208191,
0.01053621806204319,
-0.04639901593327522,
-0.016917435452342033,
-0.00213608518242836,
0... | |
corl_2024_IcOrwlXzMi | IcOrwlXzMi | corl | 2,024 | VLM-Grounder: A VLM Agent for Zero-Shot 3D Visual Grounding | 3D visual grounding is crucial for robots, requiring integration of natural language and 3D scene understanding. Traditional methods depend on supervised learning with 3D point clouds are limited by scarce datasets. Recently zero-shot methods leveraging LLMs have been proposed to address the data issue. While effective... | Runsen Xu;Zhiwei Huang;Tai Wang;Yilun Chen;Jiangmiao Pang;Dahua Lin | The Chinese University of Hong Kong;Zhejiang University;Shanghai Artificial Intelligence Laboratory;Shanghai AI Laboratory ;The Chinese University of Hong Kong;Shanghai AI Laboratory | Poster | main | 3D Visual Grounding;VLM Agent;Zero-Shot | https://github.com/OpenRobotLab/VLM-Grounder | https://openreview.net/forum?id=IcOrwlXzMi | 9 | VLM-Grounder: A VLM Agent for Zero-Shot 3D Visual Grounding
3D visual grounding is crucial for robots, requiring integration of natural language and 3D scene understanding. Traditional methods depend on supervised learning with 3D point clouds are limited by scarce datasets. Recently zero-shot methods leveraging LLMs h... | [
-0.08284973353147507,
-0.023691989481449127,
-0.015620100311934948,
0.016640372574329376,
-0.017814137041568756,
0.0271049365401268,
0.005751447286456823,
0.0028757236432284117,
0.00455059576779604,
-0.012812093831598759,
-0.03925791755318642,
-0.0043068137019872665,
0.03315433859825134,
0... | |
corl_2024_InT87E5sr4 | InT87E5sr4 | corl | 2,024 | Dreamitate: Real-World Visuomotor Policy Learning via Video Generation | A key challenge in manipulation is learning a policy that can robustly generalize to diverse visual environments. A promising mechanism for learning robust policies is to leverage video generative models, which are pretrained on large-scale datasets of internet videos. In this paper, we propose a visuomotor policy lear... | Junbang Liang;Ruoshi Liu;Ege Ozguroglu;Sruthi Sudhakar;Achal Dave;Pavel Tokmakov;Shuran Song;Carl Vondrick | Columbia University;Columbia University;Columbia University;Columbia University;Toyota Research Institute;Toyota Research Institute;Stanford University;Columbia University | Poster | main | Imitation Learning;Visuomotor Policy;Video Generation | https://dreamitate.cs.columbia.edu/ | https://openreview.net/forum?id=InT87E5sr4 | 26 | Dreamitate: Real-World Visuomotor Policy Learning via Video Generation
A key challenge in manipulation is learning a policy that can robustly generalize to diverse visual environments. A promising mechanism for learning robust policies is to leverage video generative models, which are pretrained on large-scale datasets... | [
-0.06054406613111496,
-0.024151258170604706,
0.0005628764047287405,
0.01089571975171566,
-0.04609018564224243,
-0.01533881388604641,
-0.008485203608870506,
0.04026438668370247,
0.0023390769492834806,
0.021219922229647636,
-0.011513328179717064,
-0.020464042201638222,
0.021367410197854042,
... | |
corl_2024_IsZb0wT3Kw | IsZb0wT3Kw | corl | 2,024 | ANAVI: Audio Noise Awareness using Visual of Indoor environments for NAVIgation | We propose Audio Noise Awareness using Visuals of Indoors for NAVIgation for quieter robot path planning. While humans are naturally aware of the noise they make and its impact on those around them, robots currently lack this awareness.
A key challenge in achieving audio awareness for robots is estimating how loud wi... | Vidhi Jain;Rishi Veerapaneni;Yonatan Bisk | Google;;Meta | Poster | main | Robots;Acoustic Noise;Vision;Learning | https://github.com/vidhiJain/anavi | https://openreview.net/forum?id=IsZb0wT3Kw | 0 | ANAVI: Audio Noise Awareness using Visual of Indoor environments for NAVIgation
We propose Audio Noise Awareness using Visuals of Indoors for NAVIgation for quieter robot path planning. While humans are naturally aware of the noise they make and its impact on those around them, robots currently lack this awareness.
A... | [
-0.0473044328391552,
-0.02437174692749977,
-0.011816883459687233,
0.047710321843624115,
-0.02994348481297493,
0.013034548610448837,
0.005271935369819403,
0.009155548177659512,
0.01702885888516903,
0.036400798708200455,
-0.021401382982730865,
-0.021050842478871346,
-0.0020755650475621223,
-... | |
corl_2024_Isp19rFFV4 | Isp19rFFV4 | corl | 2,024 | Multi-Strategy Deployment-Time Learning and Adaptation for Navigation under Uncertainty | We present an approach for performant point-goal navigation in unfamiliar partially-mapped environments. When deployed, our robot runs multiple strategies for deployment-time learning and visual domain adaptation in parallel and quickly selects the best-performing among them. Choosing between policies as they are learn... | Abhishek Paudel;Xuesu Xiao;Gregory J. Stein | George Mason University;George Mason University;George Mason University | Poster | main | policy selection;domain adaptation;navigation under uncertainty | https://openreview.net/forum?id=Isp19rFFV4 | 2 | Multi-Strategy Deployment-Time Learning and Adaptation for Navigation under Uncertainty
We present an approach for performant point-goal navigation in unfamiliar partially-mapped environments. When deployed, our robot runs multiple strategies for deployment-time learning and visual domain adaptation in parallel and qui... | [
-0.03185751661658287,
-0.02208687923848629,
-0.03767871856689453,
0.03674283251166344,
-0.03496465086936951,
-0.007262465078383684,
0.03704231604933739,
-0.02534375712275505,
0.029237037524580956,
0.041553281247615814,
-0.032886989414691925,
-0.028151411563158035,
-0.007136120926588774,
0.... | ||
corl_2024_IssXUYvVTg | IssXUYvVTg | corl | 2,024 | MaIL: Improving Imitation Learning with Selective State Space Models | This work introduces Mamba Imitation Learning (MaIL), a novel imitation learning (IL) architecture that offers a computationally efficient alternative to state-of-the-art (SoTA) Transformer policies. Transformer-based policies have achieved remarkable results due to their ability in handling human-recorded data with in... | Xiaogang Jia;Qian Wang;Atalay Donat;Bowen Xing;Ge Li;Hongyi Zhou;Onur Celik;Denis Blessing;Rudolf Lioutikov;Gerhard Neumann | Karlsruher Institut für Technologie;Karlsruher Institut für Technologie;Karlsruher Institut für Technologie;Karlsruher Institut für Technologie;Karlsruhe Institute of Technology;Karlsruher Institut für Technologie;Karlsruhe Institute of Technology;Karlsruher Institut für Technologie;Karlsruher Institut für Technologie;... | Poster | main | Imitation Learning;Sequence Models;Denoising Diffusion Policies | https://github.com/ALRhub/MaIL | https://openreview.net/forum?id=IssXUYvVTg | 7 | MaIL: Improving Imitation Learning with Selective State Space Models
This work introduces Mamba Imitation Learning (MaIL), a novel imitation learning (IL) architecture that offers a computationally efficient alternative to state-of-the-art (SoTA) Transformer policies. Transformer-based policies have achieved remarkable... | [
-0.06365672498941422,
-0.01910075545310974,
-0.009905479848384857,
0.009793342091143131,
-0.02992202527821064,
0.009615791030228138,
0.01953061670064926,
0.06085328757762909,
0.009400861337780952,
0.027062518522143364,
-0.039659298956394196,
-0.046424929052591324,
-0.006924491375684738,
0.... | |
corl_2024_JScswMfEQ0 | JScswMfEQ0 | corl | 2,024 | Mobility VLA: Multimodal Instruction Navigation with Long-Context VLMs and Topological Graphs | An elusive goal in navigation research is to build an intelligent agent that can understand multimodal instructions including natural language and image, and perform useful navigation. To achieve this, we study a widely useful category of navigation tasks we call Multimodal Instruction Navigation with demonstration Tou... | Zhuo Xu;Hao-Tien Lewis Chiang;Zipeng Fu;Mithun George Jacob;Tingnan Zhang;Tsang-Wei Edward Lee;Wenhao Yu;Connor Schenck;David Rendleman;Dhruv Shah;Fei Xia;Jasmine Hsu;Jonathan Hoech;Pete Florence;Sean Kirmani;Sumeet Singh;Vikas Sindhwani;Carolina Parada;Chelsea Finn;Peng Xu;Sergey Levine;Jie Tan | Google DeepMind;Google Deepmind;Stanford University;;Google;;Google;;;UC Berkeley;Google;New York University;;Google;Google DeepMind;Google Brain Robotics;Google;Google DeepMind;Google;Google;Google;Google | Poster | main | vision-language navigation;multimodal foundation models;long-context reasoning | https://openreview.net/forum?id=JScswMfEQ0 | 20 | Mobility VLA: Multimodal Instruction Navigation with Long-Context VLMs and Topological Graphs
An elusive goal in navigation research is to build an intelligent agent that can understand multimodal instructions including natural language and image, and perform useful navigation. To achieve this, we study a widely useful... | [
-0.08175727725028992,
-0.042032770812511444,
0.011801938526332378,
0.024292632937431335,
-0.0395011231303215,
0.00802308414131403,
-0.046686530113220215,
-0.026563668623566628,
0.0021430570632219315,
0.011141104623675346,
-0.0031319812405854464,
-0.004267498850822449,
0.03534996882081032,
... | ||
corl_2024_JZzaRY8m8r | JZzaRY8m8r | corl | 2,024 | KOI: Accelerating Online Imitation Learning via Hybrid Key-state Guidance | Online Imitation Learning methods struggle with the gap between extensive online exploration space and limited expert trajectories, which hinder efficient exploration due to inaccurate task-aware reward estimation.
Inspired by the findings from cognitive neuroscience that task decomposition could facilitate cogniti... | Jingxian Lu;Wenke Xia;Dong Wang;Zhigang Wang;Bin Zhao;Di Hu;Xuelong Li | Renmin University of China;Renmin University of China;Shanghai AI Laboratory;Northwest Polytechnical University Xi'an;Renmin University of China;Northwestern Polytechnical University;Shanghai AI Lab | Poster | main | Online Imitation Learning; Robotic Manipulation | https://github.com/GeWu-Lab/Keystate_Online_Imitation | https://openreview.net/forum?id=JZzaRY8m8r | 0 | KOI: Accelerating Online Imitation Learning via Hybrid Key-state Guidance
Online Imitation Learning methods struggle with the gap between extensive online exploration space and limited expert trajectories, which hinder efficient exploration due to inaccurate task-aware reward estimation.
Inspired by the findings fr... | [
-0.055256687104701996,
-0.040260013192892075,
-0.03521834686398506,
-0.012567504309117794,
-0.004347293172031641,
0.03397167846560478,
0.005357918329536915,
0.01210917066782713,
-0.013328338041901588,
0.020753340795636177,
-0.016582505777478218,
-0.05720001831650734,
-0.011595836840569973,
... | |
corl_2024_KAzku0Uyh1 | KAzku0Uyh1 | corl | 2,024 | Object-Centric Dexterous Manipulation from Human Motion Data | Manipulating objects to achieve desired goal states is a basic but important skill for dexterous manipulation. Human hand motions demonstrate proficient manipulation capability, providing valuable data for training robots with multi-finger hands. Despite this potential, substantial challenges arise due to the embodimen... | Yuanpei Chen;Chen Wang;Yaodong Yang;Karen Liu | PsiRobot;Computer Science Department, Stanford University;Peking University;Computer Science Department, Stanford University | Poster | main | Dexterous Manipulation;Reinforcement Learning;Learning from Human | https://openreview.net/forum?id=KAzku0Uyh1 | 17 | Object-Centric Dexterous Manipulation from Human Motion Data
Manipulating objects to achieve desired goal states is a basic but important skill for dexterous manipulation. Human hand motions demonstrate proficient manipulation capability, providing valuable data for training robots with multi-finger hands. Despite this... | [
-0.034400295466184616,
-0.04207150638103485,
-0.05489999055862427,
0.02118363231420517,
-0.025009995326399803,
-0.04650786891579628,
0.01036075595766306,
-0.001600094372406602,
-0.006834771484136581,
0.010194391943514347,
-0.02805999480187893,
-0.025342723354697227,
0.013336815871298313,
-... | ||
corl_2024_KPcX4jetMw | KPcX4jetMw | corl | 2,024 | Reasoning Grasping via Multimodal Large Language Model | Despite significant progress in robotic systems for operation within human-centric environments, existing models still heavily rely on explicit human commands to identify and manipulate specific objects. This limits their effectiveness in environments where understanding and acting on implicit human intentions are cruc... | Shiyu Jin;JINXUAN XU;Yutian Lei;Liangjun Zhang | ;Rutgers University, New Brunswick;;Research, Baidu | Poster | main | Robotics Grasping;Multimodal Large Language Model | https://openreview.net/forum?id=KPcX4jetMw | 25 | Reasoning Grasping via Multimodal Large Language Model
Despite significant progress in robotic systems for operation within human-centric environments, existing models still heavily rely on explicit human commands to identify and manipulate specific objects. This limits their effectiveness in environments where underst... | [
-0.05841958895325661,
-0.020245151594281197,
0.007012965157628059,
-0.012176972813904285,
-0.049903176724910736,
0.003361741080880165,
-0.011439257301390171,
-0.022953221574425697,
-0.018872439861297607,
-0.019180599600076675,
-0.020861471071839333,
-0.014016592875123024,
0.02069338411092758... | ||
corl_2024_KULBk5q24a | KULBk5q24a | corl | 2,024 | CoViS-Net: A Cooperative Visual Spatial Foundation Model for Multi-Robot Applications | Autonomous robot operation in unstructured environments is often underpinned by spatial understanding through vision. Systems composed of multiple concurrently operating robots additionally require access to frequent, accurate and reliable pose estimates. Classical vision-based methods to regress relative pose are comm... | Jan Blumenkamp;Steven Morad;Jennifer Gielis;Amanda Prorok | University of Cambridge;University of Cambridge;; | Poster | main | Multi-Robot Systems;Robot Perception;Foundation Models | https://github.com/proroklab/CoViS-Net | https://openreview.net/forum?id=KULBk5q24a | 4 | CoViS-Net: A Cooperative Visual Spatial Foundation Model for Multi-Robot Applications
Autonomous robot operation in unstructured environments is often underpinned by spatial understanding through vision. Systems composed of multiple concurrently operating robots additionally require access to frequent, accurate and rel... | [
-0.05140884220600128,
-0.022277165204286575,
-0.03504645824432373,
0.04042688012123108,
-0.0009253360913135111,
0.020195014774799347,
0.020342422649264336,
0.00415739044547081,
0.04097966477274895,
0.017983881756663322,
-0.03976354002952576,
-0.008637238293886185,
0.013718237169086933,
0.0... | |
corl_2024_KXsropnmNI | KXsropnmNI | corl | 2,024 | Transferable Tactile Transformers for Representation Learning Across Diverse Sensors and Tasks | This paper presents T3: Transferable Tactile Transformers, a framework for tactile representation learning that scales across multi-sensors and multi-tasks.T3 is designed to overcome the contemporary issue that camera-based tactile sensing is extremely heterogeneous, i.e. sensors are built into different form factors, ... | Jialiang Zhao;Yuxiang Ma;Lirui Wang;Edward Adelson | Massachusetts Institute of Technology;Massachusetts Institute of Technology;Massachusetts Institute of Technology;Massachusetts Institute of Technology | Poster | main | Tactile Sensing;Representation Learning;Heterogeneous Learning;Robot Manipulation;Robot Learning | https://github.com/alanzjl/t3 | https://openreview.net/forum?id=KXsropnmNI | 18 | Transferable Tactile Transformers for Representation Learning Across Diverse Sensors and Tasks
This paper presents T3: Transferable Tactile Transformers, a framework for tactile representation learning that scales across multi-sensors and multi-tasks.T3 is designed to overcome the contemporary issue that camera-based t... | [
-0.023850975558161736,
-0.011649435386061668,
-0.04814363643527031,
0.05962743982672691,
0.02574653923511505,
0.0033563438337296247,
-0.008824492804706097,
0.008286189287900925,
0.0045755780301988125,
0.012965287081897259,
0.00726479385048151,
0.037800852209329605,
0.02447669580578804,
0.0... | |
corl_2024_KcW31O0PtL | KcW31O0PtL | corl | 2,024 | Hint-AD: Holistically Aligned Interpretability in End-to-End Autonomous Driving | End-to-end architectures in autonomous driving (AD) face a significant challenge in interpretability, impeding human-AI trust. Human-friendly natural language has been explored for tasks such as driving explanation and 3D captioning. However, previous works primarily focused on the paradigm of declarative interpretabil... | Kairui Ding;Boyuan Chen;Yuchen Su;Huan-ang Gao;Bu Jin;Chonghao Sima;Xiaohui Li;Wuqiang Zhang;Paul Barsch;Hongyang Li;Hao Zhao | Tsinghua University;Tsinghua University;Tsinghua University;Tsinghua University;;Shanghai AI Lab;Dalian University of Technology;Karlsruher Institut für Technologie;Technische Universität Dresden;Shanghai AI Lab;Peking University | Poster | main | Interpretability;Language alignment;Autonomous driving | https://openreview.net/forum?id=KcW31O0PtL | 5 | Hint-AD: Holistically Aligned Interpretability in End-to-End Autonomous Driving
End-to-end architectures in autonomous driving (AD) face a significant challenge in interpretability, impeding human-AI trust. Human-friendly natural language has been explored for tasks such as driving explanation and 3D captioning. Howeve... | [
-0.03900745138525963,
-0.03401169553399086,
-0.007249375805258751,
-0.010056031867861748,
-0.046971164643764496,
-0.023190628737211227,
0.011134451255202293,
-0.004484197590500116,
-0.009816383011639118,
0.02359618991613388,
-0.033956389874219894,
-0.05331264063715935,
0.020075194537639618,
... | ||
corl_2024_KdVLK0Wo5z | KdVLK0Wo5z | corl | 2,024 | PoliFormer: Scaling On-Policy RL with Transformers Results in Masterful Navigators | We present PoliFormer (Policy Transformer), an RGB-only indoor navigation agent trained end-to-end with reinforcement learning at scale that generalizes to the real-world without adaptation despite being trained purely in simulation. PoliFormer uses a foundational vision transformer encoder with a causal transformer de... | Kuo-Hao Zeng;Zichen Zhang;Kiana Ehsani;Rose Hendrix;Jordi Salvador;Alvaro Herrasti;Ross Girshick;Aniruddha Kembhavi;Luca Weihs | Allen Institute for Artificial Intelligence;Allen Institute for Artificial Intelligence;Allen Institute for Artificial Intelligence;Allen Institute for Artificial Intelligence;Allen Institute for AI;Allen Institute for Artificial Intelligence;Allen Institute for Artificial Intelligence;Allen Institute for Artificial In... | Poster | main | Embodied Navigation;On-Policy RL;Transformer Policy | https://openreview.net/forum?id=KdVLK0Wo5z | 16 | PoliFormer: Scaling On-Policy RL with Transformers Results in Masterful Navigators
We present PoliFormer (Policy Transformer), an RGB-only indoor navigation agent trained end-to-end with reinforcement learning at scale that generalizes to the real-world without adaptation despite being trained purely in simulation. Pol... | [
-0.06781663745641708,
-0.0017474069027230144,
-0.03663143143057823,
0.020050298422574997,
-0.021318597719073296,
0.0018220125930383801,
-0.014827894978225231,
-0.013867346569895744,
0.008052759803831577,
0.016534501686692238,
-0.025571124628186226,
-0.02049793303012848,
0.026615606620907784,... | ||
corl_2024_Ke5xrnBFAR | Ke5xrnBFAR | corl | 2,024 | Gameplay Filters: Robust Zero-Shot Safety through Adversarial Imagination | Despite the impressive recent advances in learning-based robot control, ensuring robustness to out-of-distribution conditions remains an open challenge. Safety filters can, in principle, keep arbitrary control policies from incurring catastrophic failures by overriding unsafe actions, but existing solutions for complex... | Duy Phuong Nguyen;Kai-Chieh Hsu;Wenhao Yu;Jie Tan;Jaime Fernández Fisac | Princeton University;Princeton University;Google;Google;Princeton University | Poster | main | Robust Safety;Adversarial Reinforcement Learning;Game Theory | https://openreview.net/forum?id=Ke5xrnBFAR | 7 | Gameplay Filters: Robust Zero-Shot Safety through Adversarial Imagination
Despite the impressive recent advances in learning-based robot control, ensuring robustness to out-of-distribution conditions remains an open challenge. Safety filters can, in principle, keep arbitrary control policies from incurring catastrophic... | [
-0.05416349321603775,
-0.04557614400982857,
-0.02579922042787075,
0.009344782680273056,
-0.005576200317591429,
-0.012713736854493618,
0.016170980408787727,
0.014953509904444218,
-0.006273225415498018,
0.02697022259235382,
-0.002261846326291561,
-0.01891261339187622,
0.0030994380358606577,
... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.