Datasets:
vecId stringlengths 12 23 | id stringlengths 2 13 | conference stringclasses 11
values | year float64 2.02k 2.03k | title stringlengths 6 189 | abstract stringlengths 10 4.74k | author stringlengths 0 7.45k | aff stringlengths 0 7.16k | status stringclasses 11
values | track stringclasses 4
values | keywords stringlengths 0 804 | github stringlengths 0 141 | site stringlengths 0 193 | gsCitation float64 -1 11.1k | arxiv stringlengths 0 12 | text stringlengths 58 4.82k | vector list |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
corl_2023_-3G6_D66Aua | -3G6_D66Aua | corl | 2,023 | Simultaneous Learning of Contact and Continuous Dynamics | Robotic manipulation can greatly benefit from the data efficiency, robustness, and predictability of model-based methods if robots can quickly generate models of novel objects they encounter. This is especially difficult when effects like complex joint friction lack clear first-principles models and are usually ignored... | Bibit Bianchini;Mathew Halm;Michael Posa | University of Pennsylvania;School of Engineering and Applied Science, University of Pennsylvania;University of Pennsylvania | Poster | main | system identification;dynamics learning;contact-rich manipulation | https://github.com/ebianchi/dair_pll | https://openreview.net/forum?id=-3G6_D66Aua | 14 | Simultaneous Learning of Contact and Continuous Dynamics
Robotic manipulation can greatly benefit from the data efficiency, robustness, and predictability of model-based methods if robots can quickly generate models of novel objects they encounter. This is especially difficult when effects like complex joint friction l... | [
-0.04371373727917671,
-0.018341166898608208,
-0.028888268396258354,
0.036552123725414276,
-0.01290950272232294,
-0.016406601294875145,
0.014174411073327065,
0.010109963826835155,
-0.0027111817616969347,
0.028813861310482025,
-0.0283674243837595,
-0.039956215769052505,
-0.0002644913620315492,... | |
corl_2023_-HFJuX1uqs | -HFJuX1uqs | corl | 2,023 | Act3D: 3D Feature Field Transformers for Multi-Task Robotic Manipulation | 3D perceptual representations are well suited for robot manipulation as they easily encode occlusions and simplify spatial reasoning. Many manipulation tasks require high spatial precision in end-effector pose prediction, which typically demands high-resolution 3D feature grids that are computationally expensive to pro... | Theophile Gervet;Zhou Xian;Nikolaos Gkanatsios;Katerina Fragkiadaki | Carnegie Mellon University;Carnegie Mellon University;Carnegie Mellon University;Carnegie Mellon University | Poster | main | Learning from Demonstrations;Manipulation;Transformers | https://github.com/zhouxian/chained-diffuser | https://openreview.net/forum?id=-HFJuX1uqs | 71 | Act3D: 3D Feature Field Transformers for Multi-Task Robotic Manipulation
3D perceptual representations are well suited for robot manipulation as they easily encode occlusions and simplify spatial reasoning. Many manipulation tasks require high spatial precision in end-effector pose prediction, which typically demands h... | [
-0.034714847803115845,
-0.04591735452413559,
-0.02775711379945278,
0.006344087887555361,
-0.0011136757675558329,
-0.020153436809778214,
-0.0030197857413440943,
-0.01837247796356678,
0.03823985531926155,
0.007105378434062004,
-0.01246670912951231,
0.02094702422618866,
0.016083993017673492,
... | |
corl_2023_-K7-1WvKO3F | -K7-1WvKO3F | corl | 2,023 | ViNT: A Foundation Model for Visual Navigation | General-purpose pre-trained models (``foundation models'') have enabled practitioners to produce generalizable solutions for individual machine learning problems with datasets that are significantly smaller than those required for learning from scratch. Such models are typically trained on large and diverse datasets wi... | Dhruv Shah;Ajay Sridhar;Nitish Dashora;Kyle Stachowicz;Kevin Black;Noriaki Hirose;Sergey Levine | UC Berkeley;University of California, Berkeley;University of California, Berkeley;University of California, Berkeley;University of California, Berkeley;Toyota Central R&D Labs., Inc;Google | Oral | main | visual navigation;multi-task learning;planning;generalization | https://github.com/robodhruv/visualnav-transformer | https://openreview.net/forum?id=-K7-1WvKO3F | 154 | ViNT: A Foundation Model for Visual Navigation
General-purpose pre-trained models (``foundation models'') have enabled practitioners to produce generalizable solutions for individual machine learning problems with datasets that are significantly smaller than those required for learning from scratch. Such models are typ... | [
-0.047531452029943466,
-0.057111434638500214,
-0.03872523456811905,
0.0334378220140934,
-0.0211496539413929,
-0.010409018956124783,
0.013614628463983536,
-0.004048463888466358,
0.000013853265045327134,
0.02474214695394039,
-0.014637107960879803,
-0.02997429110109806,
0.00010247817408526316,
... | |
corl_2023_09UL1dCqf2n | 09UL1dCqf2n | corl | 2,023 | Preference learning for guiding the tree search in continuous POMDPs | A robot operating in a partially observable environment must perform sensing actions to achieve a goal, such as clearing the objects in front of a shelf to better localize a target object at the back, and estimate its shape for grasping. A POMDP is a principled framework for enabling robots to perform such information-... | Jiyong Ahn;Sanghyeon Son;Dongryung Lee;Jisu Han;Dongwon Son;Beomjoon Kim | Korea Advanced Institute of Science & Technology;;Korea Advanced Institute of Science & Technology;Korea Advanced Institute of Science & Technology;KAIST;Korea Advanced Institute of Science & Technology | Poster | main | POMDP;Online planning;Guided Search;Preference-based learning | https://openreview.net/forum?id=09UL1dCqf2n | 0 | Preference learning for guiding the tree search in continuous POMDPs
A robot operating in a partially observable environment must perform sensing actions to achieve a goal, such as clearing the objects in front of a shelf to better localize a target object at the back, and estimate its shape for grasping. A POMDP is a ... | [
-0.06318318843841553,
-0.036296725273132324,
-0.02516871690750122,
-0.0005213336553424597,
-0.006226830184459686,
-0.047387391328811646,
0.02313356101512909,
-0.006539572030305862,
0.04074046015739441,
0.02350698411464691,
-0.018521785736083984,
-0.046864598989486694,
-0.009774349629878998,
... | ||
corl_2023_0I3su3mkuL | 0I3su3mkuL | corl | 2,023 | Q-Transformer: Scalable Offline Reinforcement Learning via Autoregressive Q-Functions | In this work, we present a scalable reinforcement learning method for training multi-task policies from large offline datasets that can leverage both human demonstrations and autonomously collected data. Our method uses a Transformer to provide a scalable representation for Q-functions trained via offline temporal diff... | Yevgen Chebotar;Quan Vuong;Karol Hausman;Fei Xia;Yao Lu;Alex Irpan;Aviral Kumar;Tianhe Yu;Alexander Herzog;Karl Pertsch;Keerthana Gopalakrishnan;Julian Ibarz;Ofir Nachum;Sumedh Anand Sontakke;Grecia Salazar;Huong T Tran;Jodilyn Peralta;Clayton Tan;Deeksha Manjunath;Jaspiar Singh;Brianna Zitkovich;Tomas Jackson;Kanishka... | Google;;;Google;Google;Google DeepMind;University of California, Berkeley;Google Brain;Google;University of Southern California;Research, Google;Google;OpenAI;University of Southern California;;;;;;;;;;Google;Google | Poster | main | Reinforcement Learning;Offline RL;Transformers;Q-Learning;Robotic Manipulation | https://openreview.net/forum?id=0I3su3mkuL | 106 | Q-Transformer: Scalable Offline Reinforcement Learning via Autoregressive Q-Functions
In this work, we present a scalable reinforcement learning method for training multi-task policies from large offline datasets that can leverage both human demonstrations and autonomously collected data. Our method uses a Transformer ... | [
-0.052003927528858185,
0.010328659787774086,
-0.01656099408864975,
0.0031161669176071882,
-0.02724103257060051,
0.012779058888554573,
0.04863809421658516,
0.02093472331762314,
0.0022677744273096323,
0.015238704159855843,
-0.0039506894536316395,
0.009316137060523033,
-0.00492391362786293,
0... | ||
corl_2023_0bZaUfELuW | 0bZaUfELuW | corl | 2,023 | Goal Representations for Instruction Following: A Semi-Supervised Language Interface to Control | Our goal is for robots to follow natural language instructions like ``put the towel next to the microwave.'' But getting large amounts of labeled data, i.e. data that contains demonstrations of tasks labeled with the language instruction, is prohibitive. In contrast, obtaining policies that respond to image goals is mu... | Vivek Myers;Andre Wang He;Kuan Fang;Homer Rich Walke;Philippe Hansen-Estruch;Ching-An Cheng;Mihai Jalobeanu;Andrey Kolobov;Anca Dragan;Sergey Levine | University of California, Berkeley;UC Berkeley, University of California, Berkeley;;University of California, Berkeley;;Microsoft Research;Microsoft Research;Microsoft;University of California, Berkeley;Google | Poster | main | Instruction Following;Representation Learning;Manipulation | https://github.com/rail-berkeley/grif_release | https://openreview.net/forum?id=0bZaUfELuW | 32 | Goal Representations for Instruction Following: A Semi-Supervised Language Interface to Control
Our goal is for robots to follow natural language instructions like ``put the towel next to the microwave.'' But getting large amounts of labeled data, i.e. data that contains demonstrations of tasks labeled with the languag... | [
-0.05153782665729523,
-0.009409105405211449,
-0.010113505646586418,
0.009824281558394432,
-0.040715258568525314,
-0.03181462734937668,
-0.01010417565703392,
-0.0037855671253055334,
0.01114911399781704,
0.023455122485756874,
0.011634263209998608,
-0.027112405747175217,
0.029388876631855965,
... | |
corl_2023_0hPkttoGAf | 0hPkttoGAf | corl | 2,023 | RVT: Robotic View Transformer for 3D Object Manipulation | For 3D object manipulation, methods that build an explicit 3D representation perform better than those relying only on camera images. But using explicit 3D representations like voxels comes at large computing cost, adversely affecting scalability. In this work, we propose RVT, a multi-view transformer for 3D manipulati... | Ankit Goyal;Jie Xu;Yijie Guo;Valts Blukis;Yu-Wei Chao;Dieter Fox | NVIDIA;NVIDIA;University of Michigan;NVIDIA;NVIDIA;Department of Computer Science | Oral | main | 3D Manipulation;Multi-View;Transformer | https://github.com/nvlabs/rvt | https://openreview.net/forum?id=0hPkttoGAf | 140 | RVT: Robotic View Transformer for 3D Object Manipulation
For 3D object manipulation, methods that build an explicit 3D representation perform better than those relying only on camera images. But using explicit 3D representations like voxels comes at large computing cost, adversely affecting scalability. In this work, w... | [
-0.035961247980594635,
-0.002318314043805003,
-0.006749579682946205,
0.03733033314347267,
-0.01696750335395336,
0.057428471744060516,
-0.012476909905672073,
0.010249868966639042,
0.021832311525940895,
-0.026286395266652107,
-0.023840298876166344,
-0.009903034195303917,
0.035395361483097076,
... | |
corl_2023_0hQMcWfjG9 | 0hQMcWfjG9 | corl | 2,023 | $\alpha$-MDF: An Attention-based Multimodal Differentiable Filter for Robot State Estimation | "Differentiable Filters are recursive Bayesian estimators that derive the state transition and measu(...TRUNCATED) | Xiao Liu;Yifan Zhou;Shuhei Ikemoto;Heni Ben Amor | "Arizona State University;Arizona State University;Kyushu Institute of Technology;Arizona State Univ(...TRUNCATED) | Poster | main | Differentiable Filters;Sensor Fusion;Multimodal Learning | https://github.com/ir-lab/alpha-MDF | https://openreview.net/forum?id=0hQMcWfjG9 | 8 | "$\\alpha$-MDF: An Attention-based Multimodal Differentiable Filter for Robot State Estimation\nDiff(...TRUNCATED) | [-0.008578325621783733,-0.037264786660671234,-0.05251111835241318,-0.02297814190387726,-0.0098684690(...TRUNCATED) | |
corl_2023_0mRSANSzEK | 0mRSANSzEK | corl | 2,023 | Improving Behavioural Cloning with Positive Unlabeled Learning | "Learning control policies offline from pre-recorded datasets is a promising avenue for solving chal(...TRUNCATED) | "Qiang Wang;Robert McCarthy;David Cordova Bulens;Kevin McGuinness;Noel E. O’Connor;Francisco Rolda(...TRUNCATED) | "University College Dublin;;;Dublin City University;;Insight Centre for Data Analytics;Max Planck In(...TRUNCATED) | Poster | main | Offline policy learning;positive unlabeled learning;behavioural cloning | https://openreview.net/forum?id=0mRSANSzEK | 8 | "Improving Behavioural Cloning with Positive Unlabeled Learning\nLearning control policies offline f(...TRUNCATED) | [-0.08705994486808777,-0.03501727804541588,-0.01797330565750599,0.014562653377652168,-0.032024826854(...TRUNCATED) | ||
corl_2023_0o2JgvlzMUc | 0o2JgvlzMUc | corl | 2,023 | Deception Game: Closing the Safety-Learning Loop in Interactive Robot Autonomy | "An outstanding challenge for the widespread deployment of robotic systems like autonomous vehicles (...TRUNCATED) | Haimin Hu;Zixu Zhang;Kensuke Nakamura;Andrea Bajcsy;Jaime Fernández Fisac | "Toyota Research Institute;Princeton University;Princeton University;Princeton University;University(...TRUNCATED) | Poster | main | Learning-Aware Safety Analysis;Active Information Gathering;Adversarial Reinforcement Learning | https://openreview.net/forum?id=0o2JgvlzMUc | 16 | "Deception Game: Closing the Safety-Learning Loop in Interactive Robot Autonomy\nAn outstanding chal(...TRUNCATED) | [-0.039126694202423096,-0.019563347101211548,-0.008261033333837986,0.019830292090773582,-0.043016482(...TRUNCATED) |
End of preview. Expand in Data Studio
Robotics Papers Vector Database
Semantic search over 63,381 academic papers from 30 conference-year combinations in robotics, CV, and ML.
Contents
- Embeddings: BAAI/bge-m3 (1024 dimensions) via SiliconFlow
- Conferences: CoRL, CVPR, ECCV, ICCV, ICLR, ICML, ICRA, IROS, NeurIPS, RSS, WACV (2023-2026)
- Fields: title, abstract, author, conference, year, arxiv, github, citations, keywords
- Format: LanceDB (compacted, single fragment)
Usage
Remote query (no download needed)
import lancedb
db = lancedb.connect("hf://datasets/Litian2002/robotics-papers-vecdb/lancedb")
table = db.open_table("papers")
print(f"Papers: {table.count_rows()}")
# Semantic search (requires embedding your query first)
# See the companion repo for the full search pipeline
Local usage
git clone https://huggingface.co/datasets/Litian2002/robotics-papers-vecdb
# or use huggingface_hub to download
With vec-db CLI
# Clone the tool repo
git clone <vec-db-repo-url>
cd vec-db
VECDB_LANCE_DIR=/path/to/downloaded/lancedb npx tsx src/cli.ts search "robot grasping"
Schema
| Column | Type | Description |
|---|---|---|
| vecId | string | {conf}_{year}_{id} compound key |
| title | string | Paper title |
| abstract | string | Paper abstract |
| author | string | Authors (semicolon-separated) |
| conference | string | Conference name |
| year | float | Publication year |
| arxiv | string | arXiv ID |
| github | string | GitHub repo URL |
| gsCitation | float | Citation count (from OpenAlex) |
| vector | float[1024] | bge-m3 embedding |
- Downloads last month
- 32