metadata
license: bsd-3-clause
pipeline_tag: robotics
tags:
- ballnet
- metaball
- multimodal
- onnx
- pytorch
library_name: transformers
datasets:
- asRobotics/ballnet-100k
Model Card for BallNet
Table of Contents
Model Description
BallNet is an MLP model designed for the Metaball. It can predict both 6D force and 3D shape (mesh nodes) from the 6D motion of the ball.
Try it out on the Spaces demo.
- Developer: Xudong Han, Tianyu Wu, Fang Wan, and Chaoyang Song.
- Model type: MLP
- License: BSD-3-Clause
Intended Use
This model is intended for researchers and developers working in robotics and tactile sensing. It can be used to enhance the capabilities of robotic systems by providing accurate predictions of force and shape based on tactile data.
To load the model:
from transformers import AutoModel
model = AutoModel.from_pretrained("asRobotics/ballnet", trust_remote_code=True)
x = torch.zeros((1, 6)) # Example input: batch size of 1, 6D motion
output = model(x)
Or to load the ONNX version:
# Example code to load onnx
import onnxruntime as ort
import numpy as np
from huggingface_hub import hf_hub_download
onnx_model_path = hf_hub_download("asRobotics/ballnet", filename="model.onnx")
ort_session = ort.InferenceSession(onnx_model_path)
# Example input
x = np.zeros((1, 6), dtype=np.float32) # Batch size of 1, 6D motion
output = ort_session.run(None, {"motion": x})
Training Data
The model was trained on the BallNet-100K dataset, which includes a variety of motion, force, and shape data collected by finite element simulations.
Citation
If you use this model in your research, please cite the following papers:
@article{liu2024proprioceptive,
title={Proprioceptive learning with soft polyhedral networks},
author={Liu, Xiaobo and Han, Xudong and Hong, Wei and Wan, Fang and Song, Chaoyang},
journal={The International Journal of Robotics Research},
volume = {43},
number = {12},
pages = {1916-1935},
year = {2024},
publisher={SAGE Publications Sage UK: London, England},
doi = {10.1177/02783649241238765}
}