RLinf-logo

RLinf: Reinforcement Learning Infrastructure for Agentic AI

[RLinf](https://github.com/RLinf/RLinf) is a flexible and scalable open-source infrastructure designed for post-training foundation models (LLMs, VLMs, VLAs) via reinforcement learning. The 'inf' in RLinf stands for Infrastructure, highlighting its role as a robust backbone for next-generation training. It also stands for Infinite, symbolizing the system’s support for open-ended learning, continuous generalization, and limitless possibilities in intelligence development.
RLinf-overview
## Model Description This repository provides a PyTorch ResNet-10 pretrained checkpoint, converted from the original JAX implementation released by the [SeRL](https://github.com/rail-berkeley/serl) project. The original JAX weights can be found here: https://github.com/rail-berkeley/serl/releases/download/resnet10/resnet10_params.pkl Within RLinf, this checkpoint is used as the pretrained CNN ResNet encoder for real-world embodied learning experiments. ## How to Use Please integrate the provided model with the [RLinf](https://github.com/RLinf/RLinf) codebase. To do so, modify the ``extra_config.pretrained_ckpt_path`` as the path for this model in the configuration file ``examples/embodiment/config/model/cnn_policy.yaml``. ## License This code repository and the model weights are licensed under the MIT License.