Add model card for Robot Learning tutorial model
#1
by
nielsr
HF Staff
- opened
README.md
ADDED
|
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: mit
|
| 3 |
+
pipeline_tag: robotics
|
| 4 |
+
library_name: lerobot
|
| 5 |
+
tags:
|
| 6 |
+
- action-chunking-with-transformers
|
| 7 |
+
---
|
| 8 |
+
|
| 9 |
+
# Robot Learning: A Tutorial - ACT Model
|
| 10 |
+
|
| 11 |
+
This repository contains a model checkpoint (`act-resnet18-upside-down-side-v0.1`) associated with the paper "[Robot Learning: A Tutorial](https://huggingface.co/papers/2510.12403)". This tutorial navigates the landscape of modern robot learning, providing practical examples using the `lerobot` library developed by Hugging Face. The model provided here is an instance of the Action Chunking with Transformers (ACT) framework.
|
| 12 |
+
|
| 13 |
+
- **Paper:** [Robot Learning: A Tutorial](https://huggingface.co/papers/2510.12403)
|
| 14 |
+
- **GitHub Repository:** [https://github.com/fracapuano/robot-learning-tutorial](https://github.com/fracapuano/robot-learning-tutorial)
|
| 15 |
+
- **Project Page (Hugging Face Space):** [https://huggingface.co/spaces/lerobot/robot-learning-tutorial](https://huggingface.co/spaces/lerobot/robot-learning-tutorial)
|
| 16 |
+
|
| 17 |
+
## Abstract
|
| 18 |
+
Robot learning is at an inflection point, driven by rapid advancements in machine learning and the growing availability of large-scale robotics data. This shift from classical, model-based methods to data-driven, learning-based paradigms is unlocking unprecedented capabilities in autonomous systems. This tutorial navigates the landscape of modern robot learning, charting a course from the foundational principles of Reinforcement Learning and Behavioral Cloning to generalist, language-conditioned models capable of operating across diverse tasks and even robot embodiments. This work is intended as a guide for researchers and practitioners, and our goal is to equip the reader with the conceptual understanding and practical tools necessary to contribute to developments in robot learning, with ready-to-use examples implemented in $\texttt{lerobot}$.
|
| 19 |
+
|
| 20 |
+
## Model Details
|
| 21 |
+
This model checkpoint is configured as an Action Chunking with Transformers (ACT) policy, as defined in its `config.json`. It utilizes a `resnet18` vision backbone for processing visual observations and is designed for robotic action prediction.
|
| 22 |
+
|
| 23 |
+
- **Model Type:** `act` (Action Chunking with Transformers)
|
| 24 |
+
- **Vision Backbone:** `resnet18`
|
| 25 |
+
- **Input Features:**
|
| 26 |
+
- `observation.state`: State vector of shape `[6]`
|
| 27 |
+
- `observation.images.up`: Visual input from an 'up' camera, shape `[3, 480, 640]`
|
| 28 |
+
- `observation.images.side`: Visual input from a 'side' camera, shape `[3, 480, 640]`
|
| 29 |
+
- **Output Features:**
|
| 30 |
+
- `action`: Action vector of shape `[6]`
|
| 31 |
+
- **Uses VAE:** `true`
|
| 32 |
+
|
| 33 |
+
## Usage
|
| 34 |
+
For detailed instructions on installation, training, and utilizing robot learning models within the `lerobot` ecosystem, please refer to the comprehensive [GitHub repository](https://github.com/fracapuano/robot-learning-tutorial) of the "Robot Learning: A Tutorial". The repository provides various code examples and guidance for implementing and experimenting with such models.
|
| 35 |
+
|
| 36 |
+
## License
|
| 37 |
+
The code examples within the "Robot Learning: A Tutorial" GitHub repository are licensed under the [MIT License](https://opensource.org/licenses/MIT). This model artifact is associated with that codebase.
|