Update README.md
Browse files
README.md
CHANGED
|
@@ -1,3 +1,59 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: cdla-permissive-2.0
|
| 3 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: cdla-permissive-2.0
|
| 3 |
+
---
|
| 4 |
+
# MoCapAct Dataset
|
| 5 |
+
Control of simulated humanoid characters is a challenging benchmark for sequential decision-making methods, as it assesses a policy’s ability to drive an inherently unstable, discontinuous, and high-dimensional physical system. Motion capture (MoCap) data can be very helpful in learning sophisticated locomotion policies by teaching a humanoid agent low-level skills (e.g., standing, walking, and running) that can then be used to generate high-level behaviors. However, even with MoCap data, controlling simulated humanoids remains very hard, because this data offers only kinematic information. Finding physical control inputs to realize the MoCap-demonstrated motions has required methods like reinforcement learning that need large amounts of compute, which has effectively served as a barrier to entry for this exciting research direction.
|
| 6 |
+
|
| 7 |
+
In an effort to broaden participation and facilitate evaluation of ideas in humanoid locomotion research, we are releasing MoCapAct (Motion Capture with Actions), a library of high-quality pre-trained agents that can track over three hours of MoCap data for a simulated humanoid in the `dm_control` physics-based environment and rollouts from these experts containing proprioceptive observations and actions. MoCapAct allows researchers to sidestep the computationally intensive task of training low-level control policies from MoCap data and instead use MoCapAct's expert agents and demonstrations for learning advanced locomotion behaviors. It also allows improving on our low-level policies by using them and their demonstration data as a starting point.
|
| 8 |
+
|
| 9 |
+
In our work, we use MoCapAct to train a single hierarchical policy capable of tracking the entire MoCap dataset within `dm_control`.
|
| 10 |
+
We then re-use the learned low-level component to efficiently learn other high-level tasks.
|
| 11 |
+
Finally, we use MoCapAct to train an autoregressive GPT model and show that it can perform natural motion completion given a motion prompt.
|
| 12 |
+
We encourage the reader to visit our [project website](https://microsoft.github.io/MoCapAct/) to see videos of our results as well as get links to our paper and code.
|
| 13 |
+
|
| 14 |
+
## File Structure
|
| 15 |
+
|
| 16 |
+
The file structure of the dataset is:
|
| 17 |
+
```
|
| 18 |
+
├── all
|
| 19 |
+
│ ├── large
|
| 20 |
+
│ │ ├── large_1.tar.gz
|
| 21 |
+
│ │ ├── large_2.tar.gz
|
| 22 |
+
| │ ...
|
| 23 |
+
│ │ └── large_43.tar.gz
|
| 24 |
+
│ └── small
|
| 25 |
+
│ ├── small_1.tar.gz
|
| 26 |
+
│ ├── small_2.tar.gz
|
| 27 |
+
│ └── small_3.tar.gz
|
| 28 |
+
│
|
| 29 |
+
├── sample
|
| 30 |
+
│ ├── large.tar.gz
|
| 31 |
+
│ └── small.tar.gz
|
| 32 |
+
│
|
| 33 |
+
└── videos
|
| 34 |
+
├── full_clip_videos.tar.gz
|
| 35 |
+
└── snippet_videos.tar.gz
|
| 36 |
+
```
|
| 37 |
+
|
| 38 |
+
## MoCapAct Dataset Tarball Files
|
| 39 |
+
The dataset tarball files have the following structure:
|
| 40 |
+
- `all/small/small_*.tar.gz`: Contains HDF5 files with 20 rollouts per snippet. Due to file size limitations, we split the rollouts among multiple tarball files.
|
| 41 |
+
- `all/large/large_*.tar.gz`: Contains HDF5 files with 200 rollouts per snippet. Due to file size limitations, we split the rollouts among multiple tarball files.
|
| 42 |
+
- `sample/small.tar.gz`: Contains example HDF5 files with 20 rollouts per snippet.
|
| 43 |
+
- `sample/large.tar.gz`: Contains example HDF5 files with 200 rollouts per snippet.
|
| 44 |
+
|
| 45 |
+
The HDF5 structure is detailed in Appendix A.2 of the paper as well as https://github.com/microsoft/MoCapAct#description.
|
| 46 |
+
|
| 47 |
+
An example for loading and inspecting an HDF5 file in Python is:
|
| 48 |
+
```python
|
| 49 |
+
import h5py
|
| 50 |
+
dset = h5py.File("/path/to/small/CMU_083_33.hdf5", "r")
|
| 51 |
+
print("Expert actions from first rollout episode:")
|
| 52 |
+
print(dset["CMU_083_33-0-194/0/actions"][...])
|
| 53 |
+
```
|
| 54 |
+
|
| 55 |
+
## MoCap Videos
|
| 56 |
+
There are two tarball files containing videos of the MoCap clips in the dataset:
|
| 57 |
+
- `full_clip_videos.tar.gz` contains videos of the full MoCap clips.
|
| 58 |
+
- `snippet_videos.tar.gz` contains videos of the snippets that were used to train the experts.
|
| 59 |
+
Note that they are playbacks of the clips themselves, not rollouts of the corresponding experts.
|