Update the paper link in README.md
Browse files
README.md
CHANGED
|
@@ -17,7 +17,7 @@ CogACT is a new advanced VLA architecture derived from VLM. Unlike previous work
|
|
| 17 |
|
| 18 |
All our [code](https://github.com/microsoft/CogACT), [pretrained model weights](https://huggingface.co/CogACT), are licensed under the MIT license.
|
| 19 |
|
| 20 |
-
Please refer to our [project page](https://cogact.github.io/) and [paper](https://
|
| 21 |
|
| 22 |
|
| 23 |
## Model Summary
|
|
@@ -32,7 +32,7 @@ Please refer to our [project page](https://cogact.github.io/) and [paper](https:
|
|
| 32 |
+ **Action Model**: DiT-Small
|
| 33 |
- **Pretraining Dataset:** A subset of [Open X-Embodiment](https://robotics-transformer-x.github.io/)
|
| 34 |
- **Repository:** [https://github.com/microsoft/CogACT](https://github.com/microsoft/CogACT)
|
| 35 |
-
- **Paper:** [CogACT: A Foundational Vision-Language-Action Model for Synergizing Cognition and Action in Robotic Manipulation](https://
|
| 36 |
- **Project Page:** [https://cogact.github.io/](https://cogact.github.io/)
|
| 37 |
|
| 38 |
## Uses
|
|
|
|
| 17 |
|
| 18 |
All our [code](https://github.com/microsoft/CogACT), [pretrained model weights](https://huggingface.co/CogACT), are licensed under the MIT license.
|
| 19 |
|
| 20 |
+
Please refer to our [project page](https://cogact.github.io/) and [paper](https://arxiv.org/abs/2411.19650) for more details.
|
| 21 |
|
| 22 |
|
| 23 |
## Model Summary
|
|
|
|
| 32 |
+ **Action Model**: DiT-Small
|
| 33 |
- **Pretraining Dataset:** A subset of [Open X-Embodiment](https://robotics-transformer-x.github.io/)
|
| 34 |
- **Repository:** [https://github.com/microsoft/CogACT](https://github.com/microsoft/CogACT)
|
| 35 |
+
- **Paper:** [CogACT: A Foundational Vision-Language-Action Model for Synergizing Cognition and Action in Robotic Manipulation](https://arxiv.org/abs/2411.19650)
|
| 36 |
- **Project Page:** [https://cogact.github.io/](https://cogact.github.io/)
|
| 37 |
|
| 38 |
## Uses
|