Update the paper link in README.md.
Browse files
README.md
CHANGED
@@ -17,7 +17,7 @@ CogACT is a new advanced VLA architecture derived from VLM. Unlike previous work
|
|
17 |
|
18 |
All our [code](https://github.com/microsoft/CogACT), [pretrained model weights](https://huggingface.co/CogACT), are licensed under the MIT license.
|
19 |
|
20 |
-
Please refer to our [project page](https://cogact.github.io/) and [paper](https://
|
21 |
|
22 |
|
23 |
## Model Summary
|
@@ -32,7 +32,7 @@ Please refer to our [project page](https://cogact.github.io/) and [paper](https:
|
|
32 |
+ **Action Model**: DiT-Large
|
33 |
- **Pretraining Dataset:** A subset of [Open X-Embodiment](https://robotics-transformer-x.github.io/)
|
34 |
- **Repository:** [https://github.com/microsoft/CogACT](https://github.com/microsoft/CogACT)
|
35 |
-
- **Paper:** [CogACT: A Foundational Vision-Language-Action Model for Synergizing Cognition and Action in Robotic Manipulation](https://
|
36 |
- **Project Page:** [https://cogact.github.io/](https://cogact.github.io/)
|
37 |
|
38 |
## Uses
|
|
|
17 |
|
18 |
All our [code](https://github.com/microsoft/CogACT), [pretrained model weights](https://huggingface.co/CogACT), are licensed under the MIT license.
|
19 |
|
20 |
+
Please refer to our [project page](https://cogact.github.io/) and [paper](https://arxiv.org/abs/2411.19650) for more details.
|
21 |
|
22 |
|
23 |
## Model Summary
|
|
|
32 |
+ **Action Model**: DiT-Large
|
33 |
- **Pretraining Dataset:** A subset of [Open X-Embodiment](https://robotics-transformer-x.github.io/)
|
34 |
- **Repository:** [https://github.com/microsoft/CogACT](https://github.com/microsoft/CogACT)
|
35 |
+
- **Paper:** [CogACT: A Foundational Vision-Language-Action Model for Synergizing Cognition and Action in Robotic Manipulation](https://arxiv.org/abs/2411.19650)
|
36 |
- **Project Page:** [https://cogact.github.io/](https://cogact.github.io/)
|
37 |
|
38 |
## Uses
|