Improve Model Card: Add Metadata, Paper Link, and Code Link (#1)
Browse files- Improve Model Card: Add Metadata, Paper Link, and Code Link (602f66b0b6434e63a8676fe6541bebe13fbd1a67)
Co-authored-by: Niels Rogge <[email protected]>
README.md
CHANGED
|
@@ -1,8 +1,9 @@
|
|
| 1 |
---
|
| 2 |
license: mit
|
|
|
|
|
|
|
| 3 |
---
|
| 4 |
|
| 5 |
-
|
| 6 |
<div align="center">
|
| 7 |
|
| 8 |
# Aether: Geometric-Aware Unified World Modeling
|
|
@@ -20,6 +21,8 @@ license: mit
|
|
| 20 |
<a href='https://huggingface.co/spaces/AmberHeart/AetherV1'><img src='https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Demo-blue'></a>
|
| 21 |
</div>
|
| 22 |
|
|
|
|
|
|
|
| 23 |
Aether addresses a fundamental challenge in AI: integrating geometric reconstruction with generative modeling
|
| 24 |
for human-like spatial reasoning. Our framework unifies three core capabilities: (1) **4D dynamic reconstruction**,
|
| 25 |
(2) **action-conditioned video prediction**, and (3) **goal-conditioned visual planning**. Trained entirely on
|
|
@@ -29,6 +32,7 @@ synthetic data, Aether achieves strong zero-shot generalization to real-world sc
|
|
| 29 |
<img src="assets/teaser.png" alt="Teaser" width="800"/>
|
| 30 |
</div>
|
| 31 |
|
|
|
|
| 32 |
|
| 33 |
## 📝 Citation
|
| 34 |
If you find this work useful in your research, please consider citing:
|
|
@@ -60,4 +64,4 @@ Our work is primarily built upon
|
|
| 60 |
[DroidCalib](https://github.com/boschresearch/DroidCalib),
|
| 61 |
[Grounded-SAM-2](https://github.com/IDEA-Research/Grounded-SAM-2),
|
| 62 |
[ceres-solver](https://github.com/ceres-solver/ceres-solver), etc.
|
| 63 |
-
We extend our gratitude to all these authors for their generously open-sourced code and their significant contributions to the community.
|
|
|
|
| 1 |
---
|
| 2 |
license: mit
|
| 3 |
+
pipeline_tag: image-to-video
|
| 4 |
+
library_name: CogVideoX
|
| 5 |
---
|
| 6 |
|
|
|
|
| 7 |
<div align="center">
|
| 8 |
|
| 9 |
# Aether: Geometric-Aware Unified World Modeling
|
|
|
|
| 21 |
<a href='https://huggingface.co/spaces/AmberHeart/AetherV1'><img src='https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Demo-blue'></a>
|
| 22 |
</div>
|
| 23 |
|
| 24 |
+
This repository contains the model used in the paper [Aether: Geometric-Aware Unified World Modeling](https://arxiv.org/abs/2503.18945).
|
| 25 |
+
|
| 26 |
Aether addresses a fundamental challenge in AI: integrating geometric reconstruction with generative modeling
|
| 27 |
for human-like spatial reasoning. Our framework unifies three core capabilities: (1) **4D dynamic reconstruction**,
|
| 28 |
(2) **action-conditioned video prediction**, and (3) **goal-conditioned visual planning**. Trained entirely on
|
|
|
|
| 32 |
<img src="assets/teaser.png" alt="Teaser" width="800"/>
|
| 33 |
</div>
|
| 34 |
|
| 35 |
+
Find the code at https://github.com/OpenRobotLab/Aether.
|
| 36 |
|
| 37 |
## 📝 Citation
|
| 38 |
If you find this work useful in your research, please consider citing:
|
|
|
|
| 64 |
[DroidCalib](https://github.com/boschresearch/DroidCalib),
|
| 65 |
[Grounded-SAM-2](https://github.com/IDEA-Research/Grounded-SAM-2),
|
| 66 |
[ceres-solver](https://github.com/ceres-solver/ceres-solver), etc.
|
| 67 |
+
We extend our gratitude to all these authors for their generously open-sourced code and their significant contributions to the community.
|