Update README.md
Browse files
README.md
CHANGED
|
@@ -24,7 +24,7 @@ library_name: diffusers
|
|
| 24 |

|
| 25 |
|
| 26 |
This repository hosts the **Emuru Convolutional VAE**, described in our [CVPR2025 paper](https://arxiv.org/pdf/2503.17074). The model features a convolutional encoder and decoder, each with four layers. The output channels for these layers are 32, 64, 128, and 256, respectively. The encoder downsamples an input RGB image (with three channels and dimensions width and height) to a latent representation with a single channel and spatial dimensions that are one-eighth of the original height and width. This design compresses the style information in the image, allowing a lightweight Transformer Decoder to efficiently process the latent features.
|
| 27 |
-
Training code is released on [GitHub](https://github.com/aimagelab/
|
| 28 |
|
| 29 |
### Training Details
|
| 30 |
|
|
|
|
| 24 |

|
| 25 |
|
| 26 |
This repository hosts the **Emuru Convolutional VAE**, described in our [CVPR2025 paper](https://arxiv.org/pdf/2503.17074). The model features a convolutional encoder and decoder, each with four layers. The output channels for these layers are 32, 64, 128, and 256, respectively. The encoder downsamples an input RGB image (with three channels and dimensions width and height) to a latent representation with a single channel and spatial dimensions that are one-eighth of the original height and width. This design compresses the style information in the image, allowing a lightweight Transformer Decoder to efficiently process the latent features.
|
| 27 |
+
Training code is released on [GitHub](https://github.com/aimagelab/Emuru-autoregressive-text-img).
|
| 28 |
|
| 29 |
### Training Details
|
| 30 |
|