updates
This commit is contained in:
@@ -1,2 +1,15 @@
|
||||
# CSE5519 Advances in Computer Vision (Topic D: 2023: Image and Video Generation)
|
||||
|
||||
## Scalable Diffusion Models with Transformers
|
||||
|
||||
[link to paper](https://openaccess.thecvf.com/content/ICCV2023/papers/Peebles_Scalable_Diffusion_Models_with_Transformers_ICCV_2023_paper.pdf)
|
||||
|
||||
Create a diffusion model with transformers.
|
||||
|
||||
Train conditional DiT models over latent patches replacing the U-Net.
|
||||
|
||||
> [!TIP]
|
||||
>
|
||||
> This paper provides a scalable way to integrate the conditional DiT models over latent patches, replacing the U-Net to improve the performance of image generation.
|
||||
>
|
||||
> I wonder how classifier-free guidance is used in training the DiT and if the model also has in-context learning ability, as other transformer models do.
|
||||
Reference in New Issue
Block a user