updates
This commit is contained in:
@@ -1,2 +1,21 @@
|
||||
# CSE5519 Advances in Computer Vision (Topic D: 2024: Image and Video Generation)
|
||||
|
||||
## Autoregressive Model Beats Diffusion: Llama for Scalable Image Generation
|
||||
|
||||
[link to the paper](https://arxiv.org/pdf/2406.06525)
|
||||
|
||||
This paper shows that the autoregressive model can outperform the diffusion model in terms of image generation.
|
||||
|
||||
### Novelty in the autoregressive model
|
||||
|
||||
Use Llama 3.1 as the autoregressive model.
|
||||
|
||||
Use code book and downsampling to reduce the memory footprint.
|
||||
|
||||
> [!TIP]
|
||||
>
|
||||
> This paper shows that the autoregressive model can outperform the diffusion model in terms of image generation.
|
||||
>
|
||||
> And in later works, we showed that usually the image can be represented by a few code words; for example, 32 tokens may be enough to represent most of the images (that most humans need to annotate). However, I doubt the result if it can be generalized to more complex image generation tasks, for example, the image generation with a human face, since I found it difficult to describe people around me distinctively without calling their name.
|
||||
>
|
||||
> For more real-life videos, to ensure contextual consistency, we may need to use more code words. Is such a method scalable to video generation to produce realistic results? Or will there be an exponential memory cost for the video generation?
|
||||
Reference in New Issue
Block a user