Until now, VQGAN, the initial tokenizer is still acting an indispensible role in mainstream tasks, especially autoregressive visual generation. Limited by the bottleneck of the size of codebook and the utilization of code, the capability of AR generation with VQGAN is underestimated.
Therefore,
MAGVIT2
proposes a powerful tokenizer for visual generation task, which introduces a novel LookUpFree technique when quantization and extends the size of codebook to $2^{18}$, exhibiting promising performance in both image and video generation tasks. And it plays an important role in the recent state-of-the-art AR video generation model
VideoPoet
. However, we have no access to this strong tokenizer so far. ☹️
In the codebase, we follow the significant insights of tokenizer design in MAGVIT-2 and re-implement it with Pytorch, achieving the closest results to the original so far. We hope that our effort can foster innovation, creativity within the field of Autoregressive Visual Generation. 😄
Open-MAGVIT2 huggingface.co is an AI model on huggingface.co that provides Open-MAGVIT2's model effect (), which can be used instantly with this TencentARC Open-MAGVIT2 model. huggingface.co supports a free trial of the Open-MAGVIT2 model, and also provides paid use of the Open-MAGVIT2. Support call Open-MAGVIT2 model through api, including Node.js, Python, http.
Open-MAGVIT2 huggingface.co is an online trial and call api platform, which integrates Open-MAGVIT2's modeling effects, including api services, and provides a free online trial of Open-MAGVIT2, you can try Open-MAGVIT2 online for free by clicking the link below.
TencentARC Open-MAGVIT2 online free url in huggingface.co:
Open-MAGVIT2 is an open source model from GitHub that offers a free installation service, and any user can find Open-MAGVIT2 on GitHub to install. At the same time, huggingface.co provides the effect of Open-MAGVIT2 install, users can directly use Open-MAGVIT2 installed effect in huggingface.co for debugging and trial. It also supports api for free installation.