From b374aaead4d49180e8f0e5e8b5bf1ae57a9de6df Mon Sep 17 00:00:00 2001 From: Mr-Neutr0n <64578610+Mr-Neutr0n@users.noreply.github.com> Date: Thu, 12 Feb 2026 00:03:56 +0530 Subject: [PATCH] Fix NameError in EMAVectorQuantizer.__init__ due to undefined variables EMAVectorQuantizer.__init__ references `codebook_dim` and `num_tokens`, which are not defined in the constructor's scope. The correct parameter names are `embedding_dim` and `n_embed` respectively. This causes an immediate NameError when instantiating EMAVectorQuantizer (and by extension, the EMAVQ model), making EMA-based vector quantization completely unusable. --- taming/modules/vqvae/quantize.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/taming/modules/vqvae/quantize.py b/taming/modules/vqvae/quantize.py index d75544e4..a0d14052 100644 --- a/taming/modules/vqvae/quantize.py +++ b/taming/modules/vqvae/quantize.py @@ -362,8 +362,8 @@ class EMAVectorQuantizer(nn.Module): def __init__(self, n_embed, embedding_dim, beta, decay=0.99, eps=1e-5, remap=None, unknown_index="random"): super().__init__() - self.codebook_dim = codebook_dim - self.num_tokens = num_tokens + self.codebook_dim = embedding_dim + self.num_tokens = n_embed self.beta = beta self.embedding = EmbeddingEMA(self.num_tokens, self.codebook_dim, decay, eps)