# we provide a copy of ffhq-256 for convenience, downsampled using the same function as NVAE (https://github.com/NVlabs/NVAE) (personal communication with author) # Resizing function is this one, with the default second argument and size=256 # https://pytorch.org/docs/stable/torchvision/transforms.html#torchvision.transforms.Resize # 5-bit precision is calculated using the following lines, with num_bits=5 and for an x in [0, 1] # x = torch.floor(x * 255 / 2 ** (8 - num_bits)) # x /= (2 ** num_bits - 1) # the DMOL loss should also be adjusted to have 32 buckets instead of 256 (this code, or NVAE, can be used as reference) wget https://openaipublic.blob.core.windows.net/very-deep-vaes-assets/vdvae-assets/ffhq-256.npy