ethangoh7086cmd's picture
Update README.md
290d074 verified
metadata
license: mit
datasets:
  - open-thoughts/OpenThoughts-114k
language:
  - en
base_model:
  - Qwen/Qwen3-1.7B
  - Qwen/Qwen3-4B
  - Qwen/Qwen2.5-1.5B-Instruct
tags:
  - vae

VAE Layer for the Research Gated Latent Reasoning Loop (tentative name)

Please refer to our code: https://github.com/elliot-zzh/from-transparent-to-opaque. The project is under construction, and we will publish the paper once we are ready.

This is the pretrained VAE layer for the research Gated Latent Reasoning Loop (tentative name).

There are 3 VAEs, and they are applied to different models:

The structure of the VAE involves two linear layers, including the compressor and the uncompressor.