{ "emb_size": 768, "feedforward_size": 3072, "hidden_size": 768, "hidden_act": "gelu_fast", "heads_num": 12, "layers_num": 12, "dropout": 0.1, "max_seq_length": 1024, "embedding": ["word", "pos", "seg"], "remove_embedding_layernorm": true, "layernorm_positioning": "pre", "encoder": "transformer", "mask": "prefix", "target": ["lm"], "image_height": 256, "image_width": 256, "patch_size": 16, "seq_length": 334, "tokenizer": "text_image", "image_tokenizer": { "is_gumbel": false, "is_transformer": true, "image_vocab_size": 1024, "frame_size": 16 } }