Spaces:
Running
on
Zero
Running
on
Zero
Change model to 8B
Browse files
app.py
CHANGED
@@ -14,7 +14,7 @@ hf_token = os.getenv("HF_TOKEN")
|
|
14 |
|
15 |
|
16 |
# --- Load tokenizer ---
|
17 |
-
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.
|
18 |
vocab_size = len(tokenizer)
|
19 |
eos_token_id = tokenizer.eos_token_id
|
20 |
mask_token_id = tokenizer.encode('MASK', add_special_tokens=False)[0]
|
@@ -37,9 +37,9 @@ assistant_marker_ids = tokenizer.encode("Assistant:", add_special_tokens=False)
|
|
37 |
def load_model():
|
38 |
ckpt_path = hf_hub_download(
|
39 |
repo_id="ruurd/tini_model",
|
40 |
-
filename="diffusion-model.pth",
|
41 |
token=os.getenv("HF_TOKEN"),
|
42 |
-
revision="1ffb916dd34f442f87cf06dda74b96f86eaf1d15",
|
43 |
)
|
44 |
|
45 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
|
|
14 |
|
15 |
|
16 |
# --- Load tokenizer ---
|
17 |
+
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.1-8B", use_fast=True, token=hf_token)
|
18 |
vocab_size = len(tokenizer)
|
19 |
eos_token_id = tokenizer.eos_token_id
|
20 |
mask_token_id = tokenizer.encode('MASK', add_special_tokens=False)[0]
|
|
|
37 |
def load_model():
|
38 |
ckpt_path = hf_hub_download(
|
39 |
repo_id="ruurd/tini_model",
|
40 |
+
filename="diffusion-model-8B.pth",
|
41 |
token=os.getenv("HF_TOKEN"),
|
42 |
+
# revision="1ffb916dd34f442f87cf06dda74b96f86eaf1d15",
|
43 |
)
|
44 |
|
45 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|