ChihChiu29 commited on
Commit
2f79a00
·
1 Parent(s): fcdc17d

switich to last 128 words

Browse files
Files changed (1) hide show
  1. main.py +2 -2
main.py CHANGED
@@ -16,7 +16,7 @@ token_size_limit = None
16
  # model = AutoModelForSeq2SeqLM.from_pretrained("facebook/blenderbot-400M-distill")
17
  tokenizer = AutoTokenizer.from_pretrained("facebook/blenderbot-1B-distill")
18
  model = AutoModelForSeq2SeqLM.from_pretrained("facebook/blenderbot-1B-distill")
19
- # token_size_limit = 128
20
 
21
  # T5 model can use "any" sequence lenghth, but memory usage is O(L^2).
22
  # tokenizer = T5Tokenizer.from_pretrained("google/flan-t5-small")
@@ -25,7 +25,7 @@ model = AutoModelForSeq2SeqLM.from_pretrained("facebook/blenderbot-1B-distill")
25
  # model = T5ForConditionalGeneration.from_pretrained("google/flan-t5-base")
26
  # tokenizer = T5Tokenizer.from_pretrained("google/flan-t5-large")
27
  # model = T5ForConditionalGeneration.from_pretrained("google/flan-t5-large")
28
- token_size_limit = 512
29
 
30
  # Too large for 16GB
31
  # tokenizer = T5Tokenizer.from_pretrained("google/flan-t5-xl")
 
16
  # model = AutoModelForSeq2SeqLM.from_pretrained("facebook/blenderbot-400M-distill")
17
  tokenizer = AutoTokenizer.from_pretrained("facebook/blenderbot-1B-distill")
18
  model = AutoModelForSeq2SeqLM.from_pretrained("facebook/blenderbot-1B-distill")
19
+ token_size_limit = 128
20
 
21
  # T5 model can use "any" sequence lenghth, but memory usage is O(L^2).
22
  # tokenizer = T5Tokenizer.from_pretrained("google/flan-t5-small")
 
25
  # model = T5ForConditionalGeneration.from_pretrained("google/flan-t5-base")
26
  # tokenizer = T5Tokenizer.from_pretrained("google/flan-t5-large")
27
  # model = T5ForConditionalGeneration.from_pretrained("google/flan-t5-large")
28
+ # token_size_limit = 512
29
 
30
  # Too large for 16GB
31
  # tokenizer = T5Tokenizer.from_pretrained("google/flan-t5-xl")