sharmavaruncs commited on
Commit
87b439c
·
1 Parent(s): 4de9cc5

removed redundant directories as everything hosted on huggingface now

Browse files
EmotionDetector/Models/HubertOnly/Link.txt DELETED
@@ -1 +0,0 @@
1
- https://huggingface.co/netgvarun2005/HubertStandaloneEmoDetector
 
 
EmotionDetector/Models/MultiModal/Link.txt DELETED
@@ -1 +0,0 @@
1
- https://huggingface.co/netgvarun2005/MultiModalBertHubert
 
 
EmotionDetector/Models/Tokenizer/special_tokens_map.json DELETED
@@ -1,7 +0,0 @@
1
- {
2
- "cls_token": "[CLS]",
3
- "mask_token": "[MASK]",
4
- "pad_token": "[PAD]",
5
- "sep_token": "[SEP]",
6
- "unk_token": "[UNK]"
7
- }
 
 
 
 
 
 
 
 
EmotionDetector/Models/Tokenizer/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
EmotionDetector/Models/Tokenizer/tokenizer_config.json DELETED
@@ -1,13 +0,0 @@
1
- {
2
- "clean_up_tokenization_spaces": true,
3
- "cls_token": "[CLS]",
4
- "do_lower_case": true,
5
- "mask_token": "[MASK]",
6
- "model_max_length": 512,
7
- "pad_token": "[PAD]",
8
- "sep_token": "[SEP]",
9
- "strip_accents": null,
10
- "tokenize_chinese_chars": true,
11
- "tokenizer_class": "BertTokenizer",
12
- "unk_token": "[UNK]"
13
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
EmotionDetector/Models/Tokenizer/vocab.txt DELETED
The diff for this file is too large to render. See raw diff
 
GenAI/Model/Link.txt DELETED
@@ -1 +0,0 @@
1
- https://huggingface.co/netgvarun2005/GPTVirtualTherapist
 
 
GenAI/Tokenizer/added_tokens.json DELETED
@@ -1,4 +0,0 @@
1
- {
2
- "<|pad|>": 50258,
3
- "<|startoftext|>": 50257
4
- }
 
 
 
 
 
GenAI/Tokenizer/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
GenAI/Tokenizer/special_tokens_map.json DELETED
@@ -1,30 +0,0 @@
1
- {
2
- "bos_token": {
3
- "content": "<|startoftext|>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "<|endoftext|>",
11
- "lstrip": false,
12
- "normalized": true,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": {
17
- "content": "<|pad|>",
18
- "lstrip": false,
19
- "normalized": true,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
- "unk_token": {
24
- "content": "<|endoftext|>",
25
- "lstrip": false,
26
- "normalized": true,
27
- "rstrip": false,
28
- "single_word": false
29
- }
30
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
GenAI/Tokenizer/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
GenAI/Tokenizer/tokenizer_config.json DELETED
@@ -1,40 +0,0 @@
1
- {
2
- "add_bos_token": false,
3
- "add_prefix_space": false,
4
- "bos_token": {
5
- "__type": "AddedToken",
6
- "content": "<|startoftext|>",
7
- "lstrip": false,
8
- "normalized": true,
9
- "rstrip": false,
10
- "single_word": false
11
- },
12
- "clean_up_tokenization_spaces": true,
13
- "eos_token": {
14
- "__type": "AddedToken",
15
- "content": "<|endoftext|>",
16
- "lstrip": false,
17
- "normalized": true,
18
- "rstrip": false,
19
- "single_word": false
20
- },
21
- "errors": "replace",
22
- "model_max_length": 2048,
23
- "pad_token": {
24
- "__type": "AddedToken",
25
- "content": "<|pad|>",
26
- "lstrip": false,
27
- "normalized": true,
28
- "rstrip": false,
29
- "single_word": false
30
- },
31
- "tokenizer_class": "GPT2Tokenizer",
32
- "unk_token": {
33
- "__type": "AddedToken",
34
- "content": "<|endoftext|>",
35
- "lstrip": false,
36
- "normalized": true,
37
- "rstrip": false,
38
- "single_word": false
39
- }
40
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
GenAI/Tokenizer/vocab.json DELETED
The diff for this file is too large to render. See raw diff
 
app.py CHANGED
@@ -183,7 +183,7 @@ def preprocessWavFile(wavfile):
183
  denoised_file = noiseReduction(resampledwavfile)
184
  return removeSilence(denoised_file)
185
 
186
- @st.cache(allow_output_mutation=True)
187
  def load_model():
188
  # Load the model
189
  multiModel = MultimodalModel(bert_model_name, num_labels)
 
183
  denoised_file = noiseReduction(resampledwavfile)
184
  return removeSilence(denoised_file)
185
 
186
+ @st.cache_data(allow_output_mutation=True)
187
  def load_model():
188
  # Load the model
189
  multiModel = MultimodalModel(bert_model_name, num_labels)