Suparious commited on
Commit
42d39b3
·
verified ·
1 Parent(s): 391693d

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +34 -0
README.md ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ pipeline_tag: text-generation
4
+ ---
5
+ # mistralai/Mistral-7B-v0.2 AWQ
6
+
7
+
8
+
9
+ ## Model Summary
10
+
11
+ Mistral-7B-v0.2 has the following changes compared to Mistral-7B-v0.1
12
+
13
+ - 32k context window (vs 8k context in v0.1)
14
+ - Rope-theta = 1e6
15
+ - No Sliding-Window Attention
16
+
17
+ For full details of this model please read our [paper](https://arxiv.org/abs/2310.06825) and [release blog post](https://mistral.ai/news/la-plateforme/).
18
+
19
+ - Grouped-Query Attention
20
+ - Sliding-Window Attention
21
+ - Byte-fallback BPE tokenizer
22
+
23
+ ## Instruction format
24
+
25
+ In order to leverage instruction fine-tuning, your prompt should be surrounded by `[INST]` and `[/INST]` tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id.
26
+
27
+ E.g.
28
+ ```
29
+ text = "<s>[INST] What is your favourite condiment? [/INST]"
30
+ "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!</s> "
31
+ "[INST] Do you have mayonnaise recipes? [/INST]"
32
+ ```
33
+
34
+ This format is available as a [chat template](https://huggingface.co/docs/transformers/main/chat_templating) via the `apply_chat_template()` method.