drishanarora commited on
Commit
b5a6c9a
·
verified ·
1 Parent(s): a407075

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -0
README.md CHANGED
@@ -23,8 +23,21 @@ The Cogito LLMs are instruction tuned generative models (text in/text out). All
23
  - In both standard and reasoning modes, Cogito v1-preview models outperform their size equivalent counterparts on common industry benchmarks.
24
  - Each model is trained in over 30 languages and supports a context length of 128k.
25
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  For detailed evaluations, please refer to the [Blog Post](https://www.deepcogito.com/research/cogito-v1-preview).
27
 
 
28
  # Usage
29
  Here is a snippet below for usage with Transformers:
30
 
 
23
  - In both standard and reasoning modes, Cogito v1-preview models outperform their size equivalent counterparts on common industry benchmarks.
24
  - Each model is trained in over 30 languages and supports a context length of 128k.
25
 
26
+ # Evaluations
27
+ We compare our models against the state of the art size equivalent models in direct mode as well as the reasoning mode. For the direct mode, we compare against Llama / Qwen instruct counterparts. For reasoning, we use Deepseek's R1 distilled counterparts / Qwen's QwQ model.
28
+
29
+ <p align="left">
30
+ <img src="images/14b_benchmarks.png" alt="Logo" width="90%">
31
+ </p>
32
+
33
+ **Livebench Global Average:**
34
+ <p align="left">
35
+ <img src="images/livebench_global_average.png" alt="Logo" width="80%">
36
+ </p>
37
+
38
  For detailed evaluations, please refer to the [Blog Post](https://www.deepcogito.com/research/cogito-v1-preview).
39
 
40
+
41
  # Usage
42
  Here is a snippet below for usage with Transformers:
43