model
stringlengths 4
89
| revision
stringclasses 1
value | model_sha
stringlengths 0
40
| results
dict | commit
stringlengths 40
40
| date
timestamp[ns] | score
float64 21.8
83
⌀ |
---|---|---|---|---|---|---|
h2oai/h2ogpt-research-oasst1-llama-65b
|
main
|
a6d8676aaa2ca2c25ea99180b538f0369dc70185
|
{
"arc:challenge": 64.8,
"hellaswag": 85.9,
"hendrycksTest": 63.6,
"truthfulqa:mc": 48.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 65.8 |
h2oai/h2ogpt-oig-oasst1-256-6_9b
|
main
|
f1c9bac89b74d3487cb092788ce828fb9520c1a7
|
{
"arc:challenge": 39.9,
"hellaswag": 65.4,
"hendrycksTest": 26.4,
"truthfulqa:mc": 35
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 41.7 |
h2oai/h2ogpt-gm-oasst1-en-2048-open-llama-7b-preview-300bt
|
main
|
754e0c90ed5d9241fdfd5a188572b3ea2152eaa7
|
{
"arc:challenge": 34,
"hellaswag": 50.5,
"hendrycksTest": 24.7,
"truthfulqa:mc": 41.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 37.8 |
h2oai/h2ogpt-research-oig-oasst1-512-30b
|
main
|
4215e83b9038c9e61d979cf5223b29f860bace9b
|
{
"arc:challenge": 59,
"hellaswag": 82.6,
"hendrycksTest": 50.7,
"truthfulqa:mc": 48.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 60.2 |
h2oai/h2ogpt-gm-oasst1-en-1024-20b
|
main
|
1a5b8d25587eab67d837621a6c9423e7ef6df289
|
{
"arc:challenge": 48,
"hellaswag": 72.8,
"hendrycksTest": 26,
"truthfulqa:mc": 39.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 46.7 |
h2oai/h2ogpt-oasst1-512-12b
|
main
|
c6bb0fe363e0105839d34ca757793b61c9606f95
|
{
"arc:challenge": 42.3,
"hellaswag": 70.2,
"hendrycksTest": 26,
"truthfulqa:mc": 36.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 43.7 |
h2oai/h2ogpt-gm-oasst1-en-1024-12b
|
main
|
e547fffafb382fd39ef5de35ba3b5afc1b43e74d
|
{
"arc:challenge": 43.1,
"hellaswag": 69.7,
"hendrycksTest": 25.9,
"truthfulqa:mc": 38
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 44.2 |
minlik/chinese-alpaca-33b-merged
|
main
|
fc2535104c0b48afc42575f9fe10bbcbb7612ec3
|
{
"arc:challenge": 59.3,
"hellaswag": 78.4,
"hendrycksTest": 57.7,
"truthfulqa:mc": 52.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 62 |
xzuyn/Alpacino-SuperCOT-13B
|
main
|
3a82b04684fe99d59556421c3f96a187049a3cec
|
{
"arc:challenge": 58.4,
"hellaswag": 81.7,
"hendrycksTest": 47.9,
"truthfulqa:mc": 45.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 58.4 |
xzuyn/MedicWizard-7B
|
main
|
0b3ef975fb5e8ac1eae775160ab54c98221889df
|
{
"arc:challenge": 53.5,
"hellaswag": 78.4,
"hendrycksTest": 44.6,
"truthfulqa:mc": 41.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 54.4 |
teknium/OpenHermes-2.5-Mistral-7B
|
main
|
2a54cad766bc90828354db5c4199795aecfd0df1
|
{
"arc:challenge": 64.9,
"hellaswag": 84.3,
"hendrycksTest": 63.8,
"truthfulqa:mc": 52.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 66.3 |
teknium/OpenHermes-2-Mistral-7B
|
main
|
843a9bb94fac7d7bfc1b7c9f201efba295b6f5d6
|
{
"arc:challenge": 63.1,
"hellaswag": 83.8,
"hendrycksTest": 63.5,
"truthfulqa:mc": 50.2
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 65.2 |
teknium/OpenHermes-13B
|
main
|
f09d0fe655ad57cce9179b7b40ea6f81e07db18c
|
{
"arc:challenge": 59.8,
"hellaswag": 82.2,
"hendrycksTest": 56.3,
"truthfulqa:mc": 46
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 61.1 |
teknium/CollectiveCognition-v1.1-Mistral-7B
|
main
|
5f57f70ec99450c70da2540e94dd7fd67be4b23c
|
{
"arc:challenge": 62.1,
"hellaswag": 84.2,
"hendrycksTest": 62.3,
"truthfulqa:mc": 57.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 66.6 |
nathan0/mpt_delta_tuned_model_v2
|
main
|
6adb4cb4ba37f4ce9e9c3051d343addf1098182c
|
{
"arc:challenge": 50.7,
"hellaswag": 76.4,
"hendrycksTest": 28.7,
"truthfulqa:mc": 35.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 47.8 |
nathan0/mpt_delta_tuned_model_v3
|
main
|
6adb4cb4ba37f4ce9e9c3051d343addf1098182c
|
{
"arc:challenge": 50.7,
"hellaswag": 76.4,
"hendrycksTest": 28.7,
"truthfulqa:mc": 35.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 47.8 |
chargoddard/duplicitous-mammal-13b
|
main
|
a05d0562b8da2ac2e76aa65984e8063249bc85c8
|
{
"arc:challenge": 61.7,
"hellaswag": 83.8,
"hendrycksTest": 57.5,
"truthfulqa:mc": 52.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 63.8 |
chargoddard/llama2-22b-blocktriangular
|
main
|
40a51343ae776b5cb39f2b4343ae8f9b676ffd58
|
{
"arc:challenge": 58.5,
"hellaswag": 82.6,
"hendrycksTest": 54.6,
"truthfulqa:mc": 39.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 58.8 |
chargoddard/MelangeB-70b
|
main
|
08239fb1e30b1e42b14370f23e942bc51e76027c
|
{
"arc:challenge": 71.7,
"hellaswag": 87.5,
"hendrycksTest": 70,
"truthfulqa:mc": 59.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 72.1 |
chargoddard/ypotryll-22b-epoch2-qlora
|
main
|
26fdd8fa420d72ed835c7d17086f0441db0985d4
|
{
"arc:challenge": 59.2,
"hellaswag": 80.7,
"hendrycksTest": 54.5,
"truthfulqa:mc": 40.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 58.7 |
chargoddard/llama-2-26b-trenchcoat-stack
|
main
|
075d67c3223f4b379ab7f997c3787cd0630d80f7
|
{
"arc:challenge": 55,
"hellaswag": 79.9,
"hendrycksTest": 53.7,
"truthfulqa:mc": 40.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 57.3 |
chargoddard/llama2-22b
|
main
|
2bece0787009b4b584f49d0e0d1b49ecf4a52da9
|
{
"arc:challenge": 58.5,
"hellaswag": 82.6,
"hendrycksTest": 54.7,
"truthfulqa:mc": 39.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 58.9 |
chargoddard/platypus-2-22b-relora
|
main
|
15bca3e9b25cc2f280fec21686ef3bc445217503
|
{
"arc:challenge": 57.7,
"hellaswag": 82.4,
"hendrycksTest": 55.3,
"truthfulqa:mc": 43.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 59.8 |
chargoddard/platypus2-22b-relora
|
main
|
15bca3e9b25cc2f280fec21686ef3bc445217503
|
{
"arc:challenge": 57.5,
"hellaswag": 82.4,
"hendrycksTest": 54.9,
"truthfulqa:mc": 43.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 59.6 |
chargoddard/storytime-13b
|
main
|
233568319a636b6a7b02a4def2c51d08a3e0fbfc
|
{
"arc:challenge": 62,
"hellaswag": 84,
"hendrycksTest": 57.5,
"truthfulqa:mc": 52.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 64 |
chargoddard/Chronorctypus-Limarobormes-13b
|
main
|
75c1bf5f4b40cf61873ff6487ccd3efc4f684330
|
{
"arc:challenge": 59.9,
"hellaswag": 82.8,
"hendrycksTest": 58.4,
"truthfulqa:mc": 51.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 63.2 |
chargoddard/MelangeC-70b
|
main
|
e54a2b924dec135f3fa2373933ab8485178cde1b
|
{
"arc:challenge": 71.7,
"hellaswag": 87.6,
"hendrycksTest": 70.4,
"truthfulqa:mc": 58.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 72 |
Open-Orca/Mistral-7B-SlimOrca
|
main
|
a9744d8cf9ce4230678a891bcf8bba7cbc0aaece
|
{
"arc:challenge": 62.5,
"hellaswag": 83.9,
"hendrycksTest": 62.8,
"truthfulqa:mc": 54.2
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 65.8 |
Open-Orca/LlongOrca-13B-16k
|
main
|
8ea1fb205553cadbc90069d80a7e58281b6281c3
|
{
"arc:challenge": 62.5,
"hellaswag": 82.8,
"hendrycksTest": 55.5,
"truthfulqa:mc": 50.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 62.7 |
Open-Orca/OpenOrca-Preview1-13B
|
main
|
4c558283a98348383460939afda9cb5c54544c8f
|
{
"arc:challenge": 54.9,
"hellaswag": 78.2,
"hendrycksTest": 50.1,
"truthfulqa:mc": 49.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 58.1 |
Open-Orca/OpenOrca-Platypus2-13B
|
main
|
e7a40134f7eb687c6ab66d445dc7251257f8d391
|
{
"arc:challenge": 62.8,
"hellaswag": 83.2,
"hendrycksTest": 59.4,
"truthfulqa:mc": 53.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 64.6 |
Open-Orca/LlongOrca-7B-16k
|
main
|
1370c7c595e6c8394e6332bc535ae25e21def85b
|
{
"arc:challenge": 57.5,
"hellaswag": 79.4,
"hendrycksTest": 49.4,
"truthfulqa:mc": 49.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 59 |
grantprice/Cerebras-GPT-590M-finetuned-DND
|
main
|
a0a2fbe342cdc86433913ba5f96978e4703ff672
|
{
"arc:challenge": 24.7,
"hellaswag": 27.8,
"hendrycksTest": 23.1,
"truthfulqa:mc": 48.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 31 |
lgaalves/tinyllama-1.1b-chat-v0.3_platypus
|
main
|
0bb6ebe1d41d394bae0ed9107ec8d776d9d76a68
|
{
"arc:challenge": 30.3,
"hellaswag": 55.1,
"hendrycksTest": 26.1,
"truthfulqa:mc": 39.2
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 37.7 |
lgaalves/gpt2_guanaco-dolly-platypus
|
main
|
6bf0a8146cf255c829ec2ad83926c8b80945b431
|
{
"arc:challenge": 23.5,
"hellaswag": 31,
"hendrycksTest": 26.4,
"truthfulqa:mc": 40
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 30.2 |
lgaalves/mistral-7b_open_platypus
|
main
|
b9a60b9ad0fe06bd314ffe99d543f1df6ecd10da
|
{
"arc:challenge": 55.8,
"hellaswag": 82.1,
"hendrycksTest": 59.8,
"truthfulqa:mc": 48.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 61.6 |
lgaalves/gpt2_camel_physics-platypus
|
main
|
66165ff32ed8de6c39f3524a810f5e97ba6d3347
|
{
"arc:challenge": 23,
"hellaswag": 31.3,
"hendrycksTest": 26.9,
"truthfulqa:mc": 39.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 30.2 |
lgaalves/gpt2-xl_lima
|
main
|
f7db5b1db521abd7578b95138e737637e0037ca5
|
{
"arc:challenge": 31.1,
"hellaswag": 51.3,
"hendrycksTest": 25.4,
"truthfulqa:mc": 38.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 36.6 |
lgaalves/llama-2-7b-hf_open-platypus
|
main
|
c7e776f3f3afc0fa22cb7aff0d00522e571e9b29
|
{
"arc:challenge": 51.5,
"hellaswag": 78.6,
"hendrycksTest": 43.6,
"truthfulqa:mc": 43.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 54.4 |
NobodyExistsOnTheInternet/PuffedLIMA13bQLORA
|
main
|
7da6d235d625e16c850ccd0b947dee40071b1f89
|
{
"arc:challenge": 59.9,
"hellaswag": 84.4,
"hendrycksTest": 53.7,
"truthfulqa:mc": 39.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 59.5 |
NobodyExistsOnTheInternet/PuffedConvo13bLoraE4
|
main
|
40e4fce0c25bd23f6011b424748ee2b5374b98d5
|
{
"arc:challenge": 59.6,
"hellaswag": 84.4,
"hendrycksTest": 53.7,
"truthfulqa:mc": 39.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 59.4 |
AlpinDale/pygmalion-instruct
|
main
|
1665b271316dfee05b2a8daf8b9d6c22ed0aef60
|
{
"arc:challenge": 52.6,
"hellaswag": 77.7,
"hendrycksTest": 35.9,
"truthfulqa:mc": 42.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 52.1 |
Doctor-Shotgun/mythospice-70b
|
main
|
b00992c26604c9cd496bc41472a05e4c01cd2008
|
{
"arc:challenge": 69.3,
"hellaswag": 87.5,
"hendrycksTest": 70.1,
"truthfulqa:mc": 56.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 70.9 |
Doctor-Shotgun/mythospice-limarp-70b
|
main
|
ff29fed2a33fc050fd20d0e25b5b23c4a101b074
|
{
"arc:challenge": 69.2,
"hellaswag": 87.5,
"hendrycksTest": 70.1,
"truthfulqa:mc": 55.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 70.7 |
voidful/changpt-bart
|
main
|
e3d26f736b8b47d5275421be6133b81bef84db7d
|
{
"arc:challenge": 29.4,
"hellaswag": 26.3,
"hendrycksTest": 23.1,
"truthfulqa:mc": 47.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 31.7 |
The-Face-Of-Goonery/Huginn-22b-Prototype
|
main
|
29222b05794abb862ad0aaaf3020696c9f599810
|
{
"arc:challenge": 57.7,
"hellaswag": 80.7,
"hendrycksTest": 49.8,
"truthfulqa:mc": 52.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 60.1 |
The-Face-Of-Goonery/Huginn-v3-13b
|
main
|
6c2faf828c5380d28c51fcb4d3d0f1a420fb9a9a
|
{
"arc:challenge": 60.7,
"hellaswag": 82.3,
"hendrycksTest": 52.3,
"truthfulqa:mc": 50.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 61.5 |
The-Face-Of-Goonery/Huginn-13b-v4.5
|
main
|
f3be56d8bf71a8d3905974b1e5fcba7336b02159
|
{
"arc:challenge": 60.7,
"hellaswag": 82.3,
"hendrycksTest": 52.3,
"truthfulqa:mc": 50.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 61.5 |
The-Face-Of-Goonery/Huginn-13b-v1.2
|
main
|
cb3562e7aae05a95fe61610b7b8f4957d3529ce7
|
{
"arc:challenge": 60.9,
"hellaswag": 83.6,
"hendrycksTest": 55.3,
"truthfulqa:mc": 52
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 63 |
The-Face-Of-Goonery/huginnv1.2
|
main
|
aed4ddc951c657993939fa5b87a4088550569a3b
|
{
"arc:challenge": 62.4,
"hellaswag": 84.3,
"hendrycksTest": 57,
"truthfulqa:mc": 47.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 62.9 |
simsim314/WizardLM-70B-V1.0-HF
|
main
|
97112db6d0fae8354c13437a5e7dc99fb37b8c2e
|
{
"arc:challenge": 65.4,
"hellaswag": 84.4,
"hendrycksTest": 64,
"truthfulqa:mc": 54.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 67.2 |
HyperbeeAI/Tulpar-7b-v0
|
main
|
d7c2bc52a3ae13571357f51273ae948caf84400e
|
{
"arc:challenge": 56.3,
"hellaswag": 79,
"hendrycksTest": 52.6,
"truthfulqa:mc": 51.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 59.9 |
illuin/test-custom-llama
|
main
|
d985610bef080473e40f01c53266083c5f0c3169
|
{
"arc:challenge": 52.3,
"hellaswag": 77.5,
"hendrycksTest": 36.6,
"truthfulqa:mc": 33.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 50 |
luffycodes/mcq-vicuna-13b-v1.5
|
main
|
f769a92cfeffe8ee07beee8814ce7eca7cd62805
|
{
"arc:challenge": 56.2,
"hellaswag": 81.1,
"hendrycksTest": 53.4,
"truthfulqa:mc": 44.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 58.7 |
luffycodes/llama-shishya-7b-ep3-v2
|
main
|
679c6cb9e869df686b1ae415ed440e6cfc05f80b
|
{
"arc:challenge": 47.4,
"hellaswag": 75.9,
"hendrycksTest": 43.8,
"truthfulqa:mc": 30.2
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 49.3 |
luffycodes/nash-vicuna-13b-v1dot5-ep2-w-rag-w-simple
|
main
|
848ef91ab46a72260542283918a971347c6bfa93
|
{
"arc:challenge": 59.1,
"hellaswag": 80.6,
"hendrycksTest": 56.1,
"truthfulqa:mc": 51.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 61.8 |
luffycodes/llama-shishya-7b-ep3-v1
|
main
|
8dc109f45ef36cc7bbd0f5d83fb65ac8e768d1bd
|
{
"arc:challenge": 48,
"hellaswag": 76.6,
"hendrycksTest": 46.1,
"truthfulqa:mc": 30.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 50.4 |
luffycodes/vicuna-shishya-7b-ep3-v1
|
main
|
082cf758aa3f6d8f956056003b5b3b6cde447d88
|
{
"arc:challenge": 45.9,
"hellaswag": 76.4,
"hendrycksTest": 50,
"truthfulqa:mc": 40.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 53.2 |
RoversX/llama-2-7b-hf-small-shards-Samantha-V1-SFT
|
main
|
c39cee3821269e7fdffa690c2d0836c74dfebd25
|
{
"arc:challenge": 53.2,
"hellaswag": 77.7,
"hendrycksTest": 43.5,
"truthfulqa:mc": 45.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 54.9 |
circulus/Llama-2-13b-orca-v1
|
main
|
e77ec90f432bdffa210a0e4310d117e5d1c662df
|
{
"arc:challenge": 62.2,
"hellaswag": 82.3,
"hendrycksTest": 57.7,
"truthfulqa:mc": 49.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 63 |
mosaicml/mpt-30b-chat
|
main
|
b7957743f18845ff8695f7919420adb769ec225e
|
{
"arc:challenge": 58.4,
"hellaswag": 82.4,
"hendrycksTest": 51,
"truthfulqa:mc": 52
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 61 |
mosaicml/mpt-30b
|
main
|
0261af71d7177453889f868d26607dec8d5aaa2e
|
{
"arc:challenge": 55.9,
"hellaswag": 82.4,
"hendrycksTest": 47.9,
"truthfulqa:mc": 38.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 56.2 |
mosaicml/mpt-7b-8k-chat
|
main
|
ef97b878a279cd1765fbed7b8321fb3cff1aa5b5
|
{
"arc:challenge": 48,
"hellaswag": 77.6,
"hendrycksTest": 41.9,
"truthfulqa:mc": 43.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 52.8 |
mosaicml/mpt-7b-chat
|
main
|
64e5c9c9fb53a8e89690c2dee75a5add37f7113e
|
{
"arc:challenge": 46.5,
"hellaswag": 75.5,
"hendrycksTest": 37.6,
"truthfulqa:mc": 40.2
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 50 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.