config dict | results dict | versions null |
|---|---|---|
{
"model_dtype": "bfloat16",
"model_name": "Almawave/Velvet-14B (0-shot)",
"model_sha": "main"
} | {
"TextualEntailment": {
"acc": 67.63
},
"TextualEntailment_best": {
"acc": 78.5
},
"Sentiment Analysis": {
"acc": 60.49
},
"Sentiment Analysis_best": {
"acc": 66.59
},
"Hate Speech": {
"acc": 55.93
},
"Hate Speech_best": {
"acc": 65.64
},
"Admission Test": {
"acc":... | null |
{
"model_dtype": "bfloat16",
"model_name": "Almawave/Velvet-14B (5-shot)",
"model_sha": "main"
} | {
"TextualEntailment": {
"acc": 74.46
},
"TextualEntailment_best": {
"acc": 76.75
},
"Sentiment Analysis": {
"acc": 67.99
},
"Sentiment Analysis_best": {
"acc": 68.93
},
"Hate Speech": {
"acc": 66.35
},
"Hate Speech_best": {
"acc": 68.97
},
"Admission Test": {
"acc"... | null |
{
"model_dtype": "bfloat16",
"model_name": "CohereForAI/aya-expanse-8b",
"model_sha": "main"
} | {
"TextualEntailment": {
"acc": 60.33
},
"TextualEntailment_best": {
"acc": 75
},
"Sentiment Analysis": {
"acc": 60.16
},
"Sentiment Analysis_best": {
"acc": 70.62
},
"Hate Speech": {
"acc": 60.48
},
"Hate Speech_best": {
"acc": 62.97
},
"Admission Test": {
"acc": 3... | null |
{
"model_dtype": "bfloat16",
"model_name": "FairMind/Llama-3-8B-4bit-UltraChat-Ita",
"model_sha": "main"
} | {
"TextualEntailment": {
"acc": 58.08
},
"TextualEntailment_best": {
"acc": 64.25
},
"Sentiment Analysis": {
"acc": 46.01
},
"Sentiment Analysis_best": {
"acc": 67.3
},
"Hate Speech": {
"acc": 48.72
},
"Hate Speech_best": {
"acc": 63.38
},
"Admission Test": {
"acc":... | null |
{
"model_dtype": "bfloat16",
"model_name": "MoxoffSpA/Volare",
"model_sha": "main"
} | {
"TextualEntailment": {
"acc": 51.25
},
"TextualEntailment_best": {
"acc": 55.5
},
"Sentiment Analysis": {
"acc": 44.15
},
"Sentiment Analysis_best": {
"acc": 58.49
},
"Hate Speech": {
"acc": 43.57
},
"Hate Speech_best": {
"acc": 62.7
},
"Admission Test": {
"acc": ... | null |
{
"model_dtype": "bfloat16",
"model_name": "arcee-ai/Llama-3.1-SuperNova-Lite",
"model_sha": "main"
} | {
"TextualEntailment": {
"acc": 68.54
},
"TextualEntailment_best": {
"acc": 74.5
},
"Sentiment Analysis": {
"acc": 63.8
},
"Sentiment Analysis_best": {
"acc": 77.12
},
"Hate Speech": {
"acc": 52.21
},
"Hate Speech_best": {
"acc": 63.8
},
"Admission Test": {
"acc": 4... | null |
{
"model_dtype": "bfloat16",
"model_name": "iGeniusAI/Italia-9B-Instruct-v0.1",
"model_sha": "main"
} | {
"TextualEntailment": {
"acc": 59.33
},
"TextualEntailment_best": {
"acc": 70.25
},
"Sentiment Analysis": {
"acc": 48.41
},
"Sentiment Analysis_best": {
"acc": 59.26
},
"Hate Speech": {
"acc": 58.9
},
"Hate Speech_best": {
"acc": 63.58
},
"Admission Test": {
"acc":... | null |
{
"model_dtype": "bfloat16",
"model_name": "ibm-granite/granite-3.1-8b-instruct",
"model_sha": "main"
} | {
"TextualEntailment": {
"acc": 51.08
},
"TextualEntailment_best": {
"acc": 67
},
"Sentiment Analysis": {
"acc": 47.82
},
"Sentiment Analysis_best": {
"acc": 69.45
},
"Hate Speech": {
"acc": 39.51
},
"Hate Speech_best": {
"acc": 62.64
},
"Admission Test": {
"acc": 4... | null |
{
"model_dtype": "bfloat16",
"model_name": "microsoft/Phi-3.5-mini-instruct",
"model_sha": "main"
} | {
"TextualEntailment": {
"acc": 70.04
},
"TextualEntailment_best": {
"acc": 81.5
},
"Sentiment Analysis": {
"acc": 43.73
},
"Sentiment Analysis_best": {
"acc": 70.59
},
"Hate Speech": {
"acc": 64.49
},
"Hate Speech_best": {
"acc": 69.17
},
"Admission Test": {
"acc":... | null |
{
"model_dtype": "bfloat16",
"model_name": "microsoft/phi-4",
"model_sha": "main"
} | {
"TextualEntailment": {
"acc": 60.21
},
"TextualEntailment_best": {
"acc": 73
},
"Sentiment Analysis": {
"acc": 42.24
},
"Sentiment Analysis_best": {
"acc": 74.14
},
"Hate Speech": {
"acc": 60.44
},
"Hate Speech_best": {
"acc": 63.1
},
"Admission Test": {
"acc": 50... | null |
{
"model_dtype": "bfloat16",
"model_name": "mii-llm/maestrale-chat-v0.4-beta",
"model_sha": "main"
} | {
"TextualEntailment": {
"acc": 55.63
},
"TextualEntailment_best": {
"acc": 73.5
},
"Sentiment Analysis": {
"acc": 61.19
},
"Sentiment Analysis_best": {
"acc": 67.46
},
"Hate Speech": {
"acc": 60.67
},
"Hate Speech_best": {
"acc": 62.6
},
"Admission Test": {
"acc": ... | null |
{
"model_dtype": "bfloat16",
"model_name": "mii-llm/qwen-5588-dpo-iter1",
"model_sha": "main"
} | {
"TextualEntailment": {
"acc": 67.46
},
"TextualEntailment_best": {
"acc": 82
},
"Sentiment Analysis": {
"acc": 55.56
},
"Sentiment Analysis_best": {
"acc": 70.76
},
"Hate Speech": {
"acc": 62.69
},
"Hate Speech_best": {
"acc": 70.73
},
"Admission Test": {
"acc": 4... | null |
{
"model_dtype": "bfloat16",
"model_name": "mistralai/Mistral-7B-Instruct-v0.3",
"model_sha": "main"
} | {
"TextualEntailment": {
"acc": 60.83
},
"TextualEntailment_best": {
"acc": 71.25
},
"Sentiment Analysis": {
"acc": 58.71
},
"Sentiment Analysis_best": {
"acc": 66.14
},
"Hate Speech": {
"acc": 55.86
},
"Hate Speech_best": {
"acc": 65.39
},
"Admission Test": {
"acc"... | null |
{
"model_dtype": "bfloat16",
"model_name": "occiglot/occiglot-7b-it-en-instruct",
"model_sha": "main"
} | {
"TextualEntailment": {
"acc": 49.92
},
"TextualEntailment_best": {
"acc": 56
},
"Sentiment Analysis": {
"acc": 51.87
},
"Sentiment Analysis_best": {
"acc": 61.09
},
"Hate Speech": {
"acc": 48.49
},
"Hate Speech_best": {
"acc": 62.72
},
"Admission Test": {
"acc": 4... | null |
{
"model_dtype": "bfloat16",
"model_name": "sapienzanlp/Minerva-7B-instruct-v1.0",
"model_sha": "main"
} | {
"TextualEntailment": {
"acc": 54.13
},
"TextualEntailment_best": {
"acc": 56.5
},
"Sentiment Analysis": {
"acc": 44.52
},
"Sentiment Analysis_best": {
"acc": 59.46
},
"Hate Speech": {
"acc": 39.23
},
"Hate Speech_best": {
"acc": 60.48
},
"Admission Test": {
"acc":... | null |
{
"model_dtype": "bfloat16",
"model_name": "swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA",
"model_sha": "main"
} | {
"TextualEntailment": {
"acc": 58.92
},
"TextualEntailment_best": {
"acc": 69.1
},
"Sentiment Analysis": {
"acc": 60.95
},
"Sentiment Analysis_best": {
"acc": 72.04
},
"Hate Speech": {
"acc": 39.59
},
"Hate Speech_best": {
"acc": 66.32
},
"Admission Test": {
"acc":... | null |
README.md exists but content is empty.
- Downloads last month
- 24