Update README.md
Browse files
README.md
CHANGED
|
@@ -28,7 +28,9 @@ widget:
|
|
| 28 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/5f0b462819cb630495b814d7/DVA4MnFUs3UHBnTrX9jG6.png" style="max-width: 550px; height: auto;">
|
| 29 |
</div>
|
| 30 |
|
| 31 |
-
## Please note:
|
|
|
|
|
|
|
| 32 |
|
| 33 |
# Model Card for Minerva-7B-base-v1.0
|
| 34 |
|
|
@@ -133,9 +135,11 @@ We assessed our model using the [LM-Evaluation-Harness](https://github.com/Eleut
|
|
| 133 |
|
| 134 |
All the reported benchmark data was already present in the LM-Evaluation-Harness suite.
|
| 135 |
|
| 136 |
-
|
| 137 |
-
|
| 138 |
-
|
|
|
|
|
|
|
| 139 |
<!-- | [xcopa](https://huggingface.co/datasets/xcopa) (0-shot) | 0.694 |
|
| 140 |
| [Hellaswag](https://huggingface.co/datasets/alexandrainst/m_hellaswag) (5-shot) | 0.5293 |
|
| 141 |
| [Belebele](https://huggingface.co/datasets/facebook/belebele) (5-shot) | 0.2333 |
|
|
@@ -144,9 +148,9 @@ All the reported benchmark data was already present in the LM-Evaluation-Harness
|
|
| 144 |
| [M MMLU](https://huggingface.co/datasets/alexandrainst/m_mmlu) (5-shot) | 0.2612 |
|
| 145 |
| [arc challenge](https://huggingface.co/datasets/alexandrainst/m_arc) (5-shot) | 0.3268 | -->
|
| 146 |
|
| 147 |
-
**English** Data:
|
| 148 |
-
| Task | Accuracy |
|
| 149 |
-
| --- | --- |
|
| 150 |
<!-- | [Hellaswag](https://huggingface.co/datasets/Rowan/hellaswag) (5-shot) | 0.6168 |
|
| 151 |
| [piqa](https://huggingface.co/datasets/piqa) (5-shot) | 0.7535 |
|
| 152 |
| [sciq](https://huggingface.co/datasets/sciq) (5-shot) | 0.925 |
|
|
|
|
| 28 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/5f0b462819cb630495b814d7/DVA4MnFUs3UHBnTrX9jG6.png" style="max-width: 550px; height: auto;">
|
| 29 |
</div>
|
| 30 |
|
| 31 |
+
## Please note:
|
| 32 |
+
|
| 33 |
+
This model is currently training, the weights available at the moment are taken from the checkpoint at 10K steps out of 590K.
|
| 34 |
|
| 35 |
# Model Card for Minerva-7B-base-v1.0
|
| 36 |
|
|
|
|
| 135 |
|
| 136 |
All the reported benchmark data was already present in the LM-Evaluation-Harness suite.
|
| 137 |
|
| 138 |
+
_Scores will be available at later stage._
|
| 139 |
+
|
| 140 |
+
<!-- **Italian** Data: -->
|
| 141 |
+
<!-- | Task | Accuracy |
|
| 142 |
+
| --- | --- | -->
|
| 143 |
<!-- | [xcopa](https://huggingface.co/datasets/xcopa) (0-shot) | 0.694 |
|
| 144 |
| [Hellaswag](https://huggingface.co/datasets/alexandrainst/m_hellaswag) (5-shot) | 0.5293 |
|
| 145 |
| [Belebele](https://huggingface.co/datasets/facebook/belebele) (5-shot) | 0.2333 |
|
|
|
|
| 148 |
| [M MMLU](https://huggingface.co/datasets/alexandrainst/m_mmlu) (5-shot) | 0.2612 |
|
| 149 |
| [arc challenge](https://huggingface.co/datasets/alexandrainst/m_arc) (5-shot) | 0.3268 | -->
|
| 150 |
|
| 151 |
+
<!-- **English** Data: -->
|
| 152 |
+
<!-- | Task | Accuracy |
|
| 153 |
+
| --- | --- | -->
|
| 154 |
<!-- | [Hellaswag](https://huggingface.co/datasets/Rowan/hellaswag) (5-shot) | 0.6168 |
|
| 155 |
| [piqa](https://huggingface.co/datasets/piqa) (5-shot) | 0.7535 |
|
| 156 |
| [sciq](https://huggingface.co/datasets/sciq) (5-shot) | 0.925 |
|