Add pipeline tag, paper link, project page link and Github repo
#18
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,4 +1,6 @@
|
|
| 1 |
---
|
|
|
|
|
|
|
| 2 |
language:
|
| 3 |
- en
|
| 4 |
- fr
|
|
@@ -24,14 +26,12 @@ language:
|
|
| 24 |
- vi
|
| 25 |
- hi
|
| 26 |
- bn
|
| 27 |
-
license: apache-2.0
|
| 28 |
library_name: vllm
|
|
|
|
| 29 |
inference: false
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
|
| 33 |
-
If you want to learn more about how we process your personal data, please read
|
| 34 |
-
our <a href="https://mistral.ai/terms/">Privacy Policy</a>.
|
| 35 |
---
|
| 36 |
|
| 37 |
# Model Card for Magistral-Small-2506
|
|
@@ -42,9 +42,11 @@ Magistral Small can be deployed locally, fitting within a single RTX 4090 or a 3
|
|
| 42 |
|
| 43 |
Learn more about Magistral in our [blog post](https://mistral.ai/news/magistral/).
|
| 44 |
|
|
|
|
|
|
|
| 45 |
## Key Features
|
| 46 |
- **Reasoning:** Capable of long chains of reasoning traces before providing an answer.
|
| 47 |
-
- **Multilingual:** Supports dozens of languages, including English, French, German, Greek, Hindi, Indonesian, Italian, Japanese, Korean, Malay, Nepali, Polish, Portuguese, Romanian, Russian, Serbian, Spanish,
|
| 48 |
- **Apache 2.0 License:** Open license allowing usage and modification for both commercial and non-commercial purposes.
|
| 49 |
- **Context Window:** A 128k context window, **but** performance might degrade past **40k**. Hence we recommend setting the maximum model length to 40k.
|
| 50 |
|
|
@@ -208,7 +210,8 @@ for chunk in stream:
|
|
| 208 |
if content is not None:
|
| 209 |
if not printed_content:
|
| 210 |
printed_content = True
|
| 211 |
-
print("
|
|
|
|
| 212 |
# Extract and print the content
|
| 213 |
print(content, end="", flush=True)
|
| 214 |
|
|
|
|
| 1 |
---
|
| 2 |
+
base_model:
|
| 3 |
+
- mistralai/Mistral-Small-3.1-24B-Instruct-2503
|
| 4 |
language:
|
| 5 |
- en
|
| 6 |
- fr
|
|
|
|
| 26 |
- vi
|
| 27 |
- hi
|
| 28 |
- bn
|
|
|
|
| 29 |
library_name: vllm
|
| 30 |
+
license: apache-2.0
|
| 31 |
inference: false
|
| 32 |
+
extra_gated_description: If you want to learn more about how we process your personal
|
| 33 |
+
data, please read our <a href="https://mistral.ai/terms/">Privacy Policy</a>.
|
| 34 |
+
pipeline_tag: text-generation
|
|
|
|
|
|
|
| 35 |
---
|
| 36 |
|
| 37 |
# Model Card for Magistral-Small-2506
|
|
|
|
| 42 |
|
| 43 |
Learn more about Magistral in our [blog post](https://mistral.ai/news/magistral/).
|
| 44 |
|
| 45 |
+
The model was presented in the paper [Magistral](https://huggingface.co/papers/2506.10910).
|
| 46 |
+
|
| 47 |
## Key Features
|
| 48 |
- **Reasoning:** Capable of long chains of reasoning traces before providing an answer.
|
| 49 |
+
- **Multilingual:** Supports dozens of languages, including English, French, German, Greek, Hindi, Indonesian, Italian, Japanese, Korean, Malay, Nepali, Polish, Portuguese, Romanian, Russian, Serbian, Spanish, Turkish, Ukrainian, Vietnamese, Arabic, Bengali, Chinese, and Farsi.
|
| 50 |
- **Apache 2.0 License:** Open license allowing usage and modification for both commercial and non-commercial purposes.
|
| 51 |
- **Context Window:** A 128k context window, **but** performance might degrade past **40k**. Hence we recommend setting the maximum model length to 40k.
|
| 52 |
|
|
|
|
| 210 |
if content is not None:
|
| 211 |
if not printed_content:
|
| 212 |
printed_content = True
|
| 213 |
+
print("
|
| 214 |
+
content:", end="", flush=True)
|
| 215 |
# Extract and print the content
|
| 216 |
print(content, end="", flush=True)
|
| 217 |
|