Update README.md
Browse files
README.md
CHANGED
|
@@ -2627,7 +2627,7 @@ The model is further trained on Jina AI's collection of more than 400 millions o
|
|
| 2627 |
These pairs were obtained from various domains and were carefully selected through a thorough cleaning process.
|
| 2628 |
|
| 2629 |
The embedding model was trained using 512 sequence length, but extrapolates to 8k sequence length (or even longer) thanks to ALiBi.
|
| 2630 |
-
This makes our model useful for a range of use cases, especially when processing long documents is needed, including long document retrieval, semantic textual similarity, text reranking, recommendation, RAG and LLM-based generative search
|
| 2631 |
|
| 2632 |
This model has 33 million parameters, which enables lightning-fast and memory efficient inference, while still delivering impressive performance.
|
| 2633 |
Additionally, we provide the following embedding models:
|
|
|
|
| 2627 |
These pairs were obtained from various domains and were carefully selected through a thorough cleaning process.
|
| 2628 |
|
| 2629 |
The embedding model was trained using 512 sequence length, but extrapolates to 8k sequence length (or even longer) thanks to ALiBi.
|
| 2630 |
+
This makes our model useful for a range of use cases, especially when processing long documents is needed, including long document retrieval, semantic textual similarity, text reranking, recommendation, RAG and LLM-based generative search, etc.
|
| 2631 |
|
| 2632 |
This model has 33 million parameters, which enables lightning-fast and memory efficient inference, while still delivering impressive performance.
|
| 2633 |
Additionally, we provide the following embedding models:
|