jina-embeddings-v5-text-small-retrieval-GGUF
GGUF quantizations of jina-embeddings-v5-text-small-retrieval using llama.cpp. A 677M parameter multilingual embedding model quantized for efficient inference.
Elastic Inference Service | ArXiv | Blog
We highly recommend to first read this blog post for more technical details and customized llama.cpp build.
Overview
jina-embeddings-v5-text-small-retrieval is a task-specific embedding model for retrieval, part of the jina-embeddings-v5-text model family.
| Feature | Value |
|---|---|
| Parameters | 677M |
| Task | retrieval |
| Embedding Dimension | 1024 |
| Matryoshka Dimensions | 32, 64, 128, 256, 512, 768, 1024 |
| Pooling Strategy | Last-token pooling |
| Base Model | jina-embeddings-v5-text-small |
Usage with llama.cpp
via Elastic Inference Service
The fastest way to use v5-text in production. Elastic Inference Service (EIS) provides managed embedding inference with built-in scaling, so you can generate embeddings directly within your Elastic deployment.
PUT _inference/text_embedding/jina-v5
{
"service": "elastic",
"service_settings": {
"model_id": "jina-embeddings-v5-text-small"
}
}
See the Elastic Inference Service documentation for setup details.
# Build llama.cpp (upstream)
git clone https://github.com/ggml-org/llama.cpp
cd llama.cpp && cmake -B build && cmake --build build --config Release
# Run embedding
./build/bin/llama-embedding -m jina-embeddings-v5-text-small-retrieval-Q8_0.gguf \
--pooling last -p "Your text here"
License
CC-BY-NC-4.0. For commercial use, please contact us.
- Downloads last month
- 5,731
1-bit
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
16-bit
Model tree for jinaai/jina-embeddings-v5-text-small-retrieval-GGUF
Base model
Qwen/Qwen3-0.6B-Base