Update README.md
Browse files
README.md
CHANGED
|
@@ -29,7 +29,6 @@ Static quants of [`Qwen/Qwen3-Reranker-8B`](https://huggingface.co/Qwen/Qwen3-Re
|
|
| 29 |
| [GGUF](https://huggingface.co/giladgd/Qwen3-Reranker-8B-GGUF/resolve/main/Qwen3-Reranker-8B.Q5_K_M.gguf) | `hf:giladgd/Qwen3-Reranker-8B-GGUF:Q5_K_M` | Q5_K_M | 5.4GB |
|
| 30 |
| [GGUF](https://huggingface.co/giladgd/Qwen3-Reranker-8B-GGUF/resolve/main/Qwen3-Reranker-8B.Q6_K.gguf) | `hf:giladgd/Qwen3-Reranker-8B-GGUF:Q6_K` | Q6_K | 6.2GB |
|
| 31 |
| [GGUF](https://huggingface.co/giladgd/Qwen3-Reranker-8B-GGUF/resolve/main/Qwen3-Reranker-8B.Q8_0.gguf) | `hf:giladgd/Qwen3-Reranker-8B-GGUF:Q8_0` | Q8_0 | 8.0GB |
|
| 32 |
-
| [GGUF](https://huggingface.co/giladgd/Qwen3-Reranker-8B-GGUF/resolve/main/Qwen3-Reranker-8B.MXFP4.gguf) | `hf:giladgd/Qwen3-Reranker-8B-GGUF:MXFP4` | MXFP4 | 8.0GB |
|
| 33 |
| [GGUF](https://huggingface.co/giladgd/Qwen3-Reranker-8B-GGUF/resolve/main/Qwen3-Reranker-8B.F16.gguf) | `hf:giladgd/Qwen3-Reranker-8B-GGUF:F16` | F16 | 15.1GB |
|
| 34 |
|
| 35 |
> [!TIP]
|
|
@@ -49,7 +48,7 @@ brew install nodejs
|
|
| 49 |
### CLI
|
| 50 |
Chat with the model:
|
| 51 |
```bash
|
| 52 |
-
npx -y node-llama-cpp chat hf:giladgd/Qwen3-Reranker-8B-GGUF:
|
| 53 |
```
|
| 54 |
|
| 55 |
### Code
|
|
@@ -61,7 +60,7 @@ npm install node-llama-cpp
|
|
| 61 |
```typescript
|
| 62 |
import {getLlama, resolveModelFile, LlamaChatSession} from "node-llama-cpp";
|
| 63 |
|
| 64 |
-
const modelUri = "hf:giladgd/Qwen3-Reranker-8B-GGUF:
|
| 65 |
|
| 66 |
|
| 67 |
const llama = await getLlama();
|
|
@@ -94,10 +93,10 @@ brew install llama.cpp
|
|
| 94 |
|
| 95 |
### CLI
|
| 96 |
```bash
|
| 97 |
-
llama-cli -hf giladgd/Qwen3-Reranker-8B-GGUF:
|
| 98 |
```
|
| 99 |
|
| 100 |
### Server
|
| 101 |
```bash
|
| 102 |
-
llama-server -hf giladgd/Qwen3-Reranker-8B-GGUF:
|
| 103 |
```
|
|
|
|
| 29 |
| [GGUF](https://huggingface.co/giladgd/Qwen3-Reranker-8B-GGUF/resolve/main/Qwen3-Reranker-8B.Q5_K_M.gguf) | `hf:giladgd/Qwen3-Reranker-8B-GGUF:Q5_K_M` | Q5_K_M | 5.4GB |
|
| 30 |
| [GGUF](https://huggingface.co/giladgd/Qwen3-Reranker-8B-GGUF/resolve/main/Qwen3-Reranker-8B.Q6_K.gguf) | `hf:giladgd/Qwen3-Reranker-8B-GGUF:Q6_K` | Q6_K | 6.2GB |
|
| 31 |
| [GGUF](https://huggingface.co/giladgd/Qwen3-Reranker-8B-GGUF/resolve/main/Qwen3-Reranker-8B.Q8_0.gguf) | `hf:giladgd/Qwen3-Reranker-8B-GGUF:Q8_0` | Q8_0 | 8.0GB |
|
|
|
|
| 32 |
| [GGUF](https://huggingface.co/giladgd/Qwen3-Reranker-8B-GGUF/resolve/main/Qwen3-Reranker-8B.F16.gguf) | `hf:giladgd/Qwen3-Reranker-8B-GGUF:F16` | F16 | 15.1GB |
|
| 33 |
|
| 34 |
> [!TIP]
|
|
|
|
| 48 |
### CLI
|
| 49 |
Chat with the model:
|
| 50 |
```bash
|
| 51 |
+
npx -y node-llama-cpp chat hf:giladgd/Qwen3-Reranker-8B-GGUF:Q4_K_M
|
| 52 |
```
|
| 53 |
|
| 54 |
### Code
|
|
|
|
| 60 |
```typescript
|
| 61 |
import {getLlama, resolveModelFile, LlamaChatSession} from "node-llama-cpp";
|
| 62 |
|
| 63 |
+
const modelUri = "hf:giladgd/Qwen3-Reranker-8B-GGUF:Q4_K_M";
|
| 64 |
|
| 65 |
|
| 66 |
const llama = await getLlama();
|
|
|
|
| 93 |
|
| 94 |
### CLI
|
| 95 |
```bash
|
| 96 |
+
llama-cli -hf giladgd/Qwen3-Reranker-8B-GGUF:Q4_K_M -p "The meaning to life and the universe is"
|
| 97 |
```
|
| 98 |
|
| 99 |
### Server
|
| 100 |
```bash
|
| 101 |
+
llama-server -hf giladgd/Qwen3-Reranker-8B-GGUF:Q4_K_M -c 2048
|
| 102 |
```
|