id int64 0 1,000k | emb listlengths 1.02k 1.02k |
|---|---|
0 | [
0.06068156659603119,
0.05680697783827782,
-0.027472976595163345,
-0.0007493669982068241,
-0.006521209143102169,
-0.06852227449417114,
-0.021981433033943176,
0.027137381955981255,
0.06199343502521515,
-0.018366165459156036,
0.05345103144645691,
0.04622050002217293,
0.06431208550930023,
-0.0... |
1 | [
0.030499424785375595,
0.02477792277932167,
-0.032528650015592575,
0.021634910255670547,
0.0009092421969398856,
-0.0009650268475525081,
0.11107344180345535,
0.020994100719690323,
-0.06286024302244186,
0.018186749890446663,
-0.0016210925532504916,
-0.028424426913261414,
-0.006549213547259569,
... |
2 | [
-0.06523730605840683,
0.023403653874993324,
0.018704615533351898,
-0.04958401247859001,
0.06178930774331093,
-0.009443847462534904,
0.014425133354961872,
0.025951510295271873,
0.01737728901207447,
0.005549595225602388,
-0.07127892225980759,
-0.050651974976062775,
0.004699038807302713,
0.00... |
3 | [
0.03603912517428398,
-0.04040287807583809,
0.00981082022190094,
0.024412615224719048,
-0.01589871570467949,
-0.02981390804052353,
0.07287166267633438,
0.020598145201802254,
-0.01678367331624031,
0.025984177365899086,
-0.02741841971874237,
0.00018976994033437222,
0.05605746805667877,
0.0194... |
4 | [
-0.06897637993097305,
0.050114698708057404,
-0.03741816058754921,
-0.05399080738425255,
0.03912731260061264,
-0.013261167332530022,
-0.03381673991680145,
-0.0238365288823843,
0.035281725227832794,
0.045841824263334274,
-0.05426549166440964,
0.06229238957166672,
-0.038791585713624954,
0.054... |
5 | [
0.000014002781426825095,
0.016413405537605286,
-0.02030319906771183,
-0.03807421773672104,
-0.027594655752182007,
-0.042375873774290085,
0.09841941297054291,
0.0014319780748337507,
0.05061308294534683,
0.0006573561113327742,
-0.02028794400393963,
-0.008710087276995182,
0.05241306498646736,
... |
6 | [
0.06329439580440521,
0.008224609307944775,
0.0003826674947049469,
0.0080186128616333,
-0.04705880209803581,
0.016754398122429848,
0.0328679196536541,
0.06015104055404663,
-0.064698226749897,
-0.033539317548274994,
-0.014381621964275837,
0.04299990460276604,
-0.05102614685893059,
-0.0401312... |
7 | [
0.07851675152778625,
0.01173019502311945,
0.015126381069421768,
-0.021048717200756073,
0.009028511121869087,
-0.045608099550008774,
-0.027215274050831795,
0.06661102920770645,
-0.00927273090928793,
-0.015904832631349564,
0.0020644229371100664,
0.059833917766809464,
0.012203372083604336,
0.... |
8 | [
-0.022780608385801315,
0.02464211918413639,
0.012687258422374725,
-0.019332237541675568,
-0.000775788095779717,
0.002660662168636918,
-0.03457525745034218,
-0.002792264800518751,
-0.0007333509856835008,
0.022856900468468666,
-0.04467623680830002,
-0.007571920286864042,
0.03396492823958397,
... |
9 | [0.036049798130989075,0.04642822965979576,0.014293240383267403,0.03641609475016594,-0.00949321500957(...TRUNCATED) |
Arctic-Wiki-Arabic-1M
VDBBench-compatible vector benchmark case published as a Hugging Face dataset repository.
What’s in this repo
vdbbench/: a Parquet-only folder intended to be downloaded as-is for VDBBench.- The Dataset Viewer indexes only train/test (and optionally shuffled), and intentionally ignores
neighbors.parquet.
vdbbench/ file contract (VDBBench)
- Train vectors:
train.parquetortrain-*-of-*.parquet - Query vectors:
test.parquetortest-*-of-*.parquet - Ground truth:
neighbors.parquetwith columnsid,neighbors_id(list[int]) - Optional shuffled train:
shuffle_train.parquetorshuffle_train-*-of-*.parquet
Important: do not add non-parquet files inside vdbbench/ (VDBBench requires the folder to be parquet-only).
Download for VDBBench
Download only the parquet files and point VDBBench at the vdbbench/ folder.
hf download "AI71ai/Arctic-Wiki-Arabic-1M" \
--repo-type dataset \
--include "vdbbench/*.parquet" \
--local-dir ./my_case
Then set VDBBench /custom → Folder Path to:
./my_case/vdbbench
How this case was created
- Source:
wikimedia/wikipediaat revisionb04c8d1ceb2f5cd4588862100d08de323dccfbaa(pinned for reproducibility when set). - Embedding model:
Snowflake/snowflake-arctic-embed-l-v2.0(dimension 1024). - Languages:
ar - Train size:
1,000,000vectors - Test size:
1,000query vectors - Phase 2 seed:
42
Size filtering (chosen to embed well)
We filter Wikipedia articles by character length of the article text to avoid pages that are too short (low-signal) or too long (risking truncation / poor embedding behavior).
Approx token ranges below are computed from the Snowflake Arctic tokenizer’s observed chars/token ratios (using the per-language ~8k-token max lengths we chose).
| Language | Text length filter (chars) | Approx token range |
|---|---|---|
ar |
226–26,350 chars |
~69–~8,000 tokens |
Test set (queries)
- Queries are derived from Wikipedia titles of documents that appear in the training stream.
- Phase 2 Stage 1 selects
test_sizetitles using reservoir sampling (bounded memory) while writing train. - Phase 2 Stage 2 embeds those titles to produce
test*.parquet(query vectors).
Shuffled train (optional)
- Some vector DBs are sensitive to ingestion order; this benchmark optionally provides a shuffled training set to evaluate that behavior.
- This repo includes
shuffle_train*.parquetfor VDBBench “Use Shuffled Data” mode.
Ground truth (neighbors.parquet)
neighbors.parquetcontains the exact top-400nearest train IDs for each test query ID.- Similarity is computed with cosine (implemented as L2-normalize + inner product).
Provenance: ID map sidecars (maps/)
This repo also includes optional ID mapping sidecars under maps/ that let you trace VDBBench IDs back to the original Wikipedia/source IDs used during embedding.
maps/is intentionally not part of any Dataset Viewer split.- Files are Parquet (same sharding convention as vectors when applicable).
Files
train_id_map*.parquet: train VDBBench id → source id mappingtest_id_map*.parquet: test query id → train_id + source id mappingshuffle_train_id_map*.parquet: shuffled-train VDBBench id → source id mapping
Schemas
train_id_map:id,source_idtest_id_map:id,train_id,source_idshuffle_train_id_map:id,source_id
Notes:
source_idrefers to the upstream article ID from the embedded Wikipedia dump (see the pipeline repo for details).- For mixed-language cases, the
languagecolumn indicates which language shard the source came from.
Design decisions
- Per-case repos: keeps each benchmark case independently versionable and easily downloadable without filtering huge multi-case repos.
- Parquet-only
vdbbench/folder: matches VDBBench’s strict folder selection rule and enables “download folder → run benchmark”. - Neighbors kept next to vectors, but not indexed:
neighbors.parquetis required by VDBBench, but it’s not useful to display as a dataset split in the viewer. - Prefer single-file
test.parquet: VDBBench duplicates test queries across processes; keeping queries small and single-file reduces overhead. - Embedding model choice: Snowflake Arctic Embed L v2.0 is high on the MTEB-R leaderboard (see
https://huggingface.co/spaces/mteb/leaderboard) and is a strong fit for evaluating compressed representations (including binarized/quantized vectors).
Usage with datasets
from datasets import load_dataset
ds = load_dataset("AI71ai/Arctic-Wiki-Arabic-1M")
print(ds)
Note: neighbors.parquet is intentionally not part of any split.
- Downloads last month
- 75