Dataset Viewer
Auto-converted to Parquet Duplicate
Search is not available for this dataset
id
int64
0
1,000k
emb
listlengths
1.02k
1.02k
0
[ 0.06068156659603119, 0.05680697783827782, -0.027472976595163345, -0.0007493669982068241, -0.006521209143102169, -0.06852227449417114, -0.021981433033943176, 0.027137381955981255, 0.06199343502521515, -0.018366165459156036, 0.05345103144645691, 0.04622050002217293, 0.06431208550930023, -0.0...
1
[ 0.030499424785375595, 0.02477792277932167, -0.032528650015592575, 0.021634910255670547, 0.0009092421969398856, -0.0009650268475525081, 0.11107344180345535, 0.020994100719690323, -0.06286024302244186, 0.018186749890446663, -0.0016210925532504916, -0.028424426913261414, -0.006549213547259569, ...
2
[ -0.06523730605840683, 0.023403653874993324, 0.018704615533351898, -0.04958401247859001, 0.06178930774331093, -0.009443847462534904, 0.014425133354961872, 0.025951510295271873, 0.01737728901207447, 0.005549595225602388, -0.07127892225980759, -0.050651974976062775, 0.004699038807302713, 0.00...
3
[ 0.03603912517428398, -0.04040287807583809, 0.00981082022190094, 0.024412615224719048, -0.01589871570467949, -0.02981390804052353, 0.07287166267633438, 0.020598145201802254, -0.01678367331624031, 0.025984177365899086, -0.02741841971874237, 0.00018976994033437222, 0.05605746805667877, 0.0194...
4
[ -0.06897637993097305, 0.050114698708057404, -0.03741816058754921, -0.05399080738425255, 0.03912731260061264, -0.013261167332530022, -0.03381673991680145, -0.0238365288823843, 0.035281725227832794, 0.045841824263334274, -0.05426549166440964, 0.06229238957166672, -0.038791585713624954, 0.054...
5
[ 0.000014002781426825095, 0.016413405537605286, -0.02030319906771183, -0.03807421773672104, -0.027594655752182007, -0.042375873774290085, 0.09841941297054291, 0.0014319780748337507, 0.05061308294534683, 0.0006573561113327742, -0.02028794400393963, -0.008710087276995182, 0.05241306498646736, ...
6
[ 0.06329439580440521, 0.008224609307944775, 0.0003826674947049469, 0.0080186128616333, -0.04705880209803581, 0.016754398122429848, 0.0328679196536541, 0.06015104055404663, -0.064698226749897, -0.033539317548274994, -0.014381621964275837, 0.04299990460276604, -0.05102614685893059, -0.0401312...
7
[ 0.07851675152778625, 0.01173019502311945, 0.015126381069421768, -0.021048717200756073, 0.009028511121869087, -0.045608099550008774, -0.027215274050831795, 0.06661102920770645, -0.00927273090928793, -0.015904832631349564, 0.0020644229371100664, 0.059833917766809464, 0.012203372083604336, 0....
8
[ -0.022780608385801315, 0.02464211918413639, 0.012687258422374725, -0.019332237541675568, -0.000775788095779717, 0.002660662168636918, -0.03457525745034218, -0.002792264800518751, -0.0007333509856835008, 0.022856900468468666, -0.04467623680830002, -0.007571920286864042, 0.03396492823958397, ...
9
[0.036049798130989075,0.04642822965979576,0.014293240383267403,0.03641609475016594,-0.00949321500957(...TRUNCATED)
End of preview. Expand in Data Studio

Arctic-Wiki-Arabic-1M

VDBBench-compatible vector benchmark case published as a Hugging Face dataset repository.

What’s in this repo

  • vdbbench/: a Parquet-only folder intended to be downloaded as-is for VDBBench.
  • The Dataset Viewer indexes only train/test (and optionally shuffled), and intentionally ignores neighbors.parquet.

vdbbench/ file contract (VDBBench)

  • Train vectors: train.parquet or train-*-of-*.parquet
  • Query vectors: test.parquet or test-*-of-*.parquet
  • Ground truth: neighbors.parquet with columns id, neighbors_id (list[int])
  • Optional shuffled train: shuffle_train.parquet or shuffle_train-*-of-*.parquet

Important: do not add non-parquet files inside vdbbench/ (VDBBench requires the folder to be parquet-only).

Download for VDBBench

Download only the parquet files and point VDBBench at the vdbbench/ folder.

hf download "AI71ai/Arctic-Wiki-Arabic-1M" \
  --repo-type dataset \
  --include "vdbbench/*.parquet" \
  --local-dir ./my_case

Then set VDBBench /customFolder Path to:

  • ./my_case/vdbbench

How this case was created

  • Source: wikimedia/wikipedia at revision b04c8d1ceb2f5cd4588862100d08de323dccfbaa (pinned for reproducibility when set).
  • Embedding model: Snowflake/snowflake-arctic-embed-l-v2.0 (dimension 1024).
  • Languages: ar
  • Train size: 1,000,000 vectors
  • Test size: 1,000 query vectors
  • Phase 2 seed: 42

Size filtering (chosen to embed well)

We filter Wikipedia articles by character length of the article text to avoid pages that are too short (low-signal) or too long (risking truncation / poor embedding behavior).

Approx token ranges below are computed from the Snowflake Arctic tokenizer’s observed chars/token ratios (using the per-language ~8k-token max lengths we chose).

Language Text length filter (chars) Approx token range
ar 226–26,350 chars ~69–~8,000 tokens

Test set (queries)

  • Queries are derived from Wikipedia titles of documents that appear in the training stream.
  • Phase 2 Stage 1 selects test_size titles using reservoir sampling (bounded memory) while writing train.
  • Phase 2 Stage 2 embeds those titles to produce test*.parquet (query vectors).

Shuffled train (optional)

  • Some vector DBs are sensitive to ingestion order; this benchmark optionally provides a shuffled training set to evaluate that behavior.
  • This repo includes shuffle_train*.parquet for VDBBench “Use Shuffled Data” mode.

Ground truth (neighbors.parquet)

  • neighbors.parquet contains the exact top-400 nearest train IDs for each test query ID.
  • Similarity is computed with cosine (implemented as L2-normalize + inner product).

Provenance: ID map sidecars (maps/)

This repo also includes optional ID mapping sidecars under maps/ that let you trace VDBBench IDs back to the original Wikipedia/source IDs used during embedding.

  • maps/ is intentionally not part of any Dataset Viewer split.
  • Files are Parquet (same sharding convention as vectors when applicable).

Files

  • train_id_map*.parquet: train VDBBench id → source id mapping
  • test_id_map*.parquet: test query id → train_id + source id mapping
  • shuffle_train_id_map*.parquet: shuffled-train VDBBench id → source id mapping

Schemas

  • train_id_map: id, source_id
  • test_id_map: id, train_id, source_id
  • shuffle_train_id_map: id, source_id

Notes:

  • source_id refers to the upstream article ID from the embedded Wikipedia dump (see the pipeline repo for details).
  • For mixed-language cases, the language column indicates which language shard the source came from.

Design decisions

  • Per-case repos: keeps each benchmark case independently versionable and easily downloadable without filtering huge multi-case repos.
  • Parquet-only vdbbench/ folder: matches VDBBench’s strict folder selection rule and enables “download folder → run benchmark”.
  • Neighbors kept next to vectors, but not indexed: neighbors.parquet is required by VDBBench, but it’s not useful to display as a dataset split in the viewer.
  • Prefer single-file test.parquet: VDBBench duplicates test queries across processes; keeping queries small and single-file reduces overhead.
  • Embedding model choice: Snowflake Arctic Embed L v2.0 is high on the MTEB-R leaderboard (see https://huggingface.co/spaces/mteb/leaderboard) and is a strong fit for evaluating compressed representations (including binarized/quantized vectors).

Usage with datasets

from datasets import load_dataset

ds = load_dataset("AI71ai/Arctic-Wiki-Arabic-1M")
print(ds)

Note: neighbors.parquet is intentionally not part of any split.

Downloads last month
75