Skip to content
Athrael.net logo Athrael.net
Go back

You too can run the Vidore Benchmark with less than 32GB of GPU VRAM

Edit page

yes

TL;DR

Yes, you too can run the Vidore benchmark on Poorware GPU. Use half‑precision, small batches, and keep preprocessing lean.

Minimal Python runner (MTEB)

Install deps and run this minimal script to execute ViDoRe via MTEB with tiny batches.

pip install mteb colpali-engine --pre torch torchvision --index-url https://download.pytorch.org/whl/nightly/cu129
# mteb_run_vidore.py
import os, torch, mteb

# --- (optional but helpful) allocator + precision tweaks ---
os.environ["PYTORCH_ALLOC_CONF"] = "expandable_segments:True,max_split_size_mb:128"
os.environ["PYTORCH_CUDA_ALLOC_CONF"] = os.environ["PYTORCH_ALLOC_CONF"]  # backward compat
torch.set_float32_matmul_precision("medium")  # slightly lighter kernels

# --- load the pre-defined model from MTEB ---
model_name = "vidore/colqwen2.5-v0.2"
model = mteb.get_model(model_name)  # uses the correct wrapper internally

# --- select the ViDoRe benchmarks ---
benchmarks = mteb.get_benchmarks(names=["ViDoRe(v1)", "ViDoRe(v2)"])
evaluator = mteb.MTEB(tasks=benchmarks)

# --- run with small batches to stay under ~32 GB VRAM ---
results = evaluator.run(
    model,
    encode_kwargs={"batch_size": 1},     # This number measures your GPU poorness. A 1 means you're pretty darn poor, but better off than most. Remember, glass is half full.
    verbosity=2,
)
print(results)

Grab the Code

You can find the code here.

Troubleshooting

Interested in helping?

That’s it, have fun, bye bye!


Edit page
Share this post on:

Previous Post
The Most Beautiful RAG: Starring ColPali, Qdrant, Minio and Friends
Next Post
ColQwen2.5 FastAPI Integration