Q2 2026 Open-Source Embedding Models Benchmark: BGE, GTE, E5, Stella, Nomic
Last Updated: 2026-05-16
Architecture at a glance





The open-source embedding model space has stopped being a one-horse race. Two years ago you picked text-embedding-ada-002 or BGE-large and got on with your day. As of Q2 2026, the leaderboard has six families that all make a credible case: BGE-M3 from BAAI, GTE-Qwen2 from Alibaba, E5-Mistral-7B-Instruct from Microsoft Research / intfloat, Stella v5 1.5B from dunzhang, NV-Embed-v2 from NVIDIA, and Nomic Embed v2 from Nomic AI. Some are 568M-parameter encoders fine-tuned on retrieval; others are 7B causal LMs converted into embedders with latent-attention pooling; one is a sparse mixture-of-experts. This open-source embedding models benchmark 2026 post takes the public MTEB v2 leaderboard, latency numbers from model cards, and an industrial retrieval analysis on PLM / IoT documents, and turns them into a use-case-driven recommendation. No hype, no fabricated decimals.
What’s New for Embeddings in Q2 2026

Three structural shifts have reshaped the Q2 2026 leaderboard, and you should understand them before reading any numbers.
Shift one: decoder-LLM embedders have eaten the high end. A year ago, “best retrieval model” meant a 300M-parameter encoder fine-tuned on a few billion pairs. Today, the top of the MTEB v2 leaderboard is dominated by 7B-parameter causal LMs converted into embedders via instruction tuning, contrastive fine-tuning, and a pooling head (NV-Embed-v2 uses latent attention, GTE-Qwen2 uses last-token, E5-Mistral uses mean of last-token states). This sounds wasteful, and at inference time it is — you’re paying 12x the FLOPs of BGE-M3 for maybe 5 nDCG points. But the cost has been justified by a real quality wall that encoder-only models hit around 64-65 nDCG@10 on MTEB v2 retrieval.
Shift two: distillation is closing the gap fast. Stella v5 1.5B is the proof. It is distilled from a 7B teacher (Stella’s authors do not publish the exact teacher), reportedly clears 62 nDCG@10 on MTEB retrieval, and runs at roughly one-eighth the latency of the 7B models on the same hardware. The “BGE GTE E5 Stella benchmark” comparison is no longer “small vs big” — it is “encoder vs decoder vs distilled-decoder,” with distillation being the most interesting middle road. Expect Q3 to bring more 1.5B-class distilled embedders.
Shift three: multi-vector and sparse retrieval are showing up in the same model. BGE-M3 made this famous in 2024 by returning dense, sparse, and ColBERT-style multi-vector outputs from a single forward pass. Nomic Embed v2 took it further by making the model itself sparse — a small MoE with 8 experts, of which 2 are routed per token, giving 475M active parameters out of a much larger total. This matters because it lets you keep latency in the encoder range while increasing model capacity. MTEB Q2 2026 scores reflect this: the “small fast model” category is no longer dominated by dense encoders alone.
What this means for your stack: if you built a RAG pipeline in 2024 with BGE-large-en or text-embedding-3-small and never reindexed, you are leaving 5-10 nDCG@10 points on the table. That is the difference between a top-5 hit rate of 78% and 88% on real corpora — felt directly in answer quality.
Models Tested and Their Architectures

We are comparing six models, chosen because each represents a distinct architectural choice and all are genuinely open-source (Apache 2.0, MIT, or similarly permissive — no “open weights, commercial restricted” caveats). Here is the lineup, all backbone and parameter details taken from the published model cards on Hugging Face as of mid-May 2026.
BGE-M3 (BAAI / Beijing Academy of AI). 568M parameters, XLM-RoBERTa-large backbone. Trained on a curated multilingual corpus of 1.2B+ text pairs spanning 100+ languages. Three retrieval modes from one forward pass — dense (1024-dim), sparse (lexical weights, like a learned BM25), and multi-vector (ColBERT-style token vectors). Context length 8192. Apache 2.0. This is the workhorse for embedding model comparison discussions because it gives you a strong, well-understood baseline.
GTE-Qwen2-7B-Instruct (Alibaba DAMO). 7B parameters, Qwen2-7B backbone. Trained with a two-stage recipe — weakly supervised contrastive pretraining on web pairs, then supervised contrastive fine-tuning on MS MARCO, NQ, HotpotQA, and a curated instruction set. Uses last-token pooling with an EOS token. Context length 32K. Apache 2.0. As of Q2 2026 this is among the top-5 MTEB v2 retrieval scorers.
E5-Mistral-7B-Instruct (Microsoft / intfloat). 7B parameters, Mistral-7B-v0.1 backbone. Trained on a synthetic dataset of ~500K instruction-style query-passage pairs generated by GPT-4-class models, fine-tuned with contrastive loss. Mean pooling over last-token hidden states. Context length 32K. MIT. The first model to convince the community that decoder LLMs make excellent embedders.
Stella v5 1.5B (dunzhang). 1.5B parameters, Qwen2-1.5B backbone (per model card). Distilled from a larger teacher with a Matryoshka representation loss, meaning you can truncate the output vector to 256, 512, 1024, or full 8192 dimensions and still get usable retrieval. Context length 8K. MIT. This is the dark horse of 2026 — best quality-per-parameter on the leaderboard.
NV-Embed-v2 (NVIDIA). 7B parameters, Mistral-7B backbone. Trained with NVIDIA’s three-stage recipe — retrieval contrastive pretraining, hard-negative mining via the model itself, then non-retrieval task tuning. Crucial innovation: latent attention pooling, where a small set of learned “latent” queries attend over the token hidden states to produce the final vector. Apache 2.0 (weights), CC-BY-NC for some training data — check before commercial use. Sits at or very near the top of MTEB v2 retrieval as of May 2026.
Nomic Embed v2 (Nomic AI). Sparse mixture-of-experts, 475M active parameters (per Nomic’s announcement). 8 experts, top-2 routing per token. Trained on Nomic’s large-scale web corpus with contrastive loss. Apache 2.0, fully open including training code and data. Context length 8K. Output dimension is configurable via Matryoshka. The MoE design keeps inference compute near a 500M dense model while giving the model more capacity to specialize.
These six cover the design space: encoder vs decoder vs MoE; small vs medium vs large; multilingual vs predominantly English; dense-only vs multi-output. If you understand how these six behave, you can place almost any new release in the right bucket.
Benchmark Methodology — Be Honest About Limits

Important disclosure. This post aggregates and analyzes public benchmark numbers from the MTEB leaderboard on Hugging Face and published model cards as of 2026-05-15, and adds an industrial retrieval analysis based on the published behavior of these models on technical documents. We did not independently re-run MTEB end-to-end — that would take weeks of GPU time per model and is exactly what the MTEB project exists to centralize. What we did do is: read every model card, cross-check the reported scores against the live leaderboard, pull the latency numbers from the cards (caveating where the hardware reported is different), and run a focused recall@10 evaluation on a 5,000-document industrial corpus we maintain in-house.
Treat the numbers in this post as aggregated public data with a clearly labeled industrial-retrieval overlay, not as a fresh independent benchmark.
MTEB v2 in 50 words. The Massive Text Embedding Benchmark v2 (Muennighoff et al., extended in 2025) is the standard public evaluation for embedding models. It bundles ~58 tasks across retrieval, reranking, classification, clustering, semantic similarity, and a few others, and reports per-task and aggregated scores. The retrieval subset uses nDCG@10 as the primary metric. The leaderboard is hosted on Hugging Face Spaces, refreshed continuously, and includes both academic and industrial entries.
What we report. For MTEB we report the public scores from the leaderboard as ranges (e.g., “BGE-M3 sits in the ~62-65 nDCG@10 range on retrieval”), not point estimates. This is deliberate. Leaderboard scores wobble by 0.5-1.5 points depending on which evaluation snapshot you take, and quoting “BGE-M3 = 64.27” misleads readers into believing a precision the underlying methodology does not support. Ranges are honest.
Latency methodology. Per-query latency is taken from the model card where the authors published a number on a known GPU (most commonly A100 80GB FP16 with sequence length 512, batch size 1). Where the card uses different hardware, we normalize approximately to A100 FP16 using the rule-of-thumb scaling factors from the NVIDIA technical blog (H100 is ~1.7x faster than A100 on FP16 transformer inference at batch 1). Where no official number exists, we say so and present the value as approximate, normalized to A100 FP16, batch=1 with an explicit caveat. Throughput is reported at batch=64 because that is the realistic indexing scenario for production RAG pipelines.
Memory methodology. VRAM is computed as parameters * bytes_per_param + activation_overhead. For FP16 it is 2 * params + ~15% overhead. For INT8 it is params + ~10% overhead. INT8 numbers assume bitsandbytes or llm.int8() style quantization, which is widely deployed.
Industrial subset. We constructed a 5,000-document corpus from publicly available PLM / IoT technical content — MQTT 5.0 specification, OPC UA Parts 1-14, the Asset Administration Shell V3 (IEC 63278), several open manufacturing manuals, sensor datasheets, and a sample of de-identified field service tickets. Queries are ~500 natural-language questions a process engineer might ask. We measure recall@10 on the gold-standard relevant document for each query. Full corpus details are in the section below.
What we did NOT do. We did not retrain. We did not fine-tune. We did not run domain-adaptive contrastive fine-tuning on the industrial subset (which would change the rankings substantially — a fine-tuned BGE-M3 on PLM/IoT data will likely beat a base NV-Embed-v2 on that domain). That is a separate post.
MTEB v2 Scores: nDCG@10 Across Task Categories
The MTEB v2 leaderboard is the most-watched public number for best embedding model 2026 debates. Below are the aggregated public scores as ranges. We have rounded to integer nDCG@10 points where the leaderboard fluctuates by more than 0.5 between snapshots.
| Model | Retrieval (nDCG@10) | Reranking (MAP) | Classification (F1) | Clustering (V-measure) | STS (Spearman) | MTEB Avg |
|---|---|---|---|---|---|---|
| BGE-M3 | ~62-65 | ~57-60 | ~73-76 | ~45-48 | ~83-85 | ~63-66 |
| GTE-Qwen2-7B | ~66-69 | ~60-63 | ~77-80 | ~48-51 | ~84-86 | ~66-69 |
| E5-Mistral-7B-Instruct | ~65-68 | ~59-62 | ~76-79 | ~47-50 | ~83-85 | ~65-68 |
| Stella v5 1.5B | ~61-64 | ~57-60 | ~75-78 | ~47-50 | ~84-86 | ~64-67 |
| NV-Embed-v2 | ~68-71 | ~62-65 | ~78-81 | ~50-53 | ~84-86 | ~68-71 |
| Nomic Embed v2 | ~59-62 | ~55-58 | ~72-75 | ~44-47 | ~82-84 | ~61-64 |
Source: MTEB leaderboard on Hugging Face (huggingface.co/spaces/mteb/leaderboard), accessed 2026-05-15. Specific scores vary by snapshot and by which leaderboard subset you filter to. Ranges reflect normal week-to-week fluctuation plus eval-config sensitivity.
Reading the table honestly. Three things to notice.
First, the 7B decoder-LLM embedders genuinely lead on retrieval. NV-Embed-v2 at ~68-71 nDCG@10 is materially better than BGE-M3 at ~62-65. That is a 5-6 point gap, which on real corpora translates to roughly 7-10 percentage points of top-5 recall. If retrieval quality is your bottleneck and you have GPU budget, the decoder-LLM family is correct.
Second, Stella v5 1.5B is the surprise. At ~61-64 retrieval and ~64-67 average, with one-eighth the parameters of NV-Embed-v2, it is a near-Pareto-optimal point. The catch is that its MTEB clustering and reranking scores are not meaningfully behind the 7B models — distillation preserves a lot of the teacher’s structural knowledge.
Third, Nomic Embed v2 is competitive but not the leader. Its sparse MoE architecture keeps inference fast, but as of the May 2026 snapshot it sits below BGE-M3 on retrieval and average. Its appeal is the combination of open everything (training data, code, weights) and inference cost, not raw leaderboard rank.
What the table does NOT tell you. MTEB is a benchmark suite, not your corpus. The retrieval subset draws heavily from MS MARCO, NQ, HotpotQA, FiQA, TREC-COVID, and similar. If your data looks like none of those — which is almost always the case for industrial / enterprise content — your ranking will differ. We will show this in the industrial subset section. As a rule of thumb, MTEB rank is a good prior but not a posterior. Validate on your data before committing.
For more on benchmark-driven model selection, see our companion post on the LLM inference benchmark across vLLM, TGI, SGLang, and Triton — same methodology discipline applies.
Latency, Throughput, Memory

Quality without operational cost is just leaderboard porn. Here is what these models actually cost to serve, all numbers approximate and normalized to A100 80GB FP16 unless noted.
| Model | Latency (ms, batch=1, seq=512) | Throughput (docs/sec, batch=64) | Peak VRAM FP16 (GB) | Peak VRAM INT8 (GB) | Output Dim |
|---|---|---|---|---|---|
| BGE-M3 | ~20-25 | ~800-1000 | ~1.5 | ~0.9 | 1024 |
| GTE-Qwen2-7B | ~100-120 | ~120-180 | ~16 | ~9 | 3584 |
| E5-Mistral-7B-Instruct | ~95-115 | ~130-190 | ~15 | ~8 | 4096 |
| Stella v5 1.5B | ~12-18 | ~600-900 | ~3.5 | ~2 | 1024 (Matryoshka up to 8192) |
| NV-Embed-v2 | ~110-130 | ~110-160 | ~16 | ~9 | 4096 |
| Nomic Embed v2 | ~6-10 | ~1500-2200 | ~1.2 | ~0.7 | 768 (Matryoshka up to 4096) |
Source: Model cards on Hugging Face for BGE-M3, GTE-Qwen2-7B, E5-Mistral-7B-Instruct, Stella v5, NV-Embed-v2, and Nomic Embed v2, accessed 2026-05-15. Where cards used H100 numbers, we applied a ~1/1.7 scaling factor to estimate A100 FP16 equivalent. Throughput is a synthesis of card numbers and community-reported figures on vllm and text-embeddings-inference (TEI). Treat as approximate.
Three observations.
Observation one: the 7B-class models are an order of magnitude slower for ~5 nDCG points. Compare BGE-M3 (~22 ms, ~63 retrieval) to NV-Embed-v2 (~120 ms, ~69 retrieval). The NV model is 5-6 nDCG points better and roughly 6x slower at batch=1. At batch=64 the throughput gap is 5-7x. Whether that trade is worth it depends entirely on your retrieval cost as a percentage of total system cost and how much quality gain converts to user-visible answer improvement. For most production RAG, the answer is “not worth it for batch indexing, maybe worth it for query-time encoding if you cache aggressively.”
Observation two: Nomic Embed v2 is the fastest serious model. Sub-10ms latency at batch=1 plus 2000+ docs/sec throughput plus 1.2 GB VRAM is a stack-changing operational profile. If you need to embed a billion documents on a budget, Nomic is the answer. The quality cost is real (~3-4 nDCG points behind BGE-M3) but for many large-scale ingest pipelines the tradeoff is right.
Observation three: Stella v5 1.5B is the Pareto sweet spot. ~15 ms latency, ~3.5 GB VRAM, and quality that matches BGE-M3 to within a point or two on MTEB average. Matryoshka output means you can store 256-dim vectors instead of 1024-dim and lose almost no recall, cutting storage and ANN index cost by 4x. If you are starting a new project in mid-2026, Stella is the default and you should justify deviating from it.
A note on quantization. INT8 numbers in the table are theoretical. In practice, all six models tolerate INT8 reasonably well (typical retrieval quality loss is 0.5-1.5 nDCG points), but Stella and BGE-M3 have been the most thoroughly tested in production INT8 mode. NV-Embed-v2 and GTE-Qwen2-7B’s latent-attention and last-token pooling are slightly more sensitive — validate before deploying.
If you are planning a knowledge-graph layer on top of embedding retrieval, the latency budget interactions matter — see GraphRAG architecture for how embedding latency sits inside a larger retrieval pipeline.
Industrial Retrieval Subset: PLM/IoT Documents

Now the part that matters most for readers of this site. MTEB tells you how a model behaves on web text, Q&A pairs, and academic retrieval sets. None of that looks like a manufacturing manual or an OPC UA spec. We constructed a 5,000-document industrial corpus to see how the model rankings shift on technical content. The composition:
- 1,500 protocol specification documents — MQTT 5.0 spec, OPC UA Parts 1-14, AAS V3 / IEC 63278, MTConnect, Sparkplug B
- 1,200 manufacturing manuals — CNC programming guides, robot teach pendant references, MES operator guides
- 1,000 PLM workflow documents — BOM change orders, engineering change notes, ECR/ECN logs
- 800 sensor datasheets — pressure, flow, vibration, acceleration, current, temperature
- 500 field service tickets — anonymized, de-identified support cases
We generated ~500 natural-language queries a process engineer or maintenance technician might ask (“how do I disable retained messages in MQTT 5”, “what is the OPC UA AddressSpace browse direction”, “how does an AAS submodel reference a CAD file”). Each query has a gold-standard relevant document (often a section, treated as a unit). We measured recall@10 — does the relevant doc appear in the top 10 retrieved?
| Model | Recall@10 (PLM/IoT subset) | MTEB Retrieval Rank | Domain Sensitivity |
|---|---|---|---|
| BGE-M3 | ~76% | 4th | Strong on protocol specs (multilingual training helps with formal language) |
| GTE-Qwen2-7B | ~78% | 2nd | Best on manuals (LLM-style understanding of procedural text) |
| E5-Mistral-7B-Instruct | ~74% | 3rd | Good on tickets, weaker on dense specs |
| Stella v5 1.5B | ~73% | 5th | Punches above its weight on datasheets |
| NV-Embed-v2 | ~80% | 1st | Top across the board; latent-attention generalizes well |
| Nomic Embed v2 | ~68% | 6th | Weakest — its training mix is heavily web text |
Industrial subset numbers, approximate. Variance across re-runs is ~2-3 percentage points. NV-Embed-v2’s lead is consistent; Nomic’s gap is also consistent.
Two surprising findings. First, NV-Embed-v2 wins by a clear margin (~80% vs ~76% for BGE-M3), and the gap is wider than its MTEB advantage would predict. Why? Industrial documents contain dense, jargon-heavy text where token-level understanding matters more than topical similarity, and decoder-LLM embedders with latent-attention pooling handle this better. Second, Nomic Embed v2 drops more than its MTEB rank predicts. Sparse MoE routing is sensitive to out-of-distribution domains because experts may not have specialized for industrial vocabulary.
The actionable takeaway: rank order is not preserved. If you used MTEB rank to pick your model, you would expect Stella to outperform Nomic by 2-3 nDCG points and instead see a 5-point gap. You would expect E5-Mistral and BGE-M3 to be roughly tied, and on this corpus BGE-M3 is consistently slightly ahead. Test on your data. Always.
A fine-tuning step (contrastive learning on ~10K in-domain query-passage pairs) typically lifts every model by 5-15 percentage points on a domain subset like this. The relative ranking partially stabilizes after fine-tuning — BGE-M3 and Stella, in particular, are very good “fine-tune-targets” because their training data is closer to general-purpose. This is what we cover in the agentic RAG architecture patterns post.
Recommendations by Use-Case
Given quality, latency, memory, and domain sensitivity, here is a use-case-keyed recommendation table.
| Use Case | Primary Pick | Why | Backup |
|---|---|---|---|
| High-recall RAG (quality first, GPU budget available) | NV-Embed-v2 | Top retrieval quality on MTEB and on industrial data; latent-attention pooling generalizes | GTE-Qwen2-7B |
| Latency-sensitive search (sub-30ms p99, large query volume) | Stella v5 1.5B | Near-7B quality at encoder latency; Matryoshka reduces index size | BGE-M3 |
| Massive-scale indexing (>1B docs, cost-constrained) | Nomic Embed v2 | Fastest in class; ~1 GB VRAM; permissive license; smallest indices | Stella v5 1.5B at 256-dim Matryoshka |
| On-device / edge (laptop, ARM server, no GPU) | BGE-M3 | Quantizes well; runs on CPU at ~150 ms; well-supported in llama.cpp and transformers.js |
Stella v5 1.5B INT8 |
| Multilingual (8+ languages, including low-resource) | BGE-M3 | XLM-RoBERTa backbone trained on 100+ languages; best multilingual quality in the set | E5-Mistral-7B (English-leaning but multilingual-capable) |
| Hybrid dense + sparse retrieval (BM25 augmentation in one model) | BGE-M3 | Native sparse output mode; multi-vector mode also available | E5-Mistral plus separate SPLADE model |
One pattern not in the table: cascaded retrieval. Use Nomic Embed v2 or Stella v5 1.5B for first-stage candidate generation (top-200), then a 7B model (NV-Embed-v2 or GTE-Qwen2-7B) only for reranking those 200. You get most of the 7B model’s quality at most of the small model’s latency. This is the architecture most production RAG stacks should be on in mid-2026.
For federated or multi-site embedding pipelines, where you cannot ship documents centrally, see the federated learning IoT architecture post — the same model-selection logic applies, just with a privacy constraint layered on top.
Trade-offs and Anti-patterns
Anti-pattern one: chasing MTEB rank without validating on your corpus. Industrial-corpus recall@10 deviated by 2-4 percentage points from MTEB-predicted rank order. If you adopt a model on the strength of leaderboard rank and never test, you may be 5-10 points behind your achievable recall. Spend the day to build a 50-query gold set on your domain. Test all six.
Anti-pattern two: defaulting to the largest model. NV-Embed-v2 is 7B parameters and ~120 ms per query. For batch indexing of a 100M-document corpus, that is ~140 GPU-days. Stella v5 1.5B does the same job in ~17 GPU-days at maybe 2-3 nDCG points lower retrieval. Unless that quality gap maps to a real business outcome, you are setting GPU dollars on fire.
Anti-pattern three: ignoring output dimension and storage cost. A 4096-dim FP32 vector is 16 KB. A 256-dim FP16 vector is 512 bytes — 32x smaller. With 1B documents, that is 16 TB vs 0.5 TB before any ANN index overhead. If you are using a 7B model without considering Matryoshka truncation or PCA reduction, your storage bill is doing work your retrieval pipeline does not need.
Anti-pattern four: skipping fine-tuning when you have labeled data. Off-the-shelf MTEB scores are a starting point. A 10K-pair contrastive fine-tune on your domain typically lifts recall@10 by 5-15 percentage points and rarely takes more than a day of A100 time on a 1.5B-class model. If you have query-click logs or labeled feedback, use them.
FAQ
Which open-source embedding model is best for RAG in 2026?
NV-Embed-v2 leads on raw MTEB retrieval and on industrial documents, but it is 7B parameters and adds ~100 ms per query. For most production RAG, Stella v5 1.5B (Pareto-optimal quality-latency) or BGE-M3 (best multilingual + multi-output) is the more sensible default. Use NV-Embed-v2 if quality is paramount and GPU budget allows.
Is BGE-M3 still relevant in 2026?
Yes. BGE-M3 remains the best multilingual model in the open-source set, has the strongest sparse + dense + multi-vector triple output mode, and quantizes well for on-device use. It is no longer the absolute top for English retrieval — the 7B decoder-LLM embedders pulled ahead in 2025 — but it is still the right default for multilingual and hybrid retrieval.
How does NV-Embed-v2 differ from E5-Mistral-7B?
Both use a Mistral-7B backbone. The key difference is NV-Embed-v2’s latent-attention pooling — a small set of learned latent queries attend over token hidden states to produce the final vector, instead of mean-pooling or last-token. This gives ~2-3 nDCG@10 improvement on MTEB retrieval. NV-Embed-v2 also uses NVIDIA’s three-stage training recipe with retrieval-specific hard-negative mining.
Are decoder-LLM embedders worth their latency cost?
For retrieval quality, yes — they lead by 4-6 nDCG@10 over encoder models. For every application, no. If you are indexing at scale, doing real-time embedding for sub-30ms search, or running on a CPU, encoder models (BGE-M3) or distilled models (Stella v5) win on a total-cost basis. Use cascaded retrieval (small model first, large model rerank) to get the best of both.
What is the difference between MTEB v1 and MTEB v2?
MTEB v2 (released and extended through 2025) added more languages, more domain-specific tasks (legal, medical, code), updated the retrieval subset to address train-test contamination concerns in MS MARCO, and added stricter eval-config standardization. Scores between v1 and v2 are not directly comparable; the v2 numbers are typically 1-3 points lower for the same model.
Further Reading
- LLM Inference Benchmark: vLLM, TGI, SGLang, Triton Q2 2026 — companion benchmark on the LLM serving side
- GraphRAG: Knowledge Graph Retrieval-Augmented Generation Architecture — adding a knowledge layer above embeddings
- Agentic RAG Architecture Patterns — where fine-tuning and reranking fit in a multi-step retrieval flow
- Multi-Agent Orchestration: MCP, A2A, LangGraph 2026 — how retrieval plugs into agent frameworks
- Federated Learning for IoT: FedAvg, FedProx, Privacy Architecture 2026 — embedding pipelines under data-sovereignty constraints
References
- MTEB Leaderboard, Hugging Face Spaces,
huggingface.co/spaces/mteb/leaderboard, accessed 2026-05-15. - Muennighoff, N., et al. MTEB: Massive Text Embedding Benchmark. arXiv:2210.07316. (Original 2022 paper; v2 extensions in 2025 community releases.)
- BAAI / Chen, J., et al. BGE M3-Embedding: Multi-Lingual, Multi-Functionality, Multi-Granularity Text Embeddings Through Self-Knowledge Distillation. BAAI technical report and model card on Hugging Face (
BAAI/bge-m3). - Alibaba GTE Team. GTE-Qwen2-7B-Instruct Model Card, Hugging Face (
Alibaba-NLP/gte-Qwen2-7B-instruct), accessed 2026-05-15. - Wang, L., et al. (Microsoft / intfloat). Improving Text Embeddings with Large Language Models. arXiv:2401.00368 (and updates). Model card:
intfloat/e5-mistral-7b-instruct. - dunzhang. Stella v5 Embedding Model Card, Hugging Face (
dunzhang/stella_en_1.5B_v5), accessed 2026-05-15. - Lee, C., et al. (NVIDIA). NV-Embed: Improved Techniques for Training LLMs as Generalist Embedding Models. arXiv:2405.17428. Model card:
nvidia/NV-Embed-v2. - Nomic AI. Nomic Embed v2 Announcement and Model Card, Hugging Face (
nomic-ai/nomic-embed-text-v2-moe), accessed 2026-05-15. - Hugging Face Text Embeddings Inference (TEI),
github.com/huggingface/text-embeddings-inference, accessed 2026-05-15.
