BGE
BGE (BAAI General Embedding) is a family of open-source embedding models from the Beijing Academy of Artificial Intelligence — released under MIT license and broadly considered among the best open-weight embedding options. The family includes bge-large-en-v1.5, bge-base-en-v1.5, bge-small-en-v1.5, multilingual variants (bge-m3, bge-multilingual-gemma2), and specialized variants for various scales and use cases. Performance on MTEB benchmark consistently ranks BGE models at or near the top of open-source leaderboards, often matching commercial APIs from OpenAI and Cohere. The models produce 1024-dimensional vectors (large variant), 768-dimensional (base), or 384-dimensional (small), with reranking variants available for two-stage retrieval pipelines. Released under MIT license with weights on Hugging Face — making BGE foundational to self-hosted RAG deployments. Real-world deployments span enterprise search, on-prem RAG systems, and any application requiring open-weight embedding under permissive licensing. AI governance, AI compliance, and AI risk management programs deploy BGE widely for sovereign AI workloads supporting responsible AI through open-source, self-hosted embedding in enterprise AI environments.
Centralpoint Hosts BGE Embeddings Behind Your Firewall: Oxcyon's Centralpoint AI Governance Platform routes embeddings to self-hosted BGE alongside OpenAI, Cohere, Voyage, and other models — your perimeter, your data. Centralpoint meters consumption, keeps prompts and skills on-prem, and embeds RAG chatbots into your portals via one JavaScript line.
Related Keywords:
BGE,
,