The Linux Foundation Projects
Skip to main content

Streamlining implementation of enterprise-grade Generative AI

Efficiently integrate secure, performant, and cost-effective Generative AI workflows into business value.

The OPEA platform includes:

  • Detailed framework of composable building blocks for state-of-the-art generative AI systems including LLMs, data stores, and prompt engines
  • Architectural blueprints of retrieval-augmented generative AI component stack structure and end-to-end workflows
  • A four-step assessment for grading generative AI systems around performance, features, trustworthiness and enterprise-grade readiness

Harnesses existing infrastructure, the AI accelerator or other hardware of your choosing.

Integrates with enterprise software, with heterogeneous support and stability across system & network.

Brings together best of breed innovations and is free from proprietary vendor lock-in.

Runs everywhere through a flexible architecture built for cloud, data center, edge and PC.

Features a secure enterprise-ready pipeline and tools for responsibility, transparency, and traceability.

Access to a vibrant ecosystem of partners to help build and scale your solution.

We're thrilled to welcome OPEA to LF AI & Data with the promise to offer an open-source standardized modular and heterogenous RAG pipelines for enterprises with a focus on open model development, hardened and optimized support of various compilers and toolchains. OPEA will unlock new possibilities in AI by creating a detailed, composable framework that stands at the forefront of technology stacks. This initiative is a testament to our mission to drive open-source innovation and collaboration within the AI and data communities under a neutral and open governance model.

Ibrahim HaddadExecutive Director, LF AI & Data

With the potential that generative AI has to shape our future and the ways we do business, it’s imperative that we tap into the power of collaboration for innovation and accuracy in enterprise AI. We're excited to be a part of the newest AI & Data Sandbox project and to work with other industry leaders on OPEA.

Shadi ShahinVP, Product Strategy, SAS

Enterprises integrating their choice of cutting-edge tools into their AI platforms aren’t just ahead of the AI curve—they’re defining it. OPEA's vision aligns with our commitment to open, flexible, governed AI innovation, and we're proud to support it alongside Intel.

Thomas RobinsonCOO, Domino Data Lab

Hugging Face's mission is to democratize good machine learning and maximize its positive impact across industries and society. By joining OPEA's open-source consortium to accelerate Generative AI value to enterprise, we will be able to continue advancing open models and simplify GenAI adoption."

Julien SimonChief Evangelist, Hugging Face

Intel is at the forefront of incubating open source development to build trusted, scalable open infrastructure that enables heterogeneity and provides a platform for developer innovation. Generative AI is at this moment; OPEA, with the support of the broader community, will address critical pain points of RAG adoption and scale today. It will also define a platform for the next phases of developer innovation that harnesses the potential value generative AI can bring to enterprises and all our lives.”

Melissa EversVP of Software and Advanced Technology Group and GM of Strategy to Execution, Intel Corporation

"At DataStax, we help enterprises succeed with AI and RAG. We love to partner and collaborate with other open community companies and organizations like Intel and the Linux Foundation to develop standards to help the industry move forward. As we expand our portfolio to ease RAG Application development, we see the Open Platform for Enterprise AI (OPEA) as a foundation that will drive modularity, scale and hardening to provide a platform for future innovation."

Davor BonaciCTO, DataStax

At dstack, we're building a new approach to AI infrastructure management aimed at leveraging open source and ensuring its portability across multiple infrastructure and model vendors. We believe the mission of the OPEA initiative is crucial for the safety and democratization of enterprise AI. We're excited to be a part of it.

Andrey CheptsovCEO & Founder, dstack

Collaborative, open-source projects like OPEA fuel our excitement for the future of gen AI because of the ability it has to drive acceleration of both innovation and adoption within enterprise organizations. The power of RAG is undeniable, and its integration into gen AI creates a ballast of truth that enables businesses to confidently tap into their data and use it to grow their business.

Michael GilfixChief Product and Engineering Officer, KX

As GenAI matures, integration into existing IT is a natural and necessary step. The world needs GenAI and vectors as part of a general purpose RDBMS, and we have already demonstrated our ability to deliver this through MariaDB Server. We see huge opportunities for core MariaDB users - and users of the related MySQL Server – to build RAG solutions. It's logical to keep the source data, the AI vector data, and the output data in one and the same RDBMS. The OPEA community, as part of LF AI & Data, is an obvious entity to simplify Enterprise GenAI adoption.

Kaj ArnöCEO, MariaDB Foundation

The OPEA initiative is crucial for the future of AI development. Advocating for a foundation of open source and standards - from datasets to formats to APIs and models, enables organizations and enterprises to build transparently. The AI data infrastructure must also be built on these open principles. Only by having open source and open standard solutions, from models to infrastructure and down to the data are we able to create trust, ensure transparency and promote accountability.

AB PeriasamyCEO and co-Founder, MinIO

As the leading open-source vector database technology provider, Qdrant is excited to support the launch of the OPEA, underscoring the importance of open standards in AI for innovation and data sovereignty. Our commitment to these principles is rooted in our core, and we look forward to contributing to an ecosystem where AI thrives with a deep respect for data ownership.

Andre ZayarniCEO & co-Founder, Qdrant

As gen AI continues to advance, open source is playing a critical role in the standardization and democratization of models, frameworks, platforms and the tools needed to help enterprises realize value from AI. Red Hat is excited about the potential for AI innovation for our customers through the Open Platform for Enterprise AI.

Steven HuelsVP and GM, AI Business Unit, Red Hat

We are pleased to collaborate with the Open Platform for Enterprise AI (OPEA), which offers essential guidance in a dense and complex market. Within OPEA, Yellowbrick serves as a data provider—recognizing data as the crucial fuel for AI. Our data warehouse incorporates advanced vector capabilities, enabling seamless integration of AI with current systems and workflows. This ensures that AI augments rather than interrupts business processes, simplifying AI adoption.

Mark, CusackCTO, Yellowbrick Data

Cloudera is thrilled to join industry thought leaders like Intel in the Open Platform for Enterprise AI Alliance, embracing openness and collaboration to drive innovation and empower the future of generative AI.”

Andy MollerSVP of Global Alliances & Ecosystem, Cloudera

We firmly believe that vector databases are integral to the future of open generative AI, which is why we donated the Milvus vector database to the Linux Foundation back in 2020. Our support of OPEA is an extension of that commitment to creating a framework alongside Intel that fosters extensible, accessible, and scalable AI platforms for enterprise developers.

Charles XieCEO & Founder, Zilliz

We are seeing tremendous enthusiasm among our customer base for RAG, with organizations deploying RAG applications on-premises to empower employees and customers to find the information they need faster, creating greater efficiencies in customer service and document search. The constructs behind RAG can be universally applied to a variety of use cases, making a community-driven approach that drives consistency and interoperability for RAG applications an important step forward in helping all organizations to safely embrace the many benefits that AI has to offer.

Chris WolfGlobal Head of AI and Advanced Services, Broadcom

As pioneers of RAG-based systems, we are happy to see OPEA driving the adoption. RAG architectures are established and proven for building GenAI apps and OPEAs efforts will help to drive the standard-creation in RAG forward, thus accelerating enterprise adoption. We are proud to be part of this initiative with the Haystack-framework.

Milos RusicCEO & Founder, deepset

We live in a dynamic and exciting time where new AI technologies are turning up literally every hour. OPEA takes on two important problems: helping enterprises better understand the GenAI landscape through open architecture patterns; and accelerating quality, value, and time to market with tools and best practices. Neo4j is proud to be a member, and look forward to contributing knowledge, open software, and best practices in the field of knowledge graphs for GenAI & GraphRAG.

Philip RathleCTO, Neo4j

Contribute to OPEA

We invite like-minded industry peers to contribute to the development and standardization of enterprise-grade Retrieval Augmented Generative AI.

Contribute on GitHubContact Us

Stay Connected

Keep up to date with the latest news and initiatives from OPEA.