Open Platform for Enterprise AI (OPEA) https://opea.dev Wed, 11 Jun 2025 14:51:31 +0000 en-US hourly 1 https://wordpress.org/?v=6.9 https://opea.dev/wp-content/uploads/sites/9/2024/04/cropped-favicon-1-150x150.png Open Platform for Enterprise AI (OPEA) https://opea.dev 32 32 Nutanix Joins OPEA to Further Open Enterprise AI https://opea.dev/nutanix-joins-opea-to-further-open-enterprise-ai/ Wed, 11 Jun 2025 14:51:31 +0000 https://opea.dev/?p=728 Nutanix’s goal is to help customers achieve their GenAI objectives. As Nutanix Chief AI Officer Debo Dutta states, “Nutanix is committed to helping enterprises everywhere simplify, control and secure their GenAI initiatives…” OPEA used with the Nutanix Cloud Platform (NCP) solution can deliver exactly that, with benefits that can include:…

 

Read more at Nutatnix.

]]>
OPEA 1.3 Release Highlights https://opea.dev/opea-1-3-release-highlights/ Fri, 09 May 2025 15:09:22 +0000 https://opea.dev/?p=672 By Rachel Roumeliotis, Director, Open Source Strategy at Intel 

On our one-year anniversary we are dropping OPEA 1.3! (See our blog post celebrating the anniversary here.) While it contains lots of new and enhanced capabilities the keyword for 1.3 is AGENTS! OPEA now has advanced agent capabilities including an enhanced framework that features dynamic context-aware dialogues as well as a new FinTech Agent that automates data aggregation and leverages LLMS to generate insights. In addition, vLLM is now the default for most GenAI Examples and Haystack integration is complete. 

OPEA Release 1.3 Highlights 

You can see the entire extensive list of what is new and enhance here but let’s take a look at some of the major updates. 

Enhanced GenAI E2E Examples 

  • AgentQnA now with web search tool support and simplified run instructions 
  • EdgeCraftRAG with new UI based on Vue and Ant design, supporting concurrent multi-requests on vLLM, and JSON pipeline configuration 
  • CodeGen using RAG and Agent: Additional layer of intelligence and adaptability 

New GenAI Capabilities 

  • Struct to Graph: Transform structured data to graphs using Neo4j graph database 
  • Text to Cypher: Generate and execute cypher queries from natural language for graph database retrieval 

New Infra Functionality 

Enhanced Evaluation  

New Model Compatibility 

  • Deepseek-R1 various, Deepseek-v3, Hermes-2-Llama-3.1-8B, Granite-3.2-8b-instruct, Phi-4-mini, Phi-4-multimodal-instruct, Mistral-small-24B-Instruct-2501, Mistral-large-instruct-2411 

New Observability 

OPEA 1.4 to come in Summer 2025! 

In closing, be sure to check out our OPEA Week videos that have excellent sessions from our partners Prediction Guard, ByteDance, ArangoDB, Neo4j, and Infosys that cover graphRAG, Agents, and security. 

Want to stay up to date on OPEA? Join our mailing list by visiting OPEA.dev 

 LF AI & Data Resources 

Access other resources on LF AI & Data’s GitHub or Wiki 

]]>
NetApp and Intel Partner to Redefine AI for Enterprises (OPEA Inside!) https://opea.dev/netapp-and-intel-partner-to-redefine-ai-for-enterprises-opea-inside/ Tue, 06 May 2025 17:42:16 +0000 https://opea.dev/?p=696 “SAN JOSE, Calif. – May 6, 2025– NetApp® (NASDAQ: NTAP), the intelligent data infrastructure company, today announced the release of NetApp AIPod Mini with Intel, a joint solution designed to streamline enterprise adoption of AI inferencing. This collaboration addresses the unique challenges businesses face when deploying AI, such as cost and complexity, at the department and team level.”

“Built on an open framework powered by Open Platform for Enterprise AI (OPEA), the solution ensures modular, flexible deployments tailored to business needs. Intel Xeon processors are designed to boost computing performance and efficiency, making AI tasks more attainable and cost-effective, empowering customers to achieve more.”

Read more at NetApp

And Yahoo

]]>
Quickly Deploy GenAI Apps with the OPEA Framework on Ubuntu https://opea.dev/quickly-deploy-genai-apps-with-the-opea-framework-on-ubuntu/ Thu, 01 May 2025 18:49:34 +0000 https://opea.dev/?p=706 As the demand for AI-enabled software continues to surge, there is a rising need for platforms that can quickly provide APIs and other capabilities that are easy to consume by applications. The Open Platform for Enterprise AI (OPEA) project is designed to do just that, specifically for deployment of generative AI (GenAI) applications. In this blog post, we use the OPEA framework to demonstrate the ease of deploying ChatQnA on a laptop running Ubuntu 24.04 LTS and Canonical Kubernetes. This includes:

  • A ChatQnA web application that uses the retrieval augmented generation (RAG) architecture, complete with swappable large language model (LLM) components
  • A command-line tool for interacting with the ChatQnA service

Read more at Ubuntu.

]]>
Kubecon NA Video: AI Pipelines With OPEA: Best Practices for Cloud Native ML Operations https://opea.dev/kubecon-na-video-ai-pipelines-with-opea-best-practices-for-cloud-native-ml-operations/ Fri, 25 Apr 2025 20:39:43 +0000 https://opea.dev/?p=683 Ezequiel and Melissa will introduce you to the OPEA platform and how to empower your team to build, deploy, and manage AI pipelines more effectively. Attendees will gain insights into best practices for handling complex AI/ML workloads, automating dependency management, and integrating Kubernetes for efficient resource utilization. With a focus on real-world applications, this talk not only showcases the transformative potential of these tools but also encourages attendees to explore new ways to contribute, innovate, and collaborate in driving the future of AI adoption in enterprise environments.

 

See the video at JFrog Community Site.

]]>
One Year of OPEA: Lowering GenAI adoption barriers for enterprise https://opea.dev/one-year-of-opea-lowering-genai-adoption-barriers-for-enterprise/ Tue, 08 Apr 2025 16:59:52 +0000 https://opea.dev/?p=563 By Dr. Malini Bhandaru, OPEA Technical Steering Committee Chair, and Intel, Senior Principal Engineer & Rachel Roumeliotis, OPEA Ecosystem & Community Manager and Intel, Director, Open Source Strategy

A year ago, it felt like each week brought an exciting new AI model that promised to change how we do, well, everything. Models still feel like magic: creating art, solving math problems, generating code. But, moving from simple chatbots to enterprise-level generative AI (GenAI) solutions that unlock the full promise of AI felt like a giant leap. Intel’s 2024 Open Source Community Survey revealed that the top GenAI adoption barriers were scalability and performance challenges (37%) and a shortage of AI talent (37%). Could we lower those barriers? Help democratize AI? This is typically where open source comes to the rescue!

It all began in March 2024, as an aspirational internal whitepaper at Intel that aimed to capture all the key components necessary for an enterprise to be truly successful with GenAI. The paper covered the importance of being able to choose which software and hardware components are used to create solutions. It highlighted the need for standards around evaluation—not just for functionality, but also performance and accuracy. An effective enterprise-level GenAI solution would also need to be able to run in both development mode and at scale, and provide customizable end-to-end reference implementations for common use cases. Most importantly, the solution had to be open source and broadly available for true democratization to encourage the rich community collaborations that bring together our best minds across the technology industry.

A Year Later…

The white paper has evolved into the Open Platform For Enterprise AI (OPEA), an Apache 2.0-licensed open source project housed under the Linux AI & Data Foundation, with over 50 community partners. A year of firsts has culminated in a code base that supports multiple GenAI use cases across a variety of hardware platforms.

Some of our favorite big moments: 

OPEA Building Blocks

OPEA has  six sub-repositories, each playing a crucial role in achieving the goals of democratizing GenAI and providing scalable, enterprise-ready solutions.

  1. GenAIExamples is filled with end-to-end examples for common use cases such as a Chatbot (95% usage given its broad applicability to health care, legal services, product literature), Document Summarization (to cope with information glut),  AgenticAI, Code Generation, Code Translation, and Language Translation (increasingly global engagements), to name a few. Each of these can be run in either Docker or Kubernetes, scaled on Kubernetes, and run on your choice of cloud or on-prem, on bare metal or virtual machines, using hardware from Intel, AMD, or Nvidia. Just change an environment variable to use a different embedder or LLM to experiment and determine the best for your use case. 
  2. GenAIComps: A typical GenAI solution consists of multiple components, such as an embedder, a vector database, a model server, guardrails, and more that are chained together. For each component type, often there are multiple alternatives with their own pros and cons. OPEA supports three model serving frameworks, namely vLLM, TGI, and Ollama; three popular component chaining mechanisms, namely LangChain, LlamaIndex, and Haystack; and at least 10 vector database implementations, including graph support, with solutions differing in the amount of data they can handle, whether they auto update on underlying data changes, and more.
  3. GenAIInfra provides scripts and templates to launch workloads on popular cloud offerings such as AWS, GCP, IBM, and Intel® Tiber™ AI Cloud, and Helm charts to further ease deployment of the GenAI examples. It also contains instructions on how to integrate with authentication and authorization systems, rate limiting gateways, and more—each important to delivering an enterprise-grade solution.
  4. GenAIEval houses benchmark scripts in addition to leveraging industry best practices to evaluate accuracy. Production solutions have to comprehend cost and meet criteria such as latency, throughput, and accuracy. These scripts help us develop model deployment profiles to deliver low latency or high throughput solutions, taking us from proof-of-concept solutions to production.
  5. GenAIStudio: A low-code/no-code interface that reduces adoption barriers even further. It supports dragging and dropping components and chaining them together then, with a click of a button, building, packaging, and deploying these using Docker or Kubernetes onto your choice of infrastructure. 
  6. Docs provides crisp documentation on how to use the various OPEA components, while making few assumptions about the reader’s background and expertise to deliver a complete set of instructions. This is a living set of documents, of course, and, as such, we have tooling in place to check for link validity, perform nightly builds and publish to our doc .io site, which is constantly evolving.

Together, these repositories form a cohesive framework that empowers developers to create, scale, and deploy GenAI solutions efficiently, moving us closer to the goal of making AI practical and accessible for enterprises.

Partnerships

Strong partnerships are key to OPEA’s current and future success, enabling us to tackle enterprise AI challenges with innovative solutions and meaningful integrations.

Retrieval augmented generation (RAG), for example, reduces hallucinations and increases response relevancy in generative AI (GenAI), making vector databases a critical ingredient in AI pipelines. Recognizing this, multiple vector database vendors, including Pathway and Qdrant, have joined forces with OPEA to make their offerings first-class citizens within the platform. Similarly, Prediction Guard has integrated guardrail services into OPEA that improve response quality by addressing bias, repetition, politeness, and other factors.

OPEA’s multi-vendor approach is further strengthened by AMD’s contributions, which have enabled ROCm-based hardware-tested solutions for six GenAI examples already (AudioQnA, ChatQnA, DocSum, FaqGen, CodeTrans, and CodeGen). Partners like LlamaIndex and deepset (Haystack) have added flexibility for end-users to customize their GenAI pipelines, while AWS has brought OPEA integrations to its marketplace. Nutanix and NetApp are among the latest partners, acknowledging OPEA’s potential to simplify GenAI adoption across cloud hosting and storage domains.  

These collaborations exemplify how OPEA relies on active partnerships to deliver scalable, enterprise-ready solutions. As the project continues to grow, the contributions of engaged partners will remain critical in driving innovation, expanding adoption, and ensuring the platform meets the evolving needs of enterprises worldwide.

Powered by OPEA

A project gains adoption if it addresses some pain point or fills a gap. OPEA’s customizable GenAI examples, its curated list of GenAI components that can be mixed and matched, and its evolving end-to-end benchmark suites are what make it valuable as a one-stop-GenAI-shop.

ISTE + ASCD, to reduce costs while still maintaining security, resiliency, and performance, adopted OPEA and hosted inference services running in Intel® Tiber™ AI Cloud. The United Nations Innovation Unit (OICT) is also using OPEA as the cornerstone of its “AI Sovereign stack.”  

H3C, in partnership with Intel, unveiled its AIGC LinSeer integrated machine: a full-stack enterprise AI solution built with Intel® Xeon®, Intel® Gaudi®  2, and OPEA. Tailored for sectors like healthcare, energy, and education, the system delivers high-accuracy inference with rapid response capabilities and support for complex workflows. 

Dell recently launched a turnkey GenAI solution using the Dell PowerEdge XE9680 platform and Intel Gaudi 3 accelerators, all powered by OPEA. It demonstrates OPEA’s value in enterprise-grade infrastructure environments.

With these collaborations in place, we’re poised to explore even greater possibilities and innovations in the coming year.

What’s Next

Rapid innovation defines AI today, whether it’s more capable models, better model serving frameworks, or more efficient inference. OPEA will be integrating with AIBrix and KubeAI to reduce token latency and increase throughput, constantly integrate and test new models, publish hardware optimized images, provide more AI agents, and improve documentation. As innovation in AI accelerates, OPEA is paving the way for responsible and impactful advancements in enterprise AI.

We invite the community to learn more about OPEA, use it, and help enhance it through your contributions. Learn about and influence the OPEA project through the OPEA Developer Experience, Evaluation, and Security working groups. Be sure to stop by an OPEA community event this year—next up is our one-year-anniversary series of events, OPEA Week (April 14-18)! Join the effort to unlock the potential of GenAI, safely and responsibly, leveraging the best of open source, to create together something even more valuable and easy to use.

]]>
AI with Guy: OPEA How-to Videos: Intro, microservices architecture, and ChatQnA https://opea.dev/ai-with-guy-opea-how-to-videos-intro-microservices-architecture-and-chatqna/ Fri, 21 Mar 2025 14:55:49 +0000 https://opea.dev/?p=490 Check out the first three how-to videos on OPEA from AI with Guy, a new video series designed to educate developers on the latest AI trends, innovations, and real-world applications. 

Start with understanding how OPEA can help you build a scalable, composable, Gen-AI applications that can run on any cloud, on any device. Click here for the video. 

Find out more about the OPEA microservices architecture.  Click here for the video. 

By far the most popular E2E GenAI example within OPEA is ChatQnA, see it in action here.  

Be sure to let us know what you think of these and stay tuned for more soon. 

Join our mailing list to get the latest on OPEA. 

]]>
ISTE & ASCD: Moving from OpenAI to Open Source with OPEA https://opea.dev/powered-by-opea-solution-moving-from-openai-to-open-source-with-opea/ Wed, 19 Mar 2025 15:27:51 +0000 https://opea.dev/?p=478 With ChatGPT, OpenAI offers a quick way to get started with generative AI (GenAI). However, data confidentiality and total cost of ownership pose significant challenges. ISTE (International Society for Technology in Education)+ASCD (Association for Supervision and Curriculum Development) began exploring options to shift from OpenAI services to open source solutions to confidently handle private data and add the ability to run inference locally. Organizations can reduce costs and better control data privacy by replacing OpenAI services with open source microservices from the Open Platform for Enterprise AI (OPEA) project running on Intel hyperscale instances.

Read more at Intel.com.

]]>
Intel Blog Post: Document Summarization: A Step-by-Step Guide with OPEA™ 1.2 and Intel® Gaudi® 2 https://opea.dev/intel-blog-post-document-summarization-a-step-by-step-guide-with-opea-1-2-and-intel-gaudi-2/ Tue, 18 Mar 2025 16:25:45 +0000 https://opea.dev/?p=476 Are you drowning in an ocean of information? In today’s fast-paced world, staying updated and informed is essential for both personal and professional growth. However, the sheer volume of content available can be overwhelming. Document summarization applications offer a powerful solution by condensing information into concise summaries, capturing the key points from lengthy training videos, educational lectures, promotional content, research presentations, and podcasts. This greatly enhances your ability to keep up.

In this article, we will explore the various use cases of a Document Summarization (DocSum) application implemented using the Open Platform for Enterprise AI (OPEA™). Discover how these innovative tools can enhance productivity and efficiency across different domains. We will walk you through the steps to deploy and test drive OPEA’s DocSum application on the Intel® Gaudi® 2 AI accelerator using Intel® Tiber™ AI Cloud. From setup to execution, we’ll cover everything you need to know to unlock the potential of document summarization and transform the way you interact with information using this cutting-edge GenAI application.

Read more at Intel.com.

]]>
Intel Blog Post: Multimodal Question and Answer: A Step-by-Step Guide with OPEA™ 1.2 and Intel® Gaudi® 2 https://opea.dev/intel-blog-post-multimodal-question-and-answer-a-step-by-step-guide-with-opea-1-2-and-intel-gaudi-2/ Tue, 18 Mar 2025 15:43:52 +0000 https://opea.dev/?p=474 Imagine being able to ask a question using a picture, query instructional videos to receive a single targeted clip, or search through a diverse collection of multimedia files using your voice. All of this is possible with the Open Platform for Enterprise AI (OPEA™Multimodal Question and Answer (MMQnA) chatbot.  The MMQnA chatbot leverages the power of multimodal AI to deliver a flexible and intuitive way to interact with complex datasets. Whether you’re a developer, a data scientist, or an enterprise looking to enhance your information retrieval capabilities, this tool is designed to help you efficiently meet your needs.

In the era of Large Language Models (LLMs), we can now make use of robust and accurate models for complex datasets. Instead of being limited to a single modality, like text, we can leverage transformer architectures that support any modality type as an input. Here, we introduce a MMQnA chatbot capable of handling any mix of text, images, spoken audio, or video in a Retrieval-Augmented Generation (RAG) workflow.

This article will walk you through the steps to deploy and test drive OPEA’s MMQnA megaservice on the Intel® Gaudi® 2 AI accelerator using Intel® Tiber™ AI Cloud. From setup to execution, we’ll cover everything you need to know to get started with this multimodal GenAI application.

 

Read more at Intel.com.

]]>