By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
World of SoftwareWorld of SoftwareWorld of Software
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Search
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
Reading: NextGen Search – Where AI Meets OpenSearch Through MCP
Share
Sign In
Notification Show More
Font ResizerAa
World of SoftwareWorld of Software
Font ResizerAa
  • Software
  • Mobile
  • Computing
  • Gadget
  • Gaming
  • Videos
Search
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Have an existing account? Sign In
Follow US
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
World of Software > News > NextGen Search – Where AI Meets OpenSearch Through MCP
News

NextGen Search – Where AI Meets OpenSearch Through MCP

News Room
Last updated: 2025/12/17 at 4:43 AM
News Room Published 17 December 2025
Share
NextGen Search – Where AI Meets OpenSearch Through MCP
SHARE

Key Takeaways

  • As keyword search reaches its limits, the industry is shifting toward semantic, multi-modal, conversational, and agentic AI search that understands user’s intent, context and empowers users to get insights through natural-language queries without needing technical skills or custom application development.
  • Next-generation context-aware conversational search solutions can be built using OpenSearch and AI agents powered by Large Language Models (LLMs) and Model Context Protocol (MCP). MCP bridges AI agents and OpenSearch for creating intelligent search applications.
  • AI agents (specialized AI applications) are LLMs equipped with role, task, and context management capabilities. A typical AI agent integrates an LLM for reasoning, Memory for maintaining relevant context across interactions, Tools for extended capabilities, and Retrieval Augment Generation (RAG) for selective knowledge retrieval.
  • The proposed architecture brings these components together through three layers: an agentic layer for intelligence, an MCP protocol layer (MCP client & server) for communication, and a data layer for indexing, search, and analytics.
  • MCP server deployment patterns include local, remote, managed hybrid (on-premises/cloud), and cloud-native deployments, with each option offering different trade-offs based on organizational needs.

 

Introduction

Imagine a sales executive querying, “Show me the top ten products by revenue this quarter and predict/forecast next month’s trends” in plain English, and get comprehensive insights in seconds without waiting days for a BI team report. You could also ask your system, “What’s causing high latencies in my application?” and retrieve not just latency-related search results, but also a comprehensive analysis and correlation of error logs, metrics, and recent deployments.

The next-generation agentic search enables this conversational search functionality, where AI agents powered by LLMs interact with data systems through standardized protocols such as MCP to deliver conversational, context-aware search experiences.

In this article, we will explore how MCP bridges AI agents and OpenSearch to create intelligent search applications. We will also explore the evolution from keyword search to agentic search, understand the architectural components, and walk through practical implementations with live examples.

OpenSearch & Industry Use Cases

OpenSearch is an open-source search and analytics suite used for log analytics, real-time application monitoring, and website search use cases. With nearly nine hundred million software downloads since its inception and participation from thousands of contributors, and with over fourteen premier members like AWS, SAP, and Oracle, etc., OpenSearch is ranked in the top five search engines according to DB-Engines.

From an e-commerce product search to an observability platform, many industry verticals use OpenSearch for lexical, semantic search, and log analytics use cases. Let’s take a quick look at how search has evolved.

Search Evolution: From Keywords to Agents

Figure 1: Evolution of Search

Keyword Search

Also known as lexical search, which is a traditional method of search, i.e., using specific words or phrases (aka keywords). OpenSearch uses TF-IDF or Okapi BM25F algorithms by default, also popularly known as “Lucene” indices. Even though it is fast, deterministic, and language-agnostic, keyword search ignores user context and intent.

For example, a query for “Black jacket for men” returns any document containing these words, including “men wearing black shirts” or “jackets in other colors“. You can try the keyword search from the OpenSearch AI demos in Hugging Face by selecting the search type as “keyword search”.

Semantic Search

Semantic search is advanced compared to a keyword search; the user’s context and intent are considered during the query execution. In semantic search, data is being converted into vector embeddings (i.e., text is converted into a numerical representation). These vector embeddings are also called the vector store (aka vector database) use case. OpenSearch offers a choice of multiple pretrained models to convert the data from text to vector embeddings or other embeddings like image, audio, or video, etc.

In the same example we used in keyword search (e.g., search for “Black jacket for men“), then you will see results only related to “men wearing black jackets“. You can try the keyword search from the OpenSearch AI demos in Hugging Face by selecting the search type as “vector search”.

Multi-modal or Hybrid Search

This search approach is a combination of keyword and semantic search. A search can get you the keyword and semantic-based results. Moreover, with multi-modal search, search results use multiple models for different types of data like text with images, etc. For example, in the demo page, OpenSearch AI demos in Hugging Face, you might have witnessed the results showing both keywords and images.&

Conversational Search

Conversational Search uses natural language to query OpenSearch like Q&A by leveraging LLMs. These LLMs are stateless, but they maintain conversation context and history in two ways:

  • Built-in memory provided by modern LLM providers like OpenAI ChatGPT, Anthropic Claude, etc., for session-based retention
  • External memory storage systems for persistent enterprise-grade memory management. These storage systems include Traditional databases (e.g., PostgreSQL, Redis, Cassandra), Vector databases (e.g., OpenSearch, Pinecone), or Agentic Frameworks (e.g., LangChain, Strands, LlamaIndex)

Conversational Search with RAG augments the LLM response by connecting to external data sources (typically connected to a single data source) such as OpenSearch. Typically, users tell what exactly needs to be searched and retrieve data from OpenSearch. It is best suited for simple to moderate queries and straightforward information retrieval.

The key distinction is that memory (built-in or external) maintains conversation history for continuity of context. At the same time, RAG enhances the LLM response by retrieving relevant information from external data sources to provide more accurate and up-to-date answers.

Agentic Search (Next Generation Search)

Agentic Search in OpenSearch will help you to ask questions in natural language, like plain English.

Agentic Search is a superset of Conversational Search. Unlike conversational search, agents will have built-in memory capabilities and orchestrate a task workflow with LLM reasoning capabilities and make decisions on the query execution on OpenSearch. These tasks include Search, analyze, correlate, and execute. Agents will also iterate the workflow plan autonomously as needed.

Agentic Search can connect to multiple data sources by orchestrating multiple tools for information retrieval and augmenting responses. With Agentic Search, a user can keep the conversation intact and execute tools (aka tasks) on OpenSearch through Model Context Protocol, which will be discussed in later sections of this article.

Before diving deep into the next-generation agentic search architecture and implementation details, let’s look at how agents play a critical role in agentic AI application architecture.

What are AI Agents?

AI agents (specialized AI applications) are Large Language Models equipped with role, task, and context management capabilities. A typical AI agent integrates an LLM for reasoning, memory for maintaining relevant context across interactions, tools for extended capabilities, and RAG for selective knowledge retrieval, all designed to efficiently manage the LLM’s limited context window by retrieving only pertinent information and preserving critical details. Given a task, agents achieve objectives through iterative reasoning with available tools while dynamically managing what information enters the context window to optimize response generation.

Figure 2: Core Architecture of AI Agents

Let’s review two popular OpenSearch business use cases to understand how OpenSearch Agentic Search will help.




Search Use Case: Sales Analyst Creating an Executive Sales Report


The Sales Analyst (AI Agent) is tasked with creating a weekly sales performance report for executive leadership.


AI Agent leverages the Analytics Manager (LLM Orchestrator), which acts as a brain and directs:

What to analyze (weekly sales by category, top products, customer trends, and marketing campaign impact),

Where to look (sales database, inventory system, marketing platform, customer analytics),

How to investigate (generate queries to aggregate sales data, correlate with campaigns, and compare trends)


Once the execution plan is ready, the AI Agent uses available tools through MCP:


  • Sales Database (Salesforce) to query revenue, orders, and product performance
  • E-commerce Platform (MySQL) API to retrieve inventory levels and customer order details
  • Marketing Platform (SAP ERP) API to review campaign performance and correlate with sales spikes
  • AI Agent may also use reference documentation (knowledge bases/RAG) such as:


Sales report templates and KPI definitions


  • Database schema and field definitions
  • Historical sales reports and seasonal patterns
  • Business rules (e.g., how “active customer” is defined)


On Day 2, if the executive (user) needs to refer to the Day 1 sales summary by category, the AI Agent remembers (memory) the Day 1 findings and continues the context-aware conversation on Day 2.

Observability Use Case: DevOps Engineer Investigating a Production Outage


A DevOps Engineer (AI Agent) is tasked to investigate and resolve a production application performance issue.


AI Agent leverages the Incident Manager (LLM Orchestrator), which acts as a brain and directs,

What to investigate ( slow query logs, API latency metrics, recent deployments),

Where to look (application observability information, such as logs, metrics, traces),

How to investigate (generates a query to analyze Error logs with latency metrics & traces and correlates them with recent deployment timelines)


Once the execution plan is ready, the AI Agent uses available tools through MCP


  • OpenSearch to query application logs, metrics & traces
  • GitHub API to review recent code deployments for correlation
  • PagerDuty API (or other) to correlate related alerts


AI Agent may also use reference documentation (knowledge bases/RAG) such as


  • Troubleshooting runbooks
  • System architecture design documents
  • Historical incidents and resolutions


On Day 2, if the DevOps engineer (AI Agent) needs to refer to the patch applied for the Day 1 incident, the AI Agent remembers (Memory) the Day 1 findings and continues the context-aware conversations on Day 2.

Why do we need Agents?

LLMs: Yesterday’s Brain Problem

Large Language ModelsFunctional Models (FMs) were trained on a large corpus of data, but don’t have information on real-time data. So, working with an LLM alone is like working with yesterday’s brain. RAG addresses this issue by connecting LLMs to external data sources like OpenSearch or RDBMS, etc.

For example, if a DevOps engineer asks for real-time application performance metrics or insights into a production application. LLM alone can’t provide the information. LLM needs to augment the response using existing data stores like OpenSearch to provide real-time insights.

Traditional RAG requires users to specify exact queries and retrieves from a single source in one step. AI Agents enhance RAG by autonomously reasoning about the problem, orchestrating multiple data sources through MCP (e.g., OpenSearch, GitHub, CloudWatch), correlating findings, and iterating until a solution is found.

Memory of Conversation

LLMs alone don’t store user conversation history. LLMs process each prompt independently without retaining previous interactions. Agents can maintain conversation history through various memory mechanisms, like short-term and long-term memory.

So, there is a need for memory setup with an external database and to use the RAG technique to keep up with conversations. From OpenSearch 3.3 onwards, agentic memory is offered as a built-in feature. Modern AI agent frameworks come with built-in memory to eliminate the need to maintain separate databases.

Knowledge Bases

LLMs don’t have your company’s proprietary data. You can provide your company’s data as a knowledge base to the LLM. LLMs use this knowledge base to augment their responses with the RAG technique.

Tools

Each agent will have certain tools to execute tasks by leveraging LLMs for reasoning and planning capabilities. For example, OpenSearch provides a set of tools that perform tasks such as search, analyze, correlate, and execute. You can also implement your own agentic tools using Agentic Frameworks.

Challenges in Developing AI Agents

Building an AI agent is an easy task, but integrating it with existing systems like databases and web services, is complex. Each use case needs an implementation of a specific API or another way of integration with the respective services. For instance, databases use JDBC connections, and web services work on REST API calls.

As discussed in previous sections, the sales assistant agent connects to different data sources using distinct connectors to perform a comprehensive analysis.

Figure 3: Sales Assistant Agent using a custom connector per data source

Model Context Protocol (MCP) will help overcome this complexity by providing a single and simplified way (universal way) of connection.

Model Context Protocol (MCP): The Universal Connector

MCP provides a unified API to connect to different services, which makes AI agent integration seamless. MCP setup has two components.

  • Model Context Protocol: An open-source, standardized, and secured protocol (based on JSON-RPC 2.0) that governs communication between MCP clients and MCP servers. Think of it like a universal power adapter or travel power adapter where you can use it in different countries with different sockets, and the adapter can streamline the input power and provide the connectivity and output required. More information about MCP can be found in this article.
  • MCP Server: MCP Server is a special program that acts as a secure bridge between AI models and external data sources. It provides the tools to execute tasks on the respective service.

Figure 4: Sales Assistant Agent using MCP

How does OpenSearch Agentic Search work?

In this section, we chose the local deployment model for this demonstration to simplify setup. Production deployments should use managed hybrid or cloud-native options for better security and scalability.

Figure 5: OpenSearch Agentic Search – MCP setup and flow

Architecture Overview

  • Agentic Layer

    Claude Desktop functions as both a conversational interface (i.e., an agent AI application) and an MCP client, and can be downloaded to your local machine. As shown in the above figure, it communicates with the Claude Sonnet 4.5 LLM via the internet for reasoning and instructs the MCP to retrieve information from OpenSearch.
  • Protocol Layer (MCP Client and Server)

    MCP Client is configured through ‘claude_desktop_config.json', will hold the configuration to connect to the OpenSearch, and initiate the communication to the MCP Server via the MCP protocol. MCP Server runs as a standalone service that bridges the MCP protocol to OpenSearch. It exposes OpenSearch operations as MCP tools, translates protocol messages to REST API calls, and formats the results for LLM consumption.
  • Data Layer

    OpenSearch stores and indexes the data, exposing operations through the MCP server.

OpenSearch MCP Server Setup

OpenSearch provides the MCP server as the default from version 3.0 or greater. You can download and install the OpenSearch MCP server on your local machine or you can also follow the implementation guide provided in this article. The MCP server plays a critical role in translating/converting MCP tool queries to OpenSearch native REST HTTP API calls and submitting translated queries to OpenSearch and handling the results and formatting them as LLM-compatible responses.

The server also exposes OpenSearch operations such as Search, Analysis, etc, as MCP tools. By default, it will provide the tools to execute tasks on OpenSearch. The available default tools are:

  • ListIndexTool lists all indices in OpenSearch with full information, including docs.count, docs.deleted, and store.size.
  • IndexMappingTool retrieves index mapping and setting information for an index in OpenSearch.
  • SearchIndexTool searches an index using a query written in query domain-specific language (DSL) in OpenSearch.
  • GetShardsTool retrieves information about shards in OpenSearch.
  • ClusterHealthTool returns basic information about the health of the cluster.
  • CountTool returns the number of documents matching a query.
  • ExplainTool returns information about why a specific document matches (or doesn’t match) a query.
  • MsearchTool allows executing several search operations in one request.

MCP Server Deployment Patterns

Typically, MCP server installations come with the following deployment options.

  • Local Deployments

    MCP servers can run on an individual’s workstation alongside the Claude desktop. This deployment is suitable for development and testing.
  • Remote Deployments

    External service providers (e.g., Salesforce, SAP etc) expose their systems through MCP servers, typically for security and governance reasons.
  • Managed Hybrid (On-premises/Cloud) Deployment

    Organizations deploy a centralized “MCP Hub” on-premises or in their cloud environments. The organization’s MCP Hub will provide standardized, scalable, controlled access to multiple data sources.
  • Cloud-Native Deployments

    Major cloud providers like AWS, GCP, and Azure offer their own MCP services.

Please note that you can also implement your own MCP server tools according to your requirements.

Implementation Guide

This section demonstrates how to configure Claude Desktop with OpenSearch MCP Server for agentic search capabilities. We’ll walk through installation step-by-step guidance, configuration, and query examples using two sample datasets: e-commerce orders and observability data. The complete source code and step-by-step setup instructions are available at NextGenSearch-OpenSearch-MCP

Agentic Search – User and MCP Interaction Flow

Below is the high-level flow of user and MCP interaction steps, demonstrating when a query is issued by a user, how it’s being translated, and how MCP fetches and presents the data from OpenSearch to the user.

Figure 6: User and MCP Interaction Flow

[Click here to expand image above to full-size]

Now, let’s see how the overall architecture works in action.

Demo: Agentic Search in Action

The following examples demonstrate MCP-enabled agentic search using Claude Desktop connected to OpenSearch

Demo Environment

For this demonstration, we are using two default datasets that OpenSearch provides as an installation package. Please refer to the implementation guide or OpenSearch Dashboards Quick start guide for more details.

  • Sample eCommerce Orders: Retail transaction data for customer behavior analysis
  • Sample Observability Logs, Traces, and Metrics: Logs, traces, and metrics for system monitoring queries

Please note, we are using simple English data for this article/demo. But you can also implement the same for vector data on OpenSearch.

General Queries:

Let’s look at some general natural language queries using this setup. On first use, you may need to issue a query like “Connect to my OpenSearch using MCP” so that the MCP connection is initialized.

MCP tools Query: “List Tools“.

‘List tools’ query will give you the list of tools available under the MCP configuration to use with OpenSearch.

Index Query: “List index or list indices of sales data and observability data“

This is the NLP query where LLM understands the context of our query and goes through all available tools and selects ListIndexTool as the appropriate tool to list all the indices available in the OpenSearch.

Cluster Admin Query: “Is the cluster healthy?“

This is a platform operations query to check the OpenSearch cluster health. For this query, LLM uses ClusterHealthTool to provide the response back to the user.

Figure 7: MCP Generic Queries

Now, deep dive into some analytical insights on top of sales data.

Sales Analyst Demo: Conversational Agentic Search for Business Insights

Sales Analyst: Popular Product categories query:

“Can you find the most popular category of products ordered last quarter?“

This query aggregates and provides results for the product orders for last quarter in the most popular category.

Sales Analyst – AI Insights Query:

“Based on sales data, what is the interesting part to you?“

In this query, we are leveraging pure AI analytical insights on sales data.

Figure 8: Sales Analyst – Business Insights Queries

Sales Analyst – Executive Board BI Query

“Can you create a graph based on sales data for the executive board?“

This is a very useful scenario for executives, where they don’t need to rely or wait for their BI teams to provide sales performance reports; instead they can generate on demand with required clauses by querying in plain English.

Figure 9: Sales Analyst – Executive Board BI Query

[Click here to expand image above to full-size]

Note: Claude Desktop can create React.js code that can be converted to dashboards.

Claude Desktop can also publish public dashboards. For example, here is a quick reference to the above dashboard.

Now, let’s look at the DevOps role and how they can leverage this whole MCP setup with OpenSearch.

DevOps Demo: Conversational Insights on Observability Data

A DevOps engineer spends a significant amount of time troubleshooting a production issue by switching between different sets of dashboards and tools, as well as using custom scripts, increasing Mean Time To Detect (MTTD) and Mean Time To Recover (MTTR).

This investigative process might span multiple hours to days depending on the complexity of the issue. With OpenSearch Agentic Search with MCP, these workflows are conversational. Instead of writing complete domain-specific language (DSL) queries or navigating between different datasets and systems, engineers can ask operational questions in plain English.

DevOps Engineer – Application Performance Investigating Query

“What’s causing high latencies in my application?“

This query will scan through all the observability data available in different OpenSearch indices and automatically identify the relevant fields, and produce a summarized explanation of the latency issues.

DevOps Engineer – Monitoring and Observability Query

“Show me nodes with high CPU usage and their active tasks“

Same as a latency query, this query pick the right observability fields, and returns a clean summary of nodes with high CPU”

Figure 10: DevOps Engineer – Application Performance and Observability Queries

[Click here to expand image above to full-size]

DevOps Engineer – Observability – Correlation Analytics Query

“Give me CPU-to-Latency Correlation insights dashboard“

As you can see in the demo screenshot below, there is no need to switch between two screens or dashboard or manually correlation. Both CPU and Latency metrics are correlated, and Agent search gives a comprehensive view of correlation analytical insights.

Figure 11: DevOps Engineer – CPU-to-Latency Correlation Query and Dashboard

[Click here to expand image above to full-size]

For quick reference for the above correlation see the analysis published dashboard.

DevOps Engineer – Observability – Anomaly Detection Query

“Can you detect any anomalies in this observability data and create a dashboard?“

Traditional observability platforms require setting up and training anomaly-detection models on your data, whereas an LLM can automatically understand your observability signals and identify anomalies using simple plain English queries.

Figure 12: DevOps Engineer – Anomaly Detection Query and Dashboard

[Click here to expand image above to full-size]

For quick reference for the above see the anomaly detection published dashboard.

Conclusion

The evolution from keyword search to agentic search represents a fundamental shift in how organizations interact with data. While semantic searches understand the intent and context of a user query, with MCP and the combination of Large Language Models and OpenSearch, we’re stepping into a new era where search feels more like a conversation than a query.

MCP standardized protocol eliminates the integration complexity and makes it possible for AI agents to connect with different data sources, think through the context, and even act on what they find with reasoning. As AI continues to evolve, the combination of standardized protocols like MCP with powerful search engines like OpenSearch will make intelligent, context-aware data access accessible to every organization.

Sign Up For Daily Newsletter

Be keep up! Get the latest breaking news delivered straight to your inbox.
By signing up, you agree to our Terms of Use and acknowledge the data practices in our Privacy Policy. You may unsubscribe at any time.
Share This Article
Facebook Twitter Email Print
Share
What do you think?
Love0
Sad0
Happy0
Sleepy0
Angry0
Dead0
Wink0
Previous Article Chinese flying car startup completes first low-altitude flight in Japan · TechNode Chinese flying car startup completes first low-altitude flight in Japan · TechNode
Next Article The Apple TV app on Android adds this key missing feature The Apple TV app on Android adds this key missing feature
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Stay Connected

248.1k Like
69.1k Follow
134k Pin
54.3k Follow

Latest News

China’s NEV Sales Hit 1.395 Million in August, Up 26.8% Year-on-Year · TechNode
China’s NEV Sales Hit 1.395 Million in August, Up 26.8% Year-on-Year · TechNode
Computing
Security platform consolidation in 2026: The AI imperative | Computer Weekly
Security platform consolidation in 2026: The AI imperative | Computer Weekly
News
AI patent startup Ankar raises £15m Series A – UKTN
AI patent startup Ankar raises £15m Series A – UKTN
News
Vivo Product Manager Reveals X300 with 200MP Zeiss Camera, 6.3-Inch Display · TechNode
Vivo Product Manager Reveals X300 with 200MP Zeiss Camera, 6.3-Inch Display · TechNode
Computing

You Might also Like

Security platform consolidation in 2026: The AI imperative | Computer Weekly
News

Security platform consolidation in 2026: The AI imperative | Computer Weekly

4 Min Read
AI patent startup Ankar raises £15m Series A – UKTN
News

AI patent startup Ankar raises £15m Series A – UKTN

2 Min Read
Google’s New AI Agent Helps You Get Ahead With a Morning Briefing
News

Google’s New AI Agent Helps You Get Ahead With a Morning Briefing

4 Min Read
How to avoid AI audiobooks on Libby
News

How to avoid AI audiobooks on Libby

3 Min Read
//

World of Software is your one-stop website for the latest tech news and updates, follow us now to get the news that matters to you.

Quick Link

  • Privacy Policy
  • Terms of use
  • Advertise
  • Contact

Topics

  • Computing
  • Software
  • Press Release
  • Trending

Sign Up for Our Newsletter

Subscribe to our newsletter to get our newest articles instantly!

World of SoftwareWorld of Software
Follow US
Copyright © All Rights Reserved. World of Software.
Welcome Back!

Sign in to your account

Lost your password?