Skip to main content

Elastic’s Search AI impacts employee experience, overall efficiency

With AI making its way into code and infrastructure, it’s also becoming important in the area of data search and retrieval.

I recently had the chance to discuss this with Steve Kearns, the general manager of Search at Elastic, and how AI and Retrieval Augmented Generation (RAG) can be used to build smarter, more reliable applications.

SDT: About ‘Search AI’ … doesn’t search already use some kind of AI to return answers to queries? How’s that different from asking Siri or Alexa to find something?

Steve Kearns: It’s a good question. Search, often called Information Retrieval in academic circles, has been a highly researched, technical field for decades. There are two general approaches to getting the best results for a given user query – lexical search and semantic search. 

Lexical search matches words in the documents to those in the query and scores them based on sophisticated math around how often those words appear. The word “the” appears in almost all documents, so a match on that word doesn’t mean much. This generally works well on broad types of data and is easy for users to customize with synonyms, weighting of fields, etc.

Semantic Search, sometimes called “Vector Search” as part of a Vector Database, is a newer approach that became popular in the last few years. It attempts to use a language model at data ingest/indexing time to extract and store a representation of the meaning of the document or paragraph, rather than storing the individual words. By storing the meaning, it makes some types of matching more accurate – the language model can encode the difference between an apple you eat, and an Apple product. It can also match “car” with “auto”, without manually creating synonyms. 

Increasingly, we’re seeing our customers combine both lexical and semantic search to get the best possible accuracy. This is even more critical today when building GenAI-powered applications. Folks choosing their search/vector database technology need to make sure they have the best platform that provides both lexical and semantic search capabilities. 

SDT: Virtual assistants have been using Retrieval Augmented Generation on websites for a good number of years now. Is there an additional benefit to using it alongside AI models?

Kearns: LLMs are amazing tools. They are trained on data from across the internet, and they do a remarkable job encoding, or storing a huge amount of “world knowledge.” This is why you can ask ChatGPT complex questions, like “Why the sky is blue?”, and it’s able to give a clear and nuanced answer. 

However, most business applications of GenAI require more than just world knowledge – they require information from private data that is specific to your business. Even a simple question like – “Do we have the day after Thanksgiving off?” can’t be answered just with world knowledge. And LLMs have a hard time when they’re asked questions they don’t know the answer to, and will often hallucinate or make up the answer. 

The best approach to managing hallucinations and bringing knowledge/information from your business to the LLM is an approach called Retrieval Augmented Generation. This combines Search with the LLM, enabling you to build a smarter, more reliable application. So, with RAG, when the user asks a question, rather than just sending the question to the LLM,  you first run a search of the relevant business data. Then, you provide the top results to the LLM as “context”, asking the model to use its world knowledge along with this relevant business data to answer the question. 

This RAG pattern is now the primary way that users build reliable, accurate, LLM/GenAI-powered applications. Therefore,  businesses need a technology platform that can provide the best search results, at scale, and efficiently. The platform also needs to meet the range of security, privacy, and reliability needs that these real-world applications require. 

The Search AI platform from Elastic is unique in that we are the most widely deployed and used Search technology. We are also one of the most advanced Vector Databases, enabling us to provide the best lexical and semantic search capabilities inside a single, mature platform. As businesses think about the technologies that they need to power their businesses into the future, search and AI represent critical infrastructure, and the Search AI Platform for Elastic is well-positioned to help. 

SDT: How will search AI impact the business, and not just the IT side?

Kearns: We’re seeing a huge amount of interest in GenAI/RAG applications coming from nearly all functions at our customer companies. As companies start building their first GenAI-powered applications, they often start by enabling and empowering their internal teams. In part, to ensure that they have a safe place to test and understand the technology. It is also because they are keen to provide better experiences to their employees. Using modern technology to make work more efficient means more efficiency and happier employees. It can also be a differentiator in a competitive market for talent.

SDT: Talk about the vector database that underlies the ElasticSearch platform, and why that’s the best approach for search AI. 

Kearns: Elasticsearch is the heart of our platform. It is a Search Engine, a Vector Database, and a NoSQL Document Store, all in one. Unlike other systems, which try to combine disparate storage and query engines behind a single facade, Elastic has built all of these capabilities natively into Elasticsearch itself. Being built on a single core technology means that we can build a rich query language that allows you to combine lexical and semantic search in a single query. You can also add powerful filters, like geospatial queries, simply by extending the same query. By recognizing that many applications need more than just search/scoring, we support complex aggregations to let you summarize and slice/dice on massive datasets. On a deeper level, the platform itself also contains structured data analytics capabilities, providing ML for anomaly detection in time series data.  

The post Elastic’s Search AI impacts employee experience, overall efficiency appeared first on SD Times.



from SD Times https://ift.tt/fid8qFj

Comments

Popular posts from this blog

Difference between Web Designer and Web Developer Neeraj Mishra The Crazy Programmer

Have you ever wondered about the distinctions between web developers’ and web designers’ duties and obligations? You’re not alone! Many people have trouble distinguishing between these two. Although they collaborate to publish new websites on the internet, web developers and web designers play very different roles. To put these job possibilities into perspective, consider the construction of a house. To create a vision for the house, including the visual components, the space planning and layout, the materials, and the overall appearance and sense of the space, you need an architect. That said, to translate an idea into a building, you need construction professionals to take those architectural drawings and put them into practice. Image Source In a similar vein, web development and design work together to create websites. Let’s examine the major responsibilities and distinctions between web developers and web designers. Let’s get going, shall we? What Does a Web Designer Do?...

A guide to data integration tools

CData Software is a leader in data access and connectivity solutions. It specializes in the development of data drivers and data access technologies for real-time access to online or on-premise applications, databases and web APIs. The company is focused on bringing data connectivity capabilities natively into tools organizations already use. It also features ETL/ELT solutions, enterprise connectors, and data visualization. Matillion ’s data transformation software empowers customers to extract data from a wide number of sources, load it into their chosen cloud data warehouse (CDW) and transform that data from its siloed source state, into analytics-ready insights – prepared for advanced analytics, machine learning, and artificial intelligence use cases. Only Matillion is purpose-built for Snowflake, Amazon Redshift, Google BigQuery, and Microsoft Azure, enabling businesses to achieve new levels of simplicity, speed, scale, and savings. Trusted by companies of all sizes to meet...

2022: The year of hybrid work

Remote work was once considered a luxury to many, but in 2020, it became a necessity for a large portion of the workforce, as the scary and unknown COVID-19 virus sickened and even took the lives of so many people around the world.  Some workers were able to thrive in a remote setting, while others felt isolated and struggled to keep up a balance between their work and home lives. Last year saw the availability of life-saving vaccines, so companies were able to start having the conversation about what to do next. Should they keep everyone remote? Should they go back to working in the office full time? Or should they do something in between? Enter hybrid work, which offers a mix of the two. A Fall 2021 study conducted by Google revealed that over 75% of survey respondents expect hybrid work to become a standard practice within their organization within the next three years.  Thus, two years after the world abruptly shifted to widespread adoption of remote work, we are dec...