Got It AI Spys ChatGPT Hallucinations
January 23, 2023

Got It AI, the Autonomous Conversational AI company, introduced an innovative
new "Truth Checker" AI that can identify when ChatGPT is hallucinating
(generating fabricated answers) when answering user questions over a large set
of articles or knowledge base. This innovation makes it possible to deploy
ChatGPT-like experiences without the risk of providing incorrect responses to
users or employees. Enterprises can now confidently deploy generative
conversational AIs that leverage large scale knowledge bases such as those used
for external customer support or for internal user support queries.
The Truth Checker AI, uses a separate, advanced Large Language Model (LLM) based
AI system and a target domain of content (e.g. a large knowledge base or a
collection of articles) to train itself autonomously for one task: truth
checking. ChatGPT or an underlying LLM such as a GPT-3.5 model, provided with
the same content, can then be used to answer questions in a contextual,
multi-turn chat dialog. Each response is evaluated for truthfulness before being
presented to the user. Whenever an inaccurate response is detected, the response
is not presented to the user. Instead, a reference to relevant articles which
contain the answer is provided.
"We tested our technology with a dataset of 1000+ articles across multiple
knowledge bases using multi-turn conversations with complex linguistic
structures such as co-reference, context, and topic switches," said Chandra
Khatri, former Alexa Prize team leader and co-founder of Got It AI. "ChatGPT
produced incorrect responses for about 20% of the queries when given all the
relevant content for the query in its prompt space. Our Truth Checker AI was
able to detect 90% of the inaccurate responses, without human help. We will also
provide the customer with a simple user interface to the Truth Checking AI, to
further optimize it, identify the remaining inaccuracies and eliminate virtually
all inaccurate responses."
"Our technology is a major breakthrough in autonomous conversational AI for
'known' domains of content, such as enterprise knowledge bases, versus 'open
domain' such as the entire world wide web," said Amol Kelkar, formerly an
architect for Microsoft Office Online and co-founder of Got It AI. "It goes
beyond prompt engineering, fine tuning, or just a UI layer. It is a proprietary
model that enables us to deliver scalable, accurate and fluid conversational AI
for customers planning to leverage generative LLMs. Truth checking the generated
responses cost-effectively, is the key capability that closes the gap between an
R&D system and an enterprise ready system."
Private Beta Signups
Got
It AI's Truth Checker AI is being made available inside its ArticleBot product,
which leverages Davinci-3, a GPT-3.5 generative LLM from OpenAI. OpenAI is
expected to make ChatGPT and the much anticipated GPT-4 better and even better
with its Reinforcement Learning Human Feedback and other approaches. However,
some inaccuracies will continue to exist in any generative LLM based chatbot,
especially given specific target content, and content that is updated regularly.
Got It AI's ArticleBot with Truth Checker AI will support the latest available
generative LLMs.
To simplify evaluation of the product, Got It AI makes it possible for customers
to simply provide a knowledge base or a set of articles. ArticleBot requires no
configuration to train itself on the target content and users can start testing
it within minutes. It is ideal for enterprise-grade conversational AI customer
support, help desk and agent assist applications. |