Home > NewsRelease > Chatbots Can’t Be Trusted, and We Need Tools to Find Fact From Fiction in Them
Text
Chatbots Can’t Be Trusted, and We Need Tools to Find Fact From Fiction in Them
From:
Dr. Patricia A. Farrell -- Psychologist Dr. Patricia A. Farrell -- Psychologist
For Immediate Release:
Dateline: Tenafly, NJ
Saturday, April 27, 2024

 

AI has impacted the lives of everyone around the globe, but we can’t trust its chatbots because they make things up and spread disinformation.

Photo by Mariia Shalabaieva on Unsplash

Chatbots use AI, automatic rules, natural language processing (NLP), and machine learning (ML) to process data and answer questions. Bots that talk to people come in two main types: task-oriented (declarative) chatbots are programs that only do one thing, and virtual or digital helpers are another name for data-driven and predictive conversational chatbots. They are much smarter, more interactive, and more personalized than task-oriented chatbots.

About 1.5 billion people use chatbots, and most live in the United States, India, Germany, the United Kingdom, and Brazil. More and more people worldwide will be using chatbots in the future. By 2027, a quarter of businesses will probably use them as their primary way to contact customers. A gain of about $200 million a year is shown by this huge growth. We expect this number to hit $3 billion by the end of this decade, given its current compound annual growth rate (CAGR) of about 22%.

While chatbots are gaining importance in business and potentially healthcare, there are inherent problems that must be addressed. Ignoring issues in the chatbot can lead to biased or distorted information in chatbot algorithms. Training a chatbot is a challenging task that requires careful consideration and verification of the results. One of the shortcomings that must be overcome when creating a chatbot is the failure to carefully parse out anything that could indicate bias or failure on the part of the programmers to understand their own biases or shortcomings.

I have found with research print on chatbots that they have returned information with alleged articles and URLs that were nonexistent. If I had used them, my article would have contained many mistakes. Verifying any use of AI in medical and healthcare information searches is crucial.

The phone tree was the first chatbot. Customers found it frustrating to navigate through the computer menu to reach the automated customer service model. This model changed into pop-up, live chats on the screen as technology improved and AI, ML, and NLP became smarter. The process of change has kept going.

Although primarily aimed at business, chatbots, such as ChatGPT3, can be used for a variety of purposes, including academic research, personal interest, creative efforts, writing, or marketing. A chatbot can help with computer code, from improving it to writing code in different languages.

ChatGPT3 will allow you to prompt it to rewrite what it has provided to you and "apologize" if it has not met your expectations. Then, it will go on to provide another version of what you were seeking when you are more detailed in your prompt. The more detailed your prompt, the more likely it is that you will receive satisfactory information.

This can go on for many versions of your prompt until you are satisfied. It does not tire of attempting to satisfy your request. You can also request the number of words you wish the answer to your prompt to be in when you receive it.

Chatbots can also assist with identifying errors and generating various types of content. For example, they can summarize a play, book, or story, write a press release, write a lesson plan on a specific topic, develop a marketing plan, outline a research project or paper, and perform many other required productions.

One of the problems with research papers specifically, especially when the individual wants complete URLs for any research cited, is that the material does not exist at that computer address and, in fact, may not exist at all. The chatbot aims to follow the requested prompt, and that's one of their faults. Chatbots excel at creating fake titles or information for non-existent research articles and advertisements, and without fact-checking, they can deceive instead of providing accurate information.

While trying to please, AI chatbots can create extremely problematic situations. Take, for example, a recent interaction regarding a chatbot and elections. The GPT-4 and Google’s Gemini chatbots were trained on huge amounts of text from the internet and ready to give AI-generated answers. However, they found that the chatbots often gave voters wrong information or told them to go to polling places that did not exist 50% of the time. They also advised voters to stay home and not vote.

Remember, if you're not using the latest version of the chatbots, they won't have the most current information. For example, ChatGPT3 does not provide information after 2020, so it will tell you it can't do that if you want current information. To get current information, you must subscribe to the more current version of it. Of course, ChatGPT3 is free, which is an advantage to those who have to watch their money, but it cannot do it if you need accurate 2024 information.

Too many chatbot answers are made up, and a new tool to discover the false answers was needed. A company called Vectara, which was started by former Google workers, found that chatbots make up facts at least 3% of the time.

Cleanlab is an AI company that started as a part of MIT's quantum computing lab. They developed a new tool in 2021 that helps people understand the reliability of these models. It found errors in 10 commonly used data sets for teaching machine-learning algorithms. Data scientists may mistakenly believe that all future answers from big language models will be accurate based on a few correct responses.

Another problem, of course, is that AI has made it possible for fake people to be created on the Internet. Trolls and bots make it harder to learn online by misleading and causing skepticism about reliable information and people.

The future of AI has great promise, but it also requires careful consideration and a degree of concern that we may not have attributed to it in the past.

Website: www.drfarrell.net

Author's page: http://amzn.to/2rVYB0J

Medium page: https://medium.com/@drpatfarrell

Twitter: @drpatfarrell

Attribution of this material is appreciated.

News Media Interview Contact
Name: Dr. Patricia A. Farrell, Ph.D.
Title: Licensed Psychologist
Group: Dr. Patricia A. Farrell, Ph.D., LLC
Dateline: Tenafly, NJ United States
Cell Phone: 201-417-1827
Jump To Dr. Patricia A. Farrell -- Psychologist Jump To Dr. Patricia A. Farrell -- Psychologist
Contact Click to Contact
Other experts on these topics