Share this article
Latest news
With KB5043178 to Release Preview Channel, Microsoft advises Windows 11 users to plug in when the battery is low
Copilot in Outlook will generate personalized themes for you to customize the app
Microsoft will raise the price of its 365 Suite to include AI capabilities
Death Stranding Director’s Cut is now Xbox X|S at a huge discount
Outlook will let users create custom account icons so they can tell their accounts apart easier
The chatbot used in Robert F. Kennedy Jr’s campaign vanished
Third-party chatbots can bypass vulnerabilities and security measures and cause more harm than good
2 min. read
Published onMarch 5, 2024
published onMarch 5, 2024
Share this article
Read our disclosure page to find out how can you help Windows Report sustain the editorial teamRead more
Robert F. Kennedy Jr candidates independently to become the next United States president. However, his campaign is using various unorthodox methods, including a chatbot. In addition, the data used to train the AI contains information based on Kennedy’s data. Furthermore, a spokesperson said the chatbot’s primary purpose was to be an interactive FAQ.
Microsoft claims that using Azure OpenAI Services in Kennedy’s campaign chatbot doesn’t violate any rules. However, the chatbot is no longer available on his website and vanished overnight. So, there must be a reason for that.
Why is AI not trustworthy?
Unfortunately, AI is untrustworthy because it learns how to answer our questions using data from available sources. Thus, by accessing third-party AI like LiveChatAI, it is possible to bypass its safety measures and influence AI’s responses. For example, the campaign chatbot used by Robert F. Kennedy Jr spread information based on conspiracy theories and personal opinions. So, the chatbot led to misinformation.
According to Wired, if you tried to learn how to register to vote, the bot would give you a link to Kennedy’s We the People Party. However, Kennedy’s campaign chatbot is one of many that show that AI is unreliable. After all, OpenAI GPT-4 and Google Gemini released inaccurate answers in the past. In addition, some answers were mainly predictions.
AI has some limitations regarding voting and politicsdue to the lack of data. After all, large language models (LLMs) use their database to answer our questions. Additionally, they might need more newer data due to copyright. Furthermore,Microsoft and OpenAI are suedfor copyright infringement because ChatGPT stole and used text directly from news reports without mentioning their authors, titles, or sources.
In a nutshell, you should only trust AI if you check the source afterward. In addition, Kennedy’s campaign chatbot shows that they are unreliable. Furthermore, it also makes us doubt the security of the AI. After all, we can use third-party bots to breach security measures. On top of that, we can exploit the chatbot’s vulnerabilities through prompts.
What do you think? Did you ever find a breach in the chatbot’s vulnerabilities through prompts? Let us know in the comments.
More about the topics:AI,microsoft,OpenAI
Sebastian Filipoiu
Sebastian is a content writer with a desire to learn everything new about AI and gaming. So, he spends his time writing prompts on various LLMs to understand them better. Additionally, Sebastian has experience fixing performance-related problems in video games and knows his way around Windows. Also, he is interested in anything related to quantum technology and becomes a research freak when he wants to learn more.
User forum
0 messages
Sort by:LatestOldestMost Votes
Comment*
Name*
Email*
Commenting as.Not you?
Save information for future comments
Comment
Δ
Sebastian Filipoiu
Sebastian is a content writer with a desire to learn everything new about AI and gaming.