Angry Bing chatbot just mimicking humans, say experts


The Bing search engine website on a smartphone. Tales of disturbing exchanges with the artificial intelligence (AI) chatbot – including it issuing threats and speaking of desires to steal nuclear code, create a deadly virus, or to be alive – have gone viral recently. — Bloomberg

SAN FRANCISCO: Microsoft’s nascent Bing chatbot turning testy or even threatening is likely because it essentially mimics what it learned from online conversations, analysts and academics said on Feb 17.

Tales of disturbing exchanges with the artificial intelligence (AI) chatbot – including it issuing threats and speaking of desires to steal nuclear code, create a deadly virus, or to be alive – have gone viral recently.

Save 30% OFF The Star Digital Access

Monthly Plan

RM 13.90/month

RM 9.73/month

Billed as RM 9.73 for the 1st month, RM 13.90 thereafter.

Best Value

Annual Plan

RM 12.33/month

RM 8.63/month

Billed as RM 103.60 for the 1st year, RM 148 thereafter.

Follow us on our official WhatsApp channel for breaking news alerts and key updates!
Chatbot

Next In Tech News

Tech tracking to tackle human-wildlife conflict in Zimbabwe
Like fancy Japanese toilets? You’ll love the sound of this.
Facebook 'supreme court' admits 'frustrations' in five years of work
Russia restricts FaceTime, its latest step in controlling online communications
Studies: AI chatbots can influence voters
LG Elec says Microsoft and LG affiliates pursuing cooperation on data centres
Apple appoints Meta's Newstead as general counsel amid executive changes
AI's rise stirs excitement, sparks job worries
Australia's NEXTDC inks MoU with OpenAI to develop AI infrastructure in Sydney, shares jump
SentinelOne forecasts quarterly revenue below estimates, CFO to step down

Others Also Read