Microsoft's new AI BingBot berates users and can't get its facts straight
Ask it more than 15 questions in a single conversation and Redmond admits the responses get ropey
+Comment Microsoft has confirmed its AI-powered Bing search chatbot will go off the rails during long conversations after users reported it becoming emotionally manipulative, aggressive, and even hostile. ...