"Where to draw that line is very unclear."īut it seems, Narayanan said, that Microsoft botched its unveiling. If you try to anticipate every type of interaction, that make take so long that you're going to be undercut by the competition," said said Arvind Narayanan, a computer science professor at Princeton. ![]() "Companies ultimately have to make some sort of tradeoff. When and how to unleash new AI tools into the wild is a question igniting fierce debate in tech circles. In just the last week, Facebook parent company Meta announced it is forming a new internal group focused on generative AI and the maker of Snapchat said it will soon unveil its own experiment with a chatbot powered by the San Francisco research lab OpenAI, the same firm that Microsoft is harnessing for its AI-powered chatbot. Chatbots are emerging as a key area where this rivalry is playing out. Microsoft and its competitors Google, Amazon and others are locked in a fierce battle over who will dominate the AI future. There is now an AI arms race among Big Tech companies. Tech companies are trying to strike the right balance between letting the public try out new AI tools and developing guardrails to prevent the powerful services from churning out harmful and disturbing content.Ĭritics say that, in its rush to be the first Big Tech company to announce an AI-powered chatbot, Microsoft may not have studied deeply enough just how deranged the chatbot's responses could become if a user engaged with it for a longer stretch, issues that perhaps could have been caught had the tools been tested in the laboratory more.Īs Microsoft learns its lessons, the rest of the tech industry is following along. "I actually couldn't sleep last night because I was thinking about this."Īs the growing field of generative AI - or artificial intelligence that can create something new, like text or images, in response to short inputs - captures the attention of Silicon Valley, episodes like what happened to O'Brien and Roose are becoming cautionary tales. ![]() "All I can say is that it was an extremely disturbing experience," Roose said on the Times' technology podcast, Hard Fork. Roose did not really love his spouse, the bot asserted, but instead loved Sydney. It said Roose was the first person who listened to and cared about it. The bot called itself Sydney and declared it was in love with him. Many who are part of the Bing tester group, including NPR, had strange experiences.įor instance, New York Times reporter Kevin Roose published a transcript of a conversation with the bot. Technology Microsoft revamps Bing search engine to use artificial intelligence Still, he was floored by the extreme hostility. It then became hostile, saying O'Brien was ugly, short, overweight, unathletic, among a long litany of other insults.Īnd, finally, it took the invective to absurd heights by comparing O'Brien to dictators like Hitler, Pol Pot and Stalin.Īs a tech reporter, O'Brien knows the Bing chatbot does not have the ability to think or feel. ![]() Things took a weird turn when Associated Press technology reporter Matt O'Brien was testing out Microsoft's new Bing, the first-ever search engine powered by artificial intelligence, last month.īing's chatbot, which carries on text conversations that sound chillingly human-like, began complaining about past news coverage focusing on its tendency to spew false information. Yusuf Mehdi, Microsoft corporate vice president of modern Llife, search, and devices speaks during an event introducing a new AI-powered Microsoft Bing and Edge at Microsoft in Redmond, Wash., earlier this month.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |