As one of this century’s most powerful data brokers, Facebook is best known for its role in sucking up the personal information of users for its advertising clients. That lucrative model has led to high risks—Facebook recently shared private messages between a Nebraska mother and her teenage daughter with police probing the girl’s at-home abortion.
But in a completely different part of the business, Facebook’s exchange of information was going the other way and to good effect. Meta Platforms this month published a webpage demonstrating its chatbot, with which anyone in the US could chat about anything. While the public response was one of derision, the company had been admirably transparent about how it built the technology, publishing details about its mechanics. That’s an approach other Big Tech firms should take more often.
Facebook has been working on BlenderBot 3 for years as part of its artificial intelligence (AI) research. A precursor was called M, a digital assistant for booking restaurants or ordering flowers that could have rivalled Apple’s Siri or Amazon’s Alexa. Over time it was revealed that M was largely powered by teams of people who helped take those bookings because AI systems like chatbots were difficult to perfect. They still are. Within hours of its release, BlenderBot 3 was making anti-Semitic comments and claiming that Donald Trump had won the last US election, while saying it wanted to delete its Facebook account. The chatbot was roundly ridiculed.
Facebook’s research team seemed rankled but not defensive. A few days after the bot’s release, Meta’s managing director for fundamental AI research, Joelle Pineau, said in a blogpost that it was “painful” to read some of the bot’s offensive responses in the press. But, she added, “we also believe progress is best served by inviting a wide and diverse community to participate.” Only 0.11% of the chatbot’s responses were flagged as inappropriate, Pineau said. That suggests most people who were testing the bot were covering tamer subjects. When I asked BlenderBot 3 who the current US president was, it responded, “This sounds like a test, lol, but it’s Donald Trump right now.” The bot brought up the former president two other times, unprompted.
Why the strange answers? Facebook trained its bot on publicly available text on the internet, which is awash in conspiracy theories. Facebook tried training the bot to be more polite by using special “safer dialogue” datasets, but that wasn’t enough.
Facebook wants people to engage voluntarily with its bot, and click ‘like’ or ‘dislike’ buttons to help train it. But the company’s openness about the system and the extent to which it is showing its work are admirable at a time when tech companies have been more closed about the mechanics of AI.
Alphabet’s Google, for instance, has not offered public access to LaMDA, its most cutting-edge large language model, a series of algorithms that can predict and generate language after being trained on gigantic data sets of text. That’s in spite of the fact that one of its own engineers chatted to the system for long enough to believe it had become sentient. OpenAI, the AI research company co-founded by Elon Musk, has also become more closed about the mechanics of some of its systems. For instance, it won’t share what training data it used to create its popular image-generating system Dall-E, which can generate any image via a text prompt but has a tendency to conform to old stereotypes—all CEOs are depicted as men, nurses as women, etc. OpenAI has said that information could be put to ill use, and that it’s propriety.
Facebook, by contrast, has not only released its chatbot for public scrutiny but also published detailed information about how it was trained. Last May it also offered free, public access to a large language model it had built called OPT-175B. That approach has won it some praise from leaders in the AI community. “Meta definitely has many ups and downs, but I was happy to see that they open-sourced a large language model,” said Andrew Ng, former head of Google Brain and founder of Deeplearning.ai, referring to the company’s move in May.
Eugenia Kuyda, whose startup Replika.ai creates chatbot companions for people, said it was “really great” that Facebook had published so many details about BlenderBot 3 and praised the company’s attempts to get user feedback to train and improve the model.
Facebook deserved much of the flak it got for sharing data about the mother and daughter in Nebraska. That’s clearly a harmful consequence to collecting so much user information over the years. But the blowback over its chatbot experiment was excessive and overdone. In this case, Facebook was doing what we need to see more of from Big Tech. Let’s hope this kind of transparency continues.
Source: livemint.com