Loading...

AI now not only debates with humans but negotiates and cajoles too

AI now not only debates with humans but negotiates and cajoles too
Photo Credit: Pixabay
23 Nov, 2022
Loading...

On 18 June, 2020, the world sat up and noticed how an artificial intelligence (AI) system had engaged in the first-ever live, public debates with humans. At an event held at International Business Machines Corp.’s (IBM) Watson West site in San Francisco, a champion debater and IBM’s AI system, Project Debater, began by preparing arguments for and against the statement: “We should subsidize space exploration". IBM later held a second debate between the system and another Israeli expert debater, Dan Zafrir, that featured opposing arguments on the statement: “We should increase the use of telemedicine."

In development since 2012, Project Debater was touted as IBM’s next big milestone for AI. Aimed at helping “people make evidence-based decisions when the answers aren’t black-and-white," it doesn’t just learn a topic but can debate unfamiliar topics too, as long as these are covered in the massive corpus that the system mines, which includes hundreds of millions of articles from numerous well-known newspapers and magazines. The system uses Watson Speech to Text API (application programming interface). Project Debater’s underlying technologies are also being used in IBM Cloud and IBM Watson.

Interestingly, a year later at Think 2019 in San Francisco, IBM's Project Debater lost an argument in a live, public debate with a human champion, Harish Natarajan. They were arguing for and against the resolution, “We should subsidize preschool". Both sides had only 15 minutes to prepare their speech, following which they delivered a four-minute opening statement, a four-minute rebuttal, and a two-minute summary. The winner of the event was determined by Project Debater's ability to convince the audience of the persuasiveness of the arguments. But even though Natarajan was declared the winner, 58% of the audience said Project Debater "better enriched their knowledge about the topic at hand, compared to Harish’s 20%" ().

Loading...

Raising the bar

Meta (formerly Facebook) appears to have gone a step further. On Tuesday, it announced that CICERO is the first AI "to achieve human-level performance in the popular strategy game Diplomacy". CICERO demonstrated this by playing on webDiplomacy.net, an online version of the game, where it achieved more than double the average score of the human players and ranked in the top 10% of participants who played more than one game. Marcus Tullius Cicero was a Roman writer, orator, lawyer and politician — all bundled in one.

Meta explains that unlike games like Chess and Go, Diplomacy requires an agent to recognize that someone is likely bluffing or that another player would see a certain move as aggressive, failing which it will lose. Likewise, it has to talk like a real person, displaying empathy, building relationships, and speaking knowledgeably about the game, failing which it won't find other players willing to work with it. To achieve these goals, Meta used both strategic reasoning as used in agents that played AlphaGo and Pluribus, and natural language processing (NLP), as used in models like GPT-3, BlenderBot 3, LaMDA, and OPT-175B.

Loading...

Meta has open-sourced the code and published a paper to help the wider AI community use CICERO to "spur further progress in human-AI cooperation".

How CICERO works

CICERO continuously looks at the game board to understand and model how the other players are likely to act, following which it uses this framework to control a language model that "can generate free-form dialogue, informing other players of its plans and proposing reasonable actions for the other players that coordinate well with them". Meta started with a 2.7 billion parameter BART-like language model that is pre-trained on text from the internet and fine-tuned on over 40,000 human games on webDiplomacy.net. It also developed techniques to automatically annotate messages in the training data with corresponding planned moves in the game. The idea is to control dialogue generation while persuading other players more effectively. In short, Cicero first makes a prediction of what everyone will do; Second, it refines that prediction using planning; Third, it generates several candidate messages based on the board state, dialogue, and its intents; and fourth, it filters messages to reduce gibberish and unrelated comments.

Loading...

AI-powered machines are being continuously pitted against humans in the last decade. IBM’s Deep Blue supercomputing system, for instance, beat chess grandmaster Garry Kasparov in 1996-97 and its Watson supercomputing system even beat Jeopardy players in 2011.

In March 2016, Alphabet-owned AI firm DeepMind’s computer programme, AlphaGo, beat Go champion Lee Sedol. On 7 December 2017, AlphaZero — modelled on AlphaGo — took just four hours to learn all chess rules and master the game enough to defeat the world’s strongest open-source chess engine, Stockfish. The AlphaZero algorithm is a more generic version of the AlphaGo Zero algorithm. It uses reinforcement learning, which is an unsupervised training method that uses rewards and punishments. AlphaGo Zero does not need to train on human amateur and professional games to learn how to play the ancient Chinese game of Go. Further, the new version not only learnt from AlphaGo — the world’s strongest player of the Chinese game Go — but also defeated it in October 2017.

A year later, in July 2018, AI bots beat humans at the video game Dota 2. Published by Valve Corp., Dota 2 is a free-to-play multiplayer online battle arena video game and is one of the most popular and complex e-sports games. Professionals train throughout the year to earn part of Dota’s annual $40 million prize pool that is the largest of any e-sports game. Hence, a machine beating such players underscores the power of AI. AI bots, though, lost to professional players at Dota 2, which has been actively developed for over a decade, with the game logic implemented in hundreds of thousands of lines of code. This logic takes milliseconds per tick to execute, versus nanoseconds for Chess or Go engines. The game is updated about once every two weeks.

Loading...

Read the full story on Mint.