Wednesday, July 19, 2023

Multiple 'selves' of modular agents boost AI learning


Credit: Pixabay/CC0 Public Domain

How and why we make thousands of decisions every day has long proven to be a popular area of research and commentary.

"Predictably Irrational: The Hidden Forces That Shape Our Decisions," by Dan Ariely; "Nudge: Improving Decisions about Health, Wealth and Happiness," by Richard Thaler and Cass Sunstein; and "Simply Rational: Decision Making in the Real World," by Gerd Gigerenzer, are just a few of the scores of books analyzing the mechanics of decision-making that appear on current best-seller lists.

A team of researchers at the Princeton Neuroscience Institute has now joined the discussion with a paper examining the decision-making process when it comes to machine learning. They say they have found an approach that improves upon the commonly applied single-agent process.

In a paper published July 3 in Proceedings of the National Academy of Sciences, researchers outlined a study comparing reinforcement learning approaches used in single AI agent and modular multi-AI agent systems.

They trained deep reinforcement learning agents in a simple survival game on a two-dimensional grid. The agents were trained to seek various resources hidden around the field and to maintain sufficient supply levels to prevail.

One agent, seen as the "unified brain" or "self," operated in standard fashion, taking a step-by-step approach to evaluate each objective and, through trial-and-error, learning what the best solutions are each step of the way.

Monolithic agent in stationary environment, final 300 training steps. Top: Location of resources (yellow) and agent (moving pixel). Middle: State-value (i.e. maximum Q-value) for each agent (or sub-agent) calculated at each grid location. Bottom: Internal stat levels over time. Credit: Proceedings of the National Academy of Sciences (2023). DOI: 10.1073/pnas.2221180120

The modular agent, however, relied on input from sub-agents that had more narrowly defined goals and had their own unique experiences, successes and failures. Once input from the multiple modules were assessed in a single "brain," the agent made choices on how to proceed.

The researchers compared the setup to the principles involved in the classic longstanding debate over how the individual manages conflicting needs and objectives.

Whether a decision "relies on a single, monolithic agent (or 'self') that takes integrated account of all needs, or rather reflects an emergent process of competition among multiple modular agents (i.e., 'multiple selves') … pervades mythology and literature," lead researcher Jonathan Cohen said. "It is a focus of theoretical and empirical work in virtually every scientific discipline that studies agentic behavior, from neuroscience, psychology, economics, and sociology to artificial intelligence and machine learning."

The singular agent achieved the game's goals after 30,000 training steps. The modular agent, however, learned faster, making significant progress after only 5,000 learning steps.

"Compared to the standard monolithic approach, modular agents were much better at maintaining homeostasis of a set of internal variables in simulated environments, both static and changing," Cohen said.

The team concluded that the modular setup allowed sub-agents that focused on limited objectives to adapt to environmental challenges faster.

"The actions determined by the needs of one sub-agent served as a source of exploration for the others," Cohen said, "allowing them to discover the value of actions they may not have otherwise chosen in a given state."

He also explained that while the monolithic approach struggled with "the curse of dimensionality"—the exponentially spiraling growth of options as the complexity of the environment was increased—the modular agents, "specialists" with limited objectives, focused on smaller individual tasks and were better able to quickly adapt to environmental shifts.

"We show that designing an agent in a modular fashion as a collection of sub-agents, each dedicated to a separate need, powerfully enhanced the agent's capacity to satisfy its overall needs," the paper stated.

By more efficiently and more quickly adapting to changing environments and goals, researchers added, the modular approach "may also explain why humans have long been described as consisting of 'multiple selves.'"

by Peter Grad , Tech Xplore

Source: Multiple 'selves' of modular agents boost AI learning (techxplore.com)

No comments:

Post a Comment