AIResearch AIResearch
Back to articles
Security

AI Agents Can Now Cooperate Selfishly to Solve Problems

Researchers have developed algorithms that ensure rational AI agents, acting in their own self-interest, still compute correct solutions without needing a central authority—preventing cheating and guaranteeing fairness in decentralized networks.

AI Research
March 26, 2026
4 min read
AI Agents Can Now Cooperate Selfishly to Solve Problems

In distributed networks, where multiple agents work together to solve problems, a fundamental assumption has long been that all participants follow the rules without deviation. However, in real-world scenarios, agents often have their own incentives, leading them to cheat or manipulate outcomes for personal gain. This is particularly acute in tasks like computing a Maximal Independent Set (MIS), where each node in a network prefers to be included in the set, potentially disrupting classical algorithms. A new study addresses this by designing algorithms that are resilient to such rational behavior, ensuring that even selfish agents cooperate to achieve correct , without relying on a trusted central authority or monetary incentives.

The researchers discovered that classical MIS algorithms fail when nodes are rational agents motivated by self-interest. For example, in Luby's algorithm, nodes propose to join the MIS with a probability based on their degree, but a rational node could falsely claim it proposed successfully every time, preventing any node from joining and stalling the algorithm indefinitely. Similarly, in deterministic algorithms, a node might realize it can never join the MIS and thus have no incentive to participate. To overcome this, the study introduces two new algorithms that guarantee no node can unilaterally deviate to improve its expected payoff, while still ensuring correct MIS computation and fair participation for all nodes.

Ology involves modeling the distributed MIS problem as an extensive-form game with imperfect information, where each node is a rational player with a utility function. This function rewards nodes for being in a locally valid MIS solution, penalizes invalid solutions with negative infinity utility, and includes a third output symbol (⊥) as a credible threat against deviations. The first algorithm uses a Rock-Paper-Scissors (RPS) game between neighboring nodes: in each iteration, nodes play RPS with all undecided neighbors, and only those winning all games join the MIS. Randomness is generated uniformly, and deviations are handled by setting a BeenCheated flag, which triggers nodes to output ⊥, ensuring cheaters gain no advantage. The second algorithm employs lightweight cryptography, where each node's rank for joining the MIS is computed by combining its own random value with a signed random value from a chosen opponent, preventing unilateral manipulation and simulating a known MIS algorithm with verifiable randomness.

Show that both algorithms achieve a belief-independent sequential equilibrium, a stronger guarantee than typical game-theoretic equilibria like Nash or trembling-hand perfect equilibrium. This means that regardless of execution history or beliefs, no agent can improve its expected utility by deviating if others follow the algorithm. Under honest execution, the RPS-based algorithm terminates in O(34^Δ log n) rounds with high probability for graphs with maximum degree Δ, as proven using probabilistic analysis of edge removal rates. The cryptographic algorithm terminates in O(log n) rounds with high probability by simulating Métivier et al.'s with bit-based ranks. Both algorithms ensure correctness, producing a valid MIS where every node has a positive probability of inclusion, as demonstrated through theorems on independence, maximality, and inclusion probabilities referenced from the paper's sections 3 and 4.

Of this research are significant for decentralized systems where trust is limited, such as peer-to-peer networks, blockchain protocols, or multi-agent robotics. By ensuring rational agents cooperate without external enforcement, these algorithms could enhance security and efficiency in applications like distributed consensus, resource allocation, and collaborative AI. The study avoids speculative claims but notes that the approach addresses real-world nuances where nodes have conflicting preferences, moving beyond traditional Byzantine or obedient models. However, the authors caution that the algorithms do not prevent collusion among multiple nodes, and the guarantees may not hold if utility functions are slightly perturbed, highlighting areas for future work.

Limitations of the study include the exponential round complexity dependence on maximum degree in the RPS algorithm, making it less efficient for high-degree networks. The cryptographic algorithm relies on assumptions like digital signatures and bounded computational power, which are non-standard in distributed computing and may involve large message sizes. Additionally, the utility model is specific to nodes preferring inclusion in the MIS; alternative preferences, such as avoiding the MIS, are not addressed and remain an open problem. The paper also notes that deviations in deterministic internal computations are not modeled, though they manifest externally, and the algorithms are sensitive to utility function variations, potentially limiting robustness in broader scenarios.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn