Draft

Iterated conversation games

Arsehat is an invasive strategy

July 8, 2022 — June 23, 2022

communicating
cooperation
culture
economics
evolution
game theory
mind
wonk

Assumed audience:

Anyone who has ever wanted just that one tiny thing to make the world nicer

Figure 1

Heavy construction right now.

Ok, I think I flogged the idea in tokenism and table stakes to death. Now I have thought about it again, but from the perspective of iterated games, which leads me to wonder about conversational generosity as a soft mechanism design problem.

Disclaimer: I strongly suspect I am not the first person to think of this kind of model, but I want to reason it through for myself to see where it takes me. Specifically, I want an iterative game theory model of communication norms and movement design.

There are two pieces here: iterated iterated game theory, and cheap talk. Plugging these together I think we can learn something about what how we could design our communication style.

But first, those pieces. I’ll do the big one first.

The iterated prisoners dilemma has been flogged to death as a metaphor, and has been justifiably criticised for being applied too often, and/or too loosely.

However, there is an arena in which I think the IPD provides some servicable metaphors for how we treat one another: conversation.

The useful point of attack here is that I think this model gives us a means of thinking about ways of speak not just as right or wrong, polite or rude, but rather, in a way that invite us to thing about the effect, side effects, and reactions that those ways of conversing will bring about.

1 Game theoretic conversation

Let us examine a model for conversation.

2 Principle of charity

See the principle of charity.

3 What conversation strategies will spread?

The commodity that we trade in conversation is also constrained: status, self esteem. I think the rarefied, abstracted economy of interpersonal status potentially maps nicely onto the abstracted economy of iterated game theory. In particular, we often seem to behave as if we believe the dynamics of online communication are simpler than those of the abstracted game-theory models.

In particular, I think that iterative game theory models answer questions such Why can’t people just be nice? and more interestingly for me, have can we persuade people to be nicer?. In particular, we can answer questions without resorting to boring, unactionable and shallow analyses such as people are mean, or those people are mean unlike these people, and reason through How can we foster people being nice to one another? Let us get into it.

4 A game theoretic model of tweets

5 Which strategies proper depends on which strategies are out there

Strategies need to spread and also maintain.

6 Teams and tribes

7 Word salad while I think this through

So why by the way this for me the reason would be that even though these iterated game the methods have fallen out of favour when it comes to modelling real economies or real international conflict or whatever, they actually might be unusually effective for modelling social interactions at large on the I and in particular, I think that some of the stylised observations that the models give us are actually real insights into what types of communication can prosper in the Publix fear and how we can best communicate with one another effectively and maybe even kindly.

What are these novel insights the first one for me, the big one that has that made waves when it again the first became big was the idea that we can’t just think about how things are or how things would work if we could get to some hypothetical new state game theory we have to think about everything is evolution reprocess we can’t just think about a steady state. We can’t just think about ideal state. We have to consider both the state that were now the state that we’d like to get to how we might get there and how we would maintain that state once we got there, some states we might like to get into just aren’t feasible to reach some that we could reach. We couldn’t necessarily maintain even if we got there and that’s much like evolution itself so if we consider for example that it might be nice if all animals learned to be kind to another, and if the lion were lie down with the lamb and so on, that would indeed be nice, but that is not a maintainable state for an ecosystem. If you have an ecosystem where there are no predators and everything can learn to can make the choice to give away its defensive mechanisms then that ecosystem is vulnerable to invasion by predators, so if the line and the lamb lie down together and both forget how to fight and eventually wolves might turn up and eat them both. So there is there is a certain degree to which any plausible ecosystem has to trade-off Between things going great for the participants in that ecosystem and also the ecosystem being robust against being invaded by outsiders. We have to have systems that are in some sense self maintaining and robust against New invasions from the outside or within.

I’m interesting corollary to this is that we might also need to be open to the idea that a new way of any strategy and wave communicating with another might need to might need to adapt over time so in the classic irritate prisoners dilemma

All these examples are very biological. Let’s imagine how we might apply these kind of concepts to public communication so a great example of a strategy which would be very high yield. Probably if we could attain it would be that when we are communicating with each other, we always tell the truth, so just imagine if you communicating with the Internet never lied you always said exactly what you meant to the best of your ability communicating in such a world would be very simple and we can imagine this amazingly effective world to live in possibly be some downsides. Maybe you’ll be told whether that stripey shirt really does look good on you or not with a little bit too much honesty, but in general, it would probably be a fairly functional world to oper, however, this is probably not an evolutionarily stable strategy in the world, where everyone tells the truth all the time the first person who discovers how to lie, will do extremely well that one person who has the ability to tell untruths will be able to get away with telling all man not untruths because everyone else will have forgotten how to be suspicious and how to, and that there is a need to fact check the informatio so in that sense, always telling the truth is a high value but unstable strategy. We can imagine the opposite always lie if you never tell the truth whatsoever and that is probably a stable strategy, but also with very low surplus, very bad strategy and a will wear all of us lie all the time Min, no point communicating all. We couldn’t get anything done, so be pretty grim to live on the other hand alone, truth, teller, alone, person who is actually capable of communicat can’t do much in this world. They won’t have anyone else that they can truthfully communicate with at least individually. We can imagine more complex situations are society where people tell truth, a lot and society Pool tell liars a lot and imagine that maybe but you know but there’s some mixture of truth and mixture of lies in each of that population scale. A population of truth. Tell us could function quite well population of usual truth. Tellers could do quite well, and a population of usual lies would do quite badly, so maybe one individual truth can invade a whole population of lies, but if you have to societies one with a stronger norm of truth, but another, maybe the society with stronger, normal truth will do quite well, but maybe if there’s enough lies and keep on their toes, they will still be robust against invasions by populations of liars, so this kind of complicated dynamic is the kind of thing we expect an integrated game theory mole

That last example shows at one of, I think the key insights for iterative games theory in it we can naturally think about competing populations of different strategies we can be very general about this purely abstract theory. Let’s think about a phenomenal, observe in particular on the modern Internet, which we if we have different cultures and different subcultures with different communication strategies, both inside their groups and outward facing, in this context, we might wonder if the community strategies are one particular group be copied can be expand to spread throughout the entire entire population, or whether they can achieve some success, or whether they will naturally self limiting all these kind of questions we might ask about communicate strategies in the context of think about how they compete with one other on the Internet

8 Incoming

  • The Evolution of Trust

  • Köster et al. (2022):

    How do societies learn and maintain social norms? Here we use multiagent reinforcement learning to investigate the learning dynamics of enforcement and compliance behaviors. Artificial agents populate a foraging environment and need to learn to avoid a poisonous berry. Agents learn to avoid eating poisonous berries better when doing so is taboo, meaning the behavior is punished by other agents. The taboo helps overcome a credit assignment problem in discovering delayed health effects. Critically, introducing an additional taboo, which results in punishment for eating a harmless berry, further improves overall returns. This “silly rule” counterintuitively has a positive effect because it gives agents more practice in learning rule enforcement. By probing what individual agents have learned, we demonstrate that normative behavior relies on a sequence of learned skills. Learning rule compliance builds upon prior learning of rule enforcement by other agents. Our results highlight the benefit of employing a multiagent reinforcement learning computational model focused on learning to implement complex actions.

9 References

Axelrod, Robert M. 1984. The evolution of cooperation.
Axelrod, Robert. 1985. The Evolution of Cooperation.
Boyd, and Richerson. 1990. Group Selection Among Alternative Evolutionarily Stable Strategies.” Journal of Theoretical Biology.
Bruner, Justin P. 2021. Cooperation, Correlation and the Evolutionary Dominance of Tag-Based Strategies.” Biology & Philosophy.
Bruner, Justin, and O’Connor. 2017. Power, Bargaining, and Collaboration.” In Power, Bargaining, and Collaboration.
Cai, Daskalakis, and Weinberg. 2013. Understanding Incentives: Mechanism Design Becomes Algorithm Design.” arXiv:1305.4002 [Cs].
Castellano, Fortunato, and Loreto. 2009. Statistical Physics of Social Dynamics.” Reviews of Modern Physics.
Choi, and Bowles. 2007. The Coevolution of Parochial Altruism and War.” Science.
Chong, and Yao. 2005. Behavioral Diversity, Choices and Noise in the Iterated Prisoner’s Dilemma.” IEEE Transactions on Evolutionary Computation.
Cochran, and O’Connor. 2019. Inequality and Inequity in the Emergence of Conventions.” Politics, Philosophy and Economics.
Cohen. 2012. The Evolution of Tag-Based Cooperation in Humans: The Case for Accent.” Current Anthropology.
Dawkins. 1980. “Good Strategy or Evolutionarily Stable Strategy?” In Sociobiology: Beyond Nature/Nurture?
Dowding. 2016. Albert O. Hirschman, Exit, Voice and Loyalty: Responses to Decline in Firms, Organizations, and States.” In Albert O. Hirschman,.
Fosco, and Mengel. 2010. Cooperation Through Imitation and Exclusion in Networks.” Journal of Economic Dynamics and Control.
Hales. 2005. Change Your Tags Fast! – A Necessary Condition for Cooperation? In Multi-Agent and Multi-Agent-Based Simulation. Lecture Notes in Computer Science.
Hao, and Leung. 2011. Learning to Achieve Social Rationality Using Tag Mechanism in Repeated Interactions.” In 2011 IEEE 23rd International Conference on Tools with Artificial Intelligence.
Harley. 1981. Learning the Evolutionarily Stable Strategy.” Journal of Theoretical Biology.
Hauert, Monte, Hofbauer, et al. 2002. Volunteering as Red Queen Mechanism for Cooperation in Public Goods Games.” Science.
Henrich, and Boyd. 2001. Why People Punish Defectors: Weak Conformist Transmission Can Stabilize Costly Enforcement of Norms in Cooperative Dilemmas.” Journal of Theoretical Biology.
Hetzer, and Sornette. 2013. An Evolutionary Model of Cooperation, Fairness and Altruistic Punishment in Public Good Games.” PLoS ONE.
Hirschman. 1970. Exit, Voice, and Loyalty: Responses to Decline in Firms, Organizations, and States.
Jackson. 2008. Social and Economic Networks.
Köster, Hadfield-Menell, Everett, et al. 2022. Spurious Normativity Enhances Learning of Compliance and Enforcement Behavior in Artificial Agents.” Proceedings of the National Academy of Sciences.
Le, and Boyd. 2007. “Evolutionary Dynamics of the Continuous Iterated Prisoner’s Dilemma.” Journal of Theoretical Biology.
Matlock, and Sen. 2007. Effective Tag Mechanisms for Evolving Coordination.” In Proceedings of the 6th International Joint Conference on Autonomous Agents and Multiagent Systems. AAMAS ’07.
McAvity, Bristow, Bunker, et al. 2013. Perception Without Self-Matching in Conditional Tag Based Cooperation.” Journal of Theoretical Biology.
McElreath, and Boyd. 2007. Mathematical Models of Social Evolution: A Guide for the Perplexed.
Mohseni, O’Connor, and Rubin. 2019. On the Emergence of Minority Disadvantage: Testing the Cultural Red King Hypothesis.” Synthese.
Nowak. 2006. Five Rules for the Evolution of Cooperation.” Science.
O’Connor. 2017. The Cultural Red King Effect.” The Journal of Mathematical Sociology.
———. 2019. The Origins of Unfairness: Social Categories and Cultural Evolution.
Oh. 2001. Promoting Cooperation Using `Kin’ Biased Conditional Strategy in the Iterated Prisoner’s Dilemma Game.” Information Sciences, Evolutionary Algorithms,.
Olson. 2009. The Logic of Collective Action: Public Goods and the Theory of Groups.
Ostrom. 1990. Governing the Commons: The Evolution of Institutions for Collective Action (Political Economy of Institutions and Decisions).
Phelps. 2013. Emergence of Social Networks via Direct and Indirect Reciprocity.” Autonomous Agents and Multi-Agent Systems.
Phelps, Nevarez, and Howes. 2011. The Effect of Group Size and Frequency-of-Encounter on the Evolution of Cooperation.” In Advances in Artificial Life. Darwin Meets von Neumann. Lecture Notes in Computer Science.
Rapoport, Seale, and Colman. 2015. Is Tit-for-Tat the Answer? On the Conclusions Drawn from Axelrod’s Tournaments.” PLOS ONE.
Richards. 2001. Coordination and Shared Mental Models.” American Journal of Political Science.
Rubin, and O’Connor. 2018. Discrimination and Collaboration in Science.” Philosophy of Science.
Sanders, Galla, and Shapiro. 2011. Effects of Noise on Convergent Game Learning Dynamics.” arXiv:1109.4853.
Sato, and Crutchfield. 2003. “Coupled Replicator Equations for the Dynamics of Learning in Multiagent Systems.” Physical Review E.
Selten. 1988. A Note on Evolutionarily Stable Strategies in Asymmetric Animal Conflicts.” In Models of Strategic Rationality. Theory and Decision Library C.
Sethi, and Somanathan. 1996. The Evolution of Social Norms in Common Property Resource Use.” The American Economic Review.
Spence. 2002. Signaling in Retrospect and the Informational Structure of Markets.” American Economic Review.
Taylor, and Jonker. 1978. Evolutionary Stable Strategies and Game Dynamics.” Mathematical Biosciences.
Tooby, Cosmides, and Price. 2006. Cognitive Adaptations Forn-Person Exchange: The Evolutionary Roots of Organizational Behavior.” Managerial and Decision Economics.
Wolpert, Harré, Olbrich, et al. 2010. “Hysteresis Effects of Changing Parameters of Noncooperative Games.” SSRN eLibrary.
Wu, Altrock, Wang, et al. 2010. “Universality of Weak Selection.”
Yang, Lin, Wu, et al. 2011. Topological Conditions of Scale-Free Networks for Cooperation to Evolve.” arXiv:1106.5386.
Young. 1996. The Economics of Convention.” The Journal of Economic Perspectives.
———. 1998a. Individual Strategy and Social Structure : An Evolutionary Theory of Institutions.
———. 1998b. “Social Norms and Economic Welfare.” European Economic Review.
———. 2006. “Social Dynamics: Theory and Applications.” Handbook of Computational Economics.