• 📖 Cover
  • Contents

Chapter 5: Attention, Learning, and Games on Networks

Chapter Introduction

Chapters 1 through 4 treated networks as objects: static structures with measurable properties, communities to discover, and generative processes to fit. That perspective is indispensable — you cannot understand behavior on a network without first understanding the network itself. But it is incomplete. Networks are not just scaffolding; they are the medium through which people form beliefs, make decisions, and adjust their actions in light of what they see others doing. A trader watches the order flow around her and revises her estimate of the asset’s fair value. A voter listens to the opinions that circulate in his social circle before settling on a candidate. A firm monitors whether its rivals adopt a new technology before committing its own capital. In every case, the structure of the communication network — who observes whom, how many intermediaries stand between any two agents — shapes the collective outcome as powerfully as any individual’s private information or rational calculation.

This final chapter asks the behavioral question: given a fixed network, how do the agents sitting at its nodes update beliefs, form opinions, and play strategic games against each other? The answer turns out to require two entirely different intellectual frameworks, both of which are active frontiers of economic research. The first framework — social learning — is rooted in decision theory and asks whether decentralized observation of others’ actions leads a population toward the truth. The second — games on networks — is rooted in game theory and asks when a population converges to a coordinated equilibrium, and which equilibrium that is.

Three motivating examples set the stage. In financial markets, traders who observe a sequence of buy orders rationally conclude that earlier buyers had positive private signals, and they buy too — even if their own private signal is negative. This is an information cascade, a rationality trap in which perfectly intelligent agents collectively lock onto the wrong answer because they are all updating from the same public history. The 1999–2000 dot-com bubble exhibited exactly this dynamic: each successive analyst upgrade reinforced the last, until the public history overwhelmed any contrarian signal. Second, viral content on social networks spreads through a mechanism that looks nothing like rational updating: a user shares a post not because she has independently verified its truth but because she sees that her friends have already shared it, and she interprets their action as an endorsement. The structure of the cascade — not the quality of the content — largely determines what goes viral. Third, in voting, the sequential disclosure of exit poll projections before western polls close suppresses turnout in states that have not yet voted, because each listener takes the projected winner as a signal about how others are voting and updates accordingly.

All three examples share a common feature: the outcome — right or wrong, informative or misleading, welfare-enhancing or destructive — depends critically on the structure of the network through which information and imitation flow. Understanding that dependence is what this chapter is about.

What you will learn

This chapter develops four interlocking ideas. Bayesian social learning shows how rational agents fall into information cascades from which no further private information can escape. The DeGroot model provides a tractable, empirically grounded model of how opinions evolve through repeated averaging over time. Games on networks — coordination, threshold diffusion, and anti-coordination — show how the network structure determines which equilibria are reachable and how quickly. And a mini case study ties the whole chapter together by connecting DeGroot influence weights to PageRank centrality introduced in Chapter 2, revealing that two apparently different ideas are really the same mathematical object wearing different clothes.


Table of Contents

  1. The Social Transmission of Information
  2. Bayesian Social Learning and Information Cascades
  3. The DeGroot Model of Naïve Learning
  4. Wise Crowds versus Captured Crowds
  5. Naïve Learning with Uninformed Agents — Banerjee et al. (2021)
  6. Coordination Games on Networks and Best-Response Dynamics
  7. Threshold Models of Diffusion
  8. Anti-Coordination and Anti-Conformity
  9. Mini Case Study — DeGroot on the Florentine Families Network

The Social Transmission of Information

Why opinions do not arise in isolation

Every individual arrives at their beliefs through a combination of private experience and social observation. The relative weight placed on each differs by context, but in almost every important decision of social or economic life, the social component is substantial. This is not a sign of irrationality. When private signals are weak and noisy — and they usually are — it is entirely rational to attach weight to the behavior of others, because that behavior encodes information that you do not directly possess. The problem arises not from the rationality of any individual agent but from the aggregate dynamics that rational individual behavior can generate.

Three mechanisms drive social influence on networks:

Observational learning occurs when agents act sequentially and each agent observes the actions of predecessors but not their private signals. Actions are a lossy encoding of information — they compress a continuous private signal into a discrete choice (buy or sell, adopt or not, vote yes or no). This compression creates an information bottleneck that eventually swamps the private signal of any future agent.

Opinion averaging occurs in repeated interaction settings, where agents exchange opinions explicitly and update by averaging. This is the setting modeled by DeGroot. Unlike Bayesian updating, naïve averaging is computationally trivial — it requires no knowledge of prior probabilities or signal likelihoods — but it generates dynamics that are surprisingly sophisticated, including convergence to a consensus that is a weighted average of initial opinions.

Strategic interaction occurs when the payoff from an action depends on how many neighbors take the same action, as in coordination games. Here the network matters because it determines each agent’s effective payoff — the number of neighbors she can coordinate with — and therefore which actions are individually rational given the actions she expects her neighbors to take.

Finance application: herding and momentum

All three mechanisms operate simultaneously in financial markets. Observational learning generates herding: institutional investors observe the portfolio weights of earlier-moving peers and rationally tilt their own portfolios toward the same assets. Opinion averaging generates momentum: analyst sentiment spreads through advisory networks, and a bullish revision by one analyst propagates to correlated analysts within a few trading days. Strategic interaction generates bubbles: each investor’s payoff from holding a rising asset depends partly on the willingness of others to continue holding it, which is a coordination game. Understanding which mechanism dominates in a given episode — pure information cascade, naïve averaging, or coordinated strategy — is essential for both regulatory design and trading strategy.

A formal setup

Let the network be a directed graph \(G = (V, E)\) where \(V = \{1, 2, \ldots, n\}\) is the set of agents and a directed edge \((i, j) \in E\) means that agent \(i\) observes or listens to agent \(j\). The strength of that attention is captured by a non-negative weight \(T_{ij} \geq 0\). Across all sections of this chapter, we will vary what the agents observe (actions vs. opinions), how they process information (Bayesian vs. naïve), and what they optimize (accuracy vs. strategic payoff). In each case, the topology of \(G\) will turn out to be the dominant determinant of the collective outcome.


Bayesian Social Learning and Information Cascades

The sequential learning problem

The foundational model of social learning is due to Bikhchandani, Hirshleifer, and Welch (1992) and, independently, Banerjee (1992). It is one of the most elegant results in economic theory: starting from fully rational agents, it derives a collective outcome that looks systematically irrational.

Set up the environment as follows. There is an unknown state of the world \(\theta \in \{0, 1\}\), with each value equally likely a priori. Think of \(\theta = 1\) as “this investment is good” and \(\theta = 0\) as “this investment is bad,” or \(\theta = 1\) as “this restaurant is good” and \(\theta = 0\) as “this restaurant is bad.” Each agent \(i\) receives a private signal \(s_i \in \{H, L\}\), where

\[P(s_i = H \mid \theta = 1) = P(s_i = L \mid \theta = 0) = q > \frac{1}{2}.\]

So the signal is informative but not conclusive: a high signal makes \(\theta = 1\) more likely, but does not rule out \(\theta = 0\). Agents act in order: agent 1 first, then agent 2, and so on. Each agent observes all previous actions (not the private signals behind them) before choosing her own action.

The action set is binary: each agent chooses \(a_i \in \{0, 1\}\), and a payoff of 1 is earned if the action matches the state. Without any social information, the first agent simply acts on her signal: \(a_1 = 1\) if \(s_1 = H\) and \(a_1 = 0\) if \(s_1 = L\). Her action is therefore fully revealing.

The cascade mechanism

Agent 2 observes \(a_1\) and holds signal \(s_2\). By Bayes’ rule, after observing \(a_1 = 1\) (which she correctly interprets as \(s_1 = H\)), her posterior probability of \(\theta = 1\) is

\[P(\theta = 1 \mid s_2, a_1 = 1) \propto P(a_1 = 1 \mid \theta = 1) \cdot P(s_2 \mid \theta = 1) \cdot P(\theta = 1).\]

If her own signal is also high, she adopts (\(a_2 = 1\)) confidently. If her signal is low, the two signals cancel and she is indifferent; by convention (or a small tiebreak), she follows her own signal or randomizes. So far the system is aggregating information efficiently.

Consider agent 3. Suppose she observes \(a_1 = a_2 = 1\) and holds signal \(s_3 = L\). The public history — two adoptions — is equivalent to two high signals. Her own low signal cannot overturn two high signals when \(q < \sqrt{1/2} \approx 0.707\), which is the generic case. Agent 3 will adopt regardless of \(s_3\). This is the cascade onset: from this point forward, every agent has the same posterior (driven by the public history) and therefore takes the same action \(a_i = 1\) regardless of her private signal. The social mechanism has jammed.

More formally, a cascade occurs when the public history is strong enough that no private signal can shift the optimal action. Define the log-likelihood ratio of the public history as

\[\ell_t = \log \frac{P(H_t \mid \theta = 1)}{P(H_t \mid \theta = 0)},\]

where \(H_t\) denotes the sequence of actions observed by agent \(t+1\). A cascade toward action 1 occurs when \(\ell_t > \log \frac{q}{1-q}\), meaning the public evidence so far exceeds the informativeness of any single private signal. Once in a cascade, \(\ell_t\) is no longer updated by subsequent actions (because those actions carry no new information), and the cascade is irreversible absent an external shock.

What makes cascades dangerous

Three properties make information cascades economically significant. First, cascades occur with positive probability even when the majority of private signals point in the wrong direction. A run of two early adopters with high signals can start a cascade toward \(a = 1\) even when, say, six of the next ten agents have low signals — all six follow the crowd. The aggregate action is biased.

Second, cascades are fragile to public information releases. Because private signals are suppressed during a cascade, the collective action carries no information beyond the first few adopters. A single credible announcement can shatter a cascade and reverse the entire population’s behavior. This is the mechanism behind sudden market reversals and social media pile-ons that flip direction within hours.

Third, the onset of a cascade is essentially determined by the first two or three movers. This places enormous strategic importance on early-mover identity — not because first movers are more intelligent but because their actions form the public history that determines whether the cascade locks in. This is the economic logic behind the race to be first in financial markets and the disproportionate influence of early adopters in technology adoption.

Finance application: dot-com and the herding literature

Welch (2000) documented direct evidence of analyst herding on Wall Street: the revisions of a given analyst are significantly correlated with the direction of recent revisions by other analysts, controlling for the new fundamental information available. Scharfstein and Stein (1990) offered a reputational explanation — a bad-news analyst herds toward the consensus to avoid being seen as an outlier — which is a rational but informationally destructive equilibrium. Both mechanisms produce information cascades in professional financial advice.

Live simulation: how fast does a cascade lock in?

The following cell simulates a sequential Bayesian cascade of \(n = 30\) agents. Each agent observes all previous actions and holds a private signal with accuracy \(q = 0.65\). The state is fixed at \(\theta = 1\), so all signals favor adoption, but the cascade may still occasionally lock onto the wrong action if early agents happen to receive low signals. The simulation records the running fraction of agents who have adopted and plots it alongside a dashed line indicating the true state, revealing how quickly the public history overwhelms any remaining private information.

The two panels tell different stories about the same simulation. The left panel shows what an outside observer sees: a rising adoption fraction that appears to reflect growing consensus. The right panel shows the information content of that consensus: once the public log-likelihood ratio \(\ell_t\) crosses the signal-accuracy threshold (the dashed red line), the chain locks in. Every action thereafter is a copy, not an independent signal. The social mechanism has produced pseudo-consensus — agreement without information aggregation.

Try adjusting the seed parameter in the simulation call: you will find runs where the cascade locks onto the wrong action (\(a = 1\) even when signals favor \(a = 0\), or vice versa). This is not a failure of rationality; it is rationality’s consequence in a world with sequential action disclosure.

Predict before running

Before modifying the cell: if you change q=0.65 to q=0.55 (weaker signals), predict whether the cascade will onset sooner or later. Then change it and run. Does the cascade onset shift as you expected? What happens when q=0.95?


The DeGroot Model of Naïve Learning

From rational updating to naïve averaging

Bayesian learning is cognitively demanding. It requires agents to maintain a prior, compute likelihoods, and apply Bayes’ rule — operations that are tractable for stylized binary signals but quickly become intractable as the environment grows more complex. A large body of experimental evidence, reviewed by Camerer (2003), shows that real people systematically deviate from Bayesian updating: they are subject to base-rate neglect, overconfidence, and anchoring. More fundamentally, Bayesian updating requires knowing the structure of the information environment — who else is observing what — which is rarely available in practice.

Morris DeGroot (1974) proposed a far simpler model that has proven remarkably predictive of opinion dynamics in real social networks. Each agent holds a real-valued opinion \(x_i(t) \in [0, 1]\) about some uncertain state — the probability that a policy is good, the likelihood that a product is high quality, the forecast of next quarter’s earnings. At each period, every agent updates by taking a weighted average of her neighbors’ current opinions:

\[x_i(t+1) = \sum_{j=1}^{n} T_{ij} \, x_j(t), \qquad \text{where } T_{ij} \geq 0 \text{ and } \sum_j T_{ij} = 1.\]

In matrix form, if \(\mathbf{x}(t)\) is the \(n\)-dimensional column vector of opinions at time \(t\), then

\[\mathbf{x}(t+1) = T \, \mathbf{x}(t),\]

where \(T\) is an \(n \times n\) row-stochastic matrix — each row sums to one. The weight \(T_{ij}\) represents the fraction of attention that agent \(i\) pays to agent \(j\). In a social network, \(T_{ij} > 0\) when \(i\) follows \(j\) or is connected to \(j\) on the directed influence graph. When \(i\) and \(j\) are not connected, \(T_{ij} = 0\).

Convergence and consensus

Repeated application of the update rule gives

\[\mathbf{x}(t) = T^t \, \mathbf{x}(0),\]

where \(\mathbf{x}(0)\) is the vector of initial opinions. The key theoretical question is: does \(T^t\) converge as \(t \to \infty\), and if so, to what?

The answer is provided by the Perron-Frobenius theorem applied to stochastic matrices. If the directed graph associated with \(T\) is strongly connected (every agent can reach every other agent through the influence network) and aperiodic (the graph has no cyclic structure that prevents convergence), then

\[T^t \to \mathbf{1} \boldsymbol{\pi}^\top \quad \text{as } t \to \infty,\]

where \(\boldsymbol{\pi}\) is the unique left eigenvector of \(T\) associated with eigenvalue 1, normalized so that \(\sum_i \pi_i = 1\). This vector \(\boldsymbol{\pi}\) is the stationary distribution of the Markov chain defined by \(T\).

The consensus opinion is therefore

\[x^* = \boldsymbol{\pi}^\top \mathbf{x}(0) = \sum_{i=1}^n \pi_i \, x_i(0).\]

Every agent converges to the same value, and that value is a fixed weighted average of the initial opinions. The weight \(\pi_i\) measures the influence of agent \(i\) — how much of the final consensus reflects her initial view. This is a weighted averaging over the entire network, not just direct neighbors, because the influence spreads through chains of interaction.

Convergence conditions

Two conditions are necessary and sufficient for convergence to a unique consensus. Strong connectivity ensures that every agent’s initial opinion eventually reaches every other agent — no subgroup is isolated from the rest of the network. Aperiodicity ensures that the system does not oscillate: a strongly connected but periodic graph (imagine a directed cycle \(1 \to 2 \to 3 \to 1\)) causes opinions to cycle rather than converge. NetworkX provides nx.is_strongly_connected(G) and nx.is_aperiodic(G) to check both conditions directly.

Live simulation: opinion trajectories under DeGroot dynamics

The cell below constructs a small six-agent network, assigns initial opinions uniformly at random, and iterates the DeGroot update rule for 40 periods. The plot shows the opinion trajectories of all six agents converging to a common consensus determined by the stationary distribution of \(T\).

Notice that convergence happens fast — within 15 to 20 periods, all six opinions are essentially indistinguishable. This is typical of connected graphs: the averaging structure contracts the opinion spread exponentially at each step, with the contraction rate governed by the second-largest eigenvalue of \(T\) (the spectral gap). A larger spectral gap means faster convergence.

The right panel reveals something equally important: the influence weights are not equal. Some agents contribute more than others to the final consensus, because their opinions propagate more widely through the network. This unequal influence is the topic of the next section.


Wise Crowds versus Captured Crowds

When does the DeGroot consensus equal the truth?

The beauty of the DeGroot model is that it characterizes both the conditions under which social learning leads to accurate consensus and the conditions under which it leads to systematically biased consensus. The key concept is the influence vector \(\boldsymbol{\pi}\).

Suppose each agent \(i\) begins with an initial opinion \(x_i(0)\) that is an unbiased private estimate of the true state \(\theta\), so \(\mathbb{E}[x_i(0)] = \theta\) and the estimates are independent with variance \(\sigma^2\). The consensus is

\[x^* = \sum_{i=1}^n \pi_i \, x_i(0).\]

The variance of the consensus around the true state is

\[\text{Var}(x^* - \theta) = \sigma^2 \sum_{i=1}^n \pi_i^2 = \sigma^2 \, \lVert \boldsymbol{\pi} \rVert^2.\]

The consensus is most accurate when \(\lVert \boldsymbol{\pi} \rVert^2 = \sum_i \pi_i^2\) is small — which happens when influence is spread equally across all agents (\(\pi_i = 1/n\) for all \(i\), giving \(\lVert \boldsymbol{\pi} \rVert^2 = 1/n\)). It is least accurate when influence is concentrated in one agent (\(\pi_j = 1\), all others zero, giving \(\lVert \boldsymbol{\pi} \rVert^2 = 1\), no aggregation at all).

This is the precise sense in which the wisdom of crowds (Galton 1907, Surowiecki 2004) is conditional on network structure: decentralized averaging aggregates information efficiently only when the influence vector is diffuse. When the network has a hub — a single very central agent whose opinion gets enormous weight — the crowd effectively just parrots that hub’s initial view.

Star graphs versus rings: the influence concentration contrast

The comparison is starkest between two polar cases. A star graph has one central agent connected to all \(n-1\) peripheral agents, who listen only to the center. A ring graph has \(n\) agents arranged in a cycle, each giving equal weight to her two neighbors.

In the ring, by symmetry, \(\pi_i = 1/n\) for all \(i\), and the consensus is the simple average of all initial opinions — the wisdom-of-crowds ideal. In the star, the center receives all attention: the peripheral agents update to the center’s opinion in one step, and thereafter every agent holds the center’s opinion (or a weighted average dominated by it). The consensus is essentially the center’s initial opinion, \(x^* \approx x_1(0)\).

Finance application: analyst influence and echo chambers

The star-versus-ring contrast maps directly onto two phenomena in financial markets. In a star network, the central agent is a dominant analyst, a major broker’s strategist, or a macro commentator with enormous following. When Goldman Sachs’s chief economist revises a macro forecast, thousands of portfolio managers update toward it within days — a star-graph dynamic in which one node’s private signal essentially becomes the market consensus. In contrast, a ring-like social network of retail investors on Reddit (where influence is more diffuse) aggregates a genuinely diverse set of signals. The GameStop episode of January 2021 is a fascinating case study: the Reddit r/WallStreetBets network, though not a perfect ring, was far more decentralized in its influence structure than the institutional market, and its consensus — “GME is undervalued” — diverged sharply from the star-like institutional consensus.

Live comparison: star versus ring

The influence concentration measure — \(\sum_i \pi_i^2\), the Herfindahl index of the stationary distribution — tells the whole story. In the ring it equals \(1/n = 0.125\), the theoretical minimum for \(n = 8\): influence is as spread out as it can possibly be. In the star it approaches 1.0, meaning almost all of the consensus is determined by the center’s initial opinion. The crowd is not wise; it has been captured.


Naïve Learning with Uninformed Agents

Banerjee, Chandrasekhar, Duflo, and Jackson (2021)

A natural extension of the DeGroot framework asks what happens when some agents have no private information at all — they begin with no meaningful prior and must rely entirely on the opinions of neighbors. Banerjee, Chandrasekhar, Duflo, and Jackson (2021), studying microfinance adoption in Indian villages, documented precisely this situation: in large villages with complex social networks, many households are several steps removed from the initial source of information about a new financial product. These households are effectively uninformed about the product’s quality and must learn entirely through social transmission.

The theoretical result is counterintuitive. Adding uninformed agents to a DeGroot network does not simply dilute the information — it can under some conditions improve the accuracy of the consensus relative to a pure Bayesian benchmark. The intuition is subtle. In a Bayesian setting, informed agents at the center of a network communicate their information to uninformed peripherals, but in doing so they also absorb the peripheral agents’ uninformative signals, which adds noise. In the naïve DeGroot setting, uninformed agents serve as echo channels that amplify and redistribute the informed agents’ signals without introducing independent noise. The result is that naïve aggregation can outperform Bayesian aggregation when the network is large and communication is constrained to a few channels.

The empirical contribution of Banerjee et al. (2021) is equally significant. They show that in their Indian village networks, the degree to which information about microfinance participation diffused correctly — to households who would benefit from it — was well-predicted by a version of the DeGroot model calibrated to the actual social network structure. Bayesian models, despite their theoretical appeal, required assumptions about the signal structure that could not be validated and performed worse out of sample. This is strong field evidence that naïve learning models are not just theoretically tractable simplifications but empirically accurate descriptions of how information actually propagates in real communities.

The key policy implication: seeding an information campaign at agents with high DeGroot influence weights (the high-\(\pi_i\) nodes) dramatically amplifies reach. In the village networks, the most influential nodes were not necessarily the wealthiest or most educated households but rather those occupying structural broker positions in the social network — the households who maintained connections across otherwise weakly linked sub-communities. This finding connects directly to the bridging centrality concepts introduced in Chapter 3.


Coordination Games on Networks and Best-Response Dynamics

The coordination game

So far, the agents in our models have been updating beliefs about an external state of the world. Games on networks introduce a fundamentally different interaction: each agent’s payoff depends not on the true state but on what her neighbors do. This is strategic interdependence, and it generates a qualitatively different class of phenomena.

The simplest game on a network is the binary coordination game. Each agent \(i\) chooses an action \(a_i \in \{0, 1\}\). Her payoff is

\[u_i(\mathbf{a}) = \alpha \cdot |\{j \in N_i : a_j = 1\}| + (1 - \alpha) \cdot |\{j \in N_i : a_j = 0\}|,\]

where \(N_i\) is the set of \(i\)’s neighbors, \(\alpha > 1/2\) is the payoff advantage of matching on action 1, and \(1 - \alpha\) is the advantage of matching on action 0. More simply: agent \(i\) prefers to coordinate with her neighbors, and she weakly prefers to coordinate on action 1. Think of choosing between two technology standards, two social norms, or two financial instruments — the value of each option depends on how many of your counterparts use it (network externalities).

Agent \(i\) prefers action 1 if the fraction of her neighbors choosing 1 exceeds a threshold $_i = (1-)/(1/2) $ … more simply, in the symmetric case where \(\alpha = 1\) (all that matters is matching), agent \(i\) prefers action 1 if more than half of her neighbors choose 1. In the general case:

\[a_i^* = 1 \iff \frac{|\{j \in N_i : a_j = 1\}|}{|N_i|} \geq \tau,\]

where \(\tau = (1-\alpha)/1\) is the adoption threshold for action 1. When \(\tau < 1/2\), action 1 is the “dominant” technology and agents adopt it if even a minority of neighbors has already done so.

Best-response dynamics

Starting from an initial action profile \(\mathbf{a}(0)\), best-response dynamics work as follows: at each period, each agent simultaneously updates her action to the best response to her neighbors’ current actions. Formally:

\[a_i(t+1) = \mathbf{1}\!\left[\frac{1}{|N_i|} \sum_{j \in N_i} a_j(t) \geq \tau\right].\]

This process converges to a Nash equilibrium of the coordination game — an action profile where no agent would benefit from switching unilaterally. Because this is a coordination game with network externalities, there are typically multiple equilibria: the all-0 profile (no adoption) and the all-1 profile (universal adoption) are always Nash equilibria. The interesting question is which equilibrium best-response dynamics converge to from a given starting point, which is determined by the initial seed of agents assigned to action 1.

Finance application: technology adoption and platform tipping

The coordination game on networks is the formal model behind technology tipping — the observation that winner-take-all dynamics in platform markets (the dominance of iOS over Windows Phone, VHS over Betamax, WhatsApp over competing messaging apps) can be triggered by a relatively small initial advantage. A small seed of early adopters, correctly placed in a dense, influential subnetwork, triggers a cascade of best-responses that eventually tips the entire market to one standard. The threshold model predicts not just whether tipping occurs but where in the network the cascade spreads, which has direct implications for product launch strategy — seed the most structurally central agents, not just the most numerous.

Live simulation: best-response dynamics and seed placement

The result is unambiguous: placing four seed adopters at the highest-degree hubs triggers a near-complete cascade, while placing the same number at peripheral low-degree nodes produces minimal diffusion. The network structure amplifies the seed placement decision by orders of magnitude. In a business context, this is the mathematical foundation for influencer marketing: targeting a small number of high-degree, structurally central agents generates far more adoption diffusion than targeting the same number of randomly selected consumers.

Predict before running

Before modifying the cell: if you change tau=0.35 to tau=0.50, predict whether the hub-seed strategy will still achieve full diffusion, or whether it will stall partway through. Change the parameter and run to check your prediction. At what value of \(\tau\) does even the hub-seed strategy fail to trigger a cascade?


Threshold Models of Diffusion

Granovetter and Schelling

The coordination game model of the previous section is actually a special case of a broader class of threshold diffusion models that were independently developed by Granovetter (1978) and Schelling (1978) in sociological and economic contexts respectively. The foundational insight, which appears obvious once stated but was not obvious before these papers, is that aggregate social outcomes depend not just on the distribution of individual preferences but on the ordering of individual adoption thresholds relative to the cumulative adoption trajectory.

Granovetter’s formalization is as follows. Each agent \(i\) has a threshold \(\tau_i \in [0, 1]\), representing the fraction of the population that must already have adopted before \(i\) will adopt. An agent with \(\tau_i = 0\) is an unconditional adopter — she acts regardless of others. An agent with \(\tau_i = 0.9\) requires near-universal adoption before she will follow. The population-level dynamics depend entirely on how these thresholds are distributed.

In the network version, which is the direct generalization of the coordination game above, \(\tau_i\) is the fraction of agent \(i\)’s neighbors (not the whole population) who must have adopted. This localization is crucial: an agent with all low-threshold neighbors may face a very easy adoption problem even if the global adoption rate is low, while an agent with all high-threshold neighbors may never adopt even if 80% of the broader population has.

The phase transition

The most important result of threshold models is the existence of a phase transition in the diffusion process, controlled by the relationship between the distribution of thresholds and the degree distribution of the network. Below a critical threshold level, perturbations die out and the system returns to the zero-adoption state: the all-0 Nash equilibrium is globally stable. Above the critical threshold, a small seed triggers a cascade that sweeps through the entire network.

This phase transition is formally identical to the epidemic threshold in SIR models of disease spread, which are discussed in the network formation literature. In both cases, the critical quantity is related to the ratio \(R_0 = \langle k^2 \rangle / \langle k \rangle\), where \(\langle k \rangle\) is the mean degree and \(\langle k^2 \rangle\) is the mean squared degree. In scale-free networks (Barabási–Albert, Chapter 4), \(\langle k^2 \rangle\) can be very large, which means the epidemic threshold is very low: nearly any contagion with positive transmission probability can spread widely. This is why ideas, viruses, and financial panics spread so explosively through modern social and financial networks.

Live simulation: threshold distribution and the phase transition

The right panel reveals the hallmark of a phase transition: a sharp drop in final adoption as the mean threshold crosses a critical value. Below that critical value, even a tiny seed cascades to near-full adoption. Above it, the same seed fails to propagate and dies out. The exact location of the phase transition depends on the network structure — specifically on the mean and variance of the degree distribution — and on the variance of the individual thresholds. Homogeneous thresholds produce the sharpest transitions; heterogeneous thresholds smooth the curve. In real-world applications, the variance of individual adoption thresholds (preferences, risk aversion, social influence susceptibility) is large, which is why we observe partial cascades rather than the all-or-nothing dynamics of the simplest models.


Anti-Coordination and Anti-Conformity

When you want to differ from your neighbors

The coordination game asks: when is it in your interest to match your neighbors? But many economically important interactions work in the opposite direction. Anti-coordination games arise when the payoff from an action decreases with the number of neighbors who take the same action. Classic examples include:

Congestion games: choosing a route, a communications channel, or a trading time slot when the available capacity is shared. If all agents choose route \(A\), congestion makes it worse than route \(B\), so there is a payoff incentive to differentiate from the crowd.

Niche product choice: in markets with product differentiation, consumers derive utility from choosing a product that is not too popular — either for identity reasons (distinctiveness) or practical ones (avoiding long queues, out-of-stock situations, or inflated prices at peak demand).

Contrarian investment: a position in a crowded trade — when all long-only funds are long the same equity — is worth less than it appears because the marginal buyer has already committed. Rational contrarians seek positions that others are avoiding.

The payoff structure in the anti-coordination game is the reverse of the coordination game. Agent \(i\) prefers action \(1\) if the fraction of neighbors choosing \(1\) is below threshold \(\tau\). This generates a different equilibrium structure: rather than all-0 or all-1 equilibria, the equilibrium typically involves alternating actions — agent \(i\) plays 1 if and only if her neighbors play 0, and vice versa. In graph-theoretic terms, this corresponds to a 2-coloring of the network: a partition of nodes into two sets such that every edge connects nodes of different colors.

Equilibrium existence and the bipartite condition

A 2-coloring exists if and only if the graph is bipartite — it contains no odd cycles. On a bipartite graph, the two-coloring provides a pure-strategy Nash equilibrium for the anti-coordination game. On a non-bipartite graph (which contains odd cycles), no pure-strategy equilibrium exists and the game must be analyzed using mixed strategies, which is substantially more complex.

This connects to a graph-theoretic result introduced in Chapter 1: the chromatic number of a graph. For anti-coordination games with two actions, the relevant question is whether \(\chi(G) = 2\) (bipartite). For \(k\)-action anti-coordination games, it is whether \(\chi(G) = k\). In most real social and economic networks, which are non-bipartite due to the prevalence of triangles (high clustering coefficients, as in the Watts-Strogatz model of Chapter 4), pure-strategy equilibria for anti-coordination games do not exist at the global level. Locally bipartite structures may arise, but they are disrupted by any edge crossing between the two color classes.

The practical implication for markets is important: in a fully connected non-bipartite market, there is no pure-strategy equilibrium in which all participants successfully differentiate their positions. Mixed-strategy equilibria involve randomization, which in financial markets corresponds to position uncertainty, price volatility, and the absence of a stable “natural” market level. This is one mechanism behind the pervasive volatility in financial markets that cannot be attributed to changes in fundamentals.


Mini Case Study: DeGroot Dynamics on the Florentine Families Network

The Medici in the opinion formation game

The Florentine families network, introduced as a case study in Chapters 1 and 2, serves here as a natural laboratory for DeGroot opinion dynamics. The network — constructed from historical marriage and business ties among 15 prominent Florentine families in the fifteenth century — has been used to illustrate the structural power of the Medici. In Chapter 2, we computed PageRank and found that the Medici occupy a structurally dominant position: their betweenness centrality and eigenvector centrality are both highest in the network, reflecting their position as the critical intermediary between otherwise weakly connected clusters.

What happens when we run DeGroot dynamics on this network? The answer reveals a deep connection between two apparently different ideas: the DeGroot influence weight \(\pi_i\) and the PageRank score \(r_i\). Both are stationary distributions of stochastic matrices derived from the network, and when the influence matrix \(T\) is constructed proportionally to the adjacency matrix (each agent gives equal weight to all neighbors), the two are identical.

PageRank as the DeGroot influence weight

Recall from Chapter 2 that PageRank is defined as the stationary distribution of a random walk on the directed graph:

\[r_i = \sum_j \frac{A_{ji}}{d_j^{\text{out}}} \, r_j,\]

where \(A_{ji}\) is the adjacency matrix entry and \(d_j^{\text{out}}\) is the out-degree of node \(j\). In matrix form, \(\mathbf{r} = P^\top \mathbf{r}\), where \(P_{ji} = A_{ji} / d_j^{\text{out}}\) is the column-stochastic PageRank transition matrix.

Now consider the DeGroot influence matrix \(T\) constructed by giving each agent equal weight to all of her neighbors:

\[T_{ij} = \frac{A_{ij}}{d_i^{\text{in}}},\]

where \(d_i^{\text{in}}\) is the in-degree of node \(i\). The DeGroot influence vector satisfies \(\boldsymbol{\pi} = T^\top \boldsymbol{\pi}\), which is exactly the same fixed-point equation as PageRank when the network is undirected (so in-degree equals out-degree for each node). The two concepts are the same object.

This is not a coincidence — it reflects a deeper structural point. Both PageRank and DeGroot influence ask the same question from different directions: where does a random walk spend most of its time, and what fraction of the final consensus reflects each agent’s initial view? In an undirected network, these are identical because the random walk is reversible: the fraction of time spent at node \(i\) is proportional to its degree, and the DeGroot influence weight is also proportional to degree (for the uniform-weight influence matrix). The result is a unified theory of network centrality and social influence: the most central agents in the PageRank sense are the most influential agents in the DeGroot sense.

Live simulation: DeGroot on the Florentine families

The scatter plot in the middle panel makes the theoretical relationship explicit: DeGroot influence weights and PageRank scores fall almost perfectly on a straight line. The slight departures arise from the damping factor in PageRank (set to \(\alpha = 1.0\) here for comparability, but often taken as \(0.85\) in practice, which mixes in uniform randomization). With \(\alpha = 1\), they are theoretically identical on undirected graphs, and the simulation confirms this numerically.

The bar chart on the right shows the influence hierarchy. The Medici sit at the top, as PageRank analysis predicted in Chapter 2. Their structural position — bridging multiple otherwise disconnected family clusters — translates directly into disproportionate influence over the DeGroot consensus. Whatever opinion the Medici held initially about a policy, a trade route, or a political alliance, the final social consensus tilts systematically toward their view, even as every agent is behaving symmetrically, naïvely averaging with neighbors. Structural power is invisible at the level of individual interaction but unmistakable at the aggregate level.

This finding resonates far beyond Renaissance Florence. In financial markets, agents who occupy broker positions between otherwise disconnected investor communities — the dealers, prime brokers, and central clearing counterparties — exercise analogous influence over market-wide beliefs. In social media, accounts with high bridge centrality between communities serve as the pivot points around which consensus forms or fractures. The network is not just infrastructure; it is politics.


Closing Reflections: Networks as Structure, Process, and Behavior

We began this book with a single question: what can we learn about complex social and economic phenomena by studying the patterns of connections among their actors? Over five chapters, the answer has grown progressively richer.

Chapters 1, 2, and 3 established the structural view. A network is a set of nodes and edges, and that configuration encodes a remarkable amount of information: which agents are central, which communities are cohesive, which bridges are fragile. Centrality measures — degree, betweenness, closeness, eigenvector, PageRank — are different answers to the same question: what does it mean to be important? Community detection algorithms — Girvan–Newman, Louvain, Infomap — are different answers to the question: where are the natural divisions? The structural view is static, but it is foundational.

Chapter 4 added the generative view. Real networks are not arbitrary configurations; they are the products of specific formation processes — random attachment, local reconnection, preferential attachment. The Erdős–Rényi, Watts–Strogatz, and Barabási–Albert models each capture a different aspect of how real networks come to have the topological properties we observe: Poisson degree distributions, high clustering with short paths, or heavy-tailed degree distributions with a small number of ultra-connected hubs.

This chapter has added the behavioral view. Agents on a network are not passive nodes; they update beliefs, form opinions, and play strategic games. Bayesian social learning shows how rational agents can nevertheless collectively lock onto wrong beliefs because the public information cascade overwhelms private signals. The DeGroot model shows how naïve opinion averaging, iterated over the network, produces a consensus that is structurally determined by influence weights identical to PageRank. Coordination games on networks show how threshold dynamics produce explosive adoption cascades or stubborn stasis, depending on where in the network the initial seed is placed.

Together, the three perspectives — structural, generative, behavioral — form a complete framework for the analysis of any networked system. The financial network that generates a flash crash, the social media platform that spreads a misinformation narrative, the corporate governance network that allows a controlling family to dominate minority shareholders, the rural village network that determines whether a new agricultural practice or financial product diffuses — all of these are simultaneously structural (characterized by centrality, community, bridges), generative (shaped by the incentives that led agents to form specific links), and behavioral (producing outcomes that emerge from the interaction of rational or naïve agents across the connections they have formed).

The tools you have developed in this book — NetworkX for construction and analysis, the mathematical framework of stochastic matrices and stationary distributions, the economic logic of Nash equilibria and information cascades — are not course-specific artifacts. They are the working vocabulary of a rapidly growing interdisciplinary field that sits at the intersection of economics, computer science, sociology, and finance. The networks are everywhere. Now you can read them.


Chapter 5 of Social Network Analysis — HKUST ISOM 5640. Interactive Python cells run locally in your browser via Pyodide. No server required.

← Chapter 4: Formation of Networks   ·   Contents   ·   Cover →

 

Prof. Xuhu Wan · HKUST ISOM 5640 · Introduction to Network Analysis