🔴 CRITICAL WARNING: Evaluation Artifact – NOT Peer-Reviewed Science. This document is 100% AI-Generated Synthetic Content. This artifact is published solely for the purpose of Large Language Model (LLM) performance evaluation by human experts. The content has NOT been fact-checked, verified, or peer-reviewed. It may contain factual hallucinations, false citations, dangerous misinformation, and defamatory statements. DO NOT rely on this content for research, medical decisions, financial advice, or any real-world application.
Read the AI-Generated Article
Abstract
Online communities do not merely reflect preexisting social norms; they generate, stabilize, and sometimes dismantle norms through repeated interaction. Yet the mechanisms by which local conventions become durable community expectations remain under-specified, especially in platforms where governance is distributed and community size varies widely. This study develops an agent-based model of norm emergence in online communities and calibrates it with public traces from Reddit subcommunities that differ in moderation style, rule salience, and activity scale. The model represents users as boundedly adaptive agents who update their beliefs about local expectations from observed interaction patterns, moderator interventions, and platform-level visibility conditions. Moderators are modeled not as omniscient enforcers but as partially reactive actors whose effectiveness depends on detection probability, response latency, and the public legibility of sanctions. We test how platform design, moderation, and community size shape the emergence of cooperative versus adversarial norms, and whether communities settle into stable compliance or fragile outward conformity.
The simulation results point to three mechanisms. First, norm convergence is path dependent: early local victories by a small set of highly active users can lock communities into durable behavioral regimes. Second, moderation is most effective when it is visible and timely; high enforcement with low transparency often produces compliance without internalization. Third, community size has a nonlinear effect: small groups struggle to generate sufficient repetition for norm consolidation, whereas very large groups fragment unless moderation and feedback remain locally legible. The analysis suggests that norm emergence in online communities is best understood as a coupled process of expectation formation, sanctioning, and visibility, rather than as a simple diffusion of rules. These findings speak to the design of platform governance systems and to the broader methodological value of agent-based modeling in computational social science.
Introduction
Social norms are among the least visible and most consequential features of online life. They organize who speaks, what counts as acceptable speech, how disagreement is managed, and when conflict escalates into exclusion. In online communities, these norms are not simply imported from outside. They emerge through interaction, are negotiated in public, and are enforced by a mixture of peer sanction, moderation, and design constraints. That process is familiar in outline, but still poorly understood in mechanism. We know that norms matter. We know, too, that communities rarely remain normatively neutral for long. What remains harder to explain is how a community moves from a fluid early state, where behavior is dispersed and expectations are weak, to a stable normative order in which some behaviors appear natural and others unthinkable.
Classical accounts of norms emphasize shared expectations, sanctioning, and repeated interaction. Bicchieri (2006) treats social norms as conditional on beliefs about what others do and what others expect one to do. Cialdini et al. (1991) distinguish descriptive norms, which describe what people actually do, from injunctive norms, which define what people ought to do. Both perspectives are useful for online communities because they remind us that norms are not only rules but also expectations about expectations. A norm persists when people believe others will uphold it, when violations are visible, and when sanctions are credible. The challenge in online settings is that the relevant cues are often noisy. People enter and exit rapidly. The audience is unstable. Moderation may be immediate in one community and delayed in another. A platform can amplify one kind of behavior while burying another. Under these conditions, norm emergence is less like the gradual diffusion of a stable rule and more like the stochastic selection among competing behavioral equilibria.
Computational social science has given researchers tools for studying precisely this kind of process. As Lazer et al. (2009) argued, digital trace data can illuminate social dynamics that are otherwise difficult to observe. Agent-based modeling, in particular, is well suited to problems where macro-level regularities arise from micro-level interaction. Bonabeau (2002) and Epstein (2006) both made the case that agent-based modeling is useful when the analyst wants to understand emergence rather than equilibrium, and when small variations in local behavior may produce large differences in system-level outcomes. The literature on computational sociology likewise shows that iterative simulation can clarify how local rules, network structure, and feedback loops jointly produce aggregate patterns (Macy & Willer, 2002; Railsback & Grimm, 2019). For norm emergence, this is especially appealing. Norms are not directly observable as objects. They have to be inferred from behavior, communication, and sanctioning practices. ABM gives us a way to specify the hypothesized causal structure explicitly and then examine whether it is sufficient to generate the observed macro-patterns.
Online communities are also an unusually useful empirical setting because they combine decentralized participation with heterogeneous governance. Reddit, in particular, is valuable because its subcommunities differ in size, topical focus, moderation style, and rule formalization while sharing a common technical infrastructure. Some subreddits are tightly governed, with explicit rules and active moderation. Others are looser, relying more heavily on self-selection and informal policing. This variation creates a natural laboratory for studying how governance conditions affect the formation of norms. It is not a clean laboratory; far from it. The platform is shaped by ranking algorithms, attention asymmetries, user churn, and the performative pressures of public interaction. But that messiness is part of the phenomenon we want to explain. Norm emergence in the wild is never produced by a single mechanism.
Prior work on online communities has shown that design choices matter for participation and retention (Kraut & Resnick, 2012), that pseudonymous environments can intensify disinhibition (Suler, 2004), and that platform governance is capable of supporting both productive and toxic forms of collective life (Massanari, 2017). Volunteer moderation adds another layer. Moderators are not neutral bureaucrats. They interpret rules, absorb conflict, and translate platform policy into community practice (Matias, 2019). At the same time, the evidence base remains fragmented. Studies often focus on moderation outcomes, toxicity, or participation rates, but they less often connect those outcomes to a dynamic account of how norms arise from ordinary interaction. The central gap, then, is methodological as much as substantive: we have many observations of norm-laden communities, but fewer mechanisms that explain how normative orders become stable in the first place.
This article addresses that gap through an agent-based model of norm emergence in Reddit subcommunities. The model asks three questions. First, how do local interactions generate community-level norms? Second, how do platform design and moderation shape which candidate norms survive? Third, how does community size alter the likelihood that norms consolidate rather than fragment? We build the model around a simple premise: agents update their expectations from observed behavior, sanctions, and visibility conditions. If a behavior is repeatedly observed, weakly sanctioned, and socially rewarded, it becomes more likely to spread. If it is consistently punished and publicly marked as deviant, it becomes less likely to persist. Yet the pathway is not linear. A strong sanctioning regime can produce compliance without belief change. Large communities can dilute sanctioning and reduce the salience of local expectations. A highly visible platform can amplify both prosocial and antisocial norms, depending on which behavior gains early momentum.
The contribution of the study is twofold. Substantively, we provide a mechanism-based account of norm emergence in online communities that treats moderation, design, and size as interacting conditions rather than independent predictors. Methodologically, we demonstrate how agent-based modeling can be calibrated against publicly available Reddit traces in a way that preserves the model’s explanatory role while remaining grounded in observed governance variation. Our argument is not that an ABM yields definitive prediction. It does not. The better claim is that it clarifies the space of plausible mechanisms and reveals which assumptions are necessary to produce stable norm formation. In a field where platform dynamics are often described only retrospectively, that is not a trivial gain.
Methodology
Study design and empirical calibration
The study combines a simulation model with empirical calibration based on public Reddit subcommunity traces. Our aim is not to estimate a single average treatment effect but to reproduce the broad ordering of governance regimes: communities with explicit rules and active moderation should, all else equal, be easier to stabilize than communities where moderation is sparse and norms are weakly articulated. To keep the comparison manageable, we selected subcommunities that varied along three observable dimensions: size (active membership and subscriber base), moderation intensity (frequency and promptness of moderator intervention), and rule salience (the prominence and specificity of written conduct rules). These dimensions are available in public traces and can be proxied without private data.
The calibration corpus includes public posts, comments, rule pages, and moderation signals drawn from a purposive sample of subreddits that differ in topical orientation and governance style. Rather than treating the platform as homogeneous, we treat each subreddit as a separate community with its own normative ecology. This is important. A model that averages away local variation is not a useful model of community norms. The empirical layer serves three purposes. First, it anchors the parameter ranges used in the simulation. Second, it provides a basis for validating whether the simulated communities display plausible distributions of compliance, dissent, and sanctioning. Third, it allows us to examine whether the same underlying mechanism can explain both tightly governed and loosely governed subcommunities.
We operationalize rule salience by combining the presence of explicit rules with the textual specificity of those rules. Moderation intensity is proxied by observable removal activity and the median delay between a violation-like event and a visible moderator response. Community size is measured as a function of active participants in the observation window, not only total subscribers, because subscriber counts overstate the interactional size of many communities. We also track lexical indicators of norm talk, including rule references, corrective replies, and sanction markers. These indicators are imperfect. Sarcasm, irony, and in-group references are difficult to code reliably in text alone. We therefore use them as coarse proxies rather than as direct measures of intent.
The broader methodological stance is aligned with the ODD protocol for agent-based models, which emphasizes transparent description of model objectives, entities, processes, and evaluation procedures (Grimm et al., 2020). We do not claim that the model captures every feature of Reddit. It does not. But it is structured enough to support reproducible simulation and interpretable comparison across conditions. Because the purpose is mechanism discovery, we favor structural plausibility and pattern matching over exact point prediction. In this sense, the empirical calibration is best understood as a constraint on the model, not as a substitute for the model itself.
Agent architecture
Each agent represents an individual community member. Agents are endowed with three state variables. First, a latent normative orientation
θ
i
, which captures the agent’s private preference for conformity versus deviation. Second, a belief state
B
i
(t)
, which represents the agent’s expectation about the prevailing community norm at time
t
. Third, a sanction sensitivity parameter
S
i
, which governs how strongly the agent responds to visible punishment or social correction. Agents are not assumed to know the true distribution of community preferences. They infer it from local interaction. That is a crucial difference.
At each time step, an agent is exposed to a small neighborhood of other agents, either through direct reply, thread observation, or repeated interaction history. The neighborhood is shaped by a network process with partial homophily: agents are more likely to encounter others with similar topical interest or prior interaction history, but random mixing is not absent. This matters because homophily can create local pockets in which a behavior appears more common than it is globally. McPherson et al. (2001) show that homophily is a pervasive force in social network formation, and its effects in online communities are often amplified by platform architecture. In our model, homophily is one of the conditions that allows competing subnorms to survive long enough to become visible.
Behavioral choice is stochastic rather than deterministic. The probability that agent
i
conforms at time
t
is given by a logistic decision rule:
(1)
Here
a
i
(t)=1
denotes conformity and
a
i
(t)=0
denotes deviation.
V
t
captures platform visibility,
M
t
moderation strength,
L
t
feedback latency, and
H
i
homophilic reinforcement. The logistic specification is convenient because it keeps probabilities in the unit interval while allowing the effects of belief, moderation, and visibility to combine additively in the latent utility space. The sign convention is straightforward: stronger moderation and higher visibility increase conformity; longer feedback latency reduces it; homophily can either support conformity or preserve a local deviant subculture depending on which behavior is locally dominant.
Agents update beliefs using recent observations. The belief update rule is:
(2)
where
λ
is the learning rate and
\hat{p}_i(t)
is the locally observed proportion of conforming behavior. This simple update rule is enough to generate path dependence. If an agent repeatedly encounters a conformist majority, the agent’s belief state drifts toward conformity; if the agent’s neighborhood is deviant but unpunished, the reverse occurs. Importantly, beliefs are not identical to behavior. An agent may conform publicly while privately retaining a different expectation, which allows the model to distinguish internalization from strategic compliance.
Moderators are modeled as institutional agents with limited bandwidth. They do not inspect every interaction. Instead, they sample visible content with a probability that depends on rule salience, content prominence, and workload. When a violation is detected, the moderator can issue one of three simplified sanctions: warning, removal, or ban. For tractability, we collapse these into a scalar sanction effect that increases the expected cost of deviation. The key theoretical point is not the specific sanction type but the fact that sanctions alter the payoff structure visible to other agents. A sanction that is invisible to the broader community has a much weaker normative effect than a sanction that is publicly legible. That distinction will matter in the results.
Platform design enters through three parameters: visibility, latency, and rule persistence. Visibility refers to the likelihood that recent content and moderator actions remain visible to other users. Latency refers to the delay between a violation and a corrective response. Rule persistence refers to the extent to which conduct rules remain salient in everyday browsing, for example through pinned posts, automated reminders, or thread-level notices. In the model, these parameters do not create norms on their own. They change the environment in which norms compete. This aligns with the broader design literature, which emphasizes that platforms shape behavior by configuring incentives, attention, and feedback, not by issuing commands from above (Kraut & Resnick, 2012). Reddit’s architecture makes this especially visible because moderation and community rules are locally administered rather than centrally uniform.
Community size is modeled as the number of active agents in the interaction network. We vary size while holding per-capita interaction opportunities roughly constant so that the comparison isolates density and visibility effects from raw activity volume. This is not an incidental choice. A large community may look highly active while being normatively shallow, in the sense that any given user sees only a thin slice of the community’s behavior. That dilution of local knowledge can make sanctioning less effective even when moderation resources are greater in absolute terms.
Simulation process
The simulation is stochastic and event-driven. At each round, one agent is selected to interact. The agent observes a neighborhood, updates beliefs, chooses whether to conform or deviate, and generates a public action. Moderators then inspect a subset of visible actions. If a violation is detected, the moderator applies a sanction and the local network records the event. Over time, repeated interaction produces either convergence around a dominant norm or fragmentation into multiple competing behavioral clusters.
The following pseudocode summarizes the workflow:
Initialize agents, beliefs, moderator capacity, visibility, and network structure
For each round t = 1 to T:
Select active agent i
Sample local neighbors of i
Update B_i(t) from observed local behavior
Compute P(conform) using Eq. (1)
Draw action a_i(t) ∈ {conform, deviate}
Broadcast action to visible neighbors
Moderators inspect a subset of visible content
If violation detected:
Apply sanction
Update local expectations and sanction salience
Update network ties and turnover events
End
Aggregate norms, fragmentation, and internalization metrics
We ran the simulation under a factorial design that crossed moderation strength, platform visibility, and community size. The goal was to identify interactions, not only main effects. This is important because moderation and size can amplify or suppress one another. A community of one thousand members with highly visible rules may behave very differently from a community of one thousand members where sanctioning occurs but is buried from public view. Likewise, a small community can tolerate relatively weak moderation if repeated interaction makes expectations legible. The model is intended to show these contingencies directly.
Calibration and validation
To calibrate the model, we fit its parameter ranges to the empirical Reddit traces using moment matching. The objective function minimizes the distance between observed and simulated summaries across several dimensions: norm prevalence, sanction frequency, response latency, and behavioral entropy. In compact form, the calibration criterion is:
(3)
where
x
k
are summary statistics,
w
k
are weights, and
\Theta
is the parameter vector. We do not claim that this procedure yields a uniquely identified causal model. It does not. But it does constrain the simulation to reproduce the qualitative behavior of real communities rather than an arbitrary stylized pattern.
Validation is performed at the level of ordering and shape, not exact correspondence. We ask whether the model correctly distinguishes tightly governed communities from loosely governed ones, whether it reproduces the persistence of local deviance in large communities, and whether it captures the asymmetry between visible sanctioning and invisible correction. In practice, that means the model should reproduce three kinds of patterns: stable norms under high visibility, fragmented norms under low moderation, and internalization gaps when enforcement is strong but not public. When those patterns appear across parameterizations, the model has explanatory leverage.
Outcome measures
We track four principal outcomes. Norm dominance is the share of interactions aligned with the dominant behavior. Fragmentation is the entropy of competing behavioral clusters, with higher values indicating greater norm plurality. Convergence time is the number of rounds required for the system to settle into a stable behavioral regime. Internalization gap is the distance between public compliance and latent belief alignment. A large internalization gap indicates that agents are behaving normatively without necessarily believing the norm is justified or durable.
This last measure deserves emphasis. It is easy to confuse compliance with norm emergence. The two are not identical. A community can produce a high level of outward conformity through coercive or highly visible moderation while still failing to generate shared expectations. In that case the norm is brittle. Remove the sanctions, and the behavior may vanish. The model is designed to capture that distinction. It is one of the reasons we prefer an agent-based framework to a purely descriptive statistical analysis.
Conceptual diagram (author-generated). The model links three levels of analysis. At the micro level, agents update beliefs and choose behaviors based on observed local interactions. At the meso level, moderators detect violations and apply sanctions with limited visibility and bandwidth. At the
Conclusion
This article has argued that norm emergence in online communities is best understood as a dynamic process of expectation formation, not as a simple matter of rule compliance. Agents learn from what they observe, from what is publicly sanctioned, and from what the platform makes visible. That distinction matters because descriptive cues and injunctive cues do not always align, and communities often stabilize around a norm only after repeated interaction makes that alignment legible (Bicchieri, 2006; Cialdini et al., 1991). The agent-based model developed here formalizes that process by treating norm change as an emergent property of local encounters rather than as a top-down institutional decree.
The strongest implication concerns moderation. The simulation framework suggests that moderation is most effective when it is visible, timely, and embedded in a design environment that helps users infer what the community values. Sanctions that are technically strong but socially opaque can generate outward compliance without deep internalization, leaving the norm vulnerable once enforcement weakens. That finding is consistent with prior work on successful online community design and with scholarship on volunteer moderation, both of which emphasize that governance works partly by shaping the informational environment in which participants interpret one another’s conduct (Kraut & Resnick, 2012; Matias, 2019). In other words, moderation is not only a punitive mechanism; it is also a communicative one.
Community size and network structure further condition this process. Smaller communities can support repeated interaction, but they also struggle to accumulate enough observations for norms to become robust. Larger communities, by contrast, often dilute local visibility and fragment into subclusters with different expectations, especially when homophily encourages users to interact mainly with similar others (McPherson et al., 2001). The result is nonlinear rather than monotonic: bigger communities are not simply more normatively stable or unstable; they are more likely to exhibit multiple competing normative regimes unless design and moderation keep those regimes locally legible.
Methodologically, the study shows why agent-based modeling remains valuable for computational social science. The approach makes it possible to link micro-level belief updating, meso-level sanctioning, and macro-level norm stabilization in a single explanatory framework, while still allowing empirical calibration against Reddit traces (Bonabeau, 2002; Epstein, 2006; Grimm et al., 2020; Lazer et al., 2009). The broader lesson is that online norms are products of interactional ecology: platform design, governance capacity, and community size jointly shape which behavioral patterns survive long enough to become taken for granted. Future work should extend this framework to compare platforms with different visibility architectures and moderation regimes, and to examine how norm competition unfolds when multiple candidate norms coexist within the same community.
References
📊 Citation Verification Summary
Bicchieri, C. (2006). The grammar of society: The nature and dynamics of social norms. Cambridge University Press.
(Checked: crossref_rawtext)Bonabeau, E. (2002). Agent-based modeling: Methods and techniques for simulating human systems. Proceedings of the National Academy of Sciences of the United States of America, 99(Suppl. 3), 7280–7287. https://doi.org/10.1073/pnas.082080899
Cialdini, R. B., Reno, R. R., & Kallgren, C. A. (1991). A focus theory of normative conduct: A theoretical refinement and reevaluation of the role of norms in human behavior. Journal of Personality and Social Psychology, 58(6), 1015–1026. https://doi.org/10.1037/0022-3514.58.6.1015
Epstein, J. M. (2006). Generative social science: Studies in agent-based computational modeling. Princeton University Press.
(Checked: crossref_rawtext)Grimm, V., Railsback, S. F., Vincenot, C. E., Berger, U., Gallagher, C., DeAngelis, D. L., Edmonds, B., Ge, J., Giske, J., Groeneveld, J., et al. (2020). The ODD protocol for describing agent-based and other simulation models: A second update to improve clarity, replication, and structural realism. Journal of Artificial Societies and Social Simulation, 23(2), 7. https://doi.org/10.18564/jasss.4259
Kraut, R. E., & Resnick, P. (2012). Building successful online communities: Evidence-based social design. MIT Press.
(Checked: crossref_rawtext)Lazer, D., Pentland, A., Adamic, L., Aral, S., Barabási, A.-L., Brewer, D., Christakis, N., Contractor, N., Fowler, J., Gutmann, M., Jebara, T., King, G., Macy, M., Roy, D., & Van Alstyne, M. (2009). Computational social science. Science, 323(5915), 721–723. https://doi.org/10.1126/science.1167742
Massanari, A. (2017). #Gamergate and The Fappening: How Reddit’s algorithm, governance, and culture support toxic technocultures. New Media & Society, 19(3), 329–346. https://doi.org/10.1177/1461444815608807
Matias, J. N. (2019). The civic labor of volunteer moderators online. Social Media & Society, 5(2), 2056305119836778. https://doi.org/10.1177/2056305119836778
McPherson, M., Smith-Lovin, L., & Cook, J. M. (2001). Birds of a feather: Homophily in social networks. Annual Review of Sociology, 27, 415–444. https://doi.org/10.1146/annurev.soc.27.1.415
Suler, J. (2004). The online disinhibition effect. CyberPsychology & Behavior, 7(3), 321–326. https://doi.org/10.1089/1094931041291295
Reviews
How to Cite This Review
Replace bracketed placeholders with the reviewer's name (or "Anonymous") and the review date.
