Ethics in Complex Systems Explained
In simple systems, ethics is straightforward: you act, consequences follow, responsibility is clear. In complex systems—where actions ripple through feedback loops, interact with other forces, and produce emergent outcomes no one intended—ethical reasoning becomes radically harder. Good intentions routinely produce harm. Responsibility diffuses. Prediction fails.
This is the ethics of algorithms, markets, organizations, cities, and ecosystems. Understanding how to think ethically in complex systems is increasingly urgent.
Table of Contents
- What Makes Systems Complex
- Why Traditional Ethics Struggles
- Emergent Harm: When No One Intends the Outcome
- Second-Order and Nth-Order Effects
- The Problem of Distributed Responsibility
- Ethical Principles for Complex Systems
- System Design as Ethical Practice
- Feedback Loops and Moral Learning
- Case Studies of Systemic Ethical Failures
- Practical Decision-Making in Complexity
- References
What Makes Systems Complex
Not all systems are complex in the technical sense. Understanding the difference matters for ethical reasoning.
| System Type | Characteristics | Ethical Implications |
|---|---|---|
| Simple | Few components, linear causality, predictable | Standard ethical reasoning works: act → consequence → responsibility |
| Complicated | Many parts, but knowable relationships, predictable with expertise | Ethical analysis requires expertise but remains tractable |
| Complex | Many interacting agents, feedback loops, emergent behavior, unpredictable | Ethical outcomes emerge from interactions; prediction often fails; responsibility diffuses |
| Chaotic | Sensitive to initial conditions, no stable patterns | Ethical responsibility for setup/bounds, not specific outcomes |
Defining Features of Complex Systems
1. Emergence
Macro-level patterns arise from micro-level interactions that no single agent intends or controls. The "behavior" of the system is not reducible to the intentions of its parts.
Example: Traffic jams emerge from individual driving decisions. No one intends the jam; it arises from interactions.
2. Feedback Loops
Actions create effects that feed back to change conditions for future actions, creating reinforcing or balancing dynamics.
Example: Social media algorithms amplify engagement, which trains algorithms to amplify more, which changes user behavior, which feeds back into training data.
3. Nonlinearity
Small changes can have large effects; large changes can have small effects. Proportionality breaks down.
Example: A single subprime mortgage default is trivial. Millions of correlated defaults collapse the financial system.
4. Adaptation
Agents in the system learn and change behavior in response to the system, producing outcomes the designer never anticipated.
Example: Students "teach to the test" when assessments become high-stakes, undermining the goal of measuring learning.
5. Path Dependence
History matters. Systems can get locked into suboptimal states because early choices constrain future options.
Example: QWERTY keyboard layout persists despite more efficient alternatives because switching costs are prohibitive.
Why Traditional Ethics Struggles
Most ethical frameworks were developed for simple or complicated contexts—direct actions with foreseeable consequences and clear agents.
Ethical Framework Assumptions vs. Complex Reality
| Traditional Assumption | Reality in Complex Systems |
|---|---|
| Clear causality: Action A causes outcome B | Outcomes emerge from interactions; causality is distributed and nonlinear |
| Predictable consequences: Foresee what will happen | Emergent effects are often unpredictable |
| Identifiable agents: Know who acted | Many agents contribute; no single cause |
| Proportional impact: Small actions = small effects | Nonlinearity means tiny changes can cascade |
| Static environment: Context stays constant | Systems adapt and evolve in response |
| Reversibility: Can undo harm | Path dependence means some harms lock in |
Why Consequentialism Fails
Consequentialism judges actions by their outcomes. But in complex systems:
- Outcomes emerge over time and are often invisible at decision time
- Unintended consequences can swamp intended ones
- Attribution is ambiguous—which action caused which outcome?
Example: A well-meaning microfinance program lifts some families out of poverty but creates debt traps for others and distorts local labor markets. Is it ethical? The answer depends on which outcomes you weight and over what time horizon.
Why Deontology Fails
Deontology judges actions by adherence to rules or duties. But in complex systems:
- Rules designed for simple contexts produce perverse outcomes
- Duties conflict (duty to innovate vs. duty to avoid harm)
- "Follow the rules" becomes an excuse when the system produces harm
Example: A pharmaceutical company follows all FDA regulations but designs pricing and distribution systems that make life-saving drugs inaccessible to the poor. Rule-following doesn't equal ethical.
Why Virtue Ethics Fails
Virtue ethics judges character—are you acting with wisdom, courage, temperance, justice? But in complex systems:
- Virtuous individuals can collectively produce terrible outcomes
- Systemic harm emerges from structural features, not character flaws
- Personal virtue doesn't prevent algorithm-driven discrimination or market-driven inequality
Example: Honest, hardworking bankers following prudent local lending standards collectively fueled a subprime bubble that devastated the economy.
Implication: We need ethical frameworks adapted to complexity.
Emergent Harm: When No One Intends the Outcome
The most disturbing feature of complex systems: harm can emerge from the interactions of well-intentioned agents, none of whom intends or even sees the harm.
The Mechanism
- Individual agents act rationally (from their perspective)
- Actions interact with others' actions through the system
- Emergent patterns form at the system level
- Harmful outcomes occur that no one designed or desired
Classic Examples
| System | Individual Actions | Emergent Harm |
|---|---|---|
| Tragedy of the Commons | Farmers rationally graze more animals | Pasture collapses; all lose |
| Bank runs | Depositors rationally withdraw savings | Bank fails; depositors lose anyway |
| Arms races | Nations rationally build defenses | Everyone less secure, resources wasted |
| Social media filter bubbles | Users click what interests them; algorithms optimize engagement | Polarization, misinformation, radicalization |
| Urban sprawl | Families choose affordable housing farther out | Traffic, pollution, infrastructure strain, inequality |
The Ethical Problem
Who is responsible?
- Not individual users—they're acting reasonably given their constraints
- Not system designers—they often didn't foresee the emergent harm
- Not operators—they're following rules or incentives
Traditional responsibility frameworks break down. You can't simply blame "the system"—systems don't have agency. But you also can't pin it on individuals without ignoring systemic causality.
Second-Order and Nth-Order Effects
In complex systems, first-order effects (immediate, direct outcomes) are often swamped by second-order effects (consequences of consequences) and nth-order effects (cascading downstream impacts).
Order Effects Explained
| Order | Definition | Visibility | Ethical Weight |
|---|---|---|---|
| First-order | Direct, immediate consequence of action | High—obvious at decision time | Often emphasized but can mislead |
| Second-order | Consequences of the first-order consequence | Medium—requires thinking ahead | Often dominates long-term impact |
| Third-order+ | Further cascading effects | Low—hard to foresee | Can dwarf earlier effects; often ignored |
Example: Introducing Antibiotics
| Order | Effect |
|---|---|
| First-order (positive) | Saves millions of lives from bacterial infections; revolutionizes medicine |
| Second-order (mixed) | Overuse leads to antibiotic resistance; some bacteria become untreatable |
| Third-order (negative) | Resistant bacteria spread globally; "superbugs" kill hundreds of thousands annually |
| Fourth-order (systemic) | Evolutionary arms race; future pandemics may lack effective treatments; medical procedures become riskier |
Ethical question: Is it ethical to widely deploy antibiotics knowing resistance is inevitable? The first-order benefit is enormous. The nth-order harm could be catastrophic. How do you weigh them?
Example: Social Media Platforms
| Order | Effect |
|---|---|
| First-order (positive) | Connects people; democratizes information; enables movements like Arab Spring |
| Second-order (mixed) | Algorithms optimize engagement; viral content spreads faster than truth; filter bubbles form |
| Third-order (negative) | Polarization deepens; misinformation undermines institutions; democracies destabilize; mental health declines |
| Fourth-order (systemic) | Trust in media, science, government collapses; social cohesion erodes; authoritarian manipulation scales |
Ethical question: Are platform designers responsible for third- and fourth-order harms they didn't intend but were foreseeable? What duty do they have to model downstream effects?
The Foreseeability Problem
When are you responsible for unintended consequences?
Traditional ethics often says: "You're responsible for foreseeable harms."
But in complex systems:
- Some harms are foreseeable in principle (someone could model them) but not foreseen (no one did)
- Some harms are predictable in type (something will go wrong) but unpredictable in specifics (exactly what or when)
- Some harms are unforeseeable because they depend on interactions with future systems that don't exist yet
Emerging standard: You are responsible for harms that are foreseeable given reasonable effort to model the system. This creates a duty to think in systems terms.
The Problem of Distributed Responsibility
In complex systems, causality is distributed across many agents. This creates profound challenges for moral responsibility.
The Responsibility Dilution Problem
| Scenario | Traditional Ethics | Complex Systems Reality |
|---|---|---|
| One actor, one victim | Clear responsibility | N/A |
| One actor, many victims | Clear responsibility; actor bears full moral weight | Possible in complex systems (e.g., designer of harmful algorithm) |
| Many actors, one victim | Shared responsibility; each bears partial weight | Typical in complex systems; emergent harm |
| Many actors, many victims | Diffuse responsibility; unclear attribution | Dominant in complex systems; systemic harm |
Mechanisms of Responsibility Diffusion
1. Many Hands Problem
When many people contribute to an outcome, each can claim "I only did a small part; I'm not responsible for the whole."
Example: Financial crisis—mortgage brokers, lenders, rating agencies, regulators, investors all contributed. Each claims they're not responsible for the system-wide collapse.
2. Ignorance Defense
"I didn't know this would happen." In complex systems, genuine ignorance is common. But is ignorance an excuse?
Emerging standard: Willful ignorance or negligent failure to investigate foreseeable harms is not exculpatory.
3. Following Orders/Rules
"I was just following instructions/the law." Compliance becomes an excuse even when the system produces harm.
Ethical response: Following rules is necessary but not sufficient. You bear some responsibility for the systemic outcomes your actions enable.
4. Attribution Ambiguity
Causality is so distributed that pinning harm on specific decisions is nearly impossible.
Example: Climate change—millions of actors over decades. Who is responsible for a specific hurricane?
Toward Systemic Responsibility
New framework needed:
- Designers are responsible for foreseeable emergent harms and for building in monitoring/feedback
- Operators are responsible for detecting and responding to unintended harms
- Governors (regulators, boards) are responsible for oversight and forcing externalities to be internalized
- Participants bear responsibility proportional to their power and knowledge
Key principle: Responsibility scales with power to shape the system and knowledge of systemic effects.
Ethical Principles for Complex Systems
Given the failure of traditional frameworks, what principles can guide ethical action in complexity?
Principle 1: Precautionary Principle
Statement: When an action has uncertain but potentially catastrophic systemic effects, err on the side of caution.
Rationale: In complex systems, failures can cascade and lock in. Irreversibility demands caution.
Application:
- Don't deploy technologies with existential risk (e.g., AGI, synthetic biology) without extreme safeguards
- In ecosystems, avoid interventions that could trigger tipping points
Critique: Can stifle innovation if applied too broadly. Requires judgment about what counts as "catastrophic."
Principle 2: Transparency and Legibility
Statement: Make system behavior visible to those affected. Complex systems should be as legible as possible.
Rationale: Opacity prevents accountability and learning. Affected parties can't consent to what they can't see.
Application:
- Algorithm explainability requirements
- Public reporting of systemic risks
- Open data on system performance
Critique: Some complexity resists simplification. Full transparency can be exploited (Goodhart's Law).
Principle 3: Reversibility and Experimentation
Statement: Prefer interventions that can be undone. Test changes incrementally before scaling.
Rationale: In unpredictable systems, you learn by doing. Reversibility limits downside.
Application:
- Pilot programs before national rollout
- "Circuit breakers" that halt systems when anomalies arise
- Sunset clauses for policies
Critique: Some changes (e.g., infrastructure, cultural shifts) are inherently irreversible.
Principle 4: Monitoring and Feedback Obligations
Statement: Those who design or operate systems have a duty to monitor for unintended harms and respond.
Rationale: You can't foresee all harms ex ante. Ethical responsibility includes learning and correcting.
Application:
- Real-time monitoring dashboards
- Whistleblower protections
- Mandatory incident disclosure
Critique: Monitoring is costly. Who pays? How intrusive can it be?
Principle 5: Internalizing Externalities
Statement: Actors should bear the costs of harms their actions impose on others through the system.
Rationale: Externalities (costs borne by others) drive systemic harm. Internalization aligns incentives with ethics.
Application:
- Carbon taxes for emissions
- Liability for algorithmic discrimination
- Pollution penalties
Critique: Quantifying and assigning externalities is technically and politically difficult.
Principle 6: Diversity and Redundancy
Statement: Systems should have diverse components and redundant pathways to prevent failure cascades.
Rationale: Monocultures are brittle. Diversity buffers against systemic shocks.
Application:
- Multiple suppliers in supply chains
- Competing platforms (avoid monopoly)
- Cognitive diversity in decision-making teams
Critique: Diversity and redundancy are costly; trade off against efficiency.
System Design as Ethical Practice
If ethics in complex systems is hard, prevention is better than correction. Designing systems with ethical outcomes in mind is a form of moral action.
Design Principles for Ethical Systems
| Principle | Mechanism | Example |
|---|---|---|
| Alignment of Incentives | Structure rewards so individual rationality produces collective good | Markets with proper regulation; well-designed taxes/subsidies |
| Default to Ethical Outcomes | Make the easiest path the ethical one | Opt-out organ donation; privacy-by-default settings |
| Circuit Breakers | Automatic halts when system behaves anomalously | Stock market trading halts; medication dose limits |
| Feedback Loops | Real-time information on systemic effects | Emissions dashboards; diversity metrics; user harm reports |
| Modularity | Isolate failures so they don't cascade | Financial firewalls; modular code architecture |
| Escape Valves | Allow agents to exit or override when system fails | Emergency stops; human-in-the-loop for high-stakes decisions |
| Red Teams | Dedicated adversarial testing for failure modes | Security penetration testing; bias audits |
Case: Ethical Algorithm Design
Problem: Recommendation algorithms optimize engagement, which can amplify misinformation and extremism.
Ethical design interventions:
- Objective redesign: Optimize for long-term satisfaction or diverse exposure, not just clicks
- Friction for harmful content: Slow the spread of likely misinformation (e.g., prompt to read before sharing)
- Transparency: Show users why content was recommended
- User control: Let users adjust algorithmic parameters
- Monitoring: Track radicalization pathways and intervene
- Auditing: External review for bias and harm
Trade-off: Ethical design often reduces engagement (and revenue). This is why regulation may be necessary—market incentives alone won't produce ethical systems.
Feedback Loops and Moral Learning
In complex systems, you can't predict all outcomes ex ante. Ethical practice requires learning from feedback.
The Moral Learning Cycle
- Act (design system, launch policy, deploy technology)
- Monitor (measure outcomes, including unintended ones)
- Analyze (identify second-order effects, emergent harms, systemic failures)
- Adjust (redesign, intervene, correct)
- Iterate (repeat)
Feedback Loop Types and Ethical Implications
| Loop Type | Mechanism | Ethical Implication |
|---|---|---|
| Reinforcing | Success breeds more success (or failure breeds more failure) | Can amplify harms; creates winner-take-all dynamics |
| Balancing | System self-corrects toward equilibrium | Stabilizes; prevents runaway harm |
| Delayed | Effects take time to appear | Ethical harms invisible until it's too late to reverse |
Example: Social Media Echo Chambers (Reinforcing Loop)
- User engages with partisan content
- Algorithm learns user prefers this content
- Algorithm shows more partisan content
- User becomes more partisan
- User engages even more with partisan content
- Loop accelerates → radicalization
Ethical intervention: Break the loop—introduce balancing mechanisms (diverse content injection, engagement limits, friction).
The Challenge of Delayed Feedback
Many systemic harms have long time lags between action and consequence:
- Climate change: Emissions today cause harm decades later
- Antibiotic resistance: Overuse today creates untreatable infections years later
- Debt: Borrowing today constrains choices for decades
- Education policy: Effects on students appear years after implementation
Ethical difficulty: Humans discount future harms. Systems that create delayed harm are ethically problematic even if no one intended harm.
Implication: We need anticipatory governance—institutions that force consideration of long-term systemic effects.
Case Studies of Systemic Ethical Failures
Case Study 1: The 2008 Financial Crisis
System: Global financial markets with complex derivatives, securitization, and leverage.
Individual Actions (seemingly rational):
- Homebuyers took out mortgages they could afford under initial terms
- Mortgage brokers earned commissions selling loans
- Banks securitized mortgages to spread risk
- Rating agencies gave high ratings (paid by issuers)
- Investors bought AAA-rated securities
- Regulators relied on market discipline
Emergent Outcome: Correlated defaults triggered cascade; financial system collapsed; millions lost homes, jobs, savings.
Ethical Lessons:
- Distributed responsibility: No single villain, but system-wide failure
- Misaligned incentives: Short-term gains privatized; long-term losses socialized
- Opacity: Complexity obscured risk; even experts didn't understand system
- Externalities: Harm borne by those (homeowners, taxpayers) not party to transactions
- Lack of circuit breakers: No mechanisms to halt cascade once started
Who was responsible?
- Designers of derivatives? (Enabled complexity)
- Regulators? (Failed to govern)
- Rating agencies? (Misjudged risk)
- Bankers? (Pursued profit without considering systemic risk)
- Homebuyers? (Took on debt)
Answer: All share partial responsibility scaled to their power and knowledge.
Case Study 2: Facebook and the Rohingya Genocide
System: Facebook platform used by millions in Myanmar, where it's the primary internet for many.
Individual Actions:
- Burmese military and extremists posted anti-Rohingya propaganda
- Users shared inflammatory content
- Facebook's algorithm amplified viral, engaging content
- Moderators (understaffed, non-Burmese-speaking) missed hate speech
Emergent Outcome: Hate speech radicalized population; contributed to genocide of Rohingya Muslims.
Ethical Lessons:
- Second-order harm: Platform designed for connection enabled mass violence
- Foreseeable but not foreseen: Experts warned of risks; Facebook didn't act
- Algorithmic amplification: Engagement optimization spreads hate faster than counterspeech
- Responsibility of designers: Facebook had power to intervene; chose not to prioritize it
- Global vs. local: System designed for U.S. context failed catastrophically in Myanmar
Who was responsible?
- Military and extremists? (Direct perpetrators)
- Facebook? (Provided amplification mechanism; failed to moderate)
- Users? (Shared hate speech)
Consensus: Facebook bears significant moral responsibility for enabling and amplifying hate, even if it didn't intend genocide. Power to shape the system creates responsibility for its outcomes.
Case Study 3: The Opioid Epidemic
System: Pharmaceutical companies, doctors, patients, regulators, insurance, cultural attitudes toward pain.
Individual Actions:
- Purdue Pharma marketed OxyContin as low-risk for addiction (false claims)
- Doctors prescribed opioids to manage pain (often appropriate)
- Patients took prescribed medication (trusting doctors)
- Insurance reimbursed pills but not alternative treatments
- Regulators approved drugs based on company-funded research
Emergent Outcome: Millions addicted; hundreds of thousands dead from overdoses.
Ethical Lessons:
- Incentive corruption: Pharma profits from volume; doctors rewarded for prescribing; no one incentivized to prevent addiction
- Information asymmetry: Companies knew addiction risk; patients and doctors didn't
- Regulatory capture: Purdue influenced FDA and medical establishment
- Systemic lock-in: Once addiction took hold, people couldn't escape; treatment infrastructure inadequate
- Externalized harm: Pharma profited; patients and society bore costs
Who was responsible?
- Purdue Pharma? (Knowingly misled)
- Doctors? (Overprescribed)
- Regulators? (Failed oversight)
- Patients? (Some accountability, but power imbalance and addiction undermine agency)
Outcome: Courts held Purdue accountable (billions in fines, bankruptcy); individual executives prosecuted. Recognition that systemic harm requires systemic accountability.
Practical Decision-Making in Complexity
Given the ethical challenges of complex systems, how should individuals and organizations make decisions?
Decision Framework for Complex Systems
| Step | Action | Purpose |
|---|---|---|
| 1. Map the system | Identify key agents, feedback loops, external dependencies | Understand structure |
| 2. Model second-order effects | Ask "What happens next? And then what?" | Surface unintended consequences |
| 3. Pre-mortem analysis | Assume failure; work backward to causes | Identify failure modes |
| 4. Identify externalities | Who bears costs who isn't party to decision? | Ensure fairness |
| 5. Assess reversibility | Can this be undone if it goes wrong? | Limit downside |
| 6. Build in monitoring | How will we detect unintended harm? | Enable learning |
| 7. Stress test | What happens under extreme conditions? | Prevent catastrophic failure |
| 8. Consult affected parties | Who will live with the consequences? | Respect autonomy and justice |
| 9. Iterate and adjust | Plan to learn and correct | Commit to feedback |
Red Flags: When to Be Extra Cautious
| Warning Sign | Why It Matters |
|---|---|
| Irreversibility | Can't undo harm if wrong |
| Opacity | Can't see what's happening; accountability impossible |
| Winner-take-all dynamics | Reinforcing loops create extreme inequality |
| Long time lags | Harm appears too late to correct |
| Catastrophic downside | Low-probability, high-impact failures |
| Vulnerable populations affected | Power imbalance means they can't defend themselves |
| Novelty | No historical precedent; outcomes truly unpredictable |
Questions to Ask
Before Acting:
- What are the second- and third-order effects of this decision?
- Who benefits? Who bears the costs?
- What could go wrong? How would we know?
- Can we reverse this if it fails?
- Are we monitoring for unintended harms?
- What feedback loops does this create?
While Operating:
- Are outcomes matching expectations?
- What unintended consequences are emerging?
- Are there early warning signs of systemic failure?
- Who is being harmed that we didn't anticipate?
After Failure:
- What systemic features enabled this harm?
- Who had the power to prevent it?
- How do we redesign to prevent recurrence?
- What did we fail to monitor?
Conclusion
Ethics in complex systems is ethics in a world where:
- Good intentions routinely produce harm
- Outcomes emerge from interactions no one controls
- Responsibility is distributed and ambiguous
- Prediction is unreliable, and learning must be continuous
Traditional ethical frameworks—consequentialism, deontology, virtue ethics—were built for simpler worlds. They struggle in complexity.
What's needed:
- Anticipatory governance that models systemic effects
- System design as ethical practice
- Monitoring and feedback obligations
- Accountability scaled to power and knowledge
- Humility about limits of prediction, commitment to learning
The hardest truth: You can be individually ethical—honest, kind, rule-following—and still contribute to systemic harm. Ethics in complex systems demands thinking beyond individual virtue to structural responsibility.
The systems we build shape the outcomes we get. Designers, operators, and governors of complex systems bear moral responsibility for the worlds they create—intended and emergent.
References
Meadows, D. H. (2008). Thinking in Systems: A Primer. Chelsea Green Publishing.
Essential introduction to systems thinking and leverage points.Perrow, C. (1984). Normal Accidents: Living with High-Risk Technologies. Princeton University Press.
Explains how complex systems produce "normal" failures.Hardin, G. (1968). "The Tragedy of the Commons." Science, 162(3859), 1243–1248.
Classic articulation of emergent harm from individual rationality.Thompson, D. F. (1980). "Moral Responsibility of Public Officials: The Problem of Many Hands." American Political Science Review, 74(4), 905–916.
Foundational work on distributed responsibility.Ostrom, E. (1990). Governing the Commons: The Evolution of Institutions for Collective Action. Cambridge University Press.
How communities solve collective action problems without centralized control.Johnson, D. G., & Powers, T. M. (2005). "Computer Systems and Responsibility: A Normative Look at Technological Complexity." Ethics and Information Technology, 7(2), 99–107.
Ethics of responsibility in technological systems.O'Neil, C. (2016). Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy. Crown.
Case studies of algorithmic harm in complex systems.Tenner, E. (1996). Why Things Bite Back: Technology and the Revenge of Unintended Consequences. Knopf.
Historical examples of technological backfires.Vallor, S. (2016). Technology and the Virtues: A Philosophical Guide to a Future Worth Wanting. Oxford University Press.
Virtue ethics adapted for technological complexity.Zuboff, S. (2019). The Age of Surveillance Capitalism: The Fight for a Human Future at the New Frontier of Power. PublicAffairs.
Systemic ethical analysis of data-driven business models.Taleb, N. N. (2007). The Black Swan: The Impact of the Highly Improbable. Random House.
On unpredictability and tail risks in complex systems.Sunstein, C. R. (2019). How Change Happens. MIT Press.
Social cascades and tipping points in complex social systems.Floridi, L. (2013). The Ethics of Information. Oxford University Press.
Philosophical framework for information ethics in complex systems.Pasquale, F. (2015). The Black Box Society: The Secret Algorithms That Control Money and Information. Harvard University Press.
On opacity and accountability in algorithmic systems.United Nations Human Rights Council (2018). Report of the Independent International Fact-Finding Mission on Myanmar.
Documents Facebook's role in Rohingya genocide.
About This Series: This article is part of a larger exploration of ethics, complexity, and decision-making. For related concepts, see [Unintended Consequences], [Good Intentions, Bad Outcomes], [Systems Thinking Vocabulary], [Second-Order Thinking], and [Ethical Decision-Making Explained].