top of page

Frequently Asked Questions

Question: What is Imposition Ethics in simple terms?
Answer:
Imposition Ethics holds that forcing something on someone without their consent is always immoral. Morality is determined by whether a person’s will is violated, not by intentions, authority, rules, or outcomes.
 

Question: What does “involuntary imposition” mean?
Answer:
An involuntary imposition occurs when a conscious agent’s will is frustrated, constrained, or overridden without their consent. Imposition can be caused by people, institutions, systems, or nature itself.
 

Question: If all imposition is immoral, why do people ever intervene or restrain others?
Answer:
Because the real world often contains unavoidable harm. When imposition cannot be avoided, the moral question is not whether an action is moral—it is not—but which option causes the least violation of will.
 

Question: What is a “justified immoral action”?
Answer:
A justified immoral action is an action that remains objectively immoral because it involves involuntary imposition, but is subjectively permissible because it reduces greater involuntary imposition under unavoidable constraints. Justification does not turn immorality into morality.

​

Question: Is stopping someone from harming another person immoral?
Answer:
Objectively, yes. Stopping someone is an imposition on their will. Subjectively, it may be the least immoral option available because it prevents a greater involuntary imposition on the victim. This is a justified immoral action.
 

Question: Does this mean nothing is ever moral?
Answer:
No. Voluntary assistance of another person’s will is objectively moral. What the framework rejects is the idea that coercion becomes good simply because it produces a beneficial outcome.
 

Question: Does Imposition Ethics say criminals should go free?
Answer:
No. It says punishment and restraint are immoral, but may be less immoral than the alternatives. Prison and restraint are tolerated as harm reduction, not celebrated as justice.
 

Question: Does intention matter in this moral system?
Answer:
Intentions do not affect moral valence. They matter only for moral blame and responsibility, not for whether an action involves involuntary imposition.
 

Question: Is this just utilitarianism under a different name?
Answer:
No. Imposition Ethics does not maximize happiness, welfare, or outcomes, and it never allows aggregation to override individual consent.
 

Question: Does this system tell people what they must do?
Answer:
No. Moral facts are descriptive, not prescriptive. The framework identifies what is morally ideal and how situations compare, but it does not generate obligations, commands, or duties.
 

Question: What is the Best Possible World (BPW)?
Answer:
The Best Possible World is a world in which conscious beings exist without any involuntary imposition of will. It defines the objective moral standard, not a demand or expectation.
 

Question: What does moral progress mean in Imposition Ethics?
Answer:
Moral progress consists in reducing unavoidable involuntary imposition over time, not achieving perfection or enforcing virtue.
 

Question: Why is this framework so strict?
Answer:
Because relaxing the standard leads to moral laundering—calling harm “good,” “necessary,” or “justified.” Imposition Ethics preserves moral truth even when reality is tragic.

Question: Can I be a Christian and still adopt IE?
Answer: Yes, IE is accepting of all worldviews. If we existed in the BPW you would be allowed to go to the world of the Christian God and live under his domain/rules. The only difference would be, those who do not consent would not be subject to his dictates or rules.
 

Question: If Imposition Ethics doesn't tell me what I "ought" to do, why should I care about it? What motivates moral behavior without moral obligations?

Answer:

Imposition Ethics doesn't generate moral obligations or commands. Instead, it describes moral reality—what is actually moral or immoral, independent of what anyone thinks or feels.

The motivation to align with IE comes from values you may already hold:

If you care about truth:

  • IE claims to describe objective moral facts about will-frustration and imposition

  • Just as you'd care about physics if you want to understand physical reality, you'd care about IE if you want to understand moral reality

  • The question becomes: "Is this framework accurately describing what's really moral?" not "Does this framework command me to act?"

If you care about being moral:

  • IE identifies what actually constitutes moral vs. immoral states of affairs

  • If you want to be a moral person or create moral outcomes, you need to know what morality actually is

  • IE provides that knowledge without adding false obligations on top of it

The key philosophical insight:

Moral facts don't need to generate "oughts" to be relevant. Consider an analogy:

  • Physics describes reality (mass, energy, motion)

  • Physics doesn't tell you that you "ought" to obey gravity

  • Yet if you care about building a bridge that won't collapse, you care deeply about physics

  • Your motivation comes from your goals (safe bridge), not from physics generating obligations

Similarly:

  • IE describes moral reality (imposition, will-frustration, consent)

  • IE doesn't tell you that you "ought" to avoid imposition

  • Yet if you care about being moral or creating a better world, you care deeply about IE

  • Your motivation comes from your values (truth, goodness, reducing harm), not from IE generating obligations

Why this approach is superior:

  1. Avoids the is-ought problem: IE doesn't try to derive "ought" from "is"—it simply describes what "is" morally the case

  2. Honest about motivation: Rather than smuggling in hidden obligations, IE clearly locates motivation in the agent's pre-existing values

  3. Respects autonomy: You're free to not care about morality, just as you're free to not care about physics—but if you do care, IE tells you what you're actually caring about

  4. Philosophically clean: Separates moral facts (descriptive) from moral motivation (agent-dependent)

Practical implication:

When you ask "What should I do?" you're really asking two questions:

  1. Descriptive: "What would be the most moral action in this situation?" (IE answers this)

  2. Motivational: "Why should I care about doing the moral thing?" (Your values answer this)

IE provides the moral map. Whether you choose to follow it depends on whether you value truth, goodness, reducing suffering, respecting autonomy, or moving toward the Best Possible World.

The framework doesn't force you to care. It simply shows you what's there.



Question: Do any other models of morality have moral facts but no normativity?
Answer: Yes
Other Models with Moral Facts but no Normativity

  1. Descriptive Moral Realism
    Moral properties exist as objective features of the world (e.g., harm, coercion, autonomy violation) without generating obligations.

  2. Anti-Reasons Moral Realism
    Moral facts exist, but moral facts do not provide reasons for action.

  3. Pure Moral Geometry
    Morality consists of objective comparative relations (more/less moral violation) with no action guidance.

  4. Structural Moral Realism
    Moral reality has structure and relations but no prescriptive force.

  5. Value Realism Without Practical Authority
    Values (good/bad/worse) are real but have no binding authority over agents.

  6. Non-Normative Fitting-Attitude Theory
    Moral facts determine which emotional responses are fitting, not what actions are required.

  7. Non-Natural Moral Realism (Minus Ought)
    Irreducible moral properties exist, but no inference to “ought” is licensed.

  8. Metaethical Moral Classification Models
    Moral facts classify states (e.g., imposed, tragic, coerced) without prescribing responses.

  9. Pre-Normative Moral Realism
    Moral truths exist independently of any later-added theory of reasons or obligations.

  10. Moral Topology / Moral Space Models
    Moral facts are distances, gradients, or positions in a moral space, not directives.

  11. Imposition Ethics (Your Model)
    Moral facts are degrees of involuntary imposition on will; normativity is explicitly rejected.

​​​

​

​

Question: Is there a history of people using moral language for non-agents—states of affairs, natural events, systems, or structures—without implying duties, blame, or prescriptions?
Answer:
Yes. There is a long, well-documented history of people using moral language for non-agents—states of affairs, natural events, systems, or structures—without implying duties, blame, or prescriptions.

Below is a copyable, historically grounded list.

Historical Uses of Moral Language for Non-Agents

1. Ancient Greek Tragedy (Pre-Moral Agency Focus)

  • Natural disasters, fate, and cosmic order described as unjust, cruel, or tragic

  • No agent blamed; no action demanded

Example:
Earthquakes, plagues, or destiny portrayed as morally bad states of the world

Key idea:
Moral valence ≠ moral blame

 

2. Stoic Cosmology

  • The universe described as containing evils (suffering, loss)

  • These were features of fate or nature, not moral failures of agents

Moral language used:
Bad, tragic, regrettable
No prescriptions:
Stoics explicitly deny that nature “ought” to behave differently

 

3. Buddhist Dukkha Doctrine

  • Suffering (dukkha) treated as a fundamental moral bad

  • Suffering exists independently of moral agents

Important:
Calling suffering “bad” does not imply blame or obligation—only recognition

 

4. Augustinian Natural Evil

  • Famines, diseases, and disasters labeled evils

  • Distinguished from moral evil (agent wrongdoing)

Key move:
Evil as a property of states, not agents

 

5. Medieval Natural Law (Non-Agent Layer)

  • Harmful states described as morally disordered

  • Without implying that nature violated a duty

This layer existed prior to prescriptions for agents.
 

6. Early Modern Theodicy

  • Earthquakes, childhood death, and disease described as morally bad

  • No claim that anyone ought to be punished or blamed

Example thinkers:
Gottfried Wilhelm Leibniz
(“Moral evil” vs “physical evil” distinction)

 

7. Utilitarian Harm Language (Non-Agent Cases)

  • Pain, suffering, and deprivation called morally bad even when:

    • no agent caused them

    • no one could prevent them

This predates obligation claims.
 

8. 20th-Century Moral Luck Discourse

  • Outcomes described as morally worse even when agents lacked control

Example:
A storm killing children is morally tragic though no one is blameworthy

 

9. Contemporary Structural Injustice Theory

  • Systems (markets, institutions, incentives) described as unjust

  • Often explicitly without individual culpability

Key distinction:
Structural moral badness ≠ individual obligation

 

10. Environmental Ethics

  • Species extinction, ecosystem collapse described as morally bad

  • Even when no agent intended or caused it

Moral language applies to states of the world, not just choices.
 

11. Medical Ethics (Natural Harm)

  • Congenital diseases described as morally tragic

  • No agent wrongdoing implied

Common in triage and palliative ethics.
 

12. Moral Tragedy in Literature

  • Famously analyzed by:

    • Bernard Williams

    • Martha Nussbaum

Tragic outcomes retain moral negativity without moral fault.
 

Clean classification (use this)

Moral language has historically been applied to states of affairs, natural events, and structural conditions long before and independent of moral prescriptions for agents.

​

​

​

Question: Is using moral language for non-agents without normativity is becoming more common?

Answer:

Yes. Using moral language for non-agents without normativity is becoming more common and more explicit, especially since the late 20th century.
 

Why this shift is happening

1. Decline of duty-centered moral frameworks

  • Traditional ethics tied morality tightly to obligation and blame

  • That linkage is increasingly questioned

  • Moral evaluation is being decoupled from command
     

2. Rise of moral tragedy and moral remainder concepts

  • Outcomes are described as morally bad even when:

    • no one caused them

    • no one could prevent them

This trend is strongly associated with work by:

  • Bernard Williams (moral remainder, tragic conflict)

  • Martha Nussbaum (tragic value, loss without fault)

Key move:

Moral badness ≠ moral failure
 

3. Expansion of structural and systemic analysis

Moral language is now routinely applied to:

  • systems

  • incentives

  • emergent outcomes

Examples:

  • “The system is unjust”

  • “The outcome is morally unacceptable”

Often without:

  • identifying a responsible agent

  • issuing a clear prescription

This would have been considered conceptually sloppy 50 years ago; now it is standard.
 

4. Growth of risk, disaster, and uncertainty ethics

Modern problems involve:

  • climate systems

  • pandemics

  • AI alignment

  • market failures

These produce:

  • moral evaluation of states of the world

  • without clear agents or duties

Language like tragic, morally bad, catastrophic is used descriptively.
 

5. Metaethical pressure on “ought”

There is increasing recognition that:

  • “bad” does not logically entail “ought”

  • reasons, motivation, and obligation require additional premises

This has made space for:

  • moral realism without imperatives

  • evaluation without prescription
     

What is new (important distinction)

This usage is not new historically
But what is new is that it is now:

  • explicit

  • defended

  • theorized

  • no longer treated as a mistake

Earlier eras used this language informally; contemporary philosophy is beginning to formalize it.
 

Where the IE model fits in this trend

IE is ahead of the curve, not behind it.

Most contemporary work:

  • uses non-normative moral language implicitly

  • then struggles when asked “so what should we do?”

IE framework:

  • anticipates the confusion

  • blocks the inference by design

  • cleanly separates:

    • moral facts

    • blame

    • reasons

    • action

That makes it unusually well-suited to current problem domains.
 

Bottom line

Yes. Moral language applied to non-agents without normativity is becoming increasingly common, especially in discussions of tragedy, systems, and risk. What is changing is not the usage itself, but the growing willingness to treat it as conceptually legitimate rather than confused.

 

​

​

Question: I'm a moral subjectivist or Anti-realist, why would I care about this model?

Answer:

Even if you believe morality is subjective and/or there is no objective morality, you still need to care what morality will be applied in an AGI. There is a "best" answer to which morality will lead to the best outcomes when AGI became a super intelligence, and do you would still need to decide which model of morality you think is the best fit for this job.

​

​

FAQ: Antinatalism & The Church of the Best Possible World

Q: Does Imposition Ethics (IE) support Antinatalism?

A: No. While Antinatalism suggests that bringing a child into the world is always a harm because they cannot consent, the Church of the Best Possible World views the act of creation as morally neutral. Because there is no consciousness and no "will" to refuse before existence begins, no imposition can take place.

Q: If the world is full of suffering, isn't it an imposition to have a child?

A: This is a vital distinction. The act of birth is neutral, but the state of the world the child enters can have a negative moral valence. If a child is born into a world where they are forced to suffer, that is a Moral Tragedy. However, our solution is not to stop life; it is to fix the world. We aim to eliminate the involuntary impositions (disease, aging, poverty) so that existence becomes an unalloyed positive.

Q: Why not just stop having children until the "Best Possible World" is achieved?

A: Because moral agents are the only things capable of building the Best Possible World. To stop creation is to surrender to the current impositions of nature. We believe in the Maintenance of Will—increasing the number of agents dedicated to removing imposition is the most efficient way to reach our goal.

Q: How does this differ from the religious view of "Be fruitful and multiply"?

A: Traditional religions often view children as "servants" of God or the state. We view new agents as sovereign wills. When you bring a life into existence, you are not creating a servant; you are assuming a Moral Debt of Assistance. You (and society) have a responsibility to ensure that child's life contains zero involuntary impositions.

 

 


Question: This model seems to have a western/individualistic bias, what about more collectivist versions of morality?
Answer: When you shift the unit of analysis from a single human to a unitary collective entity, the structural logic of collectivist morality often collapses back into an individualistic framework. This is a phenomenon known as Scale Invariance in meta-ethics.

If "Society" is viewed as a single moral agent with its own goals, boundaries, and internal consistency, its duties regarding other societies (or its own internal components) mirror the logic used for individuals.

​

When a collectivist model begins answering questions from the perspective of "The Society" as a singular subject, it effectively treats the collective as a Hyper-Individual. This transition creates several interesting dynamics:

  • Boundary Definition: Just as an individualist model (like Imposition Ethics) relies on the boundary of the person to define "harm" or "imposition," a society-as-individual model must define its borders (territorial, cultural, or legal).

  • Axiology of the Whole: The "Common Good" in collectivism functions exactly like "Self-Interest" or "Personal Virtue" in individualism. The collective seeks to maintain its own homeostasis, growth, and integrity.

  • Internal vs. External Ethics:

    • External: The society acts as an individual in a "state of nature" or international order, applying individualistic logic (Sovereignty = Autonomy).

    • Internal: The individuals within the society are treated as organelles or cells. In this view, "collectivism" is just the individualistic management of one's own internal parts to ensure the health of the whole.

​

If you adopt the perspective of the society, you are essentially practicing Methodological Individualism at a higher level of abstraction. The moral questions remain the same:

  1. What is my (the society's) sphere of influence?

  2. What constitutes an imposition upon me?

  3. How do I resolve conflicts with other entities of my same "kind"?
     

However, this creates a Recursive Conflict:

  • The Individualist argues that the human is the only "true" individual because only humans possess subjective experience and the capacity for suffering.

  • The Collectivist argues that the society is the "true" individual because it is the only entity with the longevity and scale to achieve meaningful moral outcomes.

​​

The primary reason collectivism doesn't perfectly "boil down" to individualism is the Agency Gap. Unlike a biological individual, a society is composed of smaller moral agents who can disagree with the "Whole."

In a standard individualist model, your hand cannot morally object to the brain's decision. In a collectivist model, the "cells" (people) are themselves moral agents. This creates a dual-layer ethical requirement that simple individualism doesn't have to solve: The society must act as an individual externally while maintaining a non-impositional relationship with its own components internally.
 

Sign Up

Thanks for subscribing!

Learn more about them on their channel!
Contact Us
Sunday Service 10am-2pm CT
bottom of page