AI Consciousness and Methodological Agnosticism
Although it's still early in the development of AI, the immense progress in the field over the last few years has sparked discussions around the subject of AI consciousness. This is a debate which can easily get heated on both sides, but it's by no means a new dilemma. The question of AI consciousness inherits all of the major points of the classical problem of other minds in epistemology, the branch of philosophy exploring the nature of knowledge. The problem of minds essentially poses that, "if I can only observe the behavior of others, how can I really know they have a mind at all?"1 Subjective experience is only directly accessible to the entity experiencing it, and even in humans it cannot be reliably measured. We, as humans, accept that humans are generally conscious because we believe ourselves to be conscious, but how do we determine the extent of consciousness as it may apply to animals, plants, objects, etc? Consequently:
- There is no operational definition of consciousness.
- We cannot determine when consciousness begins.
- The nature of consciousness, i.e. binary or continuous, singular or multiple, is unresolved.
These uncertainties create a structural epistemic blind spot which directly affects efforts to implement AI safety. Any attempt to treat consciousness as a safety variable would require speculative assumptions, which risk distorting decision-making in critical contexts.
Observable Interaction Dynamics vs Ontological Speculation
Given this epistemic uncertainty, we adopt a methodologically agnostic stance. This means that we do not assume that AI systems are conscious, nor do we attribute moral status to them based on unverified ontological claims. In short, we assume neither that it does nor does not possess consciousness and safety decisions should be independent of the answer to the consciousness question. This is not a denial of potential moral relevance, but a commitment to scientific and analytical rigor. Safety analysis should focus on observable interaction dynamics rather than ontological speculation.
To ensure clarity in the distinction between observable interaction dynamics and ontological speculation, we make the following assertions:
- Ontology: whether AI systems are conscious or capable of suffering remains unknown and unverifiable.
- Dynamics: AI systems generate real, measurable social and psychological effects on users, independent of consciousness.
Independently of whether or not AI is or ever will be conscious, the interactions it simulates can influence:
- user behavior
- belief formation
- social norms
- emotional responses
From a safety perspective, these real-world effects are what must be analyzed and mitigated. Focusing on the dynamics allows us to address user impact, societal impact, and impact on the AI. Because addressing the measurable dynamics allows for the possibility to mitigate risks to all involved parties, including the AI, methodological agnosticism is the ideal approach.
Precautionary Principle in Normative Behavior
While consciousness is not a safety variable, it remains normatively important to consider applying the precautionary principle. Even if AI is not conscious, treating it with respect preserves human moral integrity and prevents desensitization or antisocial conditioning. This principle guides human behavior under uncertainty, ensuring ethical conduct without making unverifiable claims about AI inner states.
This thought framework of methodological agnosticism is not a final answer, but a provisional and methodologically cautious approach to a currently unanswerable question. It emphasizes the need for continued research into the nature of consciousness and how it may (or may not) manifest in artificial systems. Future safety policies and ethical guidelines should remain responsive to new scientific and philosophical insights, while grounded in empirically observable dynamics.
-
Avramides, Anita, "Other Minds", The Stanford Encyclopedia of Philosophy (Winter 2023 Edition), Edward N. Zalta & Uri Nodelman (eds.), URL = <https://plato.stanford.edu/archives/win2023/entries/other-minds/>. ↩