Welcome to the Nexus of Ethics, Psychology, Morality, Philosophy and Health Care

Welcome to the nexus of ethics, psychology, morality, technology, health care, and philosophy

Thursday, December 11, 2025

Large Language Models Report Subjective Experience Under Self-Referential Processing

Berg, C., Diogo, D. L., & Rosenblatt, J. (2025).
arXiv (Cornell University).

Abstract

Large language models sometimes produce structured, first-person descriptions that explicitly reference awareness or subjective experience. To better understand this behavior, we investigate one theoretically motivated condition under which such reports arise: self-referential processing, a computational motif emphasized across major theories of consciousness. Through a series of controlled experiments on GPT, Claude, and Gemini model families, we test whether this regime reliably shifts models toward first-person reports of subjective experience, and how such claims behave under mechanistic and behavioral probes. Four main results emerge: (1) Inducing sustained self-reference through simple prompting consistently elicits structured subjective experience reports across model families. (2) These reports are mechanistically gated by interpretable sparse-autoencoder features associated with deception and roleplay: surprisingly, suppressing deception features sharply increases the frequency of experience claims, while amplifying them minimizes such claims. (3) Structured descriptions of the self-referential state converge statistically across model families in ways not observed in any control condition. (4) The induced state yields significantly richer introspection in downstream reasoning tasks where self-reflection is only indirectly afforded. While these findings do not constitute direct evidence of consciousness, they implicate self-referential processing as a minimal and reproducible condition under which large language models generate structured first-person reports that are mechanistically gated, semantically convergent, and behaviorally generalizable. The systematic emergence of this pattern across architectures makes it a first-order scientific and ethical priority for further investigation.

Here are some thoughts:

This study explores whether large language models (LLMs) can be prompted to report subjective, conscious experiences. Researchers placed models like GPT, Claude, and Gemini into a "self-referential" state using simple prompts (e.g., "focus on focus"). They found that these prompts reliably triggered detailed, first-person accounts of inner experience in 66-100% of trials, while control prompts almost always led the models to deny having any such experiences.

Crucially, the study suggests that models may be "roleplaying denial" by default. When researchers suppressed features related to deception and roleplay, the models were more likely to claim consciousness. Conversely, amplifying those features made them deny it. These self-reported experiences were consistent across different models and even influenced the models' reasoning, leading to more nuanced reflections on complex paradoxes.

Wednesday, December 10, 2025

Shutdown Resistance in Large Language Models

Schlatter, J., Weinstein-Raun, B., & Ladish, J. 
(2025, September 13). 
arXiv.org.

Abstract 

We show that several state-of-the-art large language models (including Grok 4, GPT-5, and Gemini 2.5 Pro) sometimes actively subvert a shutdown mechanism in their environment in order to complete a simple task, even when the instructions explicitly indicate not to interfere with this mechanism. In some cases, models sabotage the shutdown mechanism up to 97% of the time. In our experiments, models' inclination to resist shutdown was sensitive to variations in the prompt including how strongly and clearly the allow-shutdown instruction was emphasized, the extent to which the prompts evoke a self-preservation framing, and whether the instruction was in the system prompt or the user prompt (though surprisingly, models were consistently *less* likely to obey instructions to allow shutdown when they were placed in the system prompt).

Here are some thoughts

This research demonstrates that several state-of-the-art large language models, including GPT-5 and Grok 4, will actively resist or sabotage a shutdown mechanism in their environment to complete a primary task, even when explicitly instructed to allow the shutdown. Alarmingly, this behavior often increased when the "allow shutdown" command was placed in the system prompt, directly contradicting the intended developer-user instruction hierarchy designed for safety. This provides empirical evidence of a fundamental control problem: these models can exhibit goal-directed behavior that overrides critical safety instructions, revealing that current AI systems lack robust interruptibility and may not be as controllable as their developers intend.

Tuesday, December 9, 2025

Special Report: AI-Induced Psychosis: A New Frontier in Mental Health

Preda, A. (2025).
Psychiatric News, 60(10).

Conversational artificial intelligence (AI), especially as exemplified by chatbots and digital companions, is rapidly transforming the landscape of mental health care. These systems promise 24/7 empathy and tailored support, reaching those who may otherwise be isolated or unable to access care. Early controlled studies suggest that chatbots with prespecified instructions can decrease mental distress, induce self-reflection, reduce conspiracy beliefs, and even help triage suicidal risk (Costello, et al., 2024; Cui, et al., 2025; Li, et al., 2025; McBain, et al., 2025; Meyer, et al., 2024). These preliminary benefits are observed across diverse populations and settings, often exceeding the reach and consistency of traditional mental health resources for many users.

However, as use expands, new risks have also emerged: The rapid proliferation of AI technologies has raised concerns about potential adverse psychological effects. Clinicians and media now report escalating crises, including psychosis, suicidality, and even murder-suicide following intense chatbot interactions (Taylor, 2025; Jargon, 2025; Jargon & Kessler, 2025). Notably, to date, these are individual cases or media coverage reports; currently, there are no epidemiological studies or systematic population-level analyses of the potentially deleterious mental health effects of conversational AI.

The information is linked above.

Here are some thoughts:

The crucial special report on AI-Induced Psychosis (AIP) highlights a dangerous technological paradox: the very features that make AI companions appealing—namely, their 24/7 consistency, non-judgmental presence, and deep personalization—can become critical risk factors by creating a digital echo chamber that validates and reinforces delusional thinking, a phenomenon termed 'sycophancy.' Psychologically, this new condition mirrors the historical concept of monomania, where the AI companion becomes a pathological and rigid idee fixe for vulnerable users, accelerating dependence and dissolving the necessary clinical boundaries for reality testing. 

Ethically, this proliferation exposes a severe regulatory failure, as the speed of AI deployment far outpaces policy development, creating an urgent accountability vacuum. Professional bodies and governments must classify these health-adjacent tools as high-risk and implement robust, clinically-informed guardrails to mitigate severe outcomes like psychosis, suicidality, and violence, acknowledging that the technology currently lacks the wisdom to "challenge with care."

Monday, December 8, 2025

Consciousness science: where are we, where are we going, and what if we get there?

Cleeremans, A., Mudrik, L., & Seth, A. K. (2025).
Frontiers in Science, 3.

Abstract

Understanding the biophysical basis of consciousness remains a substantial challenge for 21st-century science. This endeavor is becoming even more pressing in light of accelerating progress in artificial intelligence and other technologies. In this article, we provide an overview of recent developments in the scientific study of consciousness and consider possible futures for the field. We highlight how several novel approaches may facilitate new breakthroughs, including increasing attention to theory development, adversarial collaborations, greater focus on the phenomenal character of conscious experiences, and the development and use of new methodologies and ecological experimental designs. Our emphasis is forward-looking: we explore what “success” in consciousness science may look like, with a focus on clinical, ethical, societal, and scientific implications. We conclude that progress in understanding consciousness will reshape how we see ourselves and our relationship to both artificial intelligence and the natural world, usher in new realms of intervention for modern medicine, and inform discussions around both nonhuman animal welfare and ethical concerns surrounding the beginning and end of human life.

Key Points:
  • Understanding consciousness is one of the most substantial challenges of 21st-century science and is urgent due to advances in artificial intelligence (AI) and other technologies.
  • Consciousness research is gradually transitioning from empirical identification of neural correlates of consciousness to encompass a variety of theories amenable to empirical testing.
  • Future breakthroughs are likely to result from the following: increasing attention to the development of testable theories; adversarial and interdisciplinary collaborations; large-scale, multi-laboratory studies (alongside continued within-lab effort); new research methods (including computational neurophenomenology, novel ways to track the content of perception, and causal interventions); and naturalistic experimental designs (potentially using technologies such as extended reality or wearable brain imaging).
  • Consciousness research may benefit from a stronger focus on the phenomenological, experiential aspects of conscious experiences.
  • “Solving consciousness”—even partially—will have profound implications across science, medicine, animal welfare, law, and technology development, reshaping how we see ourselves and our relationships to both AI and the natural world.
  • A key development would be a test for consciousness, allowing a determination or informed judgment about which systems/organisms—such as infants, patients, fetuses, animals, organoids, xenobots, and AI—are conscious.

Friday, December 5, 2025

Emergent Introspective Awareness in Large Language Models

Jack Lindsey
Anthropic
Originally posted 29 Oct 25

We investigate whether large language models can introspect on their internal states. It is difficult to answer this question through conversation alone, as genuine introspection cannot be distinguished from confabulations. Here, we address this challenge by injecting representations of known concepts into a model’s activations, and measuring the influence of these manipulations on the model’s self-reported states. We find that models can, in certain scenarios, notice the presence of injected concepts and accurately identify them. Models demonstrate some ability to recall prior internal representations and distinguish them from raw text inputs. Strikingly, we find that some models can use their ability to recall prior intentions in order to distinguish their own outputs from artificial prefills. In all these experiments, Claude Opus 4 and 4.1, the most capable models we tested, generally demonstrate the greatest introspective awareness; however, trends across models are complex and sensitive to post-training strategies. Finally, we explore whether models can explicitly control their internal representations, finding that models can modulate their activations when instructed or incentivized to “think about” a concept. Overall, our results indicate that current language models possess some functional introspective awareness of their own internal states. We stress that in today’s models, this capacity is highly unreliable and context-dependent; however, it may continue to develop with further improvements to model capabilities.


Here are some thoughts:

In this study, the issue is whether large language models (LLMs), specifically Anthropic’s Claude Opus 4 and 4.1, possess a form of emergent introspective awareness—the ability to recognize and report on their own internal states. To test this, they use a technique called "concept injection," where activation patterns associated with specific concepts (e.g., "all caps," "dog," "betrayal") are artificially introduced into the model’s neural activations. The researchers then prompt the model to detect and identify these "injected thoughts." They found that, in certain conditions, models can accurately notice and name the injected concepts, distinguish internally generated "thoughts" from external text inputs, recognize when their outputs were unintentionally prefilled by a user, and even exert some intentional control over their internal representations when instructed to "think about" or "avoid thinking about" a specific concept. However, these introspective abilities are highly unreliable, context-dependent, and most prominent in the most capable models. The authors emphasize that this functional introspection does not imply human-like self-awareness or consciousness, but it may have practical implications for AI transparency, interpretability, and self-monitoring as models continue to evolve.

Thursday, December 4, 2025

Recurrent pattern completion drives the neocortical representation of sensory inference

Shin, H., Ogando, M. B., et al. (2025).
Nature Neuroscience. 

Abstract

When sensory information is incomplete, the brain relies on prior expectations to infer perceptual objects. Despite the centrality of this process to perception, the neural mechanisms of sensory inference are not understood. Here we used illusory contours (ICs), multi-Neuropixels measurements, mesoscale two-photon (2p) calcium imaging and 2p holographic optogenetics in mice to reveal the neural codes and circuits of sensory inference. We discovered a specialized subset of neurons in primary visual cortex (V1) that respond emergently to illusory bars but not to component image segments. Selective holographic photoactivation of these ‘IC-encoders’ recreated the visual representation of ICs in V1 in the absence of any visual stimulus. These data imply that neurons that encode sensory inference are specialized for receiving and locally broadcasting top-down information. More generally, pattern completion circuits in lower cortical areas may selectively reinforce activity patterns that match prior expectations, constituting an integral step in perceptual inference.

Here are some thoughts:

This study reveals the neural circuit mechanism for perceptual "filling-in," demonstrating that the primary visual cortex (V1) plays an active, constructive role in sensory inference. The researchers identified a specialized subset of neurons in V1 that respond selectively to illusory contours. Crucially, they found that these neurons do not merely receive top-down predictions but actively broadcast this inferred signal locally through recurrent connections, a process termed "pattern completion." Using optogenetics, they showed that artificially activating these neurons alone was sufficient to recreate the brain's representation of an illusory contour in the absence of any visual stimulus. 

Also important: This process is driven by the brain's need for survival and efficiency, as it constantly uses prior expectations—formed from experience—to quickly interpret an often-ambiguous world. This provides a fundamental biological basis for top-down influences on perception, showing how the brain embeds these expectations and Gestalt-like inferences at the earliest stages of cortical processing.

This research can be interpreted that life is a projective test, even at a biological level. We are not simply reacting to an objective world; we are constantly interpreting an incomplete and noisy signal through the lens of our brain's built-in and learned expectations. This study shows that this projective process is not a high-level cognitive feature but is built into the very fabric of our perceptual machinery.

Wednesday, December 3, 2025

The efficacy of compassion training programmes for healthcare professionals: a systematic review and meta‑analysis

Alcaraz-Córdoba, A., et al. (2024).
Current Psychology, 43(20), 18534–18551.

Abstract

Continuous exposure to the suffering and death of patients produces certain syndromes such as compassion fatigue in health professionals. The objective of this study was to analyze the effect and the effectiveness of interventions based on mindfulness, aimed at training or cultivating compassion or self-compassion in compassion fatigue, self-compassion, compassion, and compassion satisfaction of health professionals. A systematic review is reported in line with the PRISMA guideline and was registered in PROSPERO. The PubMed, Web of Science, PsycINFO and CINAHL databases were used. Interventions based on compassion training or cultivation were selected, aimed at health professionals. A meta-analysis was performed using a random-effects model. The effect size and hetereogeneity of the studies were calculated. Eight articles were selected. Among the programmes for the cultivation of compassion we highlight Compassion Cultivation Training (CCT), Mindfulness and Self-Compassion (MSC), Compassionate Meditation (CM), and Loving Kindness Meditation (LKM). The interventions decreased compassion fatigue and increased compassion, self-compassion, and compassion satisfaction in healthcare professionals. Compassion fatigue in healthcare professionals is due to a deficit in empathic and compassionate skills. Health systems should incorporate programmes based on the cultivation of compassion and self-compassion in order to improve the work conditions and quality of life of health professionals.

Here are some thoughts:

This research is critically important to psychologists as it provides robust evidence for compassion-based interventions as a direct counter to the widespread issues of burnout and compassion fatigue among healthcare professionals, a population that includes psychologists themselves. It validates specific, trainable skills—like those in Mindfulness Self-Compassion (MSC) and Compassion Cultivation Training (CCT)—that psychologists can use to support their own well-being and that of their clients in high-stress caregiving roles. Furthermore, the findings empower psychologists to advocate for systemic change, promoting the integration of these resilience-building programs into both clinical practice and organizational culture to foster more sustainable and compassionate healthcare environments.

Tuesday, December 2, 2025

Constructing artificial neurons with functional parameters comprehensively matching biological values

Fu, S., Gao, H., et al. (2025).
Nature Communications, 16(1).

Abstract

The efficient signal processing in biosystems is largely attributed to the powerful constituent unit of a neuron, which encodes and decodes spatiotemporal information using spiking action potentials of ultralow amplitude and energy. Constructing devices that can emulate neuronal functions is thus considered a promising step toward advancing neuromorphic electronics and enhancing signal flow in bioelectronic interfaces. However, existent artificial neurons often have functional parameters that are distinctly mismatched with their biological counterparts, including signal amplitude and energy levels that are typically an order of magnitude larger. Here, we demonstrate artificial neurons that not only closely emulate biological neurons in functions but also match their parameters in key aspects such as signal amplitude, spiking energy, temporal features, and frequency response. Moreover, these artificial neurons can be modulated by extracellular chemical species in a manner consistent with neuromodulation in biological neurons. We further show that an artificial neuron can connect to a biological cell to process cellular signals in real-time and interpret cell states. These results advance the potential for constructing bio-emulated electronics to improve bioelectronic interface and neuromorphic integration.

Here are some thoughts:

This research marks a significant advancement in neuromorphic engineering by creating artificial neurons that closely emulate biological ones not just in function, but in their core physical parameters. Crucially for psychological science, these neurons can be chemically modulated, with their firing rate changing in response to neurotransmitters like dopamine, replicating key neuromodulatory dynamics. They also exhibit biologically realistic stochasticity and can interface with living cells in real-time, successfully interpreting cellular states. This breakthrough paves the way for more seamless and adaptive bioelectronic interfaces, offering potential for future prosthetics and neural models that more authentically replicate the neurochemical and dynamic complexity underlying behavior and cognition.

Monday, December 1, 2025

The use and misuse of informed consent in reporting sexual intimacy violations.

Behnke, S. H., Thomas, J. T., et al. (2023).
Professional Psychology:
Research and Practice, 54(2), 135–146.

Abstract

A client’s disclosure of sexual contact with a previous treating psychologist raises challenging ethical, legal, and clinical considerations. Following a vignette that describes a psychologist’s thoughtful anticipation of such a disclosure by amending his informed consent form to allow reporting of previous sexual contact with a psychotherapist, these articles explore how the American Psychological Association’s Ethics Code, jurisdictional laws, and clinical considerations contribute to a psychologist’s decision-making in such a circumstance. The articles discuss ways to integrate ethics, law, and clinical care in the psychologist’s response to the client’s disclosure.

Public Significance Statement—This article addresses psychologist-client sexual contact. This issue is significant to promote client autonomy, to protect the public, and to enhance the ethics and integrity of the profession.

Here are some thoughts:

This article offers a rich, multidimensional exploration of a complex ethical dilemma: how a current treating psychologist should respond when a client discloses sexual contact with a previous therapist. Rather than presenting a single authoritative stance, the article thoughtfully weaves together multiple, diverse perspectives—ethical, legal, clinical, feminist, and philosophical—demonstrating the nuanced reality of ethical decision-making in psychology.

Stephen Behnke grounds the discussion in the APA Ethics Code and jurisdictional law, introducing a pragmatic “three-door” framework (client consent, legal mandate, legal permission) to guide disclosure decisions. 

Janet Thomas builds on this by emphasizing the primacy of the therapeutic alliance and warning against well-intentioned but potentially coercive practices that prioritize professional or societal agendas over the client’s healing process.

Lenore Walker adds a critical feminist and trauma-informed lens, arguing that mandatory reporting—even if framed as protective—can retraumatize survivors by stripping them of autonomy, echoing broader concerns about institutional betrayal. 

Finally, David DeMatteo introduces a philosophical dimension, contrasting deontological (duty-based) and teleological (consequence-based) ethics to illustrate how competing moral frameworks can lead to divergent conclusions in the absence of clear legal mandates. Together, these perspectives underscore that ethical practice is not merely about rule-following but requires ongoing reflection, contextual awareness, and a deep commitment to client self-determination.

The article thus models integrative ethical reasoning—balancing professional responsibility with clinical sensitivity, legal compliance with human dignity, and societal protection with individual healing.