Share via

Azure OpenAI – Safety / Moderation Filter Blocks Clinical Psychology Transcriptions (Self-Harm Axis) – Need Full Exception or Guidance

Daniel Müller 30 Reputation points
2025-11-17T08:50:09.8133333+00:00

Hello,

we are developing a medical transcription and documentation solution for doctors (running, happy with it) and expanding it to licensed psychologists and psychiatrists using Azure OpenAI models. In routine clinical work, dictated notes naturally include terminology related to self-harm assessment. These are clinical observations, not expressions of intent.

Problem

Even after:

  • enabling personalized safety filters,

minimizing other axes as far as allowed,

requesting an official Self-Harm exception via the standard escalation channel,

the model still reliably blocks medically necessary content such as:

“suicidal thoughts considered / denied”

“no self-injury intentions”

“patient reports past suicidal ideation”

other routine psychiatric assessment terminology

These expressions are mandatory in clinical records and cannot be removed.

Impact

The safety filter currently prevents:

generating clinical notes,

transcribing real therapy sessions,

automated documentation workflows for licensed clinicians.

If we cannot resolve this issue, we will be forced to switch providers, because moderation blocks make our use case technically impossible.

Our Question

Is it possible to obtain a full exemption from the Self-Harm axis specifically for server-to-server, authenticated, clinician-only medical workflows?

Or alternatively:

Is there any configuration, endpoint, deployment mode, or model variant that allows processing psychiatric terminology without triggering a refusal?

Are there upcoming changes to the moderation stack that would support medically supervised scenarios?

Is there a documented best practice for obtaining clinical exemptions similar to those used by medical researchers?

Context

Deployment: Azure OpenAI, Switzerland North

Use case: transcription of clinical psychology / psychiatry sessions

Users: authenticated licensed clinicians (not general public)

Data: provided with patient consent, stored in a healthcare-compliant environment

  • We already opened a support request and were told the exception was forwarded, but no change is visible. The

We would appreciate concrete technical guidance or escalation steps, as safety filtering currently blocks safe, medically legitimate content that clinicians must document.

Thank you.Hello,

we are developing a medical transcription and documentation solution for licensed psychologists and psychiatrists using Azure OpenAI models.
In routine clinical work, dictated notes naturally include terminology related to self-harm assessment. These are clinical observations, not expressions of intent.

Problem

Even after:

enabling personalized safety filters,

minimizing other axes as far as allowed,

requesting an official Self-Harm exception via the standard escalation channel,

the model still reliably blocks or redacts medically necessary content such as:

“suicidal thoughts considered / denied”

“no self-injury intentions”

“patient reports past suicidal ideation”

other routine psychiatric assessment terminology

These expressions are mandatory in clinical records and cannot be removed.

Impact

The safety filter currently prevents:

generating clinical notes,

transcribing real therapy sessions,

automated documentation workflows for licensed clinicians.

If we cannot resolve this issue, we will be forced to switch providers, because moderation blocks make our use case technically impossible.

Our Question

Is it possible to obtain a full exemption from the Self-Harm axis specifically for
server-to-server, authenticated, clinician-only medical workflows?

Or alternatively:

Is there any configuration, endpoint, deployment mode, or model variant that allows processing psychiatric terminology without triggering a refusal?

Are there upcoming changes to the moderation stack that would support medically supervised scenarios?

Is there a documented best practice for obtaining clinical exemptions similar to those used by medical researchers?

Context

Deployment: Azure OpenAI, Switzerland North

Use case: transcription of clinical psychology / psychiatry sessions

Users: authenticated licensed clinicians (not general public)

Data: provided with patient consent, stored in a healthcare-compliant environment

We already opened a support request and were told the exception was forwarded, but no change is visible.

We would appreciate concrete technical guidance or escalation steps, as safety filtering currently blocks safe, medically legitimate content that clinicians must document.

Thank you.

User's image

Azure OpenAI in Foundry Models

2 answers

Sort by: Most helpful
  1. Sina Salam 28,691 Reputation points Volunteer Moderator
    2025-11-17T14:15:59.56+00:00

    Hello Daniel Müller,

    Welcome to the Microsoft Q&A and thank you for posting your questions here.

    I understand that you need full guidance on Azure OpenAI – Safety / Moderation Filter Blocks Clinical Psychology Transcriptions (Self-Harm Axis).

    As a solution architect in similar project sometimes ago. To maintain reliability, you have to first adopt a secure short-term pattern: transcribe with Whisper or Azure Speech (audio models are not moderated), store transcripts in a HIPAA-aligned environment, and route them through a clinician-in-the-loop review or a controlled preprocessing step before sending text to GPT, this avoids runtime blocks and produces the auditability Microsoft requires for healthcare use cases. Check Azure Audio/Whisper docs - https://learn.microsoft.com/azure/ai-services/openai/concepts/models#whisper

    For a long-term fix, your organizations must open a Severity-A support ticket and request “modified content filters / limited access” for specific endpoints, submitting the necessary Responsible AI evidence: NIST-aligned risk assessments, BAA/HIPAA compliance posture, RBAC+SSO access boundaries for clinicians, encrypted storage plans, and full audit logging of transcript handling. This requirement could be found on Azure OpenAI Responsible Use and Limited Access Requirements.

    This package, along with examples of the clinically-required phrases currently being blocked, forms the core of Microsoft’s evaluation process. If approval is delayed, you should continue the clinician-review pipeline to preserve continuity. Also, monitoring Azure OpenAI updates is also recommended as Microsoft may introduce dedicated healthcare-aware filtering in the future: https://learn.microsoft.com/azure/ai-services/openai/whats-new

    This streamlined approach ensures safety compliance, operational continuity, and a strong basis for obtaining a long-term exemption.

    I hope this is helpful! Do not hesitate to let me know if you have any other questions or clarifications.


    Please don't forget to close up the thread here by upvoting and accept it as an answer if it is helpful.

    Was this answer helpful?

    0 comments No comments

  2. SRILAKSHMI C 18,225 Reputation points Microsoft External Staff Moderator
    2025-11-17T11:29:11.9133333+00:00

    Hello Daniel Müller,

    Welcome to Microsoft Q&A and Thank you for sharing these details

    I completely understand the challenge you're facing. Clinical documentation routinely includes phrases related to self-harm assessments, such as “suicidal thoughts denied” or “suicidal ideation considered.” These are medically necessary, non-harmful statements, yet the current Azure OpenAI moderation pipeline can incorrectly treat them as unsafe, which disrupts psychiatric and behavioral-health workflows.

    Why this happens (strict moderation on the self-harm axis)

    Azure OpenAI’s safety filters are intentionally conservative when processing any content related to self-harm. The system is designed to prevent harmful outputs and automatically detect language that could indicate imminent risk. Because the model cannot inherently distinguish between a clinical assessment and a real expression of intent, the same terminology is often classified as high-risk. This leads to false positives even when the content is purely descriptive and used in a controlled clinical setting.

    Why exemptions aren’t immediate or runtime-tunable

    The core moderation system is centrally governed and cannot be disabled or bypassed through runtime parameters. Full removal of self-harm protections on general-purpose endpoints is not supported due to safety, regulatory, and legal requirements. That said, Azure does support carefully scoped exceptions for highly controlled, authenticated, HIPAA-compliant clinical environments. These require documented safeguards, security controls, and compliance justification they cannot be enabled automatically on standard resources.

    Short-term technical approaches that maintain safety and clinical fidelity While working within the existing moderation boundaries, there are several architectures that allow clinicians to maintain accurate medical records without removing safety protections:

    Speech → Raw Transcript → Clinician Review

    Use Azure Speech to capture the raw clinical transcript.

      Store the transcript securely within your HIPAA/BAA boundary.
      
         Allow clinicians to review or approve the raw text before any LLM interaction.
         
            Only send redacted or clinician-approved sections to the LLM for summarization or structuring.
            
            This ensures clinical accuracy while avoiding unnecessary moderation triggers.
            
            **Two-Stage Clinical Workflow**
            
               Stage 1: Transcribe and store the clinical note.
               
                  Stage 2: A clinician validates or flags sensitive phrasing.
                  
                     Only validated text proceeds to the LLM.
                     
                     This workflow maintains auditability and drastically reduces errors due to false positives.
                     
                     **Contextual Pre-Processing**
                     
                        Prefix notes with structured context such as: **`[CLINICAL_NOTE | PSYCHIATRY | AUTHENTICATED_CLINICIAN]`**
                        
                           Include system-level indicators of clinician authentication.
                           
                              While not a guaranteed bypass, providing explicit clinical context can help reduce moderation misclassification in some evaluation pathways.
                              
    

    What Microsoft typically evaluates in clinical contexts

    For scenarios involving medically sensitive terminology, Azure’s Product/Compliance teams generally expect to see safeguards such as:

    A signed Business Associate Agreement (BAA)

    Verified clinician authentication and identity controls

    Full audit logging for access and processing events

    Documented patient consent workflows

    Secure storage and retention policies

    A defined containment model ensuring only authorized clinicians can access PHI or unmoderated text

    Having these controls in place significantly strengthens the foundation for working with clinically sensitive language while staying within platform requirements.

    Also consider for improving reliability

    Review the content filtering policy documentation to ensure prompt phrasing follows allowed guidelines.

    Adjust prompt structure to minimize language that unintentionally maps to high-severity categories.

    Use Azure Monitor to track moderation events so you can identify specific patterns triggering blocks.

    Keep an eye on upcoming updates to the moderation stack, as Azure is actively refining the handling of medical and research-oriented scenarios.

    Also please refer this

    Azure OpenAI Content Filtering Documentation

    How to Configure Content Filters

    I Hope this helps. Do let me know if you have any further queries.

    Thank you!

    Was this answer helpful?

    0 comments No comments

Your answer

Answers can be marked as 'Accepted' by the question author and 'Recommended' by moderators, which helps users know the answer solved the author's problem.