Hi Pierre-Yves,
Greetings & Welcome to the Microsoft Q&A forum! Thank you for sharing your query.
I understand that you're experiencing these issues with Azure OpenAI's content filtering. Here are some insights and potential solutions to address your concerns:
- Why are completely inoffensive phrases being blocked by Azure OpenAI’s content filter?
False positives can occur due to the content filtering system's sensitivity and the complexity of language nuances, especially in different languages. The models are designed to err on the side of caution to prevent harmful content, which can sometimes lead to over-blocking.
- Is filtering stricter in French than in other languages?
The content filtering models have been trained and tested on multiple languages, including French. However, variations in language structure and context can lead to differences in filtering accuracy. It is possible that the models are more sensitive in French due to these nuances.
- Is there a way to adjust filtering levels without disabling moderation entirely?
Yes, you can customize the severity settings for different harm categories. This allows you to fine-tune the sensitivity of the content filtering models to better suit your application's needs. You can also use blocklists to manage specific terms or phrases that might be causing false positives.
- How can we report a false positive to Microsoft to improve the filtering system?
You can report false positives through the Azure portal or by contacting Azure support. Providing detailed examples and context will help the team improve the filtering models.
- Have other users encountered this issue in similar chatbot use cases?
Yes, other users have reported similar issues, especially when dealing with multilingual applications. Continuous feedback and testing are essential to refine the content filtering system.
Steps to Mitigate False Positives:
Review and Verification: Confirm that the flagged content is indeed a false positive by checking the context and comparing it against content safety risk categories.
Customize Severity Settings: Adjust the severity threshold for different harm categories to reduce false positives.
Use Blocklists: Implement blocklists to manage specific terms or phrases that might be causing false positives.
Testing and Feedback: Continuously test and provide feedback on the content filtering system to improve its accuracy.
For more information, please refer: Azure AI Content Safety documentation
I hope this information helps.