Share via

Azure computer vision and text moderation scoring criteria

2JK 241 Reputation points
2021-10-25T09:24:31.693+00:00

I have a couple of questions regarding 1) the scoring criteria for the computer vision analyze image API, particularly concerning the adult, racy and gory classification scores, and 2) the text moderation scoring.

1- I tested some samples where I used an image of a flag and it gave a high racy score (0.8+) and a high adult score (0.7+) but isAdultContent and isRacyContent are both false. Why is that? Firstly, it shouldn't have score a flag that high, but is the scores not related to the boolean values at all?

2- For text moderation (in Content Moderator screen text), it's giving a very high score for offensive classification (0.98+) as soon as it encounters a word that may be used in offensive context, even if the sentence sentiment is positive. Can this be addressed or is it just a limitation?

Community Center | Not monitored
Foundry Tools
Foundry Tools

Formerly known as Azure AI Services or Azure Cognitive Services is a unified collection of prebuilt AI capabilities within the Microsoft Foundry platform

Azure Vision in Foundry Tools

Answer accepted by question author

  1. YutongTie-9091 54,021 Reputation points Moderator
    2021-10-28T17:27:15.237+00:00

    @2JK

    Thanks for the waiting. I have checked this situation with the pm of content moderator team, but unluckily, there is a limitation of this product. Below is the response I get, I am sorry for this experience but we do work on that.

    1. Adult/Racy image classifier (available in CM) – this classifier was developed by Bing, who is not actively supporting now and we do not have the details on what triggers the Boolean “isImageAdultClassified” response in relationship to the actual classification scores.

    a) The Gore score that the customer mentions is actually a response from a Custom Vision classifier that the customer has created a “connector” from CM to Custom Vision. The image is sent to CM, and then a custom workflow created by the customer sends that image to Custom Vision as well as being scanned by the classifiers available in CM. The responses from all the classifiers are then combined in the response to the customer. Because of that – we also do not have visibility/understanding as to “why” the Joker image returns a high Gore score.

    both values are provided so that the customer can make a decision on what to trigger a “violation” off of. If they feel that the Boolean doesn’t make sense (because of the high classification score), they can choose to just go with the classification score.

    b) The questions around the flag and the meme and why they are scoring high on Adult and Racy are also not something we can speak to since we didn’t develop that classifier.

    1. Text scanning – It sounds like the issue is about text that in certain context is not offensive. If that is indeed the case, the “exact match” text scanning (the default for CM) is only indicating the presence of terms that have been identified as offensive. There are two text scanners – the exact match I mention above and then the text classifier (also developed by Bing, which would return the classification score) – if a sentence is marked with a high classification score but is contextually “not offensive” – that sounds like a limitation to that classifier.

    Regards,
    Yutong

    1 person found this answer helpful.

0 additional answers

Sort by: Most helpful

Your answer

Answers can be marked as 'Accepted' by the question author and 'Recommended' by moderators, which helps users know the answer solved the author's problem.