By Default, Model deployed through serverless endpoint use content filters with medium threshold for text models and lower threshold for image models
Reference- https://learn.microsoft.com/en-us/azure/ai-foundry/concepts/model-catalog-content-safety
For Llama, It is not showing any option to configure custom content filter or set it as Non.
For other Azure AI infernece models ,We normally create custom content filters with higher severity threshold from safety and security tab and connect to a model deployment at "Deployment(Optional)" or update the deployments with custom content filter as shown in below screenshots. (Limited supported for Azure AI inference models)
You can also set Content filter as None (instead of custom content filter) for Chat completions and other models from Azure AI inference models aside text embedding, time series models
Custom Content filters
But all the model under serverless inference other than Azure AI inference is not supported, (specially timeseries and embedding models), You have to create an Azure content safety resource and pass the text inputs before/after passing it to the Serverless endpoint (llama 3.1 endpoint) to moderate it further.
Content Safety response
When a content filter is triggered, a 400-error code is returned with the description of the rule triggered.
You can use the sample code from "consume" tab of serverless endpoint and chain it with content safety codes mentioned here in content safety documentation.
Hope it addresses your queries.
Thank you.