권홍욱(AI 개발) - GTF Greetings & Welcome to Microsoft Q&A forum!
I want to use Azure openai API to provide a service utilizing LLM. I would like to know if MS(or Azure) stores any form of input data or prompt.
Microsoft takes data privacy and security very seriously, and has implemented measures to ensure that personally identifiable information (PII) is not shared with OpenAI during the usage of Azure OpenAI Service. The same has been mentioned in the FAQs as well. For more information, see the Azure OpenAI data, privacy, and security guide.
Of course, the 'Data, privacy, and security for Azure OpenAI Service' page mentions "NOT used to improve any Microsoft or 3rd party products or services.", but my client doesn't want any data to be leaked out in any form other than to improve the product (e.g. logged).
It is clearly mentioned in the page Data, privacy, and security for Azure OpenAI Service which you have already referred. Your prompts (inputs) and completions (outputs), your embeddings, and your training data:
- are NOT available to other customers.
- are NOT available to OpenAI.
- are NOT used to improve OpenAI models.
- are NOT used to improve any Microsoft or 3rd party products or services.
- are NOT used for automatically improving Azure OpenAI models for your use in your resource (The models are stateless, unless you explicitly fine-tune models with your training data).
- Your fine-tuned Azure OpenAI models are available exclusively for your use.
The Azure OpenAI Service is fully controlled by Microsoft; Microsoft hosts the OpenAI models in Microsoft’s Azure environment and the Service does NOT interact with any services operated by OpenAI (e.g. ChatGPT, or the OpenAI API).
If there is any possibility that input data or prompts could be leaked in any form to outside parties, including MS,
When you use Azure OpenAI Service, your data is encrypted in transit and at rest, and is stored in Microsoft Azure data centers that are compliant with various industry standards and regulations, such as ISO 27001, HIPAA, and GDPR. Microsoft also provides tools and services to help you manage and protect your data, such as Azure Key Vault and Azure Active Directory. See Azure OpenAI Service encryption of data at rest for more details.
You can also find more information about Microsoft's data privacy and security policies on the Microsoft Trust Center.
I would like to know if there is anything else you can suggest to prevent this from happening. (For example, if I use a fine-tuning model, it will never leak outside).
Please assure that your data is not leaked or shared as mentioned above.
Also, Azure OpenAI Service on your data empowers you to unlock the full potential of your data by running OpenAI models directly on it, eliminating the need for training or fine-tuning.
Please check Azure OpenAI on your data (preview) and the blogpost Introducing Azure OpenAI Service On Your Data in Public Preview for more details.
Hope this helps. Do let us know if you have any further queries.
If this answers your query, do click Accept Answer and Yes for was this answer helpful. Doing so would help other community members with similar issue identify the solution. I highly appreciate your contribution to the community.