Image descriptions
Azure AI Vision can analyze an image and generate a human-readable phrase that describes its contents. The algorithm returns several descriptions based on different visual features, and each description is given a confidence score. The final output is a list of descriptions ordered from highest to lowest confidence.
At this time, English is the only supported language for image description.
Try out the image captioning features quickly and easily in your browser using Vision Studio.
Image description example
The following JSON response illustrates what the Analyze API returns when describing the example image based on its visual features.
{
"description":{
"tags":[
"outdoor",
"city",
"white"
],
"captions":[
{
"text":"a city with tall buildings",
"confidence":0.48468858003616333
}
]
},
"requestId":"7e5e5cac-ef16-43ca-a0c4-02bd49d379e9",
"metadata":{
"height":300,
"width":239,
"format":"Png"
},
"modelVersion":"2021-05-01"
}
Use the API
The image description feature is part of the Analyze Image API. You can call this API through a native SDK or through REST calls. Include Description
in the visualFeatures query parameter. Then, when you get the full JSON response, parse the string for the contents of the "description"
section.
Next steps
Learn the related concepts of tagging images and categorizing images.
Σχόλια
https://aka.ms/ContentUserFeedback.
Σύντομα διαθέσιμα: Καθ' όλη τη διάρκεια του 2024 θα καταργήσουμε σταδιακά τα ζητήματα GitHub ως μηχανισμό ανάδρασης για το περιεχόμενο και θα το αντικαταστήσουμε με ένα νέο σύστημα ανάδρασης. Για περισσότερες πληροφορίες, ανατρέξτε στο θέμα:Υποβολή και προβολή σχολίων για