Train a custom model using the Sample Labeling tool
This content applies to: v2.1.
- For an enhanced experience and advanced model quality, try the Document Intelligence v3.0 Studio.
- The v3.0 Studio supports any model trained with v2.1 labeled data.
- You can refer to the API migration guide for detailed information about migrating from v2.1 to v3.0.
In this article, you use the Document Intelligence REST API with the Sample Labeling tool to train a custom model with manually labeled data.
You need the following resources to complete this project:
- Azure subscription - Create one for free
- Once you have your Azure subscription, create a Document Intelligence resource in the Azure portal to get your key and endpoint. After it deploys, select Go to resource.
- You need the key and endpoint from the resource you create to connect your application to the Document Intelligence API. You paste your key and endpoint into the code later in the quickstart.
- You can use the free pricing tier (
F0) to try the service, and upgrade later to a paid tier for production.
- A set of at least six forms of the same type. You use this data to train the model and test a form. You can use a sample data set (download and extract sample_data.zip) for this quickstart. Upload the training files to the root of a blob storage container in a standard-performance-tier Azure Storage account.
Create a Document Intelligence resource
Go to the Azure portal and create a new Document Intelligence resource . In the Create pane, provide the following information:
|Subscription||Select the Azure subscription which has been granted access.|
|Resource group||The Azure resource group that contains your resource. You can create a new group or add it to a pre-existing group.|
|Region||The location of your Azure AI services resource. Different locations may introduce latency, but have no impact on the runtime availability of your resource.|
|Name||A descriptive name for your resource. We recommend using a descriptive name, for example MyNameFormRecognizer.|
|Pricing tier||The cost of your resource depends on the pricing tier you choose and your usage. For more information, see the API pricing details.|
|Review + create||Select the Review + create button to deploy your resource on the Azure portal.|
Retrieve the key and endpoint
When your Document Intelligence resource finishes deploying, find and select it from the All resources list in the portal. Your key and endpoint will be located on the resource's Key and Endpoint page, under Resource Management. Save both of these to a temporary location before going forward.
Try it out
Try out the Document Intelligence Sample Labeling tool online:
Set up the Sample Labeling tool
If your storage data is behind a VNet or firewall, you must deploy the Document Intelligence Sample Labeling tool behind your VNet or firewall and grant access by creating a system-assigned managed identity.
You use the Docker engine to run the Sample Labeling tool. Follow these steps to set up the Docker container. For a primer on Docker and container basics, see the Docker overview.
The OCR Form Labeling Tool is also available as an open source project on GitHub. The tool is a TypeScript web application built using React + Redux. To learn more or contribute, see the OCR Form Labeling Tool repo. To try out the tool online, go to the Document Intelligence Sample Labeling tool website.
First, install Docker on a host computer. This guide shows you how to use local computer as a host. If you want to use a Docker hosting service in Azure, see the Deploy the Sample Labeling tool how-to guide.
The host computer must meet the following hardware requirements:
Container Minimum Recommended Sample Labeling tool
2core, 4-GB memory
4core, 8-GB memory
Install Docker on your machine by following the appropriate instructions for your operating system:
Get the Sample Labeling tool container with the
docker pull mcr.microsoft.com/azure-cognitive-services/custom-form/labeltool:latest-2.1
Now you're ready to run the container with
docker run -it -p 3000:80 mcr.microsoft.com/azure-cognitive-services/custom-form/labeltool:latest-2.1 eula=accept
This command makes the sample-labeling tool available through a web browser. Go to
You can also label documents and train models using the Document Intelligence REST API. To train and Analyze with the REST API, see Train with labels using the REST API and Python.
Set up input data
First, make sure all the training documents are of the same format. If you have forms in multiple formats, organize them into subfolders based on common format. When you train, you need to direct the API to a subfolder.
Configure cross-domain resource sharing (CORS)
Enable CORS on your storage account. Select your storage account in the Azure portal and then choose the CORS tab on the left pane. On the bottom line, fill in the following values. Select Save at the top.
- Allowed origins = *
- Allowed methods = [select all]
- Allowed headers = *
- Exposed headers = *
- Max age = 200
Connect to the Sample Labeling tool
The Sample Labeling tool connects to a source (your original uploaded forms) and a target (created labels and output data).
Connections can be set up and shared across projects. They use an extensible provider model, so you can easily add new source/target providers.
To create a new connection, select the New Connections (plug) icon, in the left navigation bar.
Fill in the fields with the following values:
Display Name - The connection display name.
Description - Your project description.
SAS URL - The shared access signature (SAS) URL of your Azure Blob Storage container. To retrieve the SAS URL for your custom model training data, go to your storage resource in the Azure portal and select the Storage Explorer tab. Navigate to your container, right-click, and select Get shared access signature. It's important to get the SAS for your container, not for the storage account itself. Make sure the Read, Write, Delete and List permissions are checked, and click Create. Then copy the value in the URL section to a temporary location. It should have the form:
https://<storage account>.blob.core.windows.net/<container name>?<SAS value>.
Create a new project
In the Sample Labeling tool, projects store your configurations and settings. Create a new project and fill in the fields with the following values:
- Display Name - the project display name
- Security Token - Some project settings can include sensitive values, such as keys or other shared secrets. Each project generates a security token that can be used to encrypt/decrypt sensitive project settings. You can find security tokens in the Application Settings by selecting the gear icon at the bottom of the left navigation bar.
- Source Connection - The Azure Blob Storage connection you created in the previous step that you would like to use for this project.
- Folder Path - Optional - If your source forms are located in a folder on the blob container, specify the folder name here
- Document Intelligence Service Uri - Your Document Intelligence endpoint URL.
- Key - Your Document Intelligence key.
- Description - Optional - Project description
Label your forms
When you create or open a project, the main tag editor window opens. The tag editor consists of three parts:
- A resizable v3.0 pane that contains a scrollable list of forms from the source connection.
- The main editor pane that allows you to apply tags.
- The tags editor pane that allows users to modify, lock, reorder, and delete tags.
Identify text and tables
Select Run Layout on unvisited documents on the left pane to get the text and table layout information for each document. The labeling tool draws bounding boxes around each text element.
The labeling tool also shows which tables have been automatically extracted. Select the table/grid icon on the left hand of the document to see the extracted table. In this quickstart, because the table content is automatically extracted, we don't label the table content, but rather rely on the automated extraction.
In v2.1, if your training document doesn't have a value filled in, you can draw a box where the value should be. Use Draw region on the upper left corner of the window to make the region taggable.
Apply labels to text
Next, you create tags (labels) and apply them to the text elements that you want the model to analyze.
- First, use the tags editor pane to create the tags you'd like to identify.
- Select + to create a new tag.
- Enter the tag name.
- Press Enter to save the tag.
- In the main editor, select words from the highlighted text elements or a region you drew in.
- Select the tag you want to apply, or press the corresponding keyboard key. The number keys are assigned as hotkeys for the first 10 tags. You can reorder your tags using the up and down arrow icons in the tag editor pane.
- Follow these steps to label at least five of your forms.
Keep the following tips in mind when you're labeling your forms:
- You can only apply one tag to each selected text element.
- Each tag can only be applied once per page. If a value appears multiple times on the same form, create different tags for each instance. For example: "invoice# 1", "invoice# 2" and so on.
- Tags cannot span across pages.
- Label values as they appear on the form; don't try to split a value into two parts with two different tags. For example, an address field should be labeled with a single tag even if it spans multiple lines.
- Don't include keys in your tagged fields—only the values.
- Table data should be detected automatically and will be available in the final output JSON file. However, if the model fails to detect all of your table data, you can manually tag these fields as well. Tag each cell in the table with a different label. If your forms have tables with varying numbers of rows, make sure you tag at least one form with the largest possible table.
- Use the buttons to the right of the + to search, rename, reorder, and delete your tags.
- To remove an applied tag without deleting the tag itself, select the tagged rectangle on the document view and press the delete key.
Specify tag value types
You can set the expected data type for each tag. Open the context menu to the right of a tag and select a type from the menu. This feature allows the detection algorithm to make assumptions that improve the text-detection accuracy. It also ensures that the detected values are returned in a standardized format in the final JSON output. Value type information is saved in the fields.json file in the same path as your label files.
The following value types and variations are currently supported:
- Formatted as a Floating point value.
- Example: 1234.98 on the document is formatted into 1234.98 on the output
- Formatted as an integer value.
- Example: 1234.98 on the document is formatted into 123498 on the output.
See these rules for date formatting:
You must specify a format (
ymd) for date formatting to work.
The following characters can be used as date delimiters:
, - / . \. Whitespace cannot be used as a delimiter. For example:
The day and month can each be written as one or two digits, and the year can be two or four digits:
If a date string has eight digits, the delimiter is optional:
- 01 01 2020
The month can also be written as its full or short name. If the name is used, delimiter characters are optional. However, this format may be recognized less accurately than others.
- 01 Jan 2020
Label tables (v2.1 only)
At times, your data might lend itself better to being labeled as a table rather than key-value pairs. In this case, you can create a table tag by selecting Add a new table tag. Specify whether the table has a fixed number of rows or variable number of rows depending on the document and define the schema.
Once you've defined your table tag, tag the cell values.
Train a custom model
Choose the Train icon on the left pane to open the Training page. Then select the Train button to begin training the model. Once the training process completes, you see the following information:
- Model ID - The ID of the model that was created and trained. Each training call creates a new model with its own ID. Copy this string to a secure location; you need it if you want to do prediction calls through the REST API or client library guide.
- Average Accuracy - The model's average accuracy. You can improve model accuracy by adding and labeling more forms, then retraining to create a new model. We recommend starting by labeling five forms and adding more forms as needed.
- The list of tags, and the estimated accuracy per tag.
After training finishes, examine the Average Accuracy value. If it's low, you should add more input documents and repeat the labeling steps. The documents you've already labeled remain in the project index.
You can also run the training process with a REST API call. To learn how to do this, see Train with labels using Python.
Compose trained models
With Model Compose, you can compose up to 200 models to a single model ID. When you call Analyze with the composed
modelID, Document Intelligence classifies the form you submitted, choose the best matching model, and then return results for that model. This operation is useful when incoming forms may belong to one of several templates.
- To compose models in the Sample Labeling tool, select the Model Compose (merging arrow) icon from the navigation bar.
- Select the models you wish to compose together. Models with the arrows icon are already composed models.
- Choose the Compose button. In the pop-up, name your new composed model and select Compose.
- When the operation completes, your newly composed model should appear in the list.
Analyze a form
Select the Analyze icon from the navigation bar to test your model. Select source Local file. Browse for a file and select a file from the sample dataset that you unzipped in the test folder. Then choose the Run analysis button to get key/value pairs, text and tables predictions for the form. The tool applies tags in bounding boxes and reports the confidence of each tag.
You can also run the Analyze API with a REST call. To learn how to do this, see Train with labels using Python.
Depending on the reported accuracy, you may want to do further training to improve the model. After you've done a prediction, examine the confidence values for each of the applied tags. If the average accuracy training value is high, but the confidence scores are low (or the results are inaccurate), add the prediction file to the training set, label it, and train again.
The reported average accuracy, confidence scores, and actual accuracy can be inconsistent when the analyzed documents differ from documents used in training. Keep in mind that some documents look similar when viewed by people but can look distinct to the AI model. For example, you might train with a form type that has two variations, where the training set consists of 20% variation A and 80% variation B. During prediction, the confidence scores for documents of variation A are likely to be lower.
Save a project and resume later
To resume your project at another time or in another browser, you need to save your project's security token and reenter it later.
Get project credentials
Go to your project settings page (slider icon) and take note of the security token name. Then go to your application settings (gear icon), which shows all of the security tokens in your current browser instance. Find your project's security token and copy its name and key value to a secure location.
Restore project credentials
When you want to resume your project, you first need to create a connection to the same blob storage container. To do so, repeat the steps. Then, go to the application settings page (gear icon) and see if your project's security token is there. If it isn't, add a new security token and copy over your token name and key from the previous step. Select Save to retain your settings.
Resume a project
Finally, go to the main page (house icon) and select Open Cloud Project. Then select the blob storage connection, and select your project's
.fott file. The application loads all of the project's settings because it has the security token.
In this quickstart, you've learned how to use the Document Intelligence Sample Labeling tool to train a model with manually labeled data. If you'd like to build your own utility to label training data, use the REST APIs that deal with labeled data training.