Exercise - Create a storage account using the Azure portal
In this unit, you'll use the Azure portal to create a storage account for a fictitious southern California surf report web app. The surf report site lets users upload photos and videos of local beach conditions. Viewers will use the content to help them choose the beach with the best surfing conditions.
Your list of design and feature goals is:
- Video content must load quickly.
- The site must handle unexpected spikes in upload volume.
- Outdated content must be removed as surf conditions change so the site always shows current conditions.
You decide to buffer uploaded content in an Azure Queue for processing and then transfer it to an Azure Blob for persistent storage. You need a storage account that can hold both queues and blobs while delivering low-latency access to your content.
Create a storage account using Azure portal
Sign in to the Azure portal using the same account you used to activate the sandbox.
On the resource menu, or from the Home page, select Storage accounts. The Storage accounts pane appears.
On the command bar, select Create. The Create a storage account pane appears.
On the Basics tab, enter the following values for each setting.
Setting Value Project details Subscription Concierge Subscription Resource group [sandbox resource group name]from the dropdown list. Instance details Storage account name Enter a unique name. This name will be used to generate the public URL to access the data in the account. The name must be unique across all existing storage account names in Azure. Names must have 3 to 24 characters and can contain only lowercase letters and numbers. Region Select a location near to you from the dropdown list. Performance Standard. This option decides the type of disk storage used to hold the data in the Storage account. Standard uses traditional hard disks, and Premium uses solid-state drives (SSD) for faster access. Redundancy Select Locally redundant storage (LRS) from the dropdown list. In our case, the images and videos quickly become out-of-date and are removed from the site. As a result, there's little value to paying extra for global redundancy. If a catastrophic event results in data loss, you can restart the site with fresh content from your users.
Select Next : Advanced. On the Advanced tab, enter the following values for each setting.
Setting Value Security Require secure transfer for REST API operations Check. This setting controls whether HTTP can be used for the REST APIs that access data in the storage account. Setting this option to enable forces all clients to use SSL (HTTPS). Most of the time, you'll want to set secure transfer to enable; using HTTPS over the network is considered a best practice. Enable blob public access Check. We'll allow clients to read data in that container without authorizing the request. Enable storage account key access Check. We'll allow clients to access data via SAS. Default to Azure Active Directory authorization in the Azure portal Uncheck. Clients are public, not part of an Active Directory. Minimum TLS version Select Version 1.2 from dropdown list. TLS 1.2 is a secure version of TLS, and Azure Storage uses it on public HTTPS endpoints. TLS 1.1 and 1.0 are supported for backwards compatibility. See Warning at end of table. Permitted scope for copy operations Accept default Data Lake Storage Gen2 Enable hierarchical namespace Uncheck. Data Lake hierarchical namespace is for big-data applications that aren't relevant to this module. Blob storage Enable SFTP Uncheck. SFTP is disabled by default and isn't relevant to this module. Enable network file system v3 Uncheck (default). Allow cross-tenant replication Uncheck. Active Directory is not being used for this exercise. Access tier Hot. This setting is only used for Blob storage. The Hot access tier is ideal for frequently accessed data; the Cool access tier is better for infrequently accessed data. This setting only sets the default value. When you create a Blob, you can set a different value for the data. In our case, we want the videos to load quickly, so we'll use the high-performance option for our blobs. Azure Files Enable large file shares Uncheck. Large file shares provide support up to a 100 TiB, however this type of storage account can't convert to a Geo-redundant storage offering, and upgrades are permanent.
If Enable large file shares is selected, it will enforce additional restrictions, and Azure files service connections without encryption will fail, including scenarios using SMB 2.1 or 3.0 on Linux. Because Azure storage doesn't support SSL for custom domain names, this option cannot be used with a custom domain name.
Select Next : Networking. On the Networking tab, enter the following values for each setting.
Setting Value Network connectivity Connectivity method Enable public access from all networks. We want to allow public Internet access. Our content is public facing, and we need to allow access from public clients. Network routing Routing preference Microsoft network routing. We want to make use of the Microsoft global network that is optimized for low-latency path selection.
Select Next : Data protection. On the Data protection tab, enter the following values for each setting.
Setting Value Recovery Enable point-in-time restore for containers Uncheck. Not necessary for this implementation. Enable soft delete for blobs Uncheck. Soft delete lets you recover blob data in cases where blobs or blob snapshots are deleted accidentally or overwritten. Enable soft delete for containers Uncheck. Soft delete lets you recover your containers that are deleted accidentally. Enable soft delete for file shares Uncheck. File share soft delete lets you recover your accidentally deleted file share data more easily. Tracking Enable versioning for blobs Uncheck. Not necessary for this implementation. Enable blob change feed Uncheck. Not necessary for this implementation. Access control Enable version-level immutability support Uncheck. Not necessary for this implementation.
Select Next : Encryption. Accept the defaults.
Select Next : Tags. Here, you can associate key/value pairs with the account for your categorization to determine if a feature is available to selected Azure resources.
Select Review + create to validate your options and to ensure all the required fields are selected. If there are issues, this tab will identify them so you can correct them.
When validation passes successfully, select Create to deploy the storage account.
When deployment is complete, which may take up to two minutes, select Go to resource to view Essential details about your new storage account.
You created a storage account with settings driven by your business requirements. For example, you might have selected a West US datacenter because your customers were primarily located in southern California. The typical flow for creating a storage account is: first analyze your data and goals, and then configure the storage account options to match.
Need help? See our troubleshooting guide or provide specific feedback by reporting an issue.