Edit

Share via


Storage management in Dataverse and finance and operations apps

As organizations accelerate their digital transformation journeys, the ability to manage data effectively becomes a strategic business imperative. With the rise of AI-powered applications and Copilot-driven workflows, enterprises are generating and consuming data at unprecedented rates. This data fuels innovation, allows personalized experiences, and supports critical decision-making—but only if it's governed and stored intelligently.

To support these evolving business needs, organizations must adopt a proactive, storage management strategy. This ensures that data no longer required for day-to-day operations is handled responsibly, freeing up capacity for high-value workloads, reducing operational friction, and aligning with compliance and audit requirements.

From a technical standpoint, effective storage management in Dataverse and Dynamics 365 enhances system performance, improves cost efficiency, and ensures adherence to long-term retention (LTR) policies. Both platforms offer tools and automation capabilities that empower organizations to manage storage.

By implementing the strategies outlined in this article, enterprises can reduce support overhead, streamline compliance, and unlock greater value from their business applications—turning storage from a constraint into a competitive advantage.

Key benefits

Effective storage management in Dataverse and Dynamics 365 provides several key benefits that address common, customer pain points and enhance overall operational efficiency.

  • Increased compliance with LTR: Effective storage management ensures that data is stored in compliance with LTR policies. This not only helps in meeting regulatory requirements, but also ensures that critical data is preserved and accessible when needed.

  • Improved performance: By optimizing storage management, organizations can significantly enhance the performance of their systems. Efficient storage allocation and management reduce latency and improve the speed of data retrieval, leading to smoother and faster operations.

  • Driving cost efficiency: Effective storage management empowers organizations to focus on high-value data by streamlining and decluttering their storage landscape. By retaining only what’s necessary, businesses can optimize their storage footprint, leading to smarter resource utilization and cost-effective scalability.

Background

As organizations grow and digitize more of their operations, the volume of business data stored in systems like Dataverse and Dynamics 365 increases steadily. This includes not only active, transactional data but also historical records that must be retained for audit, regulatory, or business continuity purposes. Over time, this accumulation can lead to performance degradation, increased operational overhead, and rising storage costs—especially when data that is no longer actively used remains in high-performance storage tiers.

A well-defined, storage management strategy helps organizations address these challenges by identifying data that can be archived, cleaned up, or moved to lower-cost, read-optimized storage. This is important for compliance scenarios where data must remain immutable, low-access, and read-only, such as financial records, audit logs, or regulatory filings. Ensuring that such data is retained in a compliant manner, without impacting the performance of live systems, is a key requirement for many enterprises.

By using the tools and strategies available in both platforms, organizations can gain better visibility into their storage footprint, reduce unnecessary consumption, and ensure that compliance-critical data is handled appropriately.

This article outlines practical approaches to storage management that help customers align their data retention practices with business and regulatory needs. This improves system performance, reduces operational overhead, and ensures that compliance obligations are met without compromise.

Why we store data

To select and optimize the right data retention pattern for your data, it's valuable to reflect on the reasons and uses for which we store data.

Operational data

With a business application, operational data is what is used to track sales or financial or supply chain actions.

This data needs to be accessed in real time, supporting customer and internal operational processes that record granular actions such as interactions with customers, orders, or inventory activities.

Over time, operational data may move from being actively used to infrequently used. The data might need to be accessible in near real time, to assist a customer with an order or in a support case. For example, consider the following scenarios:

  • A customer places an order, while another customer, who hasn't interacted with the business for some time, places an order.
  • Each order that has been placed and is being shipped, is being accessed constantly. There are also orders that are under a warranty period of three years that may need to be referenced for support and possibly require a refund.

This may lead to phases of operational data access needs such as:

  • Less than one year of actively accessed data.
  • Less than three years of infrequently accessed data.
  • More than three years where data is no longer operationally accessed.

The real time nature of operational storage does make that relatively expensive compared to other storage, so recognizing when data needs to be accessed operationally and when it doesn't is important for defining retention strategies.

Operational integration

As a specialized category of operational use, data may be required to be replicated between multiple operational systems, including patterns such as:

  • Banking: Customer relationship management for frontline customer interactions and replication to multiple banking systems. For example, you have current accounts, credit cards, mortgage, and credit check systems.
  • Manufacturing: Customer relationship management for frontline order taking and enterprise resource management system for supply chain management.
  • Police emergency handling: Customer relationship management for citizen interactions and a dispatch systems for police departments offer deployment management.

In these cases, while each system may have unique data it tracks, there's often common, master data that needs to be shared between the systems and kept in sync, leading to integration needs.

Audit data

A business typically has regulatory responsibility to keep data for extended periods—for example for seven years on average—for audit purposes, whether internal or external, such as supporting financial auditing, regulatory disclosure, or fraud review.

This data would typically span both data needed for operational purposes and data that is no longer needed, as it allows review across the data set from one place.

Analytics data

Organizations have a need to review and analyze the state of their business. They must measure and compare statistics over time, and spanning multiple or all parts of the business.

The large periods and breadth of data over which this analysis can occur leads to the need to replicate operational data into specialized, analytics tools. This avoids complex analytics from affecting the performance of operational systems, but also allows analysis across data sets that go beyond the period for which data is needed operationally. For example, you might need to compare data over seven years, rather than over one to two years. Differing analytics needs however may need the full data retention periods or only span the data retained in operational systems.

Analytics data typically allows for aggregation of data across multiple parts of the business and combines data from multiple systems.

Flow of data

The data of these types typically flow over time from operational data and then to transactional or historical data, as shown in the following image.

The flow of data.

Different types of storage

Dataverse storage types

Dataverse organizes storage into three main categories, each with distinct usage patterns and billing implications.

Storage type Description Common use cases
Database storage Stores structured data in tables—standard and custom. Business records, metadata, relationships, and configurations
File storage Stores attachments and binary data. Email attachments, images, documents uploaded through Power Apps
Log storage Stores audit logs and plugin trace logs. Change tracking, auditing, diagnostics, and compliance

Finance and operations platform storage types

Finance and operations storage is managed separately, but is increasingly being integrated into the Power Platform ecosystem. It includes the following storage types.

Storage type Description Common use cases
Operational database storage Core transactional data for finance, supply chain, human resources, and more Ledger entries, inventory, customer orders
Document management storage Binary large objects (Blobs) stored in Azure Blob Storage Invoices, receipts, scanned documents
Telemetry and diagnostic logs System logs and telemetry data Performance monitoring, issue diagnostics.

Shared and integrated storage scenarios

  • Dual-write storage

    • Allows real-time sync between Dataverse and finance and operations apps.
    • Requires careful role and capacity management to avoid duplication or overuse.
  • Long-term retention (LTR)

    • Moves historical data to a Managed Data Lake (MDL).
    • Reduces primary storage usage while maintaining compliance and analytics access.
    • Integrates with:
      • Quick Find (Dataverse-native search)
      • OneLake (Fabric-based analytics)
      • Synapse Link (custom lake analytics)

How your data grows over time

As organizations scale their use of Dataverse and the Dynamics 365 finance and operations platform, data growth becomes both a sign of success and a strategic challenge. What begins as a lean, transactional dataset can quickly evolve into a complex, multi-layered data estate. This section explores five key drivers of data growth and their implications for storage, performance, and governance.

Using data warehousing on operational data

To unlock insights from operational systems, many organizations use Azure Synapse Link, OneLake, or data export to replicate data from Dataverse and finance and operations apps into an analytical system. While this supports advanced reporting and AI workloads, it also introduces:

  • Redundant storage across operational and analytical layers

    Data is often duplicated between the operational and analytical environments. This redundancy increases overall storage consumption and may lead to higher costs, especially if historical data is retained indefinitely in both systems.

  • Schema duplication and versioning overhead

    To maintain consistency between systems, organizations must replicate schema changes—for example, new fields and renamed columns—across both operational and analytical layers. This adds complexity to data governance and increases the risk of schema drift, which can break downstream reports or models.

  • Increased retention of historical data for trend analysis

    Analytical systems typically retain data for longer periods to support trend analysis, forecasting, and regulatory reporting. While valuable, this long-term retention can lead to bloated datasets if not managed with proper archival and tiering strategies.

Data warehousing is essential for analytics, but without lifecycle policies, it can double or triple your storage footprint.

Using search on the data

Features like Dataverse search, Copilot indexing, and relevance search require indexing large volumes of structured and unstructured data. These indexes often:

  • Consume log and database storage

    Search indexes are stored in both log and database storage. As more tables and fields are marked as searchable, the index size grows proportionally. This can significantly impact overall storage usage, especially in environments with large volumes of records or frequent schema changes.

  • Persist even for unused or deprecated tables

    Even when certain tables are deprecated or no longer actively used, their associated search indexes may persist unless explicitly removed. This leads to unnecessary storage consumption and can complicate capacity planning.

  • Are often duplicated across environments, such as development, test, and production environments

    Search indexes are typically replicated across development, test, and production environments. While this ensures consistent search behavior, it also multiplies the storage footprint, particularly when environments are cloned or refreshed frequently.

Search improves usability and AI readiness, but index bloat is a silent contributor to storage overages.

Enabling logging on the data

Audit logs, plug-in trace logs, and telemetry are critical for compliance, debugging, and monitoring. However, note the following points:

  • Log storage grows linearly with usage and user count.

    Log data grows proportionally with:

    • The number of users and their activity levels
    • The volume of transactions and integrations
    • The complexity of business logic such as plug-ins and workflows

    In high-usage environments, this can lead to rapid expansion of log tables, consuming both database and log storage quotas.

  • Retention defaults are often too generous such as 90 days or more.

    By default, many logging features retain data for extended periods, such as 90 days or more. While this supports long-term traceability, it can result in unnecessary storage consumption, especially when logs aren't actively reviewed or exported.

  • System-generated logs are billed to the customer in Dataverse.

    In Dataverse, system-generated logs, including audit logs and plug-in trace logs, are counted against the customer’s storage entitlement. This means that without proper clean-up or export strategies, logging can directly contribute to storage overages and increased licensing costs.

Logging is non-negotiable for regulated industries, but must be paired with retention and export strategies, such as Azure Monitor or Log Analytics.

Having multiple copies of the production environment

To support development, testing, training, and troubleshooting, customers often create sandbox or cloned environments. Each copy:

  • Replicates the full data and index footprint.
  • May include nonobvious dependencies like search indexes, audit logs, and metadata.
  • Is rarely cleaned up after use.

Environment sprawl is a major driver of storage cost and complexity. Governance policies and automation are key to containment.

Optimization of queries on the data

As data volumes grow and application responsiveness becomes critical, customers and ISVs often implement various query optimization techniques to improve performance in Dataverse and Dynamics 365. These strategies are especially common in reporting, analytics, and integration-heavy scenarios.

To improve performance, customers and ISVs often create:

  • Custom indexes and materialized views

    These are used to accelerate query execution by precomputing joins or aggregations. They're helpful in scenarios involving complex filters or large datasets.

  • Denormalized tables for reporting

    To simplify reporting and reduce query complexity, developers often create flattened versions of relational data. These tables reduce the need for runtime joins and improve dashboard performance.

  • Caching layers or aggregates

    Frequently accessed data is sometimes preaggregated or cached in intermediate tables or external stores to reduce load on the primary database.

While these improve responsiveness, they also:

  • Increase storage usage

    Each optimization layer introduces more data structures, whether it's a copy of existing data in a denormalized format, a precomputed view, or a cache table. These structures often duplicate data already stored elsewhere, leading to a larger overall storage footprint. In environments with strict storage quotas or cost-based licensing models, like Dataverse, this can quickly escalate into avoidable overages.

  • Can become orphaned as apps evolve

    As applications evolve, some optimization artifacts may no longer be referenced by active reports, dashboards, or integrations. These orphaned objects continue to consume storage and may even slow down system operations, for example, during backups or indexing, if not identified and removed. Without regular audits, they can accumulate unnoticed, undermining the very performance gains they were created to support.

Query optimization is essential for scale but must be balanced with storage hygiene and telemetry-driven tuning.

Indexes and their impact on storage

Indexes are essential for improving query performance and using fast data retrieval in large datasets. In both Dataverse and Dynamics 365 finance and operations apps, indexes are automatically created for primary keys and frequently queried fields, and other custom indexes can be defined to support specific, business scenarios.

While indexes are critical for performance, they also have a direct impact on storage consumption, often underestimated during solution design.

How indexes consume storage

  • Physical duplication of data: Each index stores a copy of the indexed columns, along with pointers to the corresponding rows. The more columns and rows indexed, the larger the index size.

  • Growth with data volume: As the underlying table grows, so does the index. In high-transaction environments, indexes can grow rapidly, especially on large, denormalized tables or those with frequent inserts and updates.

  • Multiple indexes per table: It’s common for a single table to have multiple indexes, for example for search, filtering, sorting, and joins. Each other index adds to the cumulative storage footprint.

  • Search indexes in Dataverse: Features like Dataverse search and Copilot indexing create specialized indexes that span multiple fields and tables. These are stored in the DataverseSearch table and can consume significant space, especially when used across multiple environments such as development, test, and production environments.

  • System-generated indexes: Some indexes are created automatically by the platform such as for lookup fields or relationships. These may persist even if the associated tables are deprecated, unless explicitly removed.

Storage implications

  • Increased database and log storage: Indexes contribute to both database and log storage usage, which can affect licensing costs in Dataverse.
  • Environment duplication: When environments are copied or refreshed, all indexes are duplicated, amplifying storage usage across development, test, and production environments.
  • Maintenance overhead: Indexes must be updated as data changes, which can increase write latency and resource consumption.

Server-side synchronization impact on storage

Server-side synchronization in Dataverse allows seamless integration of emails, appointments, and tasks between Microsoft Exchange and Dataverse. While it enhances productivity and automation, it also contributes to storage consumption in the following ways.

  • Activity record creation: Each synchronized email or appointment generates an activity record in Dataverse, which includes metadata, body content, and potentially attachments.
  • Attachment storage: If attachments aren't filtered or offloaded, they are stored directly in Dataverse, increasing storage usage.
  • Compliance and retention: Organizations using server-side synchronization for compliance tracking may retain more data than necessary, further inflating storage.
  • Protected content: Even Purview-protected emails, though limited in content visibility, still generate placeholder records that consume space.

To manage this impact, enterprises should implement retention policies, consider offloading attachments, and monitor activity record volumes regularly.

How can I manage the ever-growing storage?

Whether you're already facing storage overages or aiming to stay ahead of them, managing data growth in Dataverse and the Dynamics 365 finance and operations platform requires a deliberate, policy-driven approach. This section outlines two strategic entry points: reactive remediation and proactive governance.

There are two possible scenarios:

  1. You want to proactively apply best practices to manage storage and avoid high costs in the future.
  2. You’re already in a situation where reducing storage size and cost is necessary.

Apply best practices to manage storage size and costs

Scenario 1: You want to proactively apply best practices to manage storage

If you're not yet in crisis mode, now is the time apply tools and techniques to manage the storage proactively.

Configure analytics for your data

As organizations grow, so does the need to extract insights from operational data, without impacting the performance of core business applications. Microsoft offers multiple ways to allow analytics on Dataverse and Dynamics 365 finance and operations data by integrating with your own data lake or warehouse.

Here are two, powerful options to consider:

Azure Synapse Link allows you to connect Dataverse directly to your own Azure Data Lake or Synapse workspace. This allows near real-time replication of operational data into an analytical environment, without writing complex ETL pipelines.

Benefits:

  • Run advanced analytics and AI models on live or near-live data.
  • Avoid performance impact on your production systems.
  • Use familiar tools like T-SQL, Spark, or Power BI for reporting.

Use case example: A retail company uses Synapse Link to analyze customer purchase behavior across regions, combining Dataverse customer relationship management data with external, market data in their own lake.

Option 2. Use OneLake – unified analytics with Microsoft Fabric

OneLake, part of Microsoft Fabric, provides a unified data lake experience where you can store and analyze data from multiple sources, including Dataverse and finance and operations apps, without duplication.

Benefits:

  • Centralized storage for all analytical workloads.
  • Native integration with Power BI, Synapse, and AI services.
  • Simplified governance and security across data domains.

Use case example: A financial services firm uses OneLake to consolidate operational data from finance and operations apps and Dataverse with external economic indicators, allowing real-time, risk modeling and executive dashboards. By doing this, you can decouple operational data from your core systems and allow scalable, cost-effective analytics by exporting that data to their own, analytical environments, without duplicating workloads or impacting performance.

Tools and techniques to reduce the storage

Dataverse offers several built-in tools and strategies to help administrators manage storage efficiently and maintain system performance.

Dataverse

Environment and data clean-up

  • Delete unused environments: You can delete an environment to recover storage space and to remove Personally Identifiable Information (PII).
  • Bulk deletion jobs: You can delete the following data in bulk:
    • Stale data or data that is irrelevant to the business.
    • Unneeded test or sample data.
    • Data that is incorrectly imported from other systems.

File and table optimization

Long-term retention (LTR) and archival

Search index optimization

  • Reduce Dataverse search: You can reduce the storage size by performing all the steps in Dataverse capacity-based storage details.
  • Reduce the size of DataverseSearch table: The DataverseSearch table is the cumulative storage used by the Dataverse search index. It includes the data from all searchable, retrievable, and filterable fields of the tables you indexed for your environment. You can reduce the table size by removing find columns, view columns, and filter conditions for one or more tables. You can turn off Dataverse search to remove all indexed data.
Finance and operations apps

Finance and operations apps provides flexible options for managing storage across production and sandbox environments.

Environment management

  • Limit the number of full production copies: You can reduce the overall storage consumption of finance and operations apps by removing full production copies in sandbox environments. For example, if you have five copies of production environments in a sandbox, your storage consumption is the sum of production plus five copies of production environments in a sandbox.
  • Trim data in sandbox environments: By trimming the data in a sandbox environment, you can reduce the overall storage footprint. You can follow the methods below to clean the data in the sandbox.
    • Restore process provides an opening and trimming execution
    • Write T-SQL
    • Write X++
  • Perform a transaction-less copy between environments: Environment copy for finance and operations apps has traditionally involved full database duplication, including configuration, master data, and transactions, which, while useful for debugging, significantly increases storage consumption across both finance and operations and Dataverse.

Custom clean-up and log management

  • Write custom clean-up routines as necessary: You can write custom clean-up routines as needed by your business to clean the unwanted data.
  • Avoid storing logs: You can move SysDatabaseLog to a less transactional database to reduce the overall storage footprint.

Archival and long-term retention

Built-in clean-up routines

  • Clean-up routines: In Dynamics 365 Finance and Dynamics 365 Supply Chain Management, clean-up routines are available in various modules. Clean-up routines provides an overview of the routines that are currently available. After copying the sandbox database, run these clean-up routines proactively to remove unnecessary tables, such as batch history, logs, and retail transaction history. Delete outdated or irrelevant data.
  • Archive credit card transaction data: Describes an archival job in Dynamics 365 Commerce that can help free up space in the database by archiving credit card payment tokens.

Reduce storage size and costs

Scenario 2: You're already in a situation where reducing storage size and cost is necessary

Assess what’s consuming storage

  • Use the Power Platform admin center and finance and operations storage reports to identify top-consuming tables, file types, and logs.
  • Use telemetry, if available, to attribute usage to specific apps, users, or business units.

Prioritize clean-up candidates

  • Focus on:
    • Staging and integration tables, such as dual-write buffers
    • Audit logs: Retain it in your own storage
    • Unused environments or sandboxes
    • Orphaned metadata and search indexes
    • Delete what you don’t need, for example bulk delete

Use Synapse Link and OneLake for analytical reporting

  • Export the analytical data to Synapse Link.
  • Use OneLake to access the retained data and business data for reporting and analytical purposes.

Apply long-term retention (LTR)

  • Move historical data to a Managed Data Lake (MDL) using LTR policies.
  • Maintain search and analytics access via Quick Find, Synapse Link, or OneLake.

Use cases

Use cases for storage management in Dataverse and finance and operations environments are critical for optimizing database space, enhancing system performance, and meeting regulatory requirements. Below are some typical scenarios that demonstrate how these strategies can be applied:

  • Managing growth of historical data

    • Scenario: A business has been live on Dynamics 365 for several years and has accumulated large volumes of historical transactions, and attachments.
    • Action: Implement long-term retention strategies to retain inactive data, reduce primary database size, and maintain compliance with audit requirements.
  • Compliance-driven data retention

    • Scenario: A regulated industry customer must retain financial or customer data for seven to ten years in a tamper-proof format.
    • Action: Use LTR to retain the immutable, read-only data in compliance with legal and regulatory requirements, while keeping business data lean without compromising on the analytics and reporting.
  • Search and Copilot index optimization

    • Scenario: Dataverse Search and Copilot indexing are enabled across all environments, including unused tables.
    • Action: Audit searchable fields and disable indexing for low-value or deprecated tables. Monitor the size of the DataverseSearch table and optimize configurations to reduce log and database storage.
  • Audit and telemetry management

    • Scenario: Plug-in trace logs and audit logs are growing rapidly, consuming storage and impacting performance.
    • Action: Export logs to external systems, like Azure Monitor, and automate clean-up of old entries to maintain visibility without bloating storage.
  • Data warehousing and analytics integration

    • Scenario: The organization replicates operational data to Azure Synapse or OneLake for analytics, leading to duplicated storage.
    • Action: Use incremental exports, apply filters, and avoid full dataset replication to minimize redundancy while allowing rich insights.
  • Reducing storage overages

    • Scenario: A customer receives a notification about exceeding their Dataverse storage quota, leading to unexpected costs.
    • Action: Use capacity reports to identify top-consuming tables, clean up obsolete environments, and remove unused attachments or logs. Consider moving cold data—typically historical or infrequently accessed records—to lower-cost storage tiers.
  • Optimizing performance in large tables

    • Scenario: Business-critical processes are slowing down due to large tables.
    • Action: Archive old records, clean up system jobs, for example AsyncOperationBase and WorkflowLogBase.
  • Environment lifecycle management

    • Scenario: Development and test environments are cloned from production, duplicating all data and indexes.
    • Action: Trim sandbox environments after a refresh, disable unnecessary search indexing, and remove test data to reduce redundant storage consumption. Delete unused sandbox environments to save storage.

Case studies

Case study 1: Reducing storage overages through index clean-up

Customer profile: A global manufacturing company using Dynamics 365 for supply chain and finance and operations apps.

Challenge: The customer was experiencing unexpected storage overages and performance degradation in their production environment. Investigation revealed that multiple custom indexes and materialized views, created during early implementation, were no longer in use but still consuming significant storage.

Solution: The team conducted a quarterly audit of all custom indexes and removed those not referenced by active queries or reports. They also implemented a governance policy to review new index requests before deployment.

Outcome:

  • Reduced database storage by 28%.
  • Improved query performance by 15%.
  • Avoided a projected $12,000 per year in other storage costs.

Case study 2: Archiving historical data to meet compliance and performance goals

Customer profile: A financial services firm using Dataverse and Dynamics 365 for customer onboarding and case management functionality.

Challenge: The firm needed to retain customer records for more than seven years to meet regulatory requirements, but the growing volume of inactive data was slowing down active workflows and increasing storage costs.

Solution: The customer implemented a long-term retention strategy using Dataverse’s archival capabilities. Inactive records were moved to a read-only, cost-optimized storage tier, while active data remained in high-performance storage.

Outcome:

  • Archived over 1.2 million records.
  • Reduced primary database size by 40%.
  • Maintained full auditability and compliance with retention policies.

Case study 3: Streamlining search indexes across environments

Customer profile: A retail organization with multiple Dataverse environments, including development, test, and production environments, supporting a Copilot-enabled customer relationship management solution.

Challenge: Search indexes were used across all environments, including unused tables and test data. This led to bloated DataverseSearch tables and unnecessary storage consumption.

Solution: The team reviewed searchable fields and stopped using indexing on noncritical tables in development and test environments. They also automated index clean-up during environment refreshes.

Outcome:

  • Reduced search index storage by 35%.
  • Improved environment refresh times by 20%.
  • Lowered overall log and database storage usage.

Case study 4: Using data export for analytics without duplicating storage

Customer profile: A healthcare provider using Dynamics 365 and Dataverse for patient engagement and billing.

Challenge: The analytics team needed access to operational data for trend analysis and AI modeling, but duplicating data into a separate warehouse was increasing storage costs and complexity.

Solution: The customer used Azure Synapse Link with incremental export and tiered storage in OneLake. They retained only essential analytical data and applied retention policies to manage historical depth.

Outcome:

  • Enabled real-time analytics without impacting operational systems.
  • Reduced redundant storage by 45%.
  • Improved governance over analytical data lifecycle.

Conclusion

Effective storage management is crucial for maintaining system performance and optimizing resource utilization in Dynamics 365 environments. The clean-up routines and archival jobs outlined in this article provide robust solutions to free up valuable database space and streamline operations. By using these tools like LTR and similar techniques, customers can address common storage challenges and create sustainable data management practices. Furthermore, real-world case studies demonstrate the efficacy of these approaches, offering insights into their practical applications. Adopting these strategies empowers organizations to proactively manage their storage needs and enhance overall efficiency.

References

Storage clean-up in Dataverse:

Storage clean-up in finance and operations:

Storage capacity: