Azure Data Factory (ADF) allows you to perform incremental data loading by capturing changes from a source system. Typically, this is done by tracking a specific column, such as an updated_at
attribute, that indicates when a record was last changed.
To perform a full load again in the future, you'd essentially need to reset the point from which the incremental load is done. I am assuming the following since you didn't provide enough information :
If you're using a watermark table or a system variable to track the last updated value, you could manually reset this value to a point in time before your data (e.g., '1970-01-01'). The next time your pipeline runs, it will see that it needs to load all changes since that early date and will, therefore, perform a full load.
You might control the incremental load through pipeline parameters. In this case, you could alter the parameter to trigger a full load. If you're using the updated_at
attribute as a parameter, you could expose a way to manually set that parameter to an earlier date, triggering a full load.
You might build your pipeline in such a way that you can manually trigger a full load through an additional trigger or manual process. This could involve a separate pipeline or an alteration to the existing one that allows you to bypass the incremental load process.
If the trigger has been set with the incremental attribute (e.g., a tumbling window trigger), you might need to change the configuration to ensure it pulls the full dataset.
If all else fails, you could recreate the pipeline or use a different pipeline specifically for full loads. Depending on your design, this could be a last-resort option if the pipeline is not easily reconfigurable.