title | description | author | ms.author | ms.service | ms.subservice | ms.topic | ms.date | ms.custom |
---|---|---|---|---|---|---|---|---|
Copy data by using the Azure Copy Data tool |
Create an Azure Data Factory and then use the Copy Data tool to copy data from one location in Azure Blob storage to another location. |
dearandyxu |
yexu |
data-factory |
tutorials |
quickstart |
07/05/2021 |
mode-other |
[!div class="op_single_selector" title1="Select the version of Data Factory service that you are using:"]
[!INCLUDEappliesto-adf-xxx-md]
In this quickstart, you use the Azure portal to create a data factory. Then, you use the Copy Data tool to create a pipeline that copies data from a folder in Azure Blob storage to another folder.
Note
If you are new to Azure Data Factory, see Introduction to Azure Data Factory before doing this quickstart.
[!INCLUDE data-factory-quickstart-prerequisites]
-
Launch Microsoft Edge or Google Chrome web browser. Currently, Data Factory UI is supported only in Microsoft Edge and Google Chrome web browsers.
-
Go to the Azure portal.
-
From the Azure portal menu, select Create a resource > Integration > Data Factory:
:::image type="content" source="./media/doc-common-process/new-azure-data-factory-menu.png" alt-text="New data factory creation":::
-
On the New data factory page, enter ADFTutorialDataFactory for Name.
The name of the Azure Data Factory must be globally unique. If you see the following error, change the name of the data factory (for example, <yourname>ADFTutorialDataFactory) and try creating again. For naming rules for Data Factory artifacts, see the Data Factory - naming rules article.
:::image type="content" source="./media/doc-common-process/name-not-available-error.png" alt-text="Error when a name is not available":::
-
For Subscription, select your Azure subscription in which you want to create the data factory.
-
For Resource Group, use one of the following steps:
- Select Use existing, and select an existing resource group from the list.
- Select Create new, and enter the name of a resource group.
To learn about resource groups, see Using resource groups to manage your Azure resources.
-
For Version, select V2.
-
For Location, select the location for the data factory.
The list shows only locations that Data Factory supports, and where your Azure Data Factory meta data will be stored. The associated data stores (like Azure Storage and Azure SQL Database) and computes (like Azure HDInsight) that Data Factory uses can run in other regions.
-
Select Create.
-
After the creation is complete, you see the Data Factory page. Select Open on the Open Azure Data Factory Studio tile to start the Azure Data Factory user interface (UI) application on a separate tab.
:::image type="content" source="./media/doc-common-process/data-factory-home-page.png" alt-text="Home page for the Azure Data Factory, with the Open Azure Data Factory Studio tile.":::
-
On the home page of Azure Data Factory, select the Ingest tile to start the Copy Data tool.
:::image type="content" source="./media/doc-common-process/get-started-page.png" alt-text="Screenshot that shows the Azure Data Factory home page.":::
-
On the Properties page of the Copy Data tool, choose Built-in copy task under Task type, then select Next.
:::image type="content" source="./media/quickstart-create-data-factory-copy-data-tool/copy-data-tool-properties-page.png" alt-text=""Properties" page":::
-
On the Source data store page, complete the following steps:
-
Click + Create new connection to add a connection.
-
Select the linked service type that you want to create for the source connection. In this tutorial, we use Azure Blob Storage. Select it from the gallery, and then select Continue.
:::image type="content" source="./media/quickstart-create-data-factory-copy-data-tool/select-blob-source.png" alt-text="Select Blob":::
-
On the New connection (Azure Blob Storage) page, specify a name for your connection. Select your Azure subscription from the Azure subscription list and your storage account from the Storage account name list, test connection, and then select Create.
:::image type="content" source="./media/quickstart-create-data-factory-copy-data-tool/configure-blob-storage.png" alt-text="Configure the Azure Blob storage account":::
-
Select the newly created connection in the Connection block.
-
In the File or folder section, select Browse to navigate to the adftutorial/input folder, select the emp.txt file, and then click OK.
-
Select the Binary copy checkbox to copy file as-is, and then select Next.
:::image type="content" source="./media/quickstart-create-data-factory-copy-data-tool/source-data-store.png" alt-text="Screenshot that shows the Source data store page.":::
-
-
On the Destination data store page, complete the following steps:
-
Select the AzureBlobStorage connection that you created in the Connection block.
-
In the Folder path section, enter adftutorial/output for the folder path.
:::image type="content" source="./media/quickstart-create-data-factory-copy-data-tool/destination-data-store.png" alt-text="Screenshot that shows the Destination data store page.":::
-
Leave other settings as default and then select Next.
-
-
On the Settings page, specify a name for the pipeline and its description, then select Next to use other default configurations.
:::image type="content" source="./media/quickstart-create-data-factory-copy-data-tool/settings.png" alt-text="Screenshot that shows the settings page.":::
-
On the Summary page, review all settings, and select Next.
-
On the Deployment complete page, select Monitor to monitor the pipeline that you created.
:::image type="content" source="./media/quickstart-create-data-factory-copy-data-tool/deployment-page.png" alt-text=""Deployment complete" page":::
-
The application switches to the Monitor tab. You see the status of the pipeline on this tab. Select Refresh to refresh the list. Click the link under Pipeline name to view activity run details or rerun the pipeline.
:::image type="content" source="./media/quickstart-create-data-factory-copy-data-tool/refresh-pipeline.png" alt-text="Refresh pipeline":::
-
On the Activity runs page, select the Details link (eyeglasses icon) under the Activity name column for more details about copy operation. For details about the properties, see Copy Activity overview.
-
To go back to the Pipeline Runs view, select the All pipeline runs link in the breadcrumb menu. To refresh the view, select Refresh.
-
Verify that the emp.txt file is created in the output folder of the adftutorial container. If the output folder doesn't exist, the Data Factory service automatically creates it.
-
Switch to the Author tab above the Monitor tab on the left panel so that you can edit linked services, datasets, and pipelines. To learn about editing them in the Data Factory UI, see Create a data factory by using the Azure portal.
:::image type="content" source="./media/quickstart-create-data-factory-copy-data-tool/select-author.png" alt-text="Select Author tab":::
The pipeline in this sample copies data from one location to another location in Azure Blob storage. To learn about using Data Factory in more scenarios, go through the tutorials.