DataWorks provides HDFS Reader and HDFS Writer for you to read data from and write data to Hadoop Distributed File System (HDFS) data sources. This topic describes how to use DataWorks to import data from HDFS to AnalyticDB for MySQL Data Lakehouse Edition (V3.0) or Data Warehouse Edition (V3.0).
Prerequisites
The required data sources are configured. Before you configure a data synchronization task, you must configure the data source from which you want to read data and the data source to which you want to write data. This way, when you configure a batch synchronization task, you can select the data sources. For information about the data source types that are supported by batch synchronization, see Supported data source types, Readers, and Writers.
NoteFor information about the items that you need to understand before you configure a data source, see Overview.
An exclusive resource group for Data Integration that meets your business requirements is purchased. For more information, see Create and use an exclusive resource group for Data Integration.
Network connections between the exclusive resource group for Data Integration and data sources are established. For more information, see Establish a network connection between a resource group and a data source.
Procedure
Configure the source database reader. For more information, see HDFS Reader.
Configure the destination database writer. For more information, see AnalyticDB for MySQL 3.0 Writer.
Configure a data synchronization job. For more information, see Configure a batch synchronization task by using the codeless UI.