All Products
Search
Document Center

DataWorks:Configure Kafka Writer

Last Updated:Nov 18, 2024

To configure Kafka Writer, you need to only select a table to which you want to write data and configure field mappings.

Prerequisites

A reader or conversion node is configured. For more information, see Data source types that support real-time synchronization.

Background information

Deduplication is not supported for data that you want to write to Kafka. If you reset the offset for your synchronization node or your synchronization node is restarted after a failover, duplicate data may be written to Kafka.

Procedure

  1. Log on to the DataWorks console. In the top navigation bar, select the desired region. In the left-side navigation pane, choose Data Development and Governance > Data Development. On the page that appears, select the desired workspace from the drop-down list and click Go to Data Development.

  2. In the Scheduled Workflow pane, move the pointer over the Create a table icon and choose Create Node > Data Integration > Real-time synchronization.

    Alternatively, right-click the required workflow, and then choose Create Node > Data Integration > Real-time synchronizationReal-time synchronization.

  3. In the Create Node dialog box, set the Sync Method parameter to End-to-end ETL and configure the Name and Path parameters.

    Important

    The node name cannot exceed 128 characters in length and can contain letters, digits, underscores (_), and periods (.).

  4. Click Confirm.

  5. On the configuration tab of the real-time synchronization node, drag Kafka in the Output section to the canvas on the right. Connect the Kafka node to the configured reader or conversion node.

  6. Click the Kafka node. In the panel that appears, configure the parameters.

    Node Configuration

    Parameter

    Description

    Data source

    The name of the Kafka data source that you added. If no data source is available, click New data source on the right to add one on the Data Source page. For more information, see Add a Kafka data source.

    Topic

    The name of the Kafka topic to which you want to write data. Kafka maintains feeds of messages in categories called topics.

    Each message that is published to a Kafka cluster is assigned to a topic. Each topic contains a group of messages.

    Note

    Kafka Reader in each synchronization node can read data from only one topic.

    Key Column

    The name of the source column whose value in each row is used as a key in the destination Kafka topic. If you select multiple columns, the column values in each row are concatenated as a key by using commas (,). If you do not select a column, empty strings are used as keys in the destination Kafka topic.

    Value Column

    The names of the source columns whose values in each row are concatenated as a value in the destination Kafka topic. If you do not select a column, the values of all source columns in each row are concatenated as a value in the destination Kafka topic. The method used to concatenate the values of source columns depend on the write mode that you specify. For more information, see the parameter description that is provided by in Kafka Writer.

    Key Type

    The data type of the keys in the Kafka topic. The value of this parameter determines the setting of key.serializer that is used to initialize a Kafka producer. Valid values: STRING, BYTEARRAY, DOUBLE, FLOAT, INTEGER, LONG, and SHORT.

    Value Type

    The data type of the values in the Kafka topic. The value of this parameter determines the setting of value.serializer that is used to initialize a Kafka producer. Valid values: STRING, BYTEARRAY, DOUBLE, FLOAT, INTEGER, LONG, and SHORT.

    Number of Bytes Written at a Time

    The number of bytes to write at a time. We recommend that you set this parameter to a value that is greater than 16000.

    Write Mode

    The write mode. You can use this parameter to specify the format in which Kafka Writer concatenates the values of value columns in the source. Valid values: text and json.

    • If you set this parameter to text, Kafka Writer concatenates the values of the columns by using the specified delimiter.

    • If you set this parameter to json, Kafka Writer concatenates the values of the columns as a JSON string.

    For example, three columns col1, col2, and col3 are obtained as value columns from the reader, and the values of the columns in a specific row are a, b, and c. If the Write Mode parameter is set to text and the Column separator parameter is set to #, the value stored in the destination Kafka topic is a#b#c. If the Write Mode parameter is set to json, the value stored in the destination Kafka topic is the string {"col1":"a","col2":"b","col3":"c"}.

    Column separator

    The delimiter that is used to concatenate column values obtained from the reader if the Write Mode parameter is set to text. The values of the columns in each row are concatenated as a value in the destination Kafka topic. You can specify one or more characters as column delimiters. You can specify Unicode characters in the format of \u0001. Escape characters such as \t and \n are supported. Default value: \t.

    Configuration parameters

    The extended parameters that you can configure when you create a Kafka consumer. For example, you can configure the bootstrap.servers, acks, and linger.ms parameters. You can configure parameters in KafkaConfig to control the data read behavior of a Kafka consumer. For a real-time synchronization node that synchronizes data to Kafka, the default value of the acks parameter for a Kafka consumer is all. If you have higher requirements for performance, you can specify a different value for the acks parameter. Valid values of the acks parameter:

    • 0: A Kafka consumer does not acknowledge whether data is written to the destination.

    • 1: A Kafka consumer acknowledges that the write operation is successful if data is written to the primary replica.

    • all: A Kafka consumer acknowledges that the write operation is successful if data is written to all replicas.

  7. Click Save in the top toolbar to save the node.