All Products
Search
Document Center

Realtime Compute for Apache Flink:April 13, 2023

Last Updated:Jun 05, 2024

This topic describes the release notes for Realtime Compute for Apache Flink and provides links to relevant references. The release notes provide the major updates and bug fixes in Realtime Compute for Apache Flink in the version that was released on April 13, 2023.

Important

A canary release is being performed for this version on the entire network. The version is expected to be released from April 13, 2023 to August 15, 2023. If you cannot find new features in the Realtime Compute for Apache Flink console, the canary release for your platform is not complete. If you want to upgrade your version at the earliest opportunity, submit a ticket to perform an upgrade based on your business requirements. For more information about the upgrade plan, see the most recent announcement on the right side of the Realtime Compute for Apache Flink console.

Overview

A new official version of Realtime Compute for Apache Flink was released on April 13, 2023. This version includes platform updates, engine updates, connector updates, performance optimization, and bug fixes.

The engine version Ververica Runtime (VVR) 6.0.6 is released, which is an enterprise-level Flink engine based on Apache Flink 1.15.3. Realtime Compute for Apache Flink that uses VVR 6.0.6 or later provides Apache Paimon (incubating), which is in invitational preview. Apache Paimon (incubating) is a streaming data lake technology that supports high-throughput and low-latency data ingestion, streaming data subscription, and real-time queries.

Multiple common features on the platform are optimized. The first-level menus in the development console of Realtime Compute for Apache Flink and the process of draft development and deployment O&M are optimized to improve user experience and enhance the alerting capability.

After the canary release is complete, the platform capabilities are upgraded. The engine version of drafts is upgraded within two weeks. After the upgrade is complete, you can view the new engine version in the Engine Version drop-down list of a draft. You can use the new capabilities of the platform and upgrade the engine that is used by your draft to the new version. We look forward to your feedback and experience.

Features

Feature

Description

References

Apache Paimon streaming data lake

The Apache Paimon streaming data lake is in invitational preview. Realtime Compute for Apache Flink supports data reading from and data writing to Apache Paimon based on Alibaba Cloud Object Storage Service (OSS).

Apache Paimon connector

Apache Paimon catalogs

Built-in Apache Paimon catalogs can be used together with Flink SQL to develop a Flink-based real-time data lake solution.

Manage Apache Paimon catalogs

Data writing to OSS-HDFS in streaming mode

Data can be written to OSS-HDFS in streaming mode. This allows you to use multiple types of data stores.

OSS connector

Adjustment of metrics

  • CPU metrics

    • The JM CPU Usage metric is added.

    • The TM CPU Load and JM CPU Load metrics are removed.

      Note

      The TM CPU Load and JM CPU Load metrics cannot reflect CPU utilization. Only the JM CPU Usage and TM CPU Usage metrics are used to monitor CPU utilization.

  • Connector metrics

    • The numRecordsOut and numRecordsOutPersecond metrics are added for the Elasticsearch connector to improve operational observability.

    • The Enumerator metric is added for the MySQL Change Data Capture (CDC) connector.

Tair result tables

Tair result tables are supported. This allows you to use multiple types of data stores.

-

Enhancement of EXPLAIN statements in streaming SQL

The EXPLAIN PLAN_ADVICE statement is supported to provide more detailed optimization suggestions.

None

Optimization of dynamic Flink complex event processing (CEP)

Conditions based on Groovy expressions are supported. The implementation of nondeterministic finite automaton (NFA) and SharedBuffer is optimized to reduce the number of times a timer is created and improve performance.

Definitions of rules in the JSON format in dynamic Flink CEP

Flink machine learning (ML)

Flink ML supported by the engine is in invitational preview and provides the real-time machine learning feature.

-

Optimization of the draft development process

The Draft Editor page is changed to the SQL Editor page to provide an SQL development platform, which optimizes the draft development process in the following aspects:

  • Provides an isolation mechanism between drafts and deployments to prevent online deployments from being affected by SQL development. The development of JAR and Python drafts is naturally isolated from online JAR and Python deployments. The development of JAR and Python drafts does not require SQL components, such as user-defined functions (UDFs), connectors, and catalogs. Therefore, the isolation mechanism is not required for JAR and Python drafts.

  • Optimizes the process of creating JAR and Python deployments. JAR and Python deployments can no longer be created on the SQL Editor page. You can directly create JAR and Python deployments on the Deployments page. You can find the JAR and Python drafts that are created on the Draft Editor page in the draft archive.

Develop an SQL draft

Optimization of the creation process of JAR and Python deployments

The process of creating JAR and Python deployments is optimized. You can click Create Deployment on the Deployments page to create a JAR or Python deployment.

Optimization of the deployment startup process

The information about the resources of a deployment and the related Flink configurations are displayed on the Configuration tab of the Deployments page. You can adjust resource configurations without the need to cancel the deployment. You need to only specify a start offset when you start a deployment.

Start a deployment

Addition of the Catalogs page

The Catalogs page is added. The Apache Flink community does not recommend the use of temporary tables. To avoid repeated use of DDL statements, we recommend that you use catalogs to create SQL drafts. This version provides enhanced catalog capabilities to support the use of catalogs in SQL drafts and allow you to manage SQL drafts in an efficient manner.

None

Optimization of the SQL deployment debugging process

The debugging process of SQL deployments is optimized. If you want to debug an SQL deployment, you can select an existing session cluster for debugging instead of configuring a new session cluster. This way, you do not need to frequently change clusters during deployment debugging if a deployment has multiple versions.

Debug a deployment

Addition of the Security page

The user authorization and key hosting features are integrated into the Security page. You can complete settings related to platform security on the Security page. In earlier versions, the key hosting feature is provided on the Key Replacement tab.

Manage a key

Addition of the Connectors page

The Connectors page is provided to allow you to view the types and versions of connectors supported by different engine versions and manage custom connectors.

Manage custom connectors

Alerts for failed deployments,

sending of alert notifications by phone, and search of contacts

The monitoring and alerting feature is optimized in the following aspects:

  • Allows you to set the Notification parameter to Phone to send alert notifications by phone and allows you to search for contacts.

  • Supports alerts for failed deployments.

Configure alert rules

Logon by using a role account

A role account can be used to log on to the management console of Realtime Compute for Apache Flink. By default, owner permissions are used. You cannot configure permissions for the role account.

-

Network detection

An IP address or a domain name can be used to check whether the running environment of a Realtime Compute for Apache Flink deployment is connected to the upstream and downstream systems.

None

Custom catalogs

Built-in catalogs are provided in the Realtime Compute for Apache Flink console. After you register metadata by using a catalog, you do not need to frequently use DDL statements to create temporary tables when you create an SQL draft. You can also create custom catalogs and use the JAR packages of the custom catalogs.

-

Enhanced intelligent deployment diagnostics capabilities

The intelligent deployment diagnostics feature is enhanced. This feature helps you analyze error logs that are generated during draft development and deployment running. When you view information on the log page, the system automatically analyzes the logs and provides executable operation suggestions.

Perform intelligent deployment diagnostics

Upgrade of the Log Service connector client

The performance and stability of the Log Service connector are improved.

None

Extended CEP SQL

The loop continuity declaration can be used together with the UNTIL syntax in CEP SQL.

None

Fixed issues

  • The following issue is fixed: NULL data is displayed as TRUE during deployment debugging.

  • The following issue is fixed: If multiple metadata services are registered, all metadata services become abnormal because one of the metadata services is unavailable.

  • The following issue is fixed: An error is returned during status restoration when a deployment is switched from a session cluster to a pre-fob cluster.

  • The following issue is fixed: A verification error occurs when a Hologres dimension table and a Hologres source table are joined.

  • The following issue is fixed: The consumer offset of a Log Service source table unexpectedly rolls back.