This topic describes the release notes for Realtime Compute for Apache Flink and provides links to relevant references. The release notes provide the major updates and bug fixes in Realtime Compute for Apache Flink in the version that was released on April 13, 2023.
A canary release is being performed for this version on the entire network. The version is expected to be released from April 13, 2023 to August 15, 2023. If you cannot find new features in the Realtime Compute for Apache Flink console, the canary release for your platform is not complete. If you want to upgrade your version at the earliest opportunity, submit a ticket to perform an upgrade based on your business requirements. For more information about the upgrade plan, see the most recent announcement on the right side of the Realtime Compute for Apache Flink console.
Overview
A new official version of Realtime Compute for Apache Flink was released on April 13, 2023. This version includes platform updates, engine updates, connector updates, performance optimization, and bug fixes.
The engine version Ververica Runtime (VVR) 6.0.6 is released, which is an enterprise-level Flink engine based on Apache Flink 1.15.3. Realtime Compute for Apache Flink that uses VVR 6.0.6 or later provides Apache Paimon (incubating), which is in invitational preview. Apache Paimon (incubating) is a streaming data lake technology that supports high-throughput and low-latency data ingestion, streaming data subscription, and real-time queries.
Multiple common features on the platform are optimized. The first-level menus in the development console of Realtime Compute for Apache Flink and the process of draft development and deployment O&M are optimized to improve user experience and enhance the alerting capability.
After the canary release is complete, the platform capabilities are upgraded. The engine version of drafts is upgraded within two weeks. After the upgrade is complete, you can view the new engine version in the Engine Version drop-down list of a draft. You can use the new capabilities of the platform and upgrade the engine that is used by your draft to the new version. We look forward to your feedback and experience.
Features
Feature | Description | References |
Apache Paimon streaming data lake | The Apache Paimon streaming data lake is in invitational preview. Realtime Compute for Apache Flink supports data reading from and data writing to Apache Paimon based on Alibaba Cloud Object Storage Service (OSS). | |
Apache Paimon catalogs | Built-in Apache Paimon catalogs can be used together with Flink SQL to develop a Flink-based real-time data lake solution. | |
Data writing to OSS-HDFS in streaming mode | Data can be written to OSS-HDFS in streaming mode. This allows you to use multiple types of data stores. | |
Adjustment of metrics |
| |
Tair result tables | Tair result tables are supported. This allows you to use multiple types of data stores. | - |
Enhancement of EXPLAIN statements in streaming SQL | The EXPLAIN PLAN_ADVICE statement is supported to provide more detailed optimization suggestions. | None |
Optimization of dynamic Flink complex event processing (CEP) | Conditions based on Groovy expressions are supported. The implementation of nondeterministic finite automaton (NFA) and SharedBuffer is optimized to reduce the number of times a timer is created and improve performance. | Definitions of rules in the JSON format in dynamic Flink CEP |
Flink machine learning (ML) | Flink ML supported by the engine is in invitational preview and provides the real-time machine learning feature. | - |
Optimization of the draft development process | The Draft Editor page is changed to the SQL Editor page to provide an SQL development platform, which optimizes the draft development process in the following aspects:
| |
Optimization of the creation process of JAR and Python deployments | The process of creating JAR and Python deployments is optimized. You can click Create Deployment on the Deployments page to create a JAR or Python deployment. | |
Optimization of the deployment startup process | The information about the resources of a deployment and the related Flink configurations are displayed on the Configuration tab of the Deployments page. You can adjust resource configurations without the need to cancel the deployment. You need to only specify a start offset when you start a deployment. | |
Addition of the Catalogs page | The Catalogs page is added. The Apache Flink community does not recommend the use of temporary tables. To avoid repeated use of DDL statements, we recommend that you use catalogs to create SQL drafts. This version provides enhanced catalog capabilities to support the use of catalogs in SQL drafts and allow you to manage SQL drafts in an efficient manner. | None |
Optimization of the SQL deployment debugging process | The debugging process of SQL deployments is optimized. If you want to debug an SQL deployment, you can select an existing session cluster for debugging instead of configuring a new session cluster. This way, you do not need to frequently change clusters during deployment debugging if a deployment has multiple versions. | |
Addition of the Security page | The user authorization and key hosting features are integrated into the Security page. You can complete settings related to platform security on the Security page. In earlier versions, the key hosting feature is provided on the Key Replacement tab. | |
Addition of the Connectors page | The Connectors page is provided to allow you to view the types and versions of connectors supported by different engine versions and manage custom connectors. | |
Alerts for failed deployments, sending of alert notifications by phone, and search of contacts | The monitoring and alerting feature is optimized in the following aspects:
| |
Logon by using a role account | A role account can be used to log on to the management console of Realtime Compute for Apache Flink. By default, owner permissions are used. You cannot configure permissions for the role account. | - |
Network detection | An IP address or a domain name can be used to check whether the running environment of a Realtime Compute for Apache Flink deployment is connected to the upstream and downstream systems. | None |
Custom catalogs | Built-in catalogs are provided in the Realtime Compute for Apache Flink console. After you register metadata by using a catalog, you do not need to frequently use DDL statements to create temporary tables when you create an SQL draft. You can also create custom catalogs and use the JAR packages of the custom catalogs. | - |
Enhanced intelligent deployment diagnostics capabilities | The intelligent deployment diagnostics feature is enhanced. This feature helps you analyze error logs that are generated during draft development and deployment running. When you view information on the log page, the system automatically analyzes the logs and provides executable operation suggestions. | |
Upgrade of the Log Service connector client | The performance and stability of the Log Service connector are improved. | None |
Extended CEP SQL | The loop continuity declaration can be used together with the UNTIL syntax in CEP SQL. | None |
Fixed issues
The following issue is fixed: NULL data is displayed as TRUE during deployment debugging.
The following issue is fixed: If multiple metadata services are registered, all metadata services become abnormal because one of the metadata services is unavailable.
The following issue is fixed: An error is returned during status restoration when a deployment is switched from a session cluster to a pre-fob cluster.
The following issue is fixed: A verification error occurs when a Hologres dimension table and a Hologres source table are joined.
The following issue is fixed: The consumer offset of a Log Service source table unexpectedly rolls back.