Migration Guide
Note Before version 1.10 for TaskManagers and before 1.11 for JobManagers, Flink did not require that memory related options are set at all as they all had default values. The new memory configuration requires that at least one subset of the following options is configured explicitly, otherwise the configuration will fail:
The shipped with Flink sets taskmanager.memory.process.size (since 1.10) and (since 1.11) to make the default memory configuration consistent.
This spreadsheet can also help to evaluate and compare the results of the legacy and new memory computations.
This chapter shortly lists all changes to Flink’s memory configuration options introduced with the 1.10 release. It also references other chapters for more details about migrating to the new configuration options.
The following options are completely removed. If they are still used, they will be ignored.
The following options are deprecated but if they are still used they will be interpreted as new options for backwards compatibility:
Although, the network memory configuration has not changed too much it is recommended to verify its configuration. It can change if other memory components have new sizes, e.g. the total memory which the network can be a fraction of. See also .
The container cut-off configuration options, and containerized.heap-cutoff-min
, have no effect anymore for TaskManagers. See also how to migrate container cut-off.
Total Memory (Previously Heap Memory)
The previous options which were responsible for the total memory used by Flink are taskmanager.heap.size
or taskmanager.heap.mb
. Despite their naming, they included not only JVM Heap but also other off-heap memory components. The options have been deprecated.
If you use the mentioned legacy options without specifying the corresponding new options, they will be directly translated into the following new options:
- Total Flink memory (taskmanager.memory.flink.size) for standalone deployments
- Total process memory () for containerized deployments (Yarn)
It is also recommended using these new options instead of the legacy ones as they might be completely removed in the following releases.
JVM Heap memory previously consisted of the managed memory (if configured to be on-heap) and the rest which included any other usages of heap memory. This rest was the remaining part of the total memory, see also how to migrate managed memory.
Now, if only total Flink memory or total process memory is configured, then the JVM Heap is the rest of what is left after subtracting all other components from the total memory, see also .
Additionally, you can now have more direct control over the JVM Heap assigned to the operator tasks (taskmanager.memory.task.heap.size), see also . The JVM Heap memory is also used by the heap state backends (MemoryStateBackend or ) if it is chosen for streaming jobs.
A part of the JVM Heap is now always reserved for the Flink framework (taskmanager.memory.framework.heap.size). See also .
Managed Memory
See also .
Explicit Size
The previous option to configure managed memory size (taskmanager.memory.size
) was renamed to and deprecated. It is recommended to use the new option because the legacy one can be removed in future releases.
Fraction
If not set explicitly, the managed memory could be previously specified as a fraction (taskmanager.memory.fraction
) of the total memory minus network memory and container cut-off (only for deployments). This option has been completely removed and will have no effect if still used. Please, use the new option taskmanager.memory.managed.fraction instead. This new option will set the to the specified fraction of the total Flink memory if its size is not set explicitly by .
RocksDB state
If the is chosen for a streaming job, its native memory consumption should now be accounted for in managed memory. The RocksDB memory allocation is limited by the size. This should prevent the killing of containers on Yarn. You can disable the RocksDB memory control by setting to . See also how to migrate container cut-off.
Other changes
Additionally, the following changes have been made:
- The managed memory is always off-heap now. The configuration option
taskmanager.memory.off-heap
is removed and will have no effect anymore. - The now uses native memory which is not direct memory. It means that the managed memory is no longer accounted for in the JVM direct memory limit.
Previously, there were options responsible for setting the JVM Heap size of the JobManager:
jobmanager.heap.size
jobmanager.heap.mb
Despite their naming, they represented the JVM Heap only for standalone deployments. For the containerized deployments ( and Yarn), they also included other off-heap memory consumption. The size of JVM Heap was additionally reduced by the container cut-off which has been completely removed after 1.11.
The mentioned legacy options have been deprecated. If they are used without specifying the corresponding new options, they will be directly translated into the following new options:
- JVM Heap () for standalone deployments
- Total process memory () for containerized deployments (Kubernetes and )
Now, if only the total Flink memory or total process memory is configured, then the JVM Heap is also derived as the rest of what is left after subtracting all other components from the total memory, see also . Additionally, you can now have more direct control over the JVM Heap by adjusting the option.
Since 1.10 release, Flink sets the JVM Metaspace and JVM Direct Memory limits for the TaskManager process by adding the corresponding JVM arguments. Since 1.11 release, Flink also sets the JVM Metaspace limit for the JobManager process. You can enable the JVM Direct Memory limit for JobManager process if you set the jobmanager.memory.enable-jvm-direct-memory-limit option. See also .
Flink sets the mentioned JVM memory limits to simplify debugging of the corresponding memory leaks and avoid the container out-of-memory errors. See also the troubleshooting guide for details about the and JVM Direct Memory OutOfMemoryErrors.
For containerized deployments, you could previously specify a cut-off memory. This memory could accommodate for unaccounted memory allocations. Dependencies which were not directly controlled by Flink were the main source of those allocations, e.g. RocksDB, JVM internals, etc. This is no longer available, and the related configuration options (containerized.heap-cutoff-ratio
and ) will have no effect anymore. The new memory model introduced more specific memory components to address these concerns.
In streaming jobs which use , the RocksDB native memory consumption should be accounted for as a part of the managed memory now. The RocksDB memory allocation is also limited by the configured size of the . See also migrating managed memory and .
The other direct or native off-heap memory consumers can now be addressed by the following new configuration options:
- Task off-heap memory (taskmanager.memory.task.off-heap.size)
- Framework off-heap memory ()
- JVM metaspace (taskmanager.memory.jvm-metaspace.size)
for JobManagers
The direct or native off-heap memory consumers can now be addressed by the following new configuration options:
- Off-heap memory ()
- JVM metaspace (jobmanager.memory.jvm-metaspace.size)
This section describes the changes of the default flink-conf.yaml
shipped with Flink.
The total memory for TaskManagers (taskmanager.heap.size
) is replaced by in the default flink-conf.yaml
. The value increased from 1024Mb to 1728Mb.
The total memory for JobManagers (jobmanager.heap.size
) is replaced by jobmanager.memory.process.size in the default flink-conf.yaml
. The value increased from 1024Mb to 1600Mb.
See also .
Warning: If you use the new default it can result in different sizes of memory components and can lead to performance changes.