name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
flink_RocksDBIncrementalRestoreOperation_readMetaData_rdh
/** * Reads Flink's state meta data file from the state handle. */ private KeyedBackendSerializationProxy<K> readMetaData(StreamStateHandle metaStateHandle) throws Exception { InputStream inputStream = null; try { inputStream = metaStateHandle.openInputStream(); cancelStreamRegistry.registerCloseable(inputStream); D...
3.26
flink_RocksDBIncrementalRestoreOperation_restore_rdh
/** * Root method that branches for different implementations of {@link KeyedStateHandle}. */ @Override public RocksDBRestoreResult restore() throws Exception { if ((restoreStateHandles == null) || restoreStateHandles.isEmpty()) { return null; } final KeyedStat...
3.26
flink_RocksDBIncrementalRestoreOperation_createColumnFamilyDescriptors_rdh
/** * This method recreates and registers all {@link ColumnFamilyDescriptor} from Flink's state * meta data snapshot. */ private List<ColumnFamilyDescriptor> createColumnFamilyDescriptors(List<StateMetaInfoSnapshot> stateMetaInfoSnapshots, boolean registerTtlCompactFilter) { List<ColumnFamilyDescriptor> columnFamily...
3.26
flink_SubtaskState_getManagedOperatorState_rdh
// -------------------------------------------------------------------------------------------- public ChainedStateHandle<OperatorStateHandle> getManagedOperatorState() { return managedOperatorState; }
3.26
flink_SubtaskState_equals_rdh
// -------------------------------------------------------------------------------------------- @Override public boolean equals(Object o) { if (this == o) {return true; } if ((o == null) || (getClass() != o.getClass())) { return false; } SubtaskState that = ((SubtaskState) (o)); if (stateSize != that.stateSize) {r...
3.26
flink_Signature_of_rdh
/** * Creates an immutable instance of {@link Signature}. */ public static Signature of(List<Argument> arguments) { return new Signature(arguments); }
3.26
flink_AggregatorRegistry_registerAggregator_rdh
// -------------------------------------------------------------------------------------------- public void registerAggregator(String name, Aggregator<?> aggregator) { if ((name == null) || (aggregator == null)) { throw new IllegalArgumentException("Name and aggregator must not be null");} if (t...
3.26
flink_PackagedProgram_deleteExtractedLibraries_rdh
/** * Deletes all temporary files created for contained packaged libraries. */ private void deleteExtractedLibraries() { deleteExtractedLibraries(this.extractedTempLibraries); this.extractedTempLibraries.clear(); }
3.26
flink_PackagedProgram_getJobJarAndDependencies_rdh
/** * Returns all provided libraries needed to run the program. */ public static List<URL> getJobJarAndDependencies(File jarFile, @Nullable String entryPointClassName) throws ProgramInvocationException { URL jarFileUrl = loadJarFile(jarFile); List<File> extractedTempLibraries = (jarFileUrl == null) ? Collec...
3.26
flink_PackagedProgram_invokeInteractiveModeForExecution_rdh
/** * This method assumes that the context environment is prepared, or the execution will be a * local execution by default. */ public void invokeInteractiveModeForExecution() throws ProgramInvocationException { FlinkSecurityManager.monitorUserSystemExitForCurrentThrea...
3.26
flink_PackagedProgram_extractContainedLibraries_rdh
/** * Takes all JAR files that are contained in this program's JAR file and extracts them to the * system's temp directory. * * @return The file names of the extracted temporary files. * @throws ProgramInvocationException * Thrown, if the extraction process failed. */ public static List<File> extractContainedL...
3.26
flink_FutureCompletingBlockingQueue_peek_rdh
/** * Get the first element from the queue without removing it. * * @return the first element in the queue, or Null if the queue is empty. */ public T peek() { lock.lock(); try { return queue.peek(); } finally { lock.unlock(); } }
3.26
flink_FutureCompletingBlockingQueue_getAvailableFuture_rdh
// ------------------------------------------------------------------------ // utilities // ------------------------------------------------------------------------ @SuppressWarnings("unchecked") private static CompletableFuture<Void> getAvailableFuture() {// this is a way to obtain the AvailabilityProvider.AVAILABLE f...
3.26
flink_FutureCompletingBlockingQueue_put_rdh
// ------------------------------------------------------------------------ // Blocking Queue Logic // ------------------------------------------------------------------------ /** * Put an element into the queue. The thread blocks if the queue is full. * * @param threadIndex * the index of the thread. * @param e...
3.26
flink_FutureCompletingBlockingQueue_wakeUpPuttingThread_rdh
/** * Gracefully wakes up the thread with the given {@code threadIndex} if it is blocked in adding * an element. to the queue. If the thread is blocked in {@link #put(int, Object)} it will * immediately return from the method with a return value of false. * * <p>If this method is called, the next time the thread w...
3.26
flink_FutureCompletingBlockingQueue_size_rdh
/** * Gets the size of the queue. */ public int size() { lock.lock(); try { return queue.size(); } finally { lock.unlock(); } }
3.26
flink_FutureCompletingBlockingQueue_remainingCapacity_rdh
/** * Checks the remaining capacity in the queue. That is the difference between the maximum * capacity and the current number of elements in the queue. */ public int remainingCapacity() { lock.lock(); try { return capacity - queue.size(); } finally { lock.unlock(); } }
3.26
flink_FutureCompletingBlockingQueue_getAvailabilityFuture_rdh
// ------------------------------------------------------------------------ // Future / Notification logic // ------------------------------------------------------------------------ /** * Returns the availability future. If the queue is non-empty, then this future will already be * complete. Otherwise the obtained f...
3.26
flink_FutureCompletingBlockingQueue_m1_rdh
/** * Checks whether the queue is empty. */ public boolean m1() { lock.lock(); try { return queue.isEmpty(); } finally { lock.unlock(); } }
3.26
flink_FutureCompletingBlockingQueue_poll_rdh
/** * Get and remove the first element from the queue. Null is returned if the queue is empty. If * this makes the queue empty (takes the last element) or finds the queue already empty, then * this resets the availability notifications. The next call to {@link #getAvailabilityFuture()} * will then return a non-comp...
3.26
flink_FutureCompletingBlockingQueue_moveToAvailable_rdh
/** * Internal utility to make sure that the current future futures are complete (until reset). */ @GuardedBy("lock") private void moveToAvailable() { final CompletableFuture<Void> current = currentFuture; if (current != AVAILABLE) { currentFuture = AVAILABLE; current.complete(null); } }
3.26
flink_FutureCompletingBlockingQueue_enqueue_rdh
// --------------- private helpers ------------------------- @GuardedBy("lock") private void enqueue(T element) { final int sizeBefore = queue.size(); queue.add(element); if (sizeBefore == 0) { moveToAvailable(); } if ((sizeBefore < (capacity - 1)) && (!notFull.isEmpty())) { signa...
3.26
flink_FutureCompletingBlockingQueue_take_rdh
/** * <b>Warning:</b> This is a dangerous method and should only be used for testing convenience. A * method that blocks until availability does not go together well with the concept of * asynchronous notifications and non-blocking polling. * * <p>Get and remove the first element from the queue. The call blocks if...
3.26
flink_FactoryUtil_createCatalogStoreFactoryHelper_rdh
/** * Creates a utility that helps validating options for a {@link CatalogStoreFactory}. * * <p>Note: This utility checks for left-over options in the final step. */public static CatalogStoreFactoryHelper createCatalogStoreFactoryHelper(CatalogStoreFactory factory, CatalogStoreFactory.Context context) { return...
3.26
flink_FactoryUtil_createModuleFactoryHelper_rdh
/** * Creates a utility that helps validating options for a {@link ModuleFactory}. * * <p>Note: This utility checks for left-over options in the final step. */ public static ModuleFactoryHelper createModuleFactoryHelper(ModuleFactory factory, ModuleFactory.Context context) { return new ModuleFactoryHelper...
3.26
flink_FactoryUtil_validateFactoryOptions_rdh
/** * Validates the required and optional {@link ConfigOption}s of a factory. * * <p>Note: It does not check for left-over options. */ public static void validateFactoryOptions(Factory factory, ReadableConfig options) { validateFactoryOptions(factory.requiredOptions(), factory.optionalOptions(), options); }
3.26
flink_FactoryUtil_discoverEncodingFormat_rdh
/** * Discovers a {@link EncodingFormat} of the given type using the given option as factory * identifier. */ public <I, F extends EncodingFormatFactory<I>> EncodingFormat<I> discoverEncodingFormat(Class<F> formatFactoryClass, ConfigOption<String> formatOption) { return discoverOptionalEncodingFormat(formatFactoryCl...
3.26
flink_FactoryUtil_discoverOptionalEncodingFormat_rdh
/** * Discovers a {@link EncodingFormat} of the given type using the given option (if present) * as factory identifier. */ public <I, F extends EncodingFormatFactory<I>> Optional<EncodingFormat<I>> discoverOptionalEncodingFormat(Class<F> formatFactoryClass, ConfigOption<String> formatOption) { return discoverOption...
3.26
flink_FactoryUtil_forwardOptions_rdh
// ---------------------------------------------------------------------------------------- /** * Forwards the options declared in {@link DynamicTableFactory#forwardOptions()} and * possibly {@link FormatFactory#forwardOptions()} from {@link DynamicTableFactory.Context#getEnrichmentOptions()} to the final options, if...
3.26
flink_FactoryUtil_discoverFactory_rdh
/** * Discovers a factory using the given factory base class and identifier. * * <p>This method is meant for cases where {@link #createTableFactoryHelper(DynamicTableFactory, * DynamicTableFactory.Context)} {@link #createTableSource(Catalog, ObjectIdentifier, * ResolvedCatalogTable, ReadableConfig, ClassLoader, bo...
3.26
flink_FactoryUtil_m1_rdh
/** * Validates the options of the factory. It checks for unconsumed option keys. */public void m1() { validateFactoryOptions(factory, allOptions); validateUnconsumedKeys(factory.factoryIdentifier(), allOptions.keySet(), consumedOptionKeys, deprecatedOptionKeys); validateWatermarkOptions(factory.factoryIdentifier(),...
3.26
flink_FactoryUtil_createDynamicTableSink_rdh
/** * Creates a {@link DynamicTableSink} from a {@link CatalogTable}. * * <p>If {@param preferredFactory} is passed, the table sink is created from that factory. * Otherwise, an attempt is made to discover a matching factory using Java SPI (see {@link Factory} for details). */ public static DynamicTableSink create...
3.26
flink_FactoryUtil_checkFormatIdentifierMatchesWithEnrichingOptions_rdh
/** * This function assumes that the format config is used only and only if the original * configuration contains the format config option. It will fail if there is a mismatch of * the identifier between the format in the plan table map and the one in enriching table * map. */ private void checkFormatIdentifierMat...
3.26
flink_FactoryUtil_createTableFactoryHelper_rdh
/** * Creates a utility that helps in discovering formats, merging options with {@link DynamicTableFactory.Context#getEnrichmentOptions()} and validating them all for a {@link DynamicTableFactory}. * * <p>The following example sketches the usage: * * <pre>{@code // in createDynamicTableSource() * helper = Factory...
3.26
flink_FactoryUtil_discoverOptionalDecodingFormat_rdh
/** * Discovers a {@link DecodingFormat} of the given type using the given option (if present) * as factory identifier. */ public <I, F extends DecodingFormatFactory<I>> Optional<DecodingFormat<I>> discoverOptionalDecodingFormat(Class<F> formatFactoryClass, ConfigOption<String> formatOption) { return discoverOptiona...
3.26
flink_FactoryUtil_validateWatermarkOptions_rdh
// -------------------------------------------------------------------------------------------- /** * Validate watermark options from table options. * * @param factoryIdentifier * identifier of table * @param conf * table options */ public static void validateWatermarkOptions(String factoryIdentifier, Readab...
3.26
flink_FactoryUtil_createCatalog_rdh
/** * Attempts to discover an appropriate catalog factory and creates an instance of the catalog. * * <p>This first uses the legacy {@link TableFactory} stack to discover a matching {@link CatalogFactory}. If none is found, it falls back to the new stack using {@link Factory} * instead. */ public static Catalog cr...
3.26
flink_FactoryUtil_createModule_rdh
/** * Discovers a matching module factory and creates an instance of it. * * <p>This first uses the legacy {@link TableFactory} stack to discover a matching {@link ModuleFactory}. If none is found, it falls back to the new stack using {@link Factory} * instead. */ public static Module createModule(String moduleNam...
3.26
flink_FactoryUtil_validateUnconsumedKeys_rdh
/** * Validates unconsumed option keys. */ public static void validateUnconsumedKeys(String factoryIdentifier, Set<String> allOptionKeys, Set<String> consumedOptionKeys) { validateUnconsumedKeys(factoryIdentifier, allOptionKeys, consumedOptionKeys, Collections.emptySet()); }
3.26
flink_FactoryUtil_validateExcept_rdh
/** * Validates the options of the factory. It checks for unconsumed option keys while ignoring * the options with given prefixes. * * <p>The option keys that have given prefix {@code prefixToSkip} would just be skipped for * validation. * * @param prefixesToSkip * Set of option key prefixes to skip validatio...
3.26
flink_FactoryUtil_getOptions_rdh
/** * Returns all options currently being consumed by the factory. This method returns the * options already merged with {@link DynamicTableFactory.Context#getEnrichmentOptions()}, * using {@link DynamicTableFactory#forwardOptions()} as reference of mergeable options. */ @Override public ReadableConfig getOptions()...
3.26
flink_FactoryUtil_getFormatPrefix_rdh
/** * Returns the required option prefix for options of the given format. */ public static String getFormatPrefix(ConfigOption<String> formatOption, String formatIdentifier) { final String formatOptionKey = formatOption.key(); if (formatOptionKey.equals(FORMAT.key())) { return formatIdentifier + "."; } else if (for...
3.26
flink_FactoryUtil_discoverDecodingFormat_rdh
/** * Discovers a {@link DecodingFormat} of the given type using the given option as factory * identifier. */ public <I, F extends DecodingFormatFactory<I>> DecodingFormat<I> discoverDecodingFormat(Class<F> formatFactoryClass, ConfigOption<String> formatOption) { return discoverOptionalDecodingFormat(formatFactoryCl...
3.26
flink_FactoryUtil_getDynamicTableFactory_rdh
// -------------------------------------------------------------------------------------------- // Helper methods // -------------------------------------------------------------------------------------------- private static <T extends DynamicTableFactory> T getDynamicTableFactory(Class<T> factoryClass, @Nullable C...
3.26
flink_FactoryUtil_createDynamicTableSource_rdh
/** * Creates a {@link DynamicTableSource} from a {@link CatalogTable}. * * <p>If {@param preferredFactory} is passed, the table source is created from that factory. * Otherwise, an attempt is made to discover a matching factory using Java SPI (see {@link Factory} for details). */public static DynamicTableSource c...
3.26
flink_FactoryUtil_checkWatermarkOptions_rdh
/** * Check watermark-related options and return error messages. * * @param conf * table options * @return Optional of error messages */ public static Optional<String> checkWatermarkOptions(ReadableConfig conf) { // try to validate watermark options by parsing it f1.forEach(option -> readOption(conf, option)); ...
3.26
flink_FactoryUtil_createCatalogFactoryHelper_rdh
/** * Creates a utility that helps validating options for a {@link CatalogFactory}. * * <p>Note: This utility checks for left-over options in the final step. */ public static CatalogFactoryHelper createCatalogFactoryHelper(CatalogFactory factory, CatalogFactory.Context context) { return new CatalogFactoryHelpe...
3.26
flink_SharingPhysicalSlotRequestBulk_clearPendingRequests_rdh
/** * Clear the pending requests. * * <p>The method can be used to make the bulk fulfilled and stop the fulfillability check in * {@link PhysicalSlotRequestBulkChecker}. */ void clearPendingRequests() { pendingRequests.clear(); }
3.26
flink_SharingPhysicalSlotRequestBulk_markFulfilled_rdh
/** * Moves a pending request to fulfilled. * * @param group * {@link ExecutionSlotSharingGroup} of the pending request * @param allocationId * {@link AllocationID} of the fulfilled request */ void markFulfilled(ExecutionSlotSharingGroup group, AllocationID allocationId) { pendingRequests.remove(group); fulf...
3.26
flink_TimeEvictor_hasTimestamp_rdh
/** * Returns true if the first element in the Iterable of {@link TimestampedValue} has a * timestamp. */ private boolean hasTimestamp(Iterable<TimestampedValue<Object>> elements) { Iterator<TimestampedValue<Object>> it = elements.iterator();if (it.hasNext()) { return it.next().hasTimestamp();} retu...
3.26
flink_TimeEvictor_m0_rdh
/** * * @param elements * The elements currently in the pane. * @return The maximum value of timestamp among the elements. */ private long m0(Iterable<TimestampedValue<Object>> elements) { long currentTime = Long.MIN_VALUE; for (Iterator<TimestampedValue<Object>> iterator = elements.iterator(); itera...
3.26
flink_TimeEvictor_of_rdh
/** * Creates a {@code TimeEvictor} that keeps the given number of elements. Eviction is done * before/after the window function based on the value of doEvictAfter. * * @param windowSize * The amount of time for which to keep elements. * @param doEvictAfter * Whether eviction is done after window function. ...
3.26
flink_TaskManagerLocation_getNodeId_rdh
/** * Return the ID of node where the task manager is located on. * * @return The ID of node where the task manager is located on. */ public String getNodeId() { return nodeId; }
3.26
flink_TaskManagerLocation_getFQDNHostname_rdh
/** * Returns the fully-qualified domain name of the TaskManager provided by {@link #hostNameSupplier}. * * @return The fully-qualified domain name of the TaskManager. */ public String getFQDNHostname() { return hostNameSupplier.getFqdnHostName(); }
3.26
flink_TaskManagerLocation_getHostName_rdh
/** * Returns the textual representation of the TaskManager's IP address as host name. * * @return The textual representation of the TaskManager's IP address. */ @Override public String getHostName() { return inetAddress.getHostAddress(); }
3.26
flink_TaskManagerLocation_getResourceID_rdh
// ------------------------------------------------------------------------ // Getters // ------------------------------------------------------------------------ /** * Gets the ID of the resource in which the TaskManager is started. The format of this depends * on how the TaskManager is start...
3.26
flink_TaskManagerLocation_getFqdnHostName_rdh
/** * Returns the textual representation of the TaskManager's IP address as FQDN host name. * * @return The textual representation of the TaskManager's IP address. */ @Overridepublic String getFqdnHostName() { return inetAddress.getHostAddress(); }
3.26
flink_IntValueComparator_m0_rdh
// -------------------------------------------------------------------------------------------- // unsupported normalization // -------------------------------------------------------------------------------------------- @Override public boolean m0() { return false; }
3.26
flink_NonClosingCheckpointOutputStream_acquireLease_rdh
/** * Returns a {@link org.apache.flink.util.ResourceGuard.Lease} that prevents closing this * stream. To allow the system to close this stream, each of the acquired leases need to call * {@link Lease#close()}, on their acquired leases. */ public final Lease acquireLease() throws IOException { return resourceGu...
3.26
flink_NonClosingCheckpointOutputStream_getDelegate_rdh
/** * This method should not be public so as to not expose internals to user code. */ CheckpointStateOutputStream getDelegate() { return delegate; }
3.26
flink_ParquetColumnarRowInputFormat_createPartitionedFormat_rdh
/** * Create a partitioned {@link ParquetColumnarRowInputFormat}, the partition columns can be * generated by {@link Path}. */ public static <SplitT extends FileSourceSplit> ParquetColumnarRowInputFormat<SplitT> createPartitionedFormat(Configuration hadoopConfig, RowType producedRowType, TypeInformation<RowData> pro...
3.26
flink_FsCheckpointStreamFactory_flush_rdh
/** * Flush buffers to file if their size is above {@link #localStateThreshold}. */ @Override public void flush() throws IOException { if ((outStream != null) || (pos > localStateThreshold)) { flushToFile(); } }
3.26
flink_FsCheckpointStreamFactory_isClosed_rdh
/** * Checks whether the stream is closed. * * @return True if the stream was closed, false if it is still open. */ public boolean isClosed() { return closed; }
3.26
flink_FsCheckpointStreamFactory_toString_rdh
// ------------------------------------------------------------------------ // utilities // ------------------------------------------------------------------------ @Override public String toString() { return "File Stream Factory @ " + checkpointDirectory; }
3.26
flink_FsCheckpointStreamFactory_createCheckpointStateOutputStream_rdh
// ------------------------------------------------------------------------ @Override public FsCheckpointStateOutputStream createCheckpointStateOutputStream(CheckpointedStateScope scope) throws IOException { Path target = getTargetPath(scope); int bufferSize = Math.max(writeBufferSize, fileStateThreshold); ...
3.26
flink_FsCheckpointStreamFactory_close_rdh
/** * If the stream is only closed, we remove the produced file (cleanup through the auto close * feature, for example). This method throws no exception if the deletion fails, but only * logs the error. */ @Override public void close() { if (!closed) {closed = true; // make sure write requests need to go to 'fl...
3.26
flink_CatalogTable_m0_rdh
/** * Serializes this instance into a map of string-based properties. * * <p>Compared to the pure table options in {@link #getOptions()}, the map includes schema, * partitioning, and other characteristics in a serialized form. * * @deprecated Only a {@link ResolvedCatalogTable} is serializable to properties. */ ...
3.26
flink_CatalogTable_getSnapshot_rdh
/** * Return the snapshot specified for the table. Return Optional.empty() if not specified. */ default Optional<Long> getSnapshot() { return Optional.empty(); }
3.26
flink_CatalogTable_of_rdh
/** * Creates an instance of {@link CatalogTable} with a specific snapshot. * * @param schema * unresolved schema * @param comment * optional comment * @param partitionKeys * list of partition keys or an empty list if not partitioned * @param options * options to configure the connector * @param snap...
3.26
flink_CatalogTable_fromProperties_rdh
/** * Creates an instance of {@link CatalogTable} from a map of string properties that were * previously created with {@link ResolvedCatalogTable#toProperties()}. * * <p>Note that the serialization and deserialization of catalog tables are not symmetric. The * framework will resolve functions and perform other val...
3.26
flink_TaskManagerServicesConfiguration_fromConfiguration_rdh
// -------------------------------------------------------------------------------------------- // Parsing of Flink configuration // -------------------------------------------------------------------------------------------- /** * Utility method to extract TaskManager config parameters from the configuration and to s...
3.26
flink_TaskManagerServicesConfiguration_getConfiguration_rdh
// -------------------------------------------------------------------------------------------- // Getter/Setter // -------------------------------------------------------------------------------------------- public Configuration getConfiguration() { return configuration; }
3.26
flink_EnumTypeInfo_toString_rdh
// ------------------------------------------------------------------------ // Standard utils // ------------------------------------------------------------------------ @Override public String toString() { return ("EnumTypeInfo<" + typeClass.getName()) + ">"; }
3.26
flink_CrossDriver_setup_rdh
// ------------------------------------------------------------------------ @Override public void setup(TaskContext<CrossFunction<T1, T2, OT>, OT> context) { this.taskContext = context; this.running = true; }
3.26
flink_SourceOutputWithWatermarks_collect_rdh
// ------------------------------------------------------------------------ // SourceOutput Methods // // Note that the two methods below are final, as a partial enforcement // of the performance design goal mentioned in the class-level comment. // ----------------------------------------------------------------------...
3.26
flink_SourceOutputWithWatermarks_createWithSeparateOutputs_rdh
// ------------------------------------------------------------------------ // Factories // ------------------------------------------------------------------------ /** * Creates a new SourceOutputWithWatermarks that emits records to the given DataOutput and * watermarks to the different WatermarkOutputs. */ public ...
3.26
flink_EndOfSuperstepEvent_hashCode_rdh
// ------------------------------------------------------------------------ @Override public int hashCode() { return 41; }
3.26
flink_CheckpointConfig_setCheckpointIdOfIgnoredInFlightData_rdh
/** * Setup the checkpoint id for which the in-flight data will be ignored for all operators in * case of the recovery from this checkpoint. * * @param checkpointIdOfIgnoredInFlightData * Checkpoint id for which in-flight data should be * ignored. * @see #setCheckpointIdOfIgnoredInFlightData */ @PublicEvolv...
3.26
flink_CheckpointConfig_getAlignedCheckpointTimeout_rdh
/** * * @return value of alignment timeout, as configured via {@link #setAlignedCheckpointTimeout(Duration)} or {@link ExecutionCheckpointingOptions#ALIGNED_CHECKPOINT_TIMEOUT}. */ @PublicEvolving public Duration getAlignedCheckpointTimeout() { return configuration.get(ExecutionCheckpointingOptions.ALIGNED_CHECK...
3.26
flink_CheckpointConfig_configure_rdh
/** * Sets all relevant options contained in the {@link ReadableConfig} such as e.g. {@link ExecutionCheckpointingOptions#CHECKPOINTING_MODE}. * * <p>It will change the value of a setting only if a corresponding option was set in the {@code configuration}. If a key is not present, the current value of a field will r...
3.26
flink_CheckpointConfig_getTolerableCheckpointFailureNumber_rdh
/** * Get the defined number of consecutive checkpoint failures that will be tolerated, before the * whole job is failed over. * * <p>If the {@link ExecutionCheckpointingOptions#TOLERABLE_FAILURE_NUMBER} has not been * configured, this method would return 0 which means the checkpoint failure manager would not * t...
3.26
flink_CheckpointConfig_setFailOnCheckpointingErrors_rdh
/** * Sets the expected behaviour for tasks in case that they encounter an error when * checkpointing. If this is set as true, which is equivalent to set * tolerableCheckpointFailureNumber as zero, job manager would fail the whole job once it * received a decline checkpoint message. If this is set as false, which i...
3.26
flink_CheckpointConfig_getMinPauseBetweenCheckpoints_rdh
/** * Gets the minimal pause between checkpointing attempts. This setting defines how soon the * checkpoint coordinator may trigger another checkpoint after it becomes possible to trigger * another checkpoint with respect to the maximum number of concurrent checkpoints (see {@link #getMaxConcurrentCheckpoints()}). ...
3.26
flink_CheckpointConfig_setCheckpointingMode_rdh
/** * Sets the checkpointing mode (exactly-once vs. at-least-once). * * @param checkpointingMode * The checkpointing mode. */ public void setCheckpointingMode(CheckpointingMode checkpointingMode) { configuration.set(ExecutionCheckpointingOptions.CHECKPOINTING_MODE, checkpointingMode); } /** * Gets the inte...
3.26
flink_CheckpointConfig_setForceCheckpointing_rdh
/** * Checks whether checkpointing is forced, despite currently non-checkpointable iteration * feedback. * * @param forceCheckpointing * The flag to force checkpointing. * @deprecated This will be removed once iterations properly participate in checkpointing. */ @Deprecated @PublicEvolving public void setForce...
3.26
flink_CheckpointConfig_setAlignedCheckpointTimeout_rdh
/** * Only relevant if {@link ExecutionCheckpointingOptions.ENABLE_UNALIGNED} is enabled. * * <p>If {@link ExecutionCheckpointingOptions#ALIGNED_CHECKPOINT_TIMEOUT} has value equal to * <code>0</code>, checkpoints will * * <p>always start unaligned. * * <p>If {@link ExecutionCheckpointingOptions#ALIGNED_CHECKPO...
3.26
flink_CheckpointConfig_setAlignmentTimeout_rdh
/** * Only relevant if {@link #isUnalignedCheckpointsEnabled} is enabled. * * <p>If {@link ExecutionCheckpointingOptions#ALIGNED_CHECKPOINT_TIMEOUT} has value equal to * <code>0</code>, checkpoints will always start unaligned. * * <p>If {@link ExecutionCheckpointingOptions#ALIGNED_CHECKPOINT_TIMEOUT} has value gr...
3.26
flink_CheckpointConfig_enableUnalignedCheckpoints_rdh
/** * Enables unaligned checkpoints, which greatly reduce checkpointing times under backpressure. * * <p>Unaligned checkpoints contain data stored in buffers as part of the checkpoint state, * which allows checkpoint barriers to overtake these buffers. Thus, the checkpoint duration * becomes independent of the cur...
3.26
flink_CheckpointConfig_getCheckpointIntervalDuringBacklog_rdh
/** * Gets the interval in which checkpoints are periodically scheduled during backlog. * * <p>This setting defines the base interval. Checkpoint triggering may be delayed by the * settings {@link #getMaxConcurrentCheckpoints()} and {@link #getMinPauseBetweenCheckpoints()}. * * <p>If not explicitly configured, ch...
3.26
flink_CheckpointConfig_setCheckpointTimeout_rdh
/** * Sets the maximum time that a checkpoint may take before being discarded. * * @param checkpointTimeout * The checkpoint timeout, in milliseconds. */ public void setCheckpointTimeout(long checkpointTimeout) { if (checkpointTimeout < MINIMAL_CHECKPOINT_TIME) { throw new IllegalArgumentException(...
3.26
flink_CheckpointConfig_getMaxSubtasksPerChannelStateFile_rdh
/** * * @return the number of subtasks to share the same channel state file, as configured via {@link #setMaxSubtasksPerChannelStateFile(int)} or {@link ExecutionCheckpointingOptions#UNALIGNED_MAX_SUBTASKS_PER_CHANNEL_STATE_FILE}. */ @PublicEvolving public int getMaxSubtasksPerChannelStateFile() { return configu...
3.26
flink_CheckpointConfig_disableCheckpointing_rdh
// ------------------------------------------------------------------------ /** * Disables checkpointing. */ public void disableCheckpointing() { configuration.removeConfig(ExecutionCheckpointingOptions.CHECKPOINTING_INTERVAL); }
3.26
flink_CheckpointConfig_setForceUnalignedCheckpoints_rdh
/** * Checks whether unaligned checkpoints are forced, despite currently non-checkpointable * iteration feedback or custom partitioners. * * @param forceUnalignedCheckpoints * The flag to force unaligned checkpoints. */ @PublicEvolving public void setForceUnalignedCheckpoints(boolean forceUnalignedCheckpoints) ...
3.26
flink_CheckpointConfig_isForceCheckpointing_rdh
/** * Checks whether checkpointing is forced, despite currently non-checkpointable iteration * feedback. * * @return True, if checkpointing is forced, false otherwise. * @deprecated This will be removed once iterations properly participate in checkpointing. */@Deprecated @PublicEvolving public boolean isForceChec...
3.26
flink_CheckpointConfig_m1_rdh
/** * CheckpointStorage defines how {@link StateBackend}'s checkpoint their state for fault * tolerance in streaming applications. Various implementations store their checkpoints in * different fashions and have different requirements and availability guarantees. * * <p>For example, {@link org.apache.flink.runtime...
3.26
flink_CheckpointConfig_isExternalizedCheckpointsEnabled_rdh
/** * Returns whether checkpoints should be persisted externally. * * @return <code>true</code> if checkpoints should be externalized. */ @PublicEvolving public boolean isExternalizedCheckpointsEnabled() { return getExternalizedCheckpointCleanup() != ExternalizedCheckpointCleanup.NO_EXTERNALIZED_CHECKPOINTS; }
3.26
flink_CheckpointConfig_isForceUnalignedCheckpoints_rdh
/** * Checks whether unaligned checkpoints are forced, despite iteration feedback. * * @return True, if unaligned checkpoints are forced, false otherwise. */ @PublicEvolving public boolean isForceUnalignedCheckpoints() { return configuration.get(ExecutionCheckpointingOptions.FORCE_UNALIGNED); }
3.26
flink_CheckpointConfig_setCheckpointIntervalDuringBacklog_rdh
/** * Sets the interval in which checkpoints are periodically scheduled during backlog. * * <p>This setting defines the base interval. Checkpoint triggering may be delayed by the * settings {@link #setMaxConcurrentCheckpoints(int)} and {@link #setMinPauseBetweenCheckpoints(long)}. * * <p>If not explicitly configu...
3.26
flink_CheckpointConfig_enableApproximateLocalRecovery_rdh
/** * Enables the approximate local recovery mode. * * <p>In this recovery mode, when a task fails, the entire downstream of the tasks (including * the failed task) restart. * * <p>Notice that 1. Approximate recovery may lead to data loss. The amount of data which leads * the failed task from the state of the la...
3.26
flink_CheckpointConfig_setMaxSubtasksPerChannelStateFile_rdh
/** * The number of subtasks to share the same channel state file. If {@link ExecutionCheckpointingOptions#UNALIGNED_MAX_SUBTASKS_PER_CHANNEL_STATE_FILE} has value equal * to <code>1</code>, each subtask will create a new channel state file. */ @PublicEvolving public void setMaxSubtasksPerChannelStateFile(int maxSub...
3.26