name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_RocksDBIncrementalRestoreOperation_readMetaData_rdh | /**
* Reads Flink's state meta data file from the state handle.
*/
private KeyedBackendSerializationProxy<K> readMetaData(StreamStateHandle metaStateHandle) throws Exception {
InputStream inputStream = null;
try {
inputStream = metaStateHandle.openInputStream();
cancelStreamRegistry.registerCloseable(inputStream);
D... | 3.26 |
flink_RocksDBIncrementalRestoreOperation_restore_rdh | /**
* Root method that branches for different implementations of {@link KeyedStateHandle}.
*/
@Override
public RocksDBRestoreResult restore() throws Exception {
if ((restoreStateHandles == null) || restoreStateHandles.isEmpty()) {
return null;
}
final KeyedStat... | 3.26 |
flink_RocksDBIncrementalRestoreOperation_createColumnFamilyDescriptors_rdh | /**
* This method recreates and registers all {@link ColumnFamilyDescriptor} from Flink's state
* meta data snapshot.
*/
private List<ColumnFamilyDescriptor> createColumnFamilyDescriptors(List<StateMetaInfoSnapshot> stateMetaInfoSnapshots, boolean registerTtlCompactFilter) {
List<ColumnFamilyDescriptor> columnFamily... | 3.26 |
flink_SubtaskState_getManagedOperatorState_rdh | // --------------------------------------------------------------------------------------------
public ChainedStateHandle<OperatorStateHandle> getManagedOperatorState() {
return managedOperatorState;
} | 3.26 |
flink_SubtaskState_equals_rdh | // --------------------------------------------------------------------------------------------
@Override
public boolean equals(Object o) {
if (this == o) {return true;
}
if ((o == null) || (getClass() != o.getClass()))
{
return false;
}
SubtaskState that = ((SubtaskState) (o));
if (stateSize != that.stateSize) {r... | 3.26 |
flink_Signature_of_rdh | /**
* Creates an immutable instance of {@link Signature}.
*/
public static Signature of(List<Argument> arguments) {
return
new Signature(arguments);
} | 3.26 |
flink_AggregatorRegistry_registerAggregator_rdh | // --------------------------------------------------------------------------------------------
public void registerAggregator(String name, Aggregator<?> aggregator) {
if ((name == null) || (aggregator == null)) {
throw new
IllegalArgumentException("Name and aggregator must not be null");}
if (t... | 3.26 |
flink_PackagedProgram_deleteExtractedLibraries_rdh | /**
* Deletes all temporary files created for contained packaged libraries.
*/
private void deleteExtractedLibraries() {
deleteExtractedLibraries(this.extractedTempLibraries);
this.extractedTempLibraries.clear();
} | 3.26 |
flink_PackagedProgram_getJobJarAndDependencies_rdh | /**
* Returns all provided libraries needed to run the program.
*/
public static List<URL> getJobJarAndDependencies(File jarFile, @Nullable
String entryPointClassName) throws ProgramInvocationException {
URL jarFileUrl = loadJarFile(jarFile);
List<File> extractedTempLibraries = (jarFileUrl == null) ? Collec... | 3.26 |
flink_PackagedProgram_invokeInteractiveModeForExecution_rdh | /**
* This method assumes that the context environment is prepared, or the execution will be a
* local execution by default.
*/
public void invokeInteractiveModeForExecution() throws ProgramInvocationException {
FlinkSecurityManager.monitorUserSystemExitForCurrentThrea... | 3.26 |
flink_PackagedProgram_extractContainedLibraries_rdh | /**
* Takes all JAR files that are contained in this program's JAR file and extracts them to the
* system's temp directory.
*
* @return The file names of the extracted temporary files.
* @throws ProgramInvocationException
* Thrown, if the extraction process failed.
*/
public static List<File> extractContainedL... | 3.26 |
flink_FutureCompletingBlockingQueue_peek_rdh | /**
* Get the first element from the queue without removing it.
*
* @return the first element in the queue, or Null if the queue is empty.
*/
public T peek() {
lock.lock();
try {
return queue.peek();
} finally {
lock.unlock();
}
} | 3.26 |
flink_FutureCompletingBlockingQueue_getAvailableFuture_rdh | // ------------------------------------------------------------------------
// utilities
// ------------------------------------------------------------------------
@SuppressWarnings("unchecked")
private static CompletableFuture<Void> getAvailableFuture() {// this is a way to obtain the AvailabilityProvider.AVAILABLE f... | 3.26 |
flink_FutureCompletingBlockingQueue_put_rdh | // ------------------------------------------------------------------------
// Blocking Queue Logic
// ------------------------------------------------------------------------
/**
* Put an element into the queue. The thread blocks if the queue is full.
*
* @param threadIndex
* the index of the thread.
* @param e... | 3.26 |
flink_FutureCompletingBlockingQueue_wakeUpPuttingThread_rdh | /**
* Gracefully wakes up the thread with the given {@code threadIndex} if it is blocked in adding
* an element. to the queue. If the thread is blocked in {@link #put(int, Object)} it will
* immediately return from the method with a return value of false.
*
* <p>If this method is called, the next time the thread w... | 3.26 |
flink_FutureCompletingBlockingQueue_size_rdh | /**
* Gets the size of the queue.
*/
public int size() {
lock.lock();
try {
return queue.size();
} finally {
lock.unlock();
}
} | 3.26 |
flink_FutureCompletingBlockingQueue_remainingCapacity_rdh | /**
* Checks the remaining capacity in the queue. That is the difference between the maximum
* capacity and the current number of elements in the queue.
*/
public int
remainingCapacity() {
lock.lock();
try {
return capacity - queue.size();
} finally {
lock.unlock();
}
} | 3.26 |
flink_FutureCompletingBlockingQueue_getAvailabilityFuture_rdh | // ------------------------------------------------------------------------
// Future / Notification logic
// ------------------------------------------------------------------------
/**
* Returns the availability future. If the queue is non-empty, then this future will already be
* complete. Otherwise the obtained f... | 3.26 |
flink_FutureCompletingBlockingQueue_m1_rdh | /**
* Checks whether the queue is empty.
*/
public boolean m1() {
lock.lock();
try {
return queue.isEmpty();
} finally {
lock.unlock();
}
} | 3.26 |
flink_FutureCompletingBlockingQueue_poll_rdh | /**
* Get and remove the first element from the queue. Null is returned if the queue is empty. If
* this makes the queue empty (takes the last element) or finds the queue already empty, then
* this resets the availability notifications. The next call to {@link #getAvailabilityFuture()}
* will then return a non-comp... | 3.26 |
flink_FutureCompletingBlockingQueue_moveToAvailable_rdh | /**
* Internal utility to make sure that the current future futures are complete (until reset).
*/
@GuardedBy("lock")
private void moveToAvailable() {
final CompletableFuture<Void> current = currentFuture;
if (current != AVAILABLE) {
currentFuture = AVAILABLE;
current.complete(null);
}
} | 3.26 |
flink_FutureCompletingBlockingQueue_enqueue_rdh | // --------------- private helpers -------------------------
@GuardedBy("lock")
private void enqueue(T element) {
final int sizeBefore = queue.size();
queue.add(element);
if (sizeBefore == 0) {
moveToAvailable();
}
if ((sizeBefore < (capacity - 1)) && (!notFull.isEmpty())) {
signa... | 3.26 |
flink_FutureCompletingBlockingQueue_take_rdh | /**
* <b>Warning:</b> This is a dangerous method and should only be used for testing convenience. A
* method that blocks until availability does not go together well with the concept of
* asynchronous notifications and non-blocking polling.
*
* <p>Get and remove the first element from the queue. The call blocks if... | 3.26 |
flink_FactoryUtil_createCatalogStoreFactoryHelper_rdh | /**
* Creates a utility that helps validating options for a {@link CatalogStoreFactory}.
*
* <p>Note: This utility checks for left-over options in the final step.
*/public static CatalogStoreFactoryHelper createCatalogStoreFactoryHelper(CatalogStoreFactory factory, CatalogStoreFactory.Context context) {
return... | 3.26 |
flink_FactoryUtil_createModuleFactoryHelper_rdh | /**
* Creates a utility that helps validating options for a {@link ModuleFactory}.
*
* <p>Note: This utility checks for left-over options in the final step.
*/
public static ModuleFactoryHelper createModuleFactoryHelper(ModuleFactory factory, ModuleFactory.Context context) {
return
new ModuleFactoryHelper... | 3.26 |
flink_FactoryUtil_validateFactoryOptions_rdh | /**
* Validates the required and optional {@link ConfigOption}s of a factory.
*
* <p>Note: It does not check for left-over options.
*/
public static void validateFactoryOptions(Factory factory, ReadableConfig options) {
validateFactoryOptions(factory.requiredOptions(), factory.optionalOptions(),
options);
} | 3.26 |
flink_FactoryUtil_discoverEncodingFormat_rdh | /**
* Discovers a {@link EncodingFormat} of the given type using the given option as factory
* identifier.
*/
public <I, F extends EncodingFormatFactory<I>> EncodingFormat<I> discoverEncodingFormat(Class<F> formatFactoryClass, ConfigOption<String> formatOption) {
return discoverOptionalEncodingFormat(formatFactoryCl... | 3.26 |
flink_FactoryUtil_discoverOptionalEncodingFormat_rdh | /**
* Discovers a {@link EncodingFormat} of the given type using the given option (if present)
* as factory identifier.
*/
public <I, F extends EncodingFormatFactory<I>> Optional<EncodingFormat<I>> discoverOptionalEncodingFormat(Class<F> formatFactoryClass, ConfigOption<String> formatOption) {
return discoverOption... | 3.26 |
flink_FactoryUtil_forwardOptions_rdh | // ----------------------------------------------------------------------------------------
/**
* Forwards the options declared in {@link DynamicTableFactory#forwardOptions()} and
* possibly {@link FormatFactory#forwardOptions()} from {@link DynamicTableFactory.Context#getEnrichmentOptions()} to the final options, if... | 3.26 |
flink_FactoryUtil_discoverFactory_rdh | /**
* Discovers a factory using the given factory base class and identifier.
*
* <p>This method is meant for cases where {@link #createTableFactoryHelper(DynamicTableFactory,
* DynamicTableFactory.Context)} {@link #createTableSource(Catalog, ObjectIdentifier,
* ResolvedCatalogTable, ReadableConfig, ClassLoader, bo... | 3.26 |
flink_FactoryUtil_m1_rdh | /**
* Validates the options of the factory. It checks for unconsumed option keys.
*/public void m1() {
validateFactoryOptions(factory, allOptions);
validateUnconsumedKeys(factory.factoryIdentifier(), allOptions.keySet(), consumedOptionKeys, deprecatedOptionKeys);
validateWatermarkOptions(factory.factoryIdentifier(),... | 3.26 |
flink_FactoryUtil_createDynamicTableSink_rdh | /**
* Creates a {@link DynamicTableSink} from a {@link CatalogTable}.
*
* <p>If {@param preferredFactory} is passed, the table sink is created from that factory.
* Otherwise, an attempt is made to discover a matching factory using Java SPI (see {@link Factory} for details).
*/
public static DynamicTableSink create... | 3.26 |
flink_FactoryUtil_checkFormatIdentifierMatchesWithEnrichingOptions_rdh | /**
* This function assumes that the format config is used only and only if the original
* configuration contains the format config option. It will fail if there is a mismatch of
* the identifier between the format in the plan table map and the one in enriching table
* map.
*/
private void checkFormatIdentifierMat... | 3.26 |
flink_FactoryUtil_createTableFactoryHelper_rdh | /**
* Creates a utility that helps in discovering formats, merging options with {@link DynamicTableFactory.Context#getEnrichmentOptions()} and validating them all for a {@link DynamicTableFactory}.
*
* <p>The following example sketches the usage:
*
* <pre>{@code // in createDynamicTableSource()
* helper = Factory... | 3.26 |
flink_FactoryUtil_discoverOptionalDecodingFormat_rdh | /**
* Discovers a {@link DecodingFormat} of the given type using the given option (if present)
* as factory identifier.
*/
public <I, F extends DecodingFormatFactory<I>> Optional<DecodingFormat<I>> discoverOptionalDecodingFormat(Class<F> formatFactoryClass, ConfigOption<String> formatOption) {
return discoverOptiona... | 3.26 |
flink_FactoryUtil_validateWatermarkOptions_rdh | // --------------------------------------------------------------------------------------------
/**
* Validate watermark options from table options.
*
* @param factoryIdentifier
* identifier of table
* @param conf
* table options
*/
public static void validateWatermarkOptions(String factoryIdentifier, Readab... | 3.26 |
flink_FactoryUtil_createCatalog_rdh | /**
* Attempts to discover an appropriate catalog factory and creates an instance of the catalog.
*
* <p>This first uses the legacy {@link TableFactory} stack to discover a matching {@link CatalogFactory}. If none is found, it falls back to the new stack using {@link Factory}
* instead.
*/ public static Catalog cr... | 3.26 |
flink_FactoryUtil_createModule_rdh | /**
* Discovers a matching module factory and creates an instance of it.
*
* <p>This first uses the legacy {@link TableFactory} stack to discover a matching {@link ModuleFactory}. If none is found, it falls back to the new stack using {@link Factory}
* instead.
*/
public static Module createModule(String moduleNam... | 3.26 |
flink_FactoryUtil_validateUnconsumedKeys_rdh | /**
* Validates unconsumed option keys.
*/
public static void validateUnconsumedKeys(String factoryIdentifier, Set<String> allOptionKeys, Set<String> consumedOptionKeys) {
validateUnconsumedKeys(factoryIdentifier, allOptionKeys, consumedOptionKeys, Collections.emptySet());
} | 3.26 |
flink_FactoryUtil_validateExcept_rdh | /**
* Validates the options of the factory. It checks for unconsumed option keys while ignoring
* the options with given prefixes.
*
* <p>The option keys that have given prefix {@code prefixToSkip} would just be skipped for
* validation.
*
* @param prefixesToSkip
* Set of option key prefixes to skip validatio... | 3.26 |
flink_FactoryUtil_getOptions_rdh | /**
* Returns all options currently being consumed by the factory. This method returns the
* options already merged with {@link DynamicTableFactory.Context#getEnrichmentOptions()},
* using {@link DynamicTableFactory#forwardOptions()} as reference of mergeable options.
*/
@Override
public ReadableConfig getOptions()... | 3.26 |
flink_FactoryUtil_getFormatPrefix_rdh | /**
* Returns the required option prefix for options of the given format.
*/
public static String getFormatPrefix(ConfigOption<String> formatOption, String formatIdentifier) {
final String formatOptionKey = formatOption.key();
if (formatOptionKey.equals(FORMAT.key())) {
return formatIdentifier + ".";
} else
if (for... | 3.26 |
flink_FactoryUtil_discoverDecodingFormat_rdh | /**
* Discovers a {@link DecodingFormat} of the given type using the given option as factory
* identifier.
*/
public <I, F extends DecodingFormatFactory<I>> DecodingFormat<I> discoverDecodingFormat(Class<F> formatFactoryClass, ConfigOption<String> formatOption) {
return discoverOptionalDecodingFormat(formatFactoryCl... | 3.26 |
flink_FactoryUtil_getDynamicTableFactory_rdh | // --------------------------------------------------------------------------------------------
// Helper methods
// --------------------------------------------------------------------------------------------
private static <T extends DynamicTableFactory> T getDynamicTableFactory(Class<T> factoryClass, @Nullable
C... | 3.26 |
flink_FactoryUtil_createDynamicTableSource_rdh | /**
* Creates a {@link DynamicTableSource} from a {@link CatalogTable}.
*
* <p>If {@param preferredFactory} is passed, the table source is created from that factory.
* Otherwise, an attempt is made to discover a matching factory using Java SPI (see {@link Factory} for details).
*/public static DynamicTableSource c... | 3.26 |
flink_FactoryUtil_checkWatermarkOptions_rdh | /**
* Check watermark-related options and return error messages.
*
* @param conf
* table options
* @return Optional of error messages
*/
public static Optional<String> checkWatermarkOptions(ReadableConfig conf) {
// try to validate watermark options by parsing it
f1.forEach(option -> readOption(conf, option));
... | 3.26 |
flink_FactoryUtil_createCatalogFactoryHelper_rdh | /**
* Creates a utility that helps validating options for a {@link CatalogFactory}.
*
* <p>Note: This utility checks for left-over options in the final step.
*/
public static CatalogFactoryHelper createCatalogFactoryHelper(CatalogFactory factory, CatalogFactory.Context context) {
return new CatalogFactoryHelpe... | 3.26 |
flink_SharingPhysicalSlotRequestBulk_clearPendingRequests_rdh | /**
* Clear the pending requests.
*
* <p>The method can be used to make the bulk fulfilled and stop the fulfillability check in
* {@link PhysicalSlotRequestBulkChecker}.
*/
void clearPendingRequests() {
pendingRequests.clear();
} | 3.26 |
flink_SharingPhysicalSlotRequestBulk_markFulfilled_rdh | /**
* Moves a pending request to fulfilled.
*
* @param group
* {@link ExecutionSlotSharingGroup} of the pending request
* @param allocationId
* {@link AllocationID} of the fulfilled request
*/
void markFulfilled(ExecutionSlotSharingGroup group, AllocationID allocationId) {
pendingRequests.remove(group);
fulf... | 3.26 |
flink_TimeEvictor_hasTimestamp_rdh | /**
* Returns true if the first element in the Iterable of {@link TimestampedValue} has a
* timestamp.
*/
private boolean hasTimestamp(Iterable<TimestampedValue<Object>> elements) {
Iterator<TimestampedValue<Object>> it = elements.iterator();if (it.hasNext()) {
return it.next().hasTimestamp();}
retu... | 3.26 |
flink_TimeEvictor_m0_rdh | /**
*
* @param elements
* The elements currently in the pane.
* @return The maximum value of timestamp among the elements.
*/
private long m0(Iterable<TimestampedValue<Object>> elements) {
long
currentTime = Long.MIN_VALUE;
for (Iterator<TimestampedValue<Object>> iterator = elements.iterator(); itera... | 3.26 |
flink_TimeEvictor_of_rdh | /**
* Creates a {@code TimeEvictor} that keeps the given number of elements. Eviction is done
* before/after the window function based on the value of doEvictAfter.
*
* @param windowSize
* The amount of time for which to keep elements.
* @param doEvictAfter
* Whether eviction is done after window function.
... | 3.26 |
flink_TaskManagerLocation_getNodeId_rdh | /**
* Return the ID of node where the task manager is located on.
*
* @return The ID of node where the task manager is located on.
*/
public String getNodeId() {
return nodeId;
} | 3.26 |
flink_TaskManagerLocation_getFQDNHostname_rdh | /**
* Returns the fully-qualified domain name of the TaskManager provided by {@link #hostNameSupplier}.
*
* @return The fully-qualified domain name of the TaskManager.
*/
public String getFQDNHostname() {
return hostNameSupplier.getFqdnHostName();
} | 3.26 |
flink_TaskManagerLocation_getHostName_rdh | /**
* Returns the textual representation of the TaskManager's IP address as host name.
*
* @return The textual representation of the TaskManager's IP address.
*/
@Override
public String getHostName() {
return inetAddress.getHostAddress();
} | 3.26 |
flink_TaskManagerLocation_getResourceID_rdh | // ------------------------------------------------------------------------
// Getters
// ------------------------------------------------------------------------
/**
* Gets the ID of the resource in which the TaskManager is started. The format of this depends
* on how the TaskManager is start... | 3.26 |
flink_TaskManagerLocation_getFqdnHostName_rdh | /**
* Returns the textual representation of the TaskManager's IP address as FQDN host name.
*
* @return The textual representation of the TaskManager's IP address.
*/
@Overridepublic String getFqdnHostName() {
return inetAddress.getHostAddress();
} | 3.26 |
flink_IntValueComparator_m0_rdh | // --------------------------------------------------------------------------------------------
// unsupported normalization
// --------------------------------------------------------------------------------------------
@Override
public boolean m0() {
return false;
} | 3.26 |
flink_NonClosingCheckpointOutputStream_acquireLease_rdh | /**
* Returns a {@link org.apache.flink.util.ResourceGuard.Lease} that prevents closing this
* stream. To allow the system to close this stream, each of the acquired leases need to call
* {@link Lease#close()}, on their acquired leases.
*/
public final Lease acquireLease() throws IOException {
return resourceGu... | 3.26 |
flink_NonClosingCheckpointOutputStream_getDelegate_rdh | /**
* This method should not be public so as to not expose internals to user code.
*/
CheckpointStateOutputStream getDelegate() {
return delegate;
} | 3.26 |
flink_ParquetColumnarRowInputFormat_createPartitionedFormat_rdh | /**
* Create a partitioned {@link ParquetColumnarRowInputFormat}, the partition columns can be
* generated by {@link Path}.
*/
public static <SplitT extends FileSourceSplit> ParquetColumnarRowInputFormat<SplitT> createPartitionedFormat(Configuration hadoopConfig, RowType producedRowType, TypeInformation<RowData> pro... | 3.26 |
flink_FsCheckpointStreamFactory_flush_rdh | /**
* Flush buffers to file if their size is above {@link #localStateThreshold}.
*/
@Override
public void flush() throws IOException {
if ((outStream != null) || (pos > localStateThreshold)) {
flushToFile();
}
} | 3.26 |
flink_FsCheckpointStreamFactory_isClosed_rdh | /**
* Checks whether the stream is closed.
*
* @return True if the stream was closed, false if it is still open.
*/
public boolean
isClosed() {
return closed;
} | 3.26 |
flink_FsCheckpointStreamFactory_toString_rdh | // ------------------------------------------------------------------------
// utilities
// ------------------------------------------------------------------------
@Override
public String toString() {
return "File Stream Factory @ " + checkpointDirectory;
} | 3.26 |
flink_FsCheckpointStreamFactory_createCheckpointStateOutputStream_rdh | // ------------------------------------------------------------------------
@Override
public FsCheckpointStateOutputStream createCheckpointStateOutputStream(CheckpointedStateScope scope) throws IOException {
Path target = getTargetPath(scope);
int bufferSize = Math.max(writeBufferSize, fileStateThreshold);
... | 3.26 |
flink_FsCheckpointStreamFactory_close_rdh | /**
* If the stream is only closed, we remove the produced file (cleanup through the auto close
* feature, for example). This method throws no exception if the deletion fails, but only
* logs the error.
*/
@Override
public void close() {
if (!closed) {closed = true;
// make sure write requests need to go to 'fl... | 3.26 |
flink_CatalogTable_m0_rdh | /**
* Serializes this instance into a map of string-based properties.
*
* <p>Compared to the pure table options in {@link #getOptions()}, the map includes schema,
* partitioning, and other characteristics in a serialized form.
*
* @deprecated Only a {@link ResolvedCatalogTable} is serializable to properties.
*/
... | 3.26 |
flink_CatalogTable_getSnapshot_rdh | /**
* Return the snapshot specified for the table. Return Optional.empty() if not specified.
*/
default Optional<Long> getSnapshot() {
return Optional.empty();
} | 3.26 |
flink_CatalogTable_of_rdh | /**
* Creates an instance of {@link CatalogTable} with a specific snapshot.
*
* @param schema
* unresolved schema
* @param comment
* optional comment
* @param partitionKeys
* list of partition keys or an empty list if not partitioned
* @param options
* options to configure the connector
* @param snap... | 3.26 |
flink_CatalogTable_fromProperties_rdh | /**
* Creates an instance of {@link CatalogTable} from a map of string properties that were
* previously created with {@link ResolvedCatalogTable#toProperties()}.
*
* <p>Note that the serialization and deserialization of catalog tables are not symmetric. The
* framework will resolve functions and perform other val... | 3.26 |
flink_TaskManagerServicesConfiguration_fromConfiguration_rdh | // --------------------------------------------------------------------------------------------
// Parsing of Flink configuration
// --------------------------------------------------------------------------------------------
/**
* Utility method to extract TaskManager config parameters from the configuration and to s... | 3.26 |
flink_TaskManagerServicesConfiguration_getConfiguration_rdh | // --------------------------------------------------------------------------------------------
// Getter/Setter
// --------------------------------------------------------------------------------------------
public Configuration getConfiguration() {
return configuration;
} | 3.26 |
flink_EnumTypeInfo_toString_rdh | // ------------------------------------------------------------------------
// Standard utils
// ------------------------------------------------------------------------
@Override
public String toString() {
return ("EnumTypeInfo<" + typeClass.getName()) + ">";
} | 3.26 |
flink_CrossDriver_setup_rdh | // ------------------------------------------------------------------------
@Override
public void setup(TaskContext<CrossFunction<T1, T2, OT>, OT> context) {
this.taskContext = context;
this.running = true;
} | 3.26 |
flink_SourceOutputWithWatermarks_collect_rdh | // ------------------------------------------------------------------------
// SourceOutput Methods
//
// Note that the two methods below are final, as a partial enforcement
// of the performance design goal mentioned in the class-level comment.
// ----------------------------------------------------------------------... | 3.26 |
flink_SourceOutputWithWatermarks_createWithSeparateOutputs_rdh | // ------------------------------------------------------------------------
// Factories
// ------------------------------------------------------------------------
/**
* Creates a new SourceOutputWithWatermarks that emits records to the given DataOutput and
* watermarks to the different WatermarkOutputs.
*/
public ... | 3.26 |
flink_EndOfSuperstepEvent_hashCode_rdh | // ------------------------------------------------------------------------
@Override
public int hashCode()
{
return 41;
} | 3.26 |
flink_CheckpointConfig_setCheckpointIdOfIgnoredInFlightData_rdh | /**
* Setup the checkpoint id for which the in-flight data will be ignored for all operators in
* case of the recovery from this checkpoint.
*
* @param checkpointIdOfIgnoredInFlightData
* Checkpoint id for which in-flight data should be
* ignored.
* @see #setCheckpointIdOfIgnoredInFlightData
*/
@PublicEvolv... | 3.26 |
flink_CheckpointConfig_getAlignedCheckpointTimeout_rdh | /**
*
* @return value of alignment timeout, as configured via {@link #setAlignedCheckpointTimeout(Duration)} or {@link ExecutionCheckpointingOptions#ALIGNED_CHECKPOINT_TIMEOUT}.
*/
@PublicEvolving
public Duration getAlignedCheckpointTimeout() {
return configuration.get(ExecutionCheckpointingOptions.ALIGNED_CHECK... | 3.26 |
flink_CheckpointConfig_configure_rdh | /**
* Sets all relevant options contained in the {@link ReadableConfig} such as e.g. {@link ExecutionCheckpointingOptions#CHECKPOINTING_MODE}.
*
* <p>It will change the value of a setting only if a corresponding option was set in the {@code configuration}. If a key is not present, the current value of a field will r... | 3.26 |
flink_CheckpointConfig_getTolerableCheckpointFailureNumber_rdh | /**
* Get the defined number of consecutive checkpoint failures that will be tolerated, before the
* whole job is failed over.
*
* <p>If the {@link ExecutionCheckpointingOptions#TOLERABLE_FAILURE_NUMBER} has not been
* configured, this method would return 0 which means the checkpoint failure manager would not
* t... | 3.26 |
flink_CheckpointConfig_setFailOnCheckpointingErrors_rdh | /**
* Sets the expected behaviour for tasks in case that they encounter an error when
* checkpointing. If this is set as true, which is equivalent to set
* tolerableCheckpointFailureNumber as zero, job manager would fail the whole job once it
* received a decline checkpoint message. If this is set as false, which i... | 3.26 |
flink_CheckpointConfig_getMinPauseBetweenCheckpoints_rdh | /**
* Gets the minimal pause between checkpointing attempts. This setting defines how soon the
* checkpoint coordinator may trigger another checkpoint after it becomes possible to trigger
* another checkpoint with respect to the maximum number of concurrent checkpoints (see {@link #getMaxConcurrentCheckpoints()}).
... | 3.26 |
flink_CheckpointConfig_setCheckpointingMode_rdh | /**
* Sets the checkpointing mode (exactly-once vs. at-least-once).
*
* @param checkpointingMode
* The checkpointing mode.
*/
public void setCheckpointingMode(CheckpointingMode checkpointingMode) {
configuration.set(ExecutionCheckpointingOptions.CHECKPOINTING_MODE, checkpointingMode);
}
/**
* Gets the inte... | 3.26 |
flink_CheckpointConfig_setForceCheckpointing_rdh | /**
* Checks whether checkpointing is forced, despite currently non-checkpointable iteration
* feedback.
*
* @param forceCheckpointing
* The flag to force checkpointing.
* @deprecated This will be removed once iterations properly participate in checkpointing.
*/
@Deprecated
@PublicEvolving
public void setForce... | 3.26 |
flink_CheckpointConfig_setAlignedCheckpointTimeout_rdh | /**
* Only relevant if {@link ExecutionCheckpointingOptions.ENABLE_UNALIGNED} is enabled.
*
* <p>If {@link ExecutionCheckpointingOptions#ALIGNED_CHECKPOINT_TIMEOUT} has value equal to
* <code>0</code>, checkpoints will
*
* <p>always start unaligned.
*
* <p>If {@link ExecutionCheckpointingOptions#ALIGNED_CHECKPO... | 3.26 |
flink_CheckpointConfig_setAlignmentTimeout_rdh | /**
* Only relevant if {@link #isUnalignedCheckpointsEnabled} is enabled.
*
* <p>If {@link ExecutionCheckpointingOptions#ALIGNED_CHECKPOINT_TIMEOUT} has value equal to
* <code>0</code>, checkpoints will always start unaligned.
*
* <p>If {@link ExecutionCheckpointingOptions#ALIGNED_CHECKPOINT_TIMEOUT} has value gr... | 3.26 |
flink_CheckpointConfig_enableUnalignedCheckpoints_rdh | /**
* Enables unaligned checkpoints, which greatly reduce checkpointing times under backpressure.
*
* <p>Unaligned checkpoints contain data stored in buffers as part of the checkpoint state,
* which allows checkpoint barriers to overtake these buffers. Thus, the checkpoint duration
* becomes independent of the cur... | 3.26 |
flink_CheckpointConfig_getCheckpointIntervalDuringBacklog_rdh | /**
* Gets the interval in which checkpoints are periodically scheduled during backlog.
*
* <p>This setting defines the base interval. Checkpoint triggering may be delayed by the
* settings {@link #getMaxConcurrentCheckpoints()} and {@link #getMinPauseBetweenCheckpoints()}.
*
* <p>If not explicitly configured, ch... | 3.26 |
flink_CheckpointConfig_setCheckpointTimeout_rdh | /**
* Sets the maximum time that a checkpoint may take before being discarded.
*
* @param checkpointTimeout
* The checkpoint timeout, in milliseconds.
*/
public void setCheckpointTimeout(long checkpointTimeout) {
if (checkpointTimeout < MINIMAL_CHECKPOINT_TIME) {
throw new IllegalArgumentException(... | 3.26 |
flink_CheckpointConfig_getMaxSubtasksPerChannelStateFile_rdh | /**
*
* @return the number of subtasks to share the same channel state file, as configured via {@link #setMaxSubtasksPerChannelStateFile(int)} or {@link ExecutionCheckpointingOptions#UNALIGNED_MAX_SUBTASKS_PER_CHANNEL_STATE_FILE}.
*/
@PublicEvolving
public int
getMaxSubtasksPerChannelStateFile() {
return configu... | 3.26 |
flink_CheckpointConfig_disableCheckpointing_rdh | // ------------------------------------------------------------------------
/**
* Disables checkpointing.
*/
public void disableCheckpointing() {
configuration.removeConfig(ExecutionCheckpointingOptions.CHECKPOINTING_INTERVAL);
} | 3.26 |
flink_CheckpointConfig_setForceUnalignedCheckpoints_rdh | /**
* Checks whether unaligned checkpoints are forced, despite currently non-checkpointable
* iteration feedback or custom partitioners.
*
* @param forceUnalignedCheckpoints
* The flag to force unaligned checkpoints.
*/
@PublicEvolving
public void setForceUnalignedCheckpoints(boolean forceUnalignedCheckpoints) ... | 3.26 |
flink_CheckpointConfig_isForceCheckpointing_rdh | /**
* Checks whether checkpointing is forced, despite currently non-checkpointable iteration
* feedback.
*
* @return True, if checkpointing is forced, false otherwise.
* @deprecated This will be removed once iterations properly participate in checkpointing.
*/@Deprecated
@PublicEvolving
public boolean isForceChec... | 3.26 |
flink_CheckpointConfig_m1_rdh | /**
* CheckpointStorage defines how {@link StateBackend}'s checkpoint their state for fault
* tolerance in streaming applications. Various implementations store their checkpoints in
* different fashions and have different requirements and availability guarantees.
*
* <p>For example, {@link org.apache.flink.runtime... | 3.26 |
flink_CheckpointConfig_isExternalizedCheckpointsEnabled_rdh | /**
* Returns whether checkpoints should be persisted externally.
*
* @return <code>true</code> if checkpoints should be externalized.
*/
@PublicEvolving
public boolean isExternalizedCheckpointsEnabled() {
return getExternalizedCheckpointCleanup() != ExternalizedCheckpointCleanup.NO_EXTERNALIZED_CHECKPOINTS;
} | 3.26 |
flink_CheckpointConfig_isForceUnalignedCheckpoints_rdh | /**
* Checks whether unaligned checkpoints are forced, despite iteration feedback.
*
* @return True, if unaligned checkpoints are forced, false otherwise.
*/
@PublicEvolving
public boolean isForceUnalignedCheckpoints() {
return configuration.get(ExecutionCheckpointingOptions.FORCE_UNALIGNED);
} | 3.26 |
flink_CheckpointConfig_setCheckpointIntervalDuringBacklog_rdh | /**
* Sets the interval in which checkpoints are periodically scheduled during backlog.
*
* <p>This setting defines the base interval. Checkpoint triggering may be delayed by the
* settings {@link #setMaxConcurrentCheckpoints(int)} and {@link #setMinPauseBetweenCheckpoints(long)}.
*
* <p>If not explicitly configu... | 3.26 |
flink_CheckpointConfig_enableApproximateLocalRecovery_rdh | /**
* Enables the approximate local recovery mode.
*
* <p>In this recovery mode, when a task fails, the entire downstream of the tasks (including
* the failed task) restart.
*
* <p>Notice that 1. Approximate recovery may lead to data loss. The amount of data which leads
* the failed task from the state of the la... | 3.26 |
flink_CheckpointConfig_setMaxSubtasksPerChannelStateFile_rdh | /**
* The number of subtasks to share the same channel state file. If {@link ExecutionCheckpointingOptions#UNALIGNED_MAX_SUBTASKS_PER_CHANNEL_STATE_FILE} has value equal
* to <code>1</code>, each subtask will create a new channel state file.
*/
@PublicEvolving
public void setMaxSubtasksPerChannelStateFile(int maxSub... | 3.26 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.