name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
flink_SingleOutputStreamOperator_returns_rdh
/** * Adds a type information hint about the return type of this operator. This method can be used * in cases where Flink cannot determine automatically what the produced type of a function is. * That can be the case if the function uses generic type variables in the return type that * cannot be inferred from the i...
3.26
flink_SingleOutputStreamOperator_setMaxParallelism_rdh
/** * Sets the maximum parallelism of this operator. * * <p>The maximum parallelism specifies the upper bound for dynamic scaling. It also defines the * number of key groups used for partitioned state. * * @param maxParallelism * Maximum parallelism * @return The operator with set maximum parallelism */ @Pub...
3.26
flink_SingleOutputStreamOperator_uid_rdh
/** * Sets an ID for this operator. * * <p>The specified ID is used to assign the same operator ID across job submissions (for * example when starting a job from a savepoint). * * <p><strong>Important</strong>: this ID needs to be unique per transformation and job. * Otherwise, jo...
3.26
flink_SingleOutputStreamOperator_setResources_rdh
/** * Sets the resources for this operator, the minimum and preferred resources are the same by * default. * * @param resources * The resources for this operator. * @return The operator with set minimum and preferred resources. */ private SingleOutputStreamOperator<T> setResources(ResourceSpec resources) { ...
3.26
flink_TypeMappingUtils_checkPhysicalLogicalTypeCompatible_rdh
/** * Checks whether the given physical field type and logical field type are compatible at the * edges of the table ecosystem. Types are still compatible if the physical type is a legacy * decimal type (converted from Types#BIG_DEC) and the logical type is DECIMAL(38, 18). This is * to support legacy TypeInformati...
3.26
flink_TypeMappingUtils_computePhysicalIndicesOrTimeAttributeMarkers_rdh
/** * Computes indices of physical fields corresponding to the selected logical fields of a {@link TableSchema}. * * <p>It puts markers (idx < 0) for time attributes extracted from {@link DefinedProctimeAttribute} and {@link DefinedRowtimeAttributes} * * <p>{@link TypeMappingUtils#computePhysicalIndices(List, Data...
3.26
flink_TypeMappingUtils_getRowtimeAttributes_rdh
/** * Returns a list with all rowtime attribute names of the [[TableSource]]. */ private static List<String> getRowtimeAttributes(TableSource<?> tableSource) {if (tableSource instanceof DefinedRowtimeAttributes) { return ((DefinedRowtimeAttributes) (tableSource)).getRowtimeAttributeDescriptors().stream().ma...
3.26
flink_TypeMappingUtils_computePhysicalIndices_rdh
/** * Computes indices of physical fields corresponding to the selected logical fields of a {@link TableSchema}. * * @param logicalColumns * Logical columns that describe the physical type. * @param physicalType * Physical type to retrieve indices from. * @param nameRemapping * Additional remapping of a l...
3.26
flink_TypeMappingUtils_getProctimeAttribute_rdh
/** * Returns the proctime attribute of the [[TableSource]] if it is defined. */ private static Optional<String> getProctimeAttribute(TableSource<?> tableSource) { if (tableSource instanceof DefinedProctimeAttribute) { return Optional.ofNullable(((DefinedProctimeAttribute) (tableSource)).getProctimeAtt...
3.26
flink_HistoryServerStaticFileServerHandler_respondWithFile_rdh
/** * Response when running with leading JobManager. */ private void respondWithFile(ChannelHandlerContext ctx, HttpRequest request, String requestPath) throws IOException, ParseException, RestHandlerException { // make sure we request the "index.html" in case there is a directory request ...
3.26
flink_HistoryServerStaticFileServerHandler_channelRead0_rdh
// ------------------------------------------------------------------------ // Responses to requests // ------------------------------------------------------------------------ @Override public void channelRead0(ChannelHandlerContext ctx, RoutedRequest routedRequest) throws Exception { String requestPath = routedR...
3.26
flink_ComponentMetricGroup_close_rdh
/** * Closes the component group by removing and closing all metrics and subgroups (inherited from * {@link AbstractMetricGroup}), plus closing and removing all dedicated component subgroups. */ @Override public void close() { synchronized(this) { if (!isClosed()) { // remove all metrics an...
3.26
flink_Grouping_getInputDataSet_rdh
/** * Returns the input DataSet of a grouping operation, that is the one before the grouping. This * means that if it is applied directly to the result of a grouping operation, it will cancel * its effect. As an example, in the following snippet: * * <pre>{@code DataSet<X> notGrouped = input.groupBy().getDataSet()...
3.26
flink_Grouping_getCustomPartitioner_rdh
/** * Gets the custom partitioner to be used for this grouping, or {@code null}, if none was * defined. * * @return The custom partitioner to be used for this grouping. */ @Internal public Partitioner<?> getCustomPartitioner() { return this.cu...
3.26
flink_ByteValue_setValue_rdh
/** * Sets the encapsulated byte to the specified value. * * @param value * the new value of the encapsulated byte. */ public void setValue(byte value) { this.value = value; }
3.26
flink_ByteValue_getBinaryLength_rdh
// -------------------------------------------------------------------------------------------- @Override public int getBinaryLength() { return 1; }
3.26
flink_ByteValue_read_rdh
// -------------------------------------------------------------------------------------------- @Override public void read(DataInputView in) throws IOException {this.value = in.readByte(); }
3.26
flink_ByteValue_getMaxNormalizedKeyLen_rdh
// -------------------------------------------------------------------------------------------- @Override public int getMaxNormalizedKeyLen() { return 1; }
3.26
flink_ByteValue_toString_rdh
// -------------------------------------------------------------------------------------------- @Override public String toString() { return String.valueOf(this.value); }
3.26
flink_HandlerRequestUtils_getQueryParameter_rdh
/** * Returns the value of a query parameter, or {@code null} if the query parameter is not set. * * @throws RestHandlerException * If the query parameter is repeated. */ public static <X, P extends MessageQueryParameter<X>, R extends RequestBody, M extends MessageParameters> X getQueryParameter(final HandlerReq...
3.26
flink_HandlerRequestUtils_fromRequestBodyOrQueryParameter_rdh
/** * Returns {@code requestValue} if it is not null, otherwise returns the query parameter value * if it is not null, otherwise returns the default value. */ public static <T> T fromRequestBodyOrQueryParameter(T requestValue, SupplierWithException<T, RestHandlerException> queryParameterExtractor, T defaultValue, L...
3.26
flink_HeartbeatManagerImpl_getOwnResourceID_rdh
// ---------------------------------------------------------------------------------------------- // Getters // ---------------------------------------------------------------------------------------------- ResourceID getOwnResourceID() {return ownResourceID; }
3.26
flink_HeartbeatManagerImpl_receiveHeartbeat_rdh
// ---------------------------------------------------------------------------------------------- // HeartbeatTarget methods // ---------------------------------------------------------------------------------------------- @Override public ...
3.26
flink_HeartbeatManagerImpl_monitorTarget_rdh
// ---------------------------------------------------------------------------------------------- // HeartbeatManager methods // ---------------------------------------------------------------------------------------------- @Override public void monitorTarget(ResourceID resourceID, HeartbeatTarget<O> heartbeatTarget) {...
3.26
flink_SessionManagerImpl_checkSessionCount_rdh
// ------------------------------------------------------------------------------------------ // Utilities // ------------------------------------------------------------------------------------------ private void checkSessionCount() throws SqlGatewayException { if (maxSessionCount <= 0) { return; } if (sessions.size(...
3.26
flink_GenericDataSourceBase_getUserCodeWrapper_rdh
/** * Gets the class describing the input format. * * <p>This method is basically identical to {@link #getFormatWrapper()}. * * @return The class describing the input format. * @see org.apache.flink.api.common.operators.Operator#getUserCodeWrapper() */ @Override public UserCodeWrapper<? extends T> getUserCodeWra...
3.26
flink_GenericDataSourceBase_getStatisticsKey_rdh
// -------------------------------------------------------------------------------------------- /** * Gets the key under which statistics about this data source may be obtained from the * statistics cache. * * @return The statistics cache key. */ public String getStatisticsKey() { return this.statisticsKey; }
3.26
flink_GenericDataSourceBase_toString_rdh
// -------------------------------------------------------------------------------------------- public String toString() { return this.name; }
3.26
flink_GenericDataSourceBase_accept_rdh
// -------------------------------------------------------------------------------------------- /** * Accepts the visitor and applies it this instance. Since the data sources have no inputs, no * recursive descend happens. The visitors pre-visit method is called and, if returning * <tt>true</tt>,...
3.26
flink_GenericDataSourceBase_setStatisticsKey_rdh
/** * Sets the key under which statistics about this data source may be obtained from the * statistics cache. Useful for testing purposes, when providing mock statistics. * * @param statisticsKey * The key for the statistics object. */ public void setStatisticsKey(String statisticsKey) { this.statisticsKey ...
3.26
flink_GenericDataSourceBase_getFormatWrapper_rdh
// -------------------------------------------------------------------------------------------- /** * Gets the class describing the input format. * * @return The class describing the input format. */ public UserCodeWrapper<? extends T> getFormatWrapper() { return this.formatWrapper; }
3.26
flink_GenericDataSourceBase_setSplitDataProperties_rdh
/** * Sets properties of input splits for this data source. Split properties can help to generate * more efficient execution plans. <br> * <b> IMPORTANT: Providing wrong split data properties can cause wrong results! </b> * * @param splitDataProperties * The data properties of this data source's splits. */ pub...
3.26
flink_GenericDataSourceBase_executeOnCollections_rdh
// -------------------------------------------------------------------------------------------- protected List<OUT> executeOnCollections(RuntimeContext ctx, ExecutionConfig executionConfig) throws Exception { @SuppressWarnings("unchecked") InputFormat<OUT, InputSplit> inputFormat = ((InputFormat<OUT, InputS...
3.26
flink_ChannelStateWriteRequestExecutorFactory_getOrCreateExecutor_rdh
/** * * @param startExecutor * It is for test to prevent create too many threads when some unit tests * create executor frequently. */ ChannelStateWriteRequestExecutor getOrCreateExecutor(JobVertexID jobVertexID, int subtaskIndex, CheckpointStorage checkpointStorage, int maxSubtasksPerChannelStateFile, boolean...
3.26
flink_Task_getExecutionState_rdh
// ------------------------------------------------------------------------ // Task Execution // ------------------------------------------------------------------------ /** * Returns the current execution state of the task. * * @return The current execution state of the task. */ public ExecutionState getExecutionS...
3.26
flink_Task_triggerCheckpointBarrier_rdh
// ------------------------------------------------------------------------ // Notifications on the invokable // ------------------------------------------------------------------------ /** * Calls the invokable to trigger a checkpoint. * * @param checkpointID * The ID identifying the checkpoint. * @param checkp...
3.26
flink_Task_getJobID_rdh
// ------------------------------------------------------------------------ // Accessors // ------------------------------------------------------------------------ @Override public JobID getJobID() { return jobId; }
3.26
flink_Task_run_rdh
/** * The core work method that bootstraps the task and executes its code. */ @Override public void run() { try { doRun(); } finally { terminationFuture.complete(executionState); } }
3.26
flink_Task_cancelExecution_rdh
// ---------------------------------------------------------------------------------------------------------------- // Canceling / Failing the task from the outside // ---------------------------------------------------------------------------------------------------------------- /** * Cancels the task execution. If t...
3.26
flink_Task_loadAndInstantiateInvokable_rdh
/** * Instantiates the given task invokable class, passing the given environment (and possibly the * initial task state) to the task's constructor. * * <p>The method will first try to instantiate the task via a constructor accepting both the * Environment and the TaskStateSnapshot. If no such constructor exists, a...
3.26
flink_Task_runWithSystemExitMonitoring_rdh
/** * Monitor user codes from exiting JVM covering user function invocation. This can be done in a * finer-grained way like enclosing user callback functions individually, but as exit triggered * by framework is not performed and expected in this invoke function anyhow, we can monitor * exiting JVM for entire scope...
3.26
flink_Task_m1_rdh
/** * Unwrap, enrich and handle fatal errors. */private Throwable m1(Throwable t) { // unwrap wrapped exceptions to make stack traces more compact if (t instanceof WrappingRuntimeException) { t = ((WrappingRuntimeException) (t)).unwrap(); } TaskManagerExceptionUtils.tryEnrichTaskManagerError...
3.26
flink_Task_transitionState_rdh
/** * Try to transition the execution state from the current state to the new state. * * @param currentState * of the execution * @param newState * of the execution * @return true if the transition was successful, otherwise false */ private boolean transitionState(ExecutionState currentState, ExecutionState...
3.26
flink_Task_isCanceledOrFailed_rdh
/** * Checks whether the task has failed, is canceled, or is being canceled at the moment. * * @return True is the task in state FAILED, CANCELING, or CANCELED, false otherwise. */ public boolean isCanceledOrFailed() { return ((executionState == ExecutionState.CANCELING) || (executionState == ExecutionState.C...
3.26
flink_Task_deliverOperatorEvent_rdh
/** * Dispatches an operator event to the invokable task. * * <p>If the event delivery did not succeed, this method throws an exception. Callers can use * that exception for error reporting, but need not react with failing this task (this method * takes care of that). * * @throws FlinkException * This method ...
3.26
flink_Task_requestPartitionProducerState_rdh
// ------------------------------------------------------------------------ // Partition State Listeners // ------------------------------------------------------------------------ @Override public void requestPartitionProducerState(final IntermediateDataSetID intermediateDataSetId, final ResultPartitionID resultPartit...
3.26
flink_Task_failExternally_rdh
/** * Marks task execution failed for an external reason (a reason other than the task code itself * throwing an exception). If the task is already in a terminal state (such as FINISHED, * CANCELED, FAILED), or if the task is already canceling this does nothing. Otherwise it sets * the state to FAILED, and, if the ...
3.26
flink_Task_releaseResources_rdh
/** * Releases resources before task exits. We should also fail the partition to release if the * task has failed, is canceled, or is being canceled at the moment. */ private void releaseResources() { LOG.debug("Release task {} network resources (state: {}).", taskNameWithSubtask, getExecutionState()); for (...
3.26
flink_Task_m0_rdh
/** * If the task has failed, this method gets the exception that caused this task to fail. * Otherwise this method returns null. * * @return The exception that caused the task to fail, or null, if the task has not failed. */public Throwable m0() { return failureCause; }
3.26
flink_EnvironmentSettings_getUserClassLoader_rdh
/** * Returns the user {@link ClassLoader} to use for code generation, UDF loading and other * operations requiring reflections on user code. */ @Internal public ClassLoader getUserClassLoader() { return classLoader; }
3.26
flink_EnvironmentSettings_inBatchMode_rdh
/** * Sets that the components should work in a batch mode. Streaming mode by default. */ public Builder inBatchMode() { configuration.set(RUNTIME_MODE, BATCH); return this; }
3.26
flink_EnvironmentSettings_m0_rdh
/** * Tells if the {@link TableEnvironment} should work in a batch or streaming mode. */public boolean m0() { return configuration.get(RUNTIME_MODE) == STREAMING; }
3.26
flink_EnvironmentSettings_inStreamingMode_rdh
/** * Sets that the components should work in a streaming mode. Enabled by default. */ public Builder inStreamingMode() { configuration.set(RUNTIME_MODE, STREAMING); return this; }
3.26
flink_EnvironmentSettings_m1_rdh
/** * Specifies the classloader to use in the planner for operations related to code * generation, UDF loading, operations requiring reflections on user classes, discovery of * factories. * * <p>By default, this is configured using {@code Thread.currentThread().getContextClassLoader()}. * * <p>Modify the {@link ...
3.26
flink_EnvironmentSettings_newInstance_rdh
/** * Creates a builder for creating an instance of {@link EnvironmentSettings}. */ public static Builder newInstance() { return new Builder(); } /** * Creates an instance of {@link EnvironmentSettings} from configuration. * * @deprecated use {@link Builder#withConfiguration(Configuration)}
3.26
flink_EnvironmentSettings_getConfiguration_rdh
/** * Get the underlying {@link Configuration}. */ public Configuration getConfiguration() { return configuration;}
3.26
flink_EnvironmentSettings_withConfiguration_rdh
/** * Add extra configuration to {@link EnvironmentSettings}. */ public Builder withConfiguration(Configuration configuration) { this.configuration.addAll(configuration); return this; }
3.26
flink_EnvironmentSettings_withBuiltInCatalogName_rdh
/** * Specifies the name of the initial catalog to be created when instantiating a {@link TableEnvironment}. * * <p>This catalog is an in-memory catalog that will be used to store all temporary objects * (e.g. from {@link TableEnvironment#createTemporaryView(String, Table)} or {@link TableEnvironment#createTemporar...
3.26
flink_EnvironmentSettings_getBuiltInCatalogName_rdh
/** * Gets the specified name of the initial catalog to be created when instantiating a {@link TableEnvironment}. */ public String getBuiltInCatalogName() { return configuration.get(TABLE_CATALOG_NAME); }
3.26
flink_EnvironmentSettings_build_rdh
/** * Returns an immutable instance of {@link EnvironmentSettings}. */ public EnvironmentSettings build() { if (classLoader == null) { classLoader = Thread.currentThread().getContextClassLoader(); } return new EnvironmentSettings(configuration, classLoader, catalogStore); }
3.26
flink_EnvironmentSettings_withBuiltInDatabaseName_rdh
/** * Specifies the name of the default database in the initial catalog to be created when * instantiating a {@link TableEnvironment}. * * <p>This database is an in-memory database that will be used to store all temporary * objects (e.g. from {@link TableEnvironment#createTemporaryView(String, Table)} or {@link Ta...
3.26
flink_EnvironmentSettings_getBuiltInDatabaseName_rdh
/** * Gets the specified name of the default database in the initial catalog to be created when * instantiating a {@link TableEnvironment}. */ public String getBuiltInDatabaseName() {return configuration.get(TABLE_DATABASE_NAME); }
3.26
flink_FileLock_unlock_rdh
/** * Release the file lock. * * @throws IOException * If the FileChannel is closed */ public void unlock() throws IOException { if ((lock != null) && lock.channel().isOpen()) { lock.release(); } }
3.26
flink_FileLock_inTempFolder_rdh
/** * Initialize a FileLock using a file located inside temp folder. * * @param fileName * The name of the locking file * @return The initialized FileLock */ public static FileLock inTempFolder(String fileName) {return new FileLock(TEMP_DIR, fileName); }
3.26
flink_FileLock_tryLock_rdh
/** * Try to acquire a lock on the locking file. This method immediately returns whenever the lock * is acquired or not. * * @return True if successfully acquired the lock * @throws IOException * If the file path is invalid */ public boolean tryLock() throws IOException { if (outputStream == null) { ...
3.26
flink_FileLock_unlockAndDestroy_rdh
/** * Release the file lock, close the fileChannel and FileOutputStream then try deleting the * locking file if other file lock does not need it, which means the lock will not be used * anymore. * * @throws IOException * If an I/O error occurs */ public void unlockAndDestroy() throws IOException { try { ...
3.26
flink_FileLock_init_rdh
/** * Check whether the locking file exists in the file system. Create it if it does not exist. * Then create a FileOutputStream for it. * * @throws IOException * If the file path is invalid or the parent dir does not exist */ private void init() throws IOException { if (!this.file.exists()) { thi...
3.26
flink_FileLock_normalizeFileName_rdh
/** * Normalize the file name, which only allows slash, backslash, digits and letters. * * @param fileName * Original file name * @return File name with illegal characters stripped */ private static String normalizeFileName(String fileName) {return fileName.replaceAll("[^\\w/\\\\]", ""); }
3.26
flink_SerdeContext_get_rdh
/** * Retrieve context from {@link SerializerProvider} and {@link DeserializationContext}. */public static SerdeContext get(DatabindContext databindContext) { final SerdeContext serdeContext = ((SerdeContext) (databindContext.getAttribute(SERDE_CONTEXT_KEY))); assert serdeContext != null; return serdeCont...
3.26
flink_RawObjectConverter_create_rdh
// -------------------------------------------------------------------------------------------- // Factory method // -------------------------------------------------------------------------------------------- public static RawObjectConverter<?> create(DataType dataType) { final LogicalType logicalType = dataType.g...
3.26
flink_DeltaEvictor_of_rdh
/** * Creates a {@code DeltaEvictor} from the given threshold and {@code DeltaFunction}. Eviction * is done before the window function. * * @param threshold * The threshold * @param deltaFunction * The {@code DeltaFunction} */ p...
3.26
flink_PrimitiveArrayTypeInfo_getComponentClass_rdh
/** * Gets the class that represents the component type. * * @return The class of the component type. */ @PublicEvolving public Class<?> getComponentClass() { return this.arrayClass.getComponentType(); }
3.26
flink_PrimitiveArrayTypeInfo_getComponentType_rdh
/** * Gets the type information of the component type. * * @return The type information of the component type. */ @PublicEvolving public TypeInformation<?> getComponentType() { return BasicTypeInfo.getInfoFor(getComponentClass()); }
3.26
flink_PrimitiveArrayTypeInfo_isBasicType_rdh
// -------------------------------------------------------------------------------------------- @Override @PublicEvolving public boolean isBasicType() { return false; }
3.26
flink_ExecNodeBase_inputsContainSingleton_rdh
/** * Whether singleton distribution is required. */ protected boolean inputsContainSingleton() { return getInputProperties().stream().anyMatch(p -> p.getRequiredDistribution().getType() == InputProperty.DistributionType.SINGLETON); }
3.26
flink_ExecNodeBase_getPersistedConfig_rdh
// Custom filter to exclude node configuration if no consumed options are used @JsonProperty(value = FIELD_NAME_CONFIGURATION, access = Access.READ_ONLY, index = 2) @JsonInclude(value = Include.CUSTOM, valueFilter = ConfigurationJsonSerializerFilter.class) public ReadableConfig getPersistedConfig() { return persistedC...
3.26
flink_ExecNodeBase_translateToFusionCodegenSpecInternal_rdh
/** * Internal method, translates this node into a operator codegen spec generator. * * @param planner * The planner. * @param config * per-{@link ExecNode} configuration that contains the merged configuration from * various layers which all the nodes implementing this method should use, instead of * re...
3.26
flink_ReaderInfo_getLocation_rdh
/** * * @return the location of the subtask that runs this source reader. */ public String getLocation() { return location; }
3.26
flink_ReaderInfo_getSubtaskId_rdh
/** * * @return the ID of the subtask that runs the source reader. */ public int getSubtaskId() { return subtaskId; }
3.26
flink_RowTimeIntervalJoin_getMaxOutputDelay_rdh
/** * Get the maximum interval between receiving a row and emitting it (as part of a joined * result). This is the time interval by which watermarks need to be held back. * * @return the maximum delay for the outputs */ public long getMaxOutputDelay() { return Math.max(leftRelativeSize, rightRelativeSize) + a...
3.26
flink_CliInputView_init_rdh
// -------------------------------------------------------------------------------------------- @Override protected void init() { // nothing to do }
3.26
flink_CliInputView_insert_rdh
// -------------------------------------------------------------------------------------------- private void insert(String binding) { currentInput.insert(cursorPos, binding); cursorPos += binding.length(); // reset view resetMainPart(); }
3.26
flink_CalculatedTableFactory_create_rdh
/** * Creates a valid {@link CalculatedQueryOperation} operation. * * @param callExpr * call to table function as expression * @return valid calculated table */ QueryOperation create(ResolvedExpression callExpr, List<String> leftTableFieldNames) { FunctionTableCallVisitor calculatedTableCreator = new Functi...
3.26
flink_StatePartitionStreamProvider_getStream_rdh
/** * Returns a stream with the data of one state partition. */ public InputStream getStream() throws IOException { if (creationException != null) { throw new IOException(creationException); } return stream; }
3.26
flink_SupportsProjectionPushDown_applyProjection_rdh
/** * Provides the field index paths that should be used for a projection. The indices are 0-based * and support fields within (possibly nested) structures if this is enabled via {@link #supportsNestedProjection()}. * * <p>In the example mentioned in {@link SupportsProjectionPushDown}, this method would receive: *...
3.26
flink_ArchivedExecutionVertex_getTaskNameWithSubtaskIndex_rdh
// -------------------------------------------------------------------------------------------- // Accessors // -------------------------------------------------------------------------------------------- @Override public String getTaskNameWithSubtaskIndex() { return this.taskNameWithSubtask; }
3.26
flink_FailureEnricherUtils_labelFailure_rdh
/** * Enriches a Throwable by returning the merged label output of a Set of FailureEnrichers. * * @param cause * the Throwable to label * @param context * the context of the Throwable * @param mainThreadExecutor * the executor to comp...
3.26
flink_FailureEnricherUtils_m0_rdh
/** * Filters out invalid {@link FailureEnricher} objects that have duplicate output keys. * * @param failureEnrichers * a set of {@link FailureEnricher} objects to filter * @return a filtered collection without any duplicate output keys */ @VisibleForTesting static Collection<FailureEnricher> m0(final Set<Fail...
3.26
flink_FailureEnricherUtils_getIncludedFailureEnrichers_rdh
/** * Returns a set of failure enricher names included in the given configuration. * * @param configuration * the configuration to get the failure enricher names from * @return failure enricher names */ @VisibleForTesting static Set<String> getIncludedFailureEnrichers(final Configuration configuration) { fi...
3.26
flink_FailureEnricherUtils_getFailureEnrichers_rdh
/** * Returns a set of validated FailureEnrichers for a given configuration. * * @param configuration * the configuration for the job * @return a collection of validated FailureEnrichers */ public static Collection<FailureEnricher> getFailureEnrichers(final Configuration configuration) { final PluginMana...
3.26
flink_StateUtil_bestEffortDiscardAllStateObjects_rdh
/** * Iterates through the passed state handles and calls discardState() on each handle that is not * null. All occurring exceptions are suppressed and collected until the iteration is over and * emitted as a single exception. * * @param handlesToDiscard * State handles to discard. Passed iterable is allowed to...
3.26
flink_StateUtil_discardStateFuture_rdh
/** * Discards the given state future by first trying to cancel it. If this is not possible, then * the state object contained in the future is calculated and afterwards discarded. * * @param stateFuture * to be discarded * @throws Exception * if the discard operation failed ...
3.26
flink_StateUtil_getStateSize_rdh
/** * Returns the size of a state object. * * @param handle * The handle to the retrieved state */ public static long getStateSize(StateObject handle) { return handle == null ? 0 : handle.getStateSize(); }
3.26
flink_StateUtil_unexpectedStateHandleException_rdh
/** * Creates a {@link RuntimeException} that signals that an operation did not get the type of * {@link StateObject} that was expected. This can mostly happen when a different {@link StateBackend} from the one that was used for taking a checkpoint/savepoint is used when * restoring. */ public static RuntimeExcepti...
3.26
flink_JvmUtils_createThreadInfoSample_rdh
/** * Creates a {@link ThreadInfoSample} for a specific thread. Contains thread traces if * maxStackTraceDepth > 0. * * @param threadId * The ID of the thread to create the thread dump for. * @param maxStackTraceDepth * The maximum number of entries in the stack trace to be collected. * @return The thread i...
3.26
flink_JvmUtils_createThreadDump_rdh
/** * Creates a thread dump of the current JVM. * * @return the thread dump of current JVM */ public static Collection<ThreadInfo> createThreadDump() { ThreadMXBean threadMxBean = ManagementFactory.getThreadMXBean(); return Arrays.asList(threadMxBean.dumpAllThreads(true, true)); }
3.26
flink_ReduceTransformation_getInputType_rdh
/** * Returns the {@code TypeInformation} for the elements of the input. */ public TypeInformation<IN> getInputType() { return input.getOutputType(); }
3.26
flink_CatalogFactory_supportedProperties_rdh
/** * * @deprecated Implement the {@link Factory} based stack instead. */ @Deprecated default List<String> supportedProperties() { // Default implementation for catalogs implementing the new {@link Factory} stack instead. return null; }
3.26
flink_CatalogFactory_requiredContext_rdh
// -------------------------------------------------------------------------------------------- // Default implementations for legacy {@link TableFactory} stack. // -------------------------------------------------------------------------------------------- /** * * @deprecated Implement the {@link Factory} based stac...
3.26
flink_CatalogFactory_createCatalog_rdh
/** * Creates and configures a {@link Catalog} using the given context. * * <p>An implementation should perform validation and the discovery of further (nested) * factories in this method. */ default Catalog createCatalog(Context context) { throw new CatalogException("Catalog fact...
3.26