name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_SingleOutputStreamOperator_returns_rdh | /**
* Adds a type information hint about the return type of this operator. This method can be used
* in cases where Flink cannot determine automatically what the produced type of a function is.
* That can be the case if the function uses generic type variables in the return type that
* cannot be inferred from the i... | 3.26 |
flink_SingleOutputStreamOperator_setMaxParallelism_rdh | /**
* Sets the maximum parallelism of this operator.
*
* <p>The maximum parallelism specifies the upper bound for dynamic scaling. It also defines the
* number of key groups used for partitioned state.
*
* @param maxParallelism
* Maximum parallelism
* @return The operator with set maximum parallelism
*/
@Pub... | 3.26 |
flink_SingleOutputStreamOperator_uid_rdh | /**
* Sets an ID for this operator.
*
* <p>The specified ID is used to assign the same operator ID across job submissions (for
* example when starting a job from a savepoint).
*
* <p><strong>Important</strong>: this ID needs to be unique per transformation and job.
* Otherwise, jo... | 3.26 |
flink_SingleOutputStreamOperator_setResources_rdh | /**
* Sets the resources for this operator, the minimum and preferred resources are the same by
* default.
*
* @param resources
* The resources for this operator.
* @return The operator with set minimum and preferred resources.
*/
private SingleOutputStreamOperator<T> setResources(ResourceSpec resources) {
... | 3.26 |
flink_TypeMappingUtils_checkPhysicalLogicalTypeCompatible_rdh | /**
* Checks whether the given physical field type and logical field type are compatible at the
* edges of the table ecosystem. Types are still compatible if the physical type is a legacy
* decimal type (converted from Types#BIG_DEC) and the logical type is DECIMAL(38, 18). This is
* to support legacy TypeInformati... | 3.26 |
flink_TypeMappingUtils_computePhysicalIndicesOrTimeAttributeMarkers_rdh | /**
* Computes indices of physical fields corresponding to the selected logical fields of a {@link TableSchema}.
*
* <p>It puts markers (idx < 0) for time attributes extracted from {@link DefinedProctimeAttribute} and {@link DefinedRowtimeAttributes}
*
* <p>{@link TypeMappingUtils#computePhysicalIndices(List, Data... | 3.26 |
flink_TypeMappingUtils_getRowtimeAttributes_rdh | /**
* Returns a list with all rowtime attribute names of the [[TableSource]].
*/
private static List<String> getRowtimeAttributes(TableSource<?> tableSource) {if (tableSource instanceof DefinedRowtimeAttributes) {
return ((DefinedRowtimeAttributes) (tableSource)).getRowtimeAttributeDescriptors().stream().ma... | 3.26 |
flink_TypeMappingUtils_computePhysicalIndices_rdh | /**
* Computes indices of physical fields corresponding to the selected logical fields of a {@link TableSchema}.
*
* @param logicalColumns
* Logical columns that describe the physical type.
* @param physicalType
* Physical type to retrieve indices from.
* @param nameRemapping
* Additional remapping of a l... | 3.26 |
flink_TypeMappingUtils_getProctimeAttribute_rdh | /**
* Returns the proctime attribute of the [[TableSource]] if it is defined.
*/
private static Optional<String> getProctimeAttribute(TableSource<?> tableSource) {
if (tableSource instanceof DefinedProctimeAttribute) {
return Optional.ofNullable(((DefinedProctimeAttribute) (tableSource)).getProctimeAtt... | 3.26 |
flink_HistoryServerStaticFileServerHandler_respondWithFile_rdh | /**
* Response when running with leading JobManager.
*/
private void respondWithFile(ChannelHandlerContext ctx, HttpRequest request, String requestPath) throws IOException, ParseException, RestHandlerException {
// make sure we request the "index.html" in case there is a directory request
... | 3.26 |
flink_HistoryServerStaticFileServerHandler_channelRead0_rdh | // ------------------------------------------------------------------------
// Responses to requests
// ------------------------------------------------------------------------
@Override
public void channelRead0(ChannelHandlerContext ctx, RoutedRequest routedRequest) throws Exception {
String requestPath = routedR... | 3.26 |
flink_ComponentMetricGroup_close_rdh | /**
* Closes the component group by removing and closing all metrics and subgroups (inherited from
* {@link AbstractMetricGroup}), plus closing and removing all dedicated component subgroups.
*/
@Override
public void close() {
synchronized(this) {
if (!isClosed()) {
// remove all metrics an... | 3.26 |
flink_Grouping_getInputDataSet_rdh | /**
* Returns the input DataSet of a grouping operation, that is the one before the grouping. This
* means that if it is applied directly to the result of a grouping operation, it will cancel
* its effect. As an example, in the following snippet:
*
* <pre>{@code DataSet<X> notGrouped = input.groupBy().getDataSet()... | 3.26 |
flink_Grouping_getCustomPartitioner_rdh | /**
* Gets the custom partitioner to be used for this grouping, or {@code null}, if none was
* defined.
*
* @return The custom partitioner to be used for this grouping.
*/
@Internal
public Partitioner<?> getCustomPartitioner() {
return this.cu... | 3.26 |
flink_ByteValue_setValue_rdh | /**
* Sets the encapsulated byte to the specified value.
*
* @param value
* the new value of the encapsulated byte.
*/
public void setValue(byte value) {
this.value = value;
} | 3.26 |
flink_ByteValue_getBinaryLength_rdh | // --------------------------------------------------------------------------------------------
@Override
public int getBinaryLength() {
return 1;
} | 3.26 |
flink_ByteValue_read_rdh | // --------------------------------------------------------------------------------------------
@Override
public void read(DataInputView in) throws IOException {this.value = in.readByte();
} | 3.26 |
flink_ByteValue_getMaxNormalizedKeyLen_rdh | // --------------------------------------------------------------------------------------------
@Override
public int getMaxNormalizedKeyLen() {
return 1;
} | 3.26 |
flink_ByteValue_toString_rdh | // --------------------------------------------------------------------------------------------
@Override
public String toString() {
return String.valueOf(this.value);
} | 3.26 |
flink_HandlerRequestUtils_getQueryParameter_rdh | /**
* Returns the value of a query parameter, or {@code null} if the query parameter is not set.
*
* @throws RestHandlerException
* If the query parameter is repeated.
*/
public static <X, P extends MessageQueryParameter<X>, R extends RequestBody, M extends MessageParameters> X getQueryParameter(final HandlerReq... | 3.26 |
flink_HandlerRequestUtils_fromRequestBodyOrQueryParameter_rdh | /**
* Returns {@code requestValue} if it is not null, otherwise returns the query parameter value
* if it is not null, otherwise returns the default value.
*/
public static <T> T fromRequestBodyOrQueryParameter(T requestValue, SupplierWithException<T, RestHandlerException> queryParameterExtractor, T defaultValue, L... | 3.26 |
flink_HeartbeatManagerImpl_getOwnResourceID_rdh | // ----------------------------------------------------------------------------------------------
// Getters
// ----------------------------------------------------------------------------------------------
ResourceID getOwnResourceID() {return ownResourceID;
} | 3.26 |
flink_HeartbeatManagerImpl_receiveHeartbeat_rdh | // ----------------------------------------------------------------------------------------------
// HeartbeatTarget methods
// ----------------------------------------------------------------------------------------------
@Override
public ... | 3.26 |
flink_HeartbeatManagerImpl_monitorTarget_rdh | // ----------------------------------------------------------------------------------------------
// HeartbeatManager methods
// ----------------------------------------------------------------------------------------------
@Override
public void monitorTarget(ResourceID resourceID, HeartbeatTarget<O> heartbeatTarget) {... | 3.26 |
flink_SessionManagerImpl_checkSessionCount_rdh | // ------------------------------------------------------------------------------------------
// Utilities
// ------------------------------------------------------------------------------------------
private void checkSessionCount() throws SqlGatewayException {
if (maxSessionCount <= 0) {
return;
}
if (sessions.size(... | 3.26 |
flink_GenericDataSourceBase_getUserCodeWrapper_rdh | /**
* Gets the class describing the input format.
*
* <p>This method is basically identical to {@link #getFormatWrapper()}.
*
* @return The class describing the input format.
* @see org.apache.flink.api.common.operators.Operator#getUserCodeWrapper()
*/
@Override
public UserCodeWrapper<? extends T> getUserCodeWra... | 3.26 |
flink_GenericDataSourceBase_getStatisticsKey_rdh | // --------------------------------------------------------------------------------------------
/**
* Gets the key under which statistics about this data source may be obtained from the
* statistics cache.
*
* @return The statistics cache key.
*/
public String getStatisticsKey() {
return this.statisticsKey;
} | 3.26 |
flink_GenericDataSourceBase_toString_rdh | // --------------------------------------------------------------------------------------------
public String toString() {
return this.name;
} | 3.26 |
flink_GenericDataSourceBase_accept_rdh | // --------------------------------------------------------------------------------------------
/**
* Accepts the visitor and applies it this instance. Since the data sources have no inputs, no
* recursive descend happens. The visitors pre-visit method is called and, if returning
* <tt>true</tt>,... | 3.26 |
flink_GenericDataSourceBase_setStatisticsKey_rdh | /**
* Sets the key under which statistics about this data source may be obtained from the
* statistics cache. Useful for testing purposes, when providing mock statistics.
*
* @param statisticsKey
* The key for the statistics object.
*/
public void setStatisticsKey(String statisticsKey) {
this.statisticsKey ... | 3.26 |
flink_GenericDataSourceBase_getFormatWrapper_rdh | // --------------------------------------------------------------------------------------------
/**
* Gets the class describing the input format.
*
* @return The class describing the input format.
*/
public UserCodeWrapper<? extends T> getFormatWrapper() {
return this.formatWrapper;
} | 3.26 |
flink_GenericDataSourceBase_setSplitDataProperties_rdh | /**
* Sets properties of input splits for this data source. Split properties can help to generate
* more efficient execution plans. <br>
* <b> IMPORTANT: Providing wrong split data properties can cause wrong results! </b>
*
* @param splitDataProperties
* The data properties of this data source's splits.
*/
pub... | 3.26 |
flink_GenericDataSourceBase_executeOnCollections_rdh | // --------------------------------------------------------------------------------------------
protected List<OUT> executeOnCollections(RuntimeContext ctx, ExecutionConfig executionConfig) throws Exception {
@SuppressWarnings("unchecked")
InputFormat<OUT, InputSplit> inputFormat = ((InputFormat<OUT, InputS... | 3.26 |
flink_ChannelStateWriteRequestExecutorFactory_getOrCreateExecutor_rdh | /**
*
* @param startExecutor
* It is for test to prevent create too many threads when some unit tests
* create executor frequently.
*/
ChannelStateWriteRequestExecutor getOrCreateExecutor(JobVertexID jobVertexID, int subtaskIndex, CheckpointStorage checkpointStorage, int maxSubtasksPerChannelStateFile, boolean... | 3.26 |
flink_Task_getExecutionState_rdh | // ------------------------------------------------------------------------
// Task Execution
// ------------------------------------------------------------------------
/**
* Returns the current execution state of the task.
*
* @return The current execution state of the task.
*/
public ExecutionState getExecutionS... | 3.26 |
flink_Task_triggerCheckpointBarrier_rdh | // ------------------------------------------------------------------------
// Notifications on the invokable
// ------------------------------------------------------------------------
/**
* Calls the invokable to trigger a checkpoint.
*
* @param checkpointID
* The ID identifying the checkpoint.
* @param checkp... | 3.26 |
flink_Task_getJobID_rdh | // ------------------------------------------------------------------------
// Accessors
// ------------------------------------------------------------------------
@Override
public JobID getJobID() {
return jobId;
} | 3.26 |
flink_Task_run_rdh | /**
* The core work method that bootstraps the task and executes its code.
*/
@Override
public void run() {
try {
doRun();
} finally {
terminationFuture.complete(executionState);
}
} | 3.26 |
flink_Task_cancelExecution_rdh | // ----------------------------------------------------------------------------------------------------------------
// Canceling / Failing the task from the outside
// ----------------------------------------------------------------------------------------------------------------
/**
* Cancels the task execution. If t... | 3.26 |
flink_Task_loadAndInstantiateInvokable_rdh | /**
* Instantiates the given task invokable class, passing the given environment (and possibly the
* initial task state) to the task's constructor.
*
* <p>The method will first try to instantiate the task via a constructor accepting both the
* Environment and the TaskStateSnapshot. If no such constructor exists, a... | 3.26 |
flink_Task_runWithSystemExitMonitoring_rdh | /**
* Monitor user codes from exiting JVM covering user function invocation. This can be done in a
* finer-grained way like enclosing user callback functions individually, but as exit triggered
* by framework is not performed and expected in this invoke function anyhow, we can monitor
* exiting JVM for entire scope... | 3.26 |
flink_Task_m1_rdh | /**
* Unwrap, enrich and handle fatal errors.
*/private Throwable m1(Throwable t) {
// unwrap wrapped exceptions to make stack traces more compact
if (t instanceof WrappingRuntimeException) {
t = ((WrappingRuntimeException) (t)).unwrap();
}
TaskManagerExceptionUtils.tryEnrichTaskManagerError... | 3.26 |
flink_Task_transitionState_rdh | /**
* Try to transition the execution state from the current state to the new state.
*
* @param currentState
* of the execution
* @param newState
* of the execution
* @return true if the transition was successful, otherwise false
*/
private boolean transitionState(ExecutionState currentState, ExecutionState... | 3.26 |
flink_Task_isCanceledOrFailed_rdh | /**
* Checks whether the task has failed, is canceled, or is being canceled at the moment.
*
* @return True is the task in state FAILED, CANCELING, or CANCELED, false otherwise.
*/
public boolean isCanceledOrFailed() {
return ((executionState == ExecutionState.CANCELING) || (executionState == ExecutionState.C... | 3.26 |
flink_Task_deliverOperatorEvent_rdh | /**
* Dispatches an operator event to the invokable task.
*
* <p>If the event delivery did not succeed, this method throws an exception. Callers can use
* that exception for error reporting, but need not react with failing this task (this method
* takes care of that).
*
* @throws FlinkException
* This method ... | 3.26 |
flink_Task_requestPartitionProducerState_rdh | // ------------------------------------------------------------------------
// Partition State Listeners
// ------------------------------------------------------------------------
@Override
public void requestPartitionProducerState(final IntermediateDataSetID intermediateDataSetId, final ResultPartitionID resultPartit... | 3.26 |
flink_Task_failExternally_rdh | /**
* Marks task execution failed for an external reason (a reason other than the task code itself
* throwing an exception). If the task is already in a terminal state (such as FINISHED,
* CANCELED, FAILED), or if the task is already canceling this does nothing. Otherwise it sets
* the state to FAILED, and, if the ... | 3.26 |
flink_Task_releaseResources_rdh | /**
* Releases resources before task exits. We should also fail the partition to release if the
* task has failed, is canceled, or is being canceled at the moment.
*/
private void releaseResources() {
LOG.debug("Release task {} network resources (state: {}).", taskNameWithSubtask, getExecutionState());
for (... | 3.26 |
flink_Task_m0_rdh | /**
* If the task has failed, this method gets the exception that caused this task to fail.
* Otherwise this method returns null.
*
* @return The exception that caused the task to fail, or null, if the task has not failed.
*/public Throwable m0() {
return failureCause;
} | 3.26 |
flink_EnvironmentSettings_getUserClassLoader_rdh | /**
* Returns the user {@link ClassLoader} to use for code generation, UDF loading and other
* operations requiring reflections on user code.
*/
@Internal
public ClassLoader getUserClassLoader()
{
return classLoader;
} | 3.26 |
flink_EnvironmentSettings_inBatchMode_rdh | /**
* Sets that the components should work in a batch mode. Streaming mode by default.
*/
public Builder inBatchMode() {
configuration.set(RUNTIME_MODE, BATCH);
return this;
} | 3.26 |
flink_EnvironmentSettings_m0_rdh | /**
* Tells if the {@link TableEnvironment} should work in a batch or streaming mode.
*/public boolean m0() {
return configuration.get(RUNTIME_MODE) == STREAMING;
} | 3.26 |
flink_EnvironmentSettings_inStreamingMode_rdh | /**
* Sets that the components should work in a streaming mode. Enabled by default.
*/
public Builder inStreamingMode() {
configuration.set(RUNTIME_MODE, STREAMING);
return this;
} | 3.26 |
flink_EnvironmentSettings_m1_rdh | /**
* Specifies the classloader to use in the planner for operations related to code
* generation, UDF loading, operations requiring reflections on user classes, discovery of
* factories.
*
* <p>By default, this is configured using {@code Thread.currentThread().getContextClassLoader()}.
*
* <p>Modify the {@link ... | 3.26 |
flink_EnvironmentSettings_newInstance_rdh | /**
* Creates a builder for creating an instance of {@link EnvironmentSettings}.
*/
public static Builder newInstance() {
return new Builder();
}
/**
* Creates an instance of {@link EnvironmentSettings} from configuration.
*
* @deprecated use {@link Builder#withConfiguration(Configuration)} | 3.26 |
flink_EnvironmentSettings_getConfiguration_rdh | /**
* Get the underlying {@link Configuration}.
*/
public Configuration getConfiguration() {
return configuration;} | 3.26 |
flink_EnvironmentSettings_withConfiguration_rdh | /**
* Add extra configuration to {@link EnvironmentSettings}.
*/
public Builder withConfiguration(Configuration configuration) {
this.configuration.addAll(configuration);
return
this;
} | 3.26 |
flink_EnvironmentSettings_withBuiltInCatalogName_rdh | /**
* Specifies the name of the initial catalog to be created when instantiating a {@link TableEnvironment}.
*
* <p>This catalog is an in-memory catalog that will be used to store all temporary objects
* (e.g. from {@link TableEnvironment#createTemporaryView(String, Table)} or {@link TableEnvironment#createTemporar... | 3.26 |
flink_EnvironmentSettings_getBuiltInCatalogName_rdh | /**
* Gets the specified name of the initial catalog to be created when instantiating a {@link TableEnvironment}.
*/
public String getBuiltInCatalogName() {
return configuration.get(TABLE_CATALOG_NAME);
} | 3.26 |
flink_EnvironmentSettings_build_rdh | /**
* Returns an immutable instance of {@link EnvironmentSettings}.
*/
public EnvironmentSettings
build() {
if (classLoader == null) {
classLoader = Thread.currentThread().getContextClassLoader();
}
return new EnvironmentSettings(configuration, classLoader, catalogStore);
} | 3.26 |
flink_EnvironmentSettings_withBuiltInDatabaseName_rdh | /**
* Specifies the name of the default database in the initial catalog to be created when
* instantiating a {@link TableEnvironment}.
*
* <p>This database is an in-memory database that will be used to store all temporary
* objects (e.g. from {@link TableEnvironment#createTemporaryView(String, Table)} or {@link Ta... | 3.26 |
flink_EnvironmentSettings_getBuiltInDatabaseName_rdh | /**
* Gets the specified name of the default database in the initial catalog to be created when
* instantiating a {@link TableEnvironment}.
*/
public String getBuiltInDatabaseName() {return configuration.get(TABLE_DATABASE_NAME);
} | 3.26 |
flink_FileLock_unlock_rdh | /**
* Release the file lock.
*
* @throws IOException
* If the FileChannel is closed
*/
public void unlock() throws IOException {
if ((lock != null) && lock.channel().isOpen()) {
lock.release();
}
} | 3.26 |
flink_FileLock_inTempFolder_rdh | /**
* Initialize a FileLock using a file located inside temp folder.
*
* @param fileName
* The name of the locking file
* @return The initialized FileLock
*/
public static FileLock inTempFolder(String fileName) {return new FileLock(TEMP_DIR, fileName);
} | 3.26 |
flink_FileLock_tryLock_rdh | /**
* Try to acquire a lock on the locking file. This method immediately returns whenever the lock
* is acquired or not.
*
* @return True if successfully acquired the lock
* @throws IOException
* If the file path is invalid
*/
public boolean tryLock() throws IOException {
if (outputStream == null) {
... | 3.26 |
flink_FileLock_unlockAndDestroy_rdh | /**
* Release the file lock, close the fileChannel and FileOutputStream then try deleting the
* locking file if other file lock does not need it, which means the lock will not be used
* anymore.
*
* @throws IOException
* If an I/O error occurs
*/
public void unlockAndDestroy() throws IOException {
try {
... | 3.26 |
flink_FileLock_init_rdh | /**
* Check whether the locking file exists in the file system. Create it if it does not exist.
* Then create a FileOutputStream for it.
*
* @throws IOException
* If the file path is invalid or the parent dir does not exist
*/ private void init() throws IOException {
if (!this.file.exists()) {
thi... | 3.26 |
flink_FileLock_normalizeFileName_rdh | /**
* Normalize the file name, which only allows slash, backslash, digits and letters.
*
* @param fileName
* Original file name
* @return File name with illegal characters stripped
*/
private static String normalizeFileName(String fileName) {return fileName.replaceAll("[^\\w/\\\\]", "");
} | 3.26 |
flink_SerdeContext_get_rdh | /**
* Retrieve context from {@link SerializerProvider} and {@link DeserializationContext}.
*/public static SerdeContext get(DatabindContext databindContext) {
final SerdeContext serdeContext = ((SerdeContext) (databindContext.getAttribute(SERDE_CONTEXT_KEY)));
assert serdeContext != null;
return serdeCont... | 3.26 |
flink_RawObjectConverter_create_rdh | // --------------------------------------------------------------------------------------------
// Factory method
// --------------------------------------------------------------------------------------------
public static RawObjectConverter<?> create(DataType dataType) {
final LogicalType logicalType = dataType.g... | 3.26 |
flink_DeltaEvictor_of_rdh | /**
* Creates a {@code DeltaEvictor} from the given threshold and {@code DeltaFunction}. Eviction
* is done before the window function.
*
* @param threshold
* The threshold
* @param deltaFunction
* The {@code DeltaFunction}
*/
p... | 3.26 |
flink_PrimitiveArrayTypeInfo_getComponentClass_rdh | /**
* Gets the class that represents the component type.
*
* @return The class of the component type.
*/
@PublicEvolving
public Class<?> getComponentClass() {
return this.arrayClass.getComponentType();
} | 3.26 |
flink_PrimitiveArrayTypeInfo_getComponentType_rdh | /**
* Gets the type information of the component type.
*
* @return The type information of the component type.
*/
@PublicEvolving
public TypeInformation<?>
getComponentType()
{
return BasicTypeInfo.getInfoFor(getComponentClass());
} | 3.26 |
flink_PrimitiveArrayTypeInfo_isBasicType_rdh | // --------------------------------------------------------------------------------------------
@Override
@PublicEvolving
public boolean isBasicType() {
return false;
} | 3.26 |
flink_ExecNodeBase_inputsContainSingleton_rdh | /**
* Whether singleton distribution is required.
*/
protected boolean inputsContainSingleton() {
return getInputProperties().stream().anyMatch(p -> p.getRequiredDistribution().getType() == InputProperty.DistributionType.SINGLETON);
} | 3.26 |
flink_ExecNodeBase_getPersistedConfig_rdh | // Custom filter to exclude node configuration if no consumed options are used
@JsonProperty(value = FIELD_NAME_CONFIGURATION, access = Access.READ_ONLY, index = 2)
@JsonInclude(value = Include.CUSTOM, valueFilter = ConfigurationJsonSerializerFilter.class)
public ReadableConfig getPersistedConfig() {
return persistedC... | 3.26 |
flink_ExecNodeBase_translateToFusionCodegenSpecInternal_rdh | /**
* Internal method, translates this node into a operator codegen spec generator.
*
* @param planner
* The planner.
* @param config
* per-{@link ExecNode} configuration that contains the merged configuration from
* various layers which all the nodes implementing this method should use, instead of
* re... | 3.26 |
flink_ReaderInfo_getLocation_rdh | /**
*
* @return the location of the subtask that runs this source reader.
*/
public String getLocation() {
return location;
} | 3.26 |
flink_ReaderInfo_getSubtaskId_rdh | /**
*
* @return the ID of the subtask that runs the source reader.
*/
public int getSubtaskId() {
return subtaskId;
} | 3.26 |
flink_RowTimeIntervalJoin_getMaxOutputDelay_rdh | /**
* Get the maximum interval between receiving a row and emitting it (as part of a joined
* result). This is the time interval by which watermarks need to be held back.
*
* @return the maximum delay for the outputs
*/
public long getMaxOutputDelay() {
return Math.max(leftRelativeSize, rightRelativeSize) + a... | 3.26 |
flink_CliInputView_init_rdh | // --------------------------------------------------------------------------------------------
@Override
protected void init() {
// nothing to do
} | 3.26 |
flink_CliInputView_insert_rdh | // --------------------------------------------------------------------------------------------
private void insert(String binding) {
currentInput.insert(cursorPos, binding);
cursorPos += binding.length();
// reset view
resetMainPart();
} | 3.26 |
flink_CalculatedTableFactory_create_rdh | /**
* Creates a valid {@link CalculatedQueryOperation} operation.
*
* @param callExpr
* call to table function as expression
* @return valid calculated table
*/
QueryOperation create(ResolvedExpression callExpr, List<String> leftTableFieldNames) {
FunctionTableCallVisitor calculatedTableCreator = new Functi... | 3.26 |
flink_StatePartitionStreamProvider_getStream_rdh | /**
* Returns a stream with the data of one state partition.
*/
public InputStream getStream() throws IOException {
if (creationException != null) {
throw new IOException(creationException);
}
return stream;
} | 3.26 |
flink_SupportsProjectionPushDown_applyProjection_rdh | /**
* Provides the field index paths that should be used for a projection. The indices are 0-based
* and support fields within (possibly nested) structures if this is enabled via {@link #supportsNestedProjection()}.
*
* <p>In the example mentioned in {@link SupportsProjectionPushDown}, this method would receive:
*... | 3.26 |
flink_ArchivedExecutionVertex_getTaskNameWithSubtaskIndex_rdh | // --------------------------------------------------------------------------------------------
// Accessors
// --------------------------------------------------------------------------------------------
@Override
public String getTaskNameWithSubtaskIndex() {
return this.taskNameWithSubtask;
} | 3.26 |
flink_FailureEnricherUtils_labelFailure_rdh | /**
* Enriches a Throwable by returning the merged label output of a Set of FailureEnrichers.
*
* @param cause
* the Throwable to label
* @param context
* the context of the Throwable
* @param mainThreadExecutor
* the executor to comp... | 3.26 |
flink_FailureEnricherUtils_m0_rdh | /**
* Filters out invalid {@link FailureEnricher} objects that have duplicate output keys.
*
* @param failureEnrichers
* a set of {@link FailureEnricher} objects to filter
* @return a filtered collection without any duplicate output keys
*/
@VisibleForTesting
static Collection<FailureEnricher> m0(final Set<Fail... | 3.26 |
flink_FailureEnricherUtils_getIncludedFailureEnrichers_rdh | /**
* Returns a set of failure enricher names included in the given configuration.
*
* @param configuration
* the configuration to get the failure enricher names from
* @return failure enricher names
*/
@VisibleForTesting
static Set<String> getIncludedFailureEnrichers(final Configuration configuration) {
fi... | 3.26 |
flink_FailureEnricherUtils_getFailureEnrichers_rdh | /**
* Returns a set of validated FailureEnrichers for a given configuration.
*
* @param configuration
* the configuration for the job
* @return a collection of validated FailureEnrichers
*/
public static Collection<FailureEnricher> getFailureEnrichers(final Configuration configuration) {
final PluginMana... | 3.26 |
flink_StateUtil_bestEffortDiscardAllStateObjects_rdh | /**
* Iterates through the passed state handles and calls discardState() on each handle that is not
* null. All occurring exceptions are suppressed and collected until the iteration is over and
* emitted as a single exception.
*
* @param handlesToDiscard
* State handles to discard. Passed iterable is allowed to... | 3.26 |
flink_StateUtil_discardStateFuture_rdh | /**
* Discards the given state future by first trying to cancel it. If this is not possible, then
* the state object contained in the future is calculated and afterwards discarded.
*
* @param stateFuture
* to be discarded
* @throws Exception
* if the discard operation failed
... | 3.26 |
flink_StateUtil_getStateSize_rdh | /**
* Returns the size of a state object.
*
* @param handle
* The handle to the retrieved state
*/
public static long getStateSize(StateObject handle) {
return handle == null ? 0 : handle.getStateSize();
} | 3.26 |
flink_StateUtil_unexpectedStateHandleException_rdh | /**
* Creates a {@link RuntimeException} that signals that an operation did not get the type of
* {@link StateObject} that was expected. This can mostly happen when a different {@link StateBackend} from the one that was used for taking a checkpoint/savepoint is used when
* restoring.
*/
public static RuntimeExcepti... | 3.26 |
flink_JvmUtils_createThreadInfoSample_rdh | /**
* Creates a {@link ThreadInfoSample} for a specific thread. Contains thread traces if
* maxStackTraceDepth > 0.
*
* @param threadId
* The ID of the thread to create the thread dump for.
* @param maxStackTraceDepth
* The maximum number of entries in the stack trace to be collected.
* @return The thread i... | 3.26 |
flink_JvmUtils_createThreadDump_rdh | /**
* Creates a thread dump of the current JVM.
*
* @return the thread dump of current JVM
*/
public static Collection<ThreadInfo> createThreadDump() {
ThreadMXBean threadMxBean = ManagementFactory.getThreadMXBean();
return Arrays.asList(threadMxBean.dumpAllThreads(true, true));
} | 3.26 |
flink_ReduceTransformation_getInputType_rdh | /**
* Returns the {@code TypeInformation} for the elements of the input.
*/
public TypeInformation<IN> getInputType() {
return input.getOutputType();
} | 3.26 |
flink_CatalogFactory_supportedProperties_rdh | /**
*
* @deprecated Implement the {@link Factory} based stack instead.
*/
@Deprecated
default List<String> supportedProperties() {
// Default implementation for catalogs implementing the new {@link Factory} stack instead.
return null;
} | 3.26 |
flink_CatalogFactory_requiredContext_rdh | // --------------------------------------------------------------------------------------------
// Default implementations for legacy {@link TableFactory} stack.
// --------------------------------------------------------------------------------------------
/**
*
* @deprecated Implement the {@link Factory} based stac... | 3.26 |
flink_CatalogFactory_createCatalog_rdh | /**
* Creates and configures a {@link Catalog} using the given context.
*
* <p>An implementation should perform validation and the discovery of further (nested)
* factories in this method.
*/
default Catalog createCatalog(Context context) {
throw new CatalogException("Catalog fact... | 3.26 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.