name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_SlotID_getDynamicSlotID_rdh | /**
* Get a SlotID without actual slot index for dynamic slot allocation.
*/
public static SlotID getDynamicSlotID(ResourceID resourceID) {
return new SlotID(resourceID);
} | 3.26 |
flink_SlotID_getResourceID_rdh | // ------------------------------------------------------------------------
@Override
public ResourceID getResourceID() {
return resourceId;} | 3.26 |
flink_SlotID_equals_rdh | // ------------------------------------------------------------------------
@Override
public boolean equals(Object o) {
if (this == o) {
return true;
}
if ((o == null) || (getClass() != o.getClass())) {
return false;
}
SlotID slotID = ((SlotID) (o));
return (slotNumber == slotID.... | 3.26 |
flink_GenericValueMetricGroup_putVariables_rdh | // ------------------------------------------------------------------------
@Override
protected void putVariables(Map<String, String> variables) {
variables.put(ScopeFormat.asVariable(this.key), value);
} | 3.26 |
flink_SqlNodeConverter_supportedSqlKinds_rdh | /**
* Returns the {@link SqlKind SqlKinds} of {@link SqlNode SqlNodes} that the {@link SqlNodeConverter} supports to convert.
*
* <p>If a {@link SqlNodeConverter} returns a non-empty SqlKinds, The conversion framework will
* find the corresponding converter by matching the SqlKind of SqlNode instead of the class of... | 3.26 |
flink_PartitionSpec_getFieldIndices_rdh | /**
* Gets field index of all fields in input.
*/
@JsonIgnore
public int[] getFieldIndices() {
return fields;
} | 3.26 |
flink_FunctionIdentifier_normalizeName_rdh | /**
* Normalize a function name.
*/
public static String normalizeName(String name) {
return name.toLowerCase(); } | 3.26 |
flink_FunctionIdentifier_asSummaryString_rdh | /**
* Returns a string that summarizes this instance for printing to a console or log.
*/
public String asSummaryString() {
if (objectIdentifier != null) {
return String.join(".", objectIdentifier.getCatalogName(), objectIdentifier.getDatabaseName(), objectIdentifier.getObjectName());
} else {return f... | 3.26 |
flink_FunctionIdentifier_toList_rdh | /**
* List of the component names of this function identifier.
*/
public List<String> toList() {
if (objectIdentifier != null) {
return objectIdentifier.toList();} else if (functionName != null)
{
return Collections.singletonList(functionName);
} else { throw new Illega... | 3.26 |
flink_FunctionIdentifier_normalizeObjectIdentifier_rdh | /**
* Normalize an object identifier by only normalizing the function name.
*/
public static ObjectIdentifier normalizeObjectIdentifier(ObjectIdentifier oi) {
return ObjectIdentifier.of(oi.getCatalogName(), oi.getDatabaseName(), normalizeName(oi.getObjectName()));
} | 3.26 |
flink_RefCountedFileWithStream_newFile_rdh | // ------------------------------ Factory methods for initializing a temporary file
// ------------------------------
public static RefCountedFileWithStream newFile(final File file, final OutputStream currentOut) throws IOException {
return new RefCountedFileWithStream(file, currentOut, 0L);
} | 3.26 |
flink_SplitEnumerator_notifyCheckpointComplete_rdh | /**
* We have an empty default implementation here because most source readers do not have to
* implement the method.
*
* @see CheckpointListener#notifyCheckpointComplete(long)
*/
@Override
default void notifyCheckpointComplete(long checkpointId) throws Exception {
} | 3.26 |
flink_SplitEnumerator_handleSourceEvent_rdh | /**
* Handles a custom source event from the source reader.
*
* <p>This method has a default implementation that does nothing, because it is only required to
* be implemented by some sources, which have a custom event protocol between reader and
* enumerator. The common events for reader registration and split req... | 3.26 |
flink_TGetQueryIdResp_isSet_rdh | /**
* Returns true if field corresponding to fieldID is set (has been assigned a value) and false
* otherwise
*/
public boolean isSet(_Fields field) {
if (field == null) {
throw new IllegalArgumentException();
}
switch (field... | 3.26 |
flink_TGetQueryIdResp_findByThriftId_rdh | /**
* Find the _Fields constant that matches fieldId, or null if its not found.
*/
public static _Fields findByThriftId(int fieldId) {
switch (fieldId) {
case 1 :
// QUERY_ID
return QUERY_ID;default :return null;
}
} | 3.26 |
flink_TGetQueryIdResp_findByThriftIdOrThrow_rdh | /**
* Find the _Fields constant that matches fieldId, throwing an exception if it is not found.
*/
public static _Fields findByThriftIdOrThrow(int fieldId) {
_Fields fields = findByThriftId(fieldId);
if (fields == null)
throw new IllegalArgumentException(("Field " + fieldId) + " doesn't exist!");
... | 3.26 |
flink_TGetQueryIdResp_isSetQueryId_rdh | /**
* Returns true if field queryId is set (has been assigned a value) and false otherwise
*/
public boolean isSetQueryId() {
return this.queryId != null;
} | 3.26 |
flink_TGetQueryIdResp_findByName_rdh | /**
* Find the _Fields constant that matches name, or null if its not found.
*/
public static _Fields findByName(String name) {
return byName.get(name);
} | 3.26 |
flink_RequestJobsOverview_hashCode_rdh | // ------------------------------------------------------------------------
@Override
public int hashCode() {
return RequestJobsOverview.class.hashCode();
} | 3.26 |
flink_RequestJobsOverview_readResolve_rdh | /**
* Preserve the singleton property by returning the singleton instance
*/
private Object readResolve() {
return INSTANCE;
} | 3.26 |
flink_GenericInputSplit_getSplitNumber_rdh | // --------------------------------------------------------------------------------------------
@Override
public int getSplitNumber() {
return this.partitionNumber;
} | 3.26 |
flink_GenericInputSplit_hashCode_rdh | // --------------------------------------------------------------------------------------------
@Override
public int hashCode() {
return this.partitionNumber ^ this.totalNumberOfPartitions;
} | 3.26 |
flink_LocalInputChannel_checkpointStarted_rdh | // ------------------------------------------------------------------------
// Consume
// ------------------------------------------------------------------------
public void checkpointStarted(CheckpointBarrier barrier) throws CheckpointException {
channelStatePersister.startPersisting(barrier.getId(), Collections... | 3.26 |
flink_LocalInputChannel_releaseAllResources_rdh | /**
* Releases the partition reader.
*/
@Override
void releaseAllResources() throws IOException {
if (!isReleased) {
isReleased = true;
ResultSubpartitionView view = subpartitionView;
if (view != null) {
view.releas... | 3.26 |
flink_LocalInputChannel_getSubpartitionView_rdh | // ------------------------------------------------------------------------
@VisibleForTesting
ResultSubpartitionView getSubpartitionView() {
return subpartitionView;
} | 3.26 |
flink_LocalInputChannel_isReleased_rdh | // ------------------------------------------------------------------------
// Life cycle
// ------------------------------------------------------------------------
@Override
boolean isReleased() {
return isReleased;
} | 3.26 |
flink_LocalInputChannel_retriggerSubpartitionRequest_rdh | /**
* Retriggers a subpartition request.
*/
void retriggerSubpartitionRequest(Timer timer) {
synchronized(requestLock) {
checkState(subpartitionView == null, "already requested partition");
timer.schedule(new TimerTask() {
@Override
public void run() {
try {... | 3.26 |
flink_ChannelStateCheckpointWriter_m0_rdh | /**
* The throwable is just used for specific subtask that triggered the failure. Other subtasks
* should fail by {@link CHANNEL_STATE_SHARED_STREAM_EXCEPTION}.
*/
public void m0(JobVertexID jobVertexID, int subtaskIndex, Throwable throwable) {
if (isDone()) {
return;
}
this.throwable = throwable;
ChannelStatePendi... | 3.26 |
flink_EndOfChannelStateEvent_read_rdh | // ------------------------------------------------------------------------
@Override
public void read(DataInputView in) {
// Nothing to do here
} | 3.26 |
flink_EndOfChannelStateEvent_hashCode_rdh | // ------------------------------------------------------------------------
@Override
public int hashCode() {return 1965146670;
} | 3.26 |
flink_DuplicatingFileSystem_of_rdh | /**
* A factory method for creating a simple pair of source/destination.
*/
static CopyRequest of(Path source, Path destination) {
return new CopyRequest() {
@Override
public Path getSource()
{
return source;
}
... | 3.26 |
flink_HashTableBloomFilter_addHash_rdh | /**
*
* @return false if the accuracy of the BloomFilter is not high.
*/
boolean addHash(int hash) {
setLocation(hash);
filter.addHash(hash); size++;
return size <= maxSize;
} | 3.26 |
flink_EvictingWindowReader_process_rdh | /**
* Reads window state generated without any preaggregation such as {@code WindowedStream#apply}
* and {@code WindowedStream#process}.
*
* @param uid
* The uid of the operator.
* @param readerFunction
* The window reader function.
* @param keyType
* The key type of the window.
* @param stateType
* ... | 3.26 |
flink_EvictingWindowReader_aggregate_rdh | /**
* Reads window state generated using an {@link AggregateFunction}.
*
* @param uid
* The uid of the operator.
* @param aggregateFunction
* The aggregate function used to create the window.
* @param readerFunction
* The window reader function.
* @param keyType
* The key type of the window.
* @param... | 3.26 |
flink_EvictingWindowReader_reduce_rdh | /**
* Reads window state generated using a {@link ReduceFunction}.
*
* @param uid
* The uid of the operator.
* @param function
* The reduce function used to create the window.
* @param readerFunction
* The window reader function.
* @param keyType
* The key type of the window.
* @param reduceType
* ... | 3.26 |
flink_HashSubpartitionBufferAccumulator_m0_rdh | // ------------------------------------------------------------------------
// Called by HashBufferAccumulator
// ------------------------------------------------------------------------
public void m0(ByteBuffer record, Buffer.DataType dataType) throws IOException {
if (dataType.isEvent()) {
writeEvent(re... | 3.26 |
flink_HashSubpartitionBufferAccumulator_writeEvent_rdh | // ------------------------------------------------------------------------
// Internal Methods
// ------------------------------------------------------------------------
private void writeEvent(ByteBuffer event, Buffer.DataType dataType) {checkArgument(dataType.isEvent());
// Each event should take an exclusive... | 3.26 |
flink_Ordering_getNumberOfFields_rdh | // --------------------------------------------------------------------------------------------
public int getNumberOfFields() {
return this.indexes.size();
} | 3.26 |
flink_Ordering_clone_rdh | // --------------------------------------------------------------------------------------------
public Ordering clone() {
Ordering newOrdering = new Ordering();
newOrdering.indexes = this.indexes;
newOrdering.types.addAll(this.types);
newOrdering.orders.addAll(this.orders);
return newOrdering;
} | 3.26 |
flink_Ordering_isMetBy_rdh | // --------------------------------------------------------------------------------------------
public boolean isMetBy(Ordering otherOrdering) {
if
((otherOrdering == null) || (this.indexes.size() > otherOrdering.indexes.size())) {
return false;
} for... | 3.26 |
flink_Ordering_appendOrdering_rdh | /**
* Extends this ordering by appending an additional order requirement. If the index has been
* previously appended then the unmodified Ordering is returned.
*
* @param index
* Field index of the appended order requirement.
* @param type
* Type of the appended order requirement.
* @param order
* Order ... | 3.26 |
flink_Ordering_getTypes_rdh | // --------------------------------------------------------------------------------------------
@SuppressWarnings("unchecked")
public Class<? extends Comparable<?>>[] getTypes() {
return this.types.toArray(new Class[this.types.size()]);
} | 3.26 |
flink_BinaryMapData_valueOf_rdh | // ------------------------------------------------------------------------------------------
// Construction Utilities
// ------------------------------------------------------------------------------------------
public static BinaryMapData valueOf(BinaryArrayData key, BinaryArrayData value) {
checkArgument((key.s... | 3.26 |
flink_DefaultLeaderRetrievalService_notifyLeaderAddress_rdh | /**
* Called by specific {@link LeaderRetrievalDriver} to notify leader address.
*
* @param leaderInformation
* new notified leader information address. The exception will be
* handled by leader listener.
*/
@Override
@GuardedBy("lock")
public void notifyLeaderAddress(LeaderInformation leaderInformation) {
... | 3.26 |
flink_HeapPriorityQueue_clear_rdh | /**
* Clears the queue.
*/
public void clear() {
final int arrayOffset = getHeadElementIndex();
Arrays.fill(queue, arrayOffset, arrayOffset + size, null);
size = 0;
} | 3.26 |
flink_DemultiplexingRecordDeserializer_getNextRecord_rdh | /**
* Summarizes the status and watermarks of all virtual channels.
*/
@Override
public DeserializationResult getNextRecord(DeserializationDelegate<StreamElement> delegate) throws IOException {
DeserializationResult result;
do {
result = currentVirtualChannel.getNextRecord(delegate);
if (resul... | 3.26 |
flink_UnsortedGrouping_min_rdh | /**
* Syntactic sugar for aggregate (MIN, field).
*
* @param field
* The index of the Tuple field on which the aggregation function is applied.
* @return An AggregateOperator that represents the min'ed DataSet.
* @see org.apache.flink.api.java.operators.AggregateOperator
*/
public AggregateOperator<T> min(int ... | 3.26 |
flink_UnsortedGrouping_max_rdh | /**
* Syntactic sugar for aggregate (MAX, field).
*
* @param field
* The index of the Tuple field on which the aggregation function is applied.
* @return An AggregateOperator that represents the max'ed DataSet.
* @see org.apache.flink.api.java.operators.AggregateOperator
*/
public AggregateOperator<T> max(int... | 3.26 |
flink_UnsortedGrouping_reduce_rdh | /**
* Applies a Reduce transformation on a grouped {@link DataSet}.
*
* <p>For each group, the transformation consecutively calls a {@link org.apache.flink.api.common.functions.RichReduceFunction} until only a single element for
* each group remains. A ReduceFunction combines two elements into one new element of th... | 3.26 |
flink_UnsortedGrouping_withPartitioner_rdh | /**
* Uses a custom partitioner for the grouping.
*
* @param partitioner
* The custom partitioner.
* @return The grouping object itself, to allow for method chaining.
*/
public UnsortedGrouping<T> withPartitioner(Partitioner<?> partitioner) {
Preconditions.checkNotNull(partitioner);
getKeys().validateCu... | 3.26 |
flink_UnsortedGrouping_sortGroup_rdh | /**
* Sorts elements within a group on a key extracted by the specified {@link org.apache.flink.api.java.functions.KeySelector} in the specified {@link Order}.
*
* <p>Chaining {@link #sortGroup(KeySelector, Order)} calls is not supported.
*
* @param keySelector
* The KeySelector with which the group is sorted.
... | 3.26 |
flink_UnsortedGrouping_maxBy_rdh | /**
* Applies a special case of a reduce transformation (maxBy) on a grouped {@link DataSet}.
*
* <p>The transformation consecutively calls a {@link ReduceFunction} until only a single
* element remains which is the result of the transformation. A ReduceFunction combines two
* elements into one new element of the ... | 3.26 |
flink_UnsortedGrouping_sum_rdh | /**
* Syntactic sugar for aggregate (SUM, field).
*
* @param field
* The index of the Tuple field on which the aggregation function is applied.
* @return An AggregateOperator that represents the summed DataSet.
* @see org.apache.flink.api.java.operators.AggregateOperator
*/
public AggregateOperator<T> sum(int... | 3.26 |
flink_UnsortedGrouping_first_rdh | /**
* Returns a new set containing the first n elements in this grouped {@link DataSet}.
*
* @param n
* The desired number of elements for each group.
* @return A GroupReduceOperator that represents the DataSet containing the elements.
*/
public GroupReduceOperator<T, T> first(int n) {
if (n < 1) {
... | 3.26 |
flink_UnsortedGrouping_reduceGroup_rdh | /**
* Applies a GroupReduce transformation on a grouped {@link DataSet}.
*
* <p>The transformation calls a {@link org.apache.flink.api.common.functions.RichGroupReduceFunction} for each group of the DataSet.
* A GroupReduceFunction can iterate over all elements of a group and emit any number of output
* elements i... | 3.26 |
flink_UnsortedGrouping_combineGroup_rdh | /**
* Applies a GroupCombineFunction on a grouped {@link DataSet}. A GroupCombineFunction is
* similar to a GroupReduceFunction but does not perform a full data exchange. Instead, the
* CombineFunction calls the combine method once per partition for combining a group of results.
* This operator is suitable for comb... | 3.26 |
flink_UnsortedGrouping_minBy_rdh | /**
* Applies a special case of a reduce transformation (minBy) on a grouped {@link DataSet}.
*
* <p>The transformation consecutively calls a {@link ReduceFunction} until only a single
* element remains which is the result of the transformation. A ReduceFunction combines two
* elements into one new element of the ... | 3.26 |
flink_FileMergingSnapshotManagerBase_createManagedDirectory_rdh | // ------------------------------------------------------------------------
// utilities
// ------------------------------------------------------------------------
private void createManagedDirectory(Path
managedPath) {
try {
FileStatus fileStatus = n... | 3.26 |
flink_FileMergingSnapshotManagerBase_createPhysicalFile_rdh | /**
* Create a physical file in right location (managed directory), which is specified by scope of
* this checkpoint and current subtask.
*
* @param subtaskKey
* the {@link SubtaskKey} of current subtask.
* @param scope
* the scope of the checkpoint.
* @return the created physical file.
* @throws IOExcepti... | 3.26 |
flink_FileMergingSnapshotManagerBase_generatePhysicalFilePath_rdh | /**
* Generate a file path for a physical file.
*
* @param dirPath
* the parent directory path for the physical file.
* @return the generated file path for a physical file.
*/
protected Path generatePhysicalFilePath(Path dirPath) {
// this must be called after initFileSystem() is called
// so th... | 3.26 |
flink_FileMergingSnapshotManagerBase_createLogicalFile_rdh | // ------------------------------------------------------------------------
// logical & physical file
// ------------------------------------------------------------------------
/**
* Create a logical file on a physical file.
*
* @param physicalFile
* the underlying physical file.
* @param startOffset
* the ... | 3.26 |
flink_FileMergingSnapshotManagerBase_deletePhysicalFile_rdh | /**
* Delete a physical file by given file path. Use the io executor to do the deletion.
*
* @param filePath
* the given file path to delete.
*/
protected final void deletePhysicalFile(Path filePath) {
f0.execute(() -> {
try {
fs.delete(filePath, false);
LOG.debug("Physical ... | 3.26 |
flink_FileMergingSnapshotManagerBase_getManagedDir_rdh | // ------------------------------------------------------------------------
// file system
// ------------------------------------------------------------------------
@Overridepublic Path getManagedDir(SubtaskKey subtaskKey, CheckpointedStateScope scope) {
if (scope.equals(CheckpointedStateScope.SHARED)) {
... | 3.26 |
flink_TemplateUtils_extractProcedureLocalFunctionTemplates_rdh | /**
* Retrieve local templates from procedure method.
*/static Set<FunctionTemplate> extractProcedureLocalFunctionTemplates(DataTypeFactory typeFactory, Method method) {
return m0(typeFactory, collectAnnotationsOfMethod(ProcedureHint.class,
method));
} | 3.26 |
flink_TemplateUtils_extractGlobalFunctionTemplates_rdh | /**
* Retrieve global templates from function class.
*/
static Set<FunctionTemplate> extractGlobalFunctionTemplates(DataTypeFactory typeFactory, Class<? extends UserDefinedFunction> function) {
return asFunctionTemplates(typeFactory, collectAnnotationsOfClass(FunctionHint.class, function));
} | 3.26 |
flink_TemplateUtils_findInputOnlyTemplates_rdh | /**
* Hints that only declare an input.
*/
static Set<FunctionSignatureTemplate> findInputOnlyTemplates(Set<FunctionTemplate> global, Set<FunctionTemplate> local, Function<FunctionTemplate, FunctionResultTemplate> accessor) {
return Stream.concat(global.stream(), local.stream()).filter(t -> (t.getSignatureTempla... | 3.26 |
flink_TemplateUtils_findResultOnlyTemplates_rdh | /**
* Find a template that only specifies a result.
*/
static Set<FunctionResultTemplate> findResultOnlyTemplates(Set<FunctionTemplate> functionTemplates, Function<FunctionTemplate, FunctionResultTemplate> accessor) {
return functionTemplates.stream().filter(t -> (t.getSignatureTemplate() == null) && (accessor... | 3.26 |
flink_TemplateUtils_m0_rdh | /**
* Converts {@link ProcedureHint}s to {@link FunctionTemplate}.
*/
static Set<FunctionTemplate> m0(DataTypeFactory typeFactory, Set<ProcedureHint> hints) {
return hints.stream().map(hint -> {
try {
return FunctionTemplate.fromAnnotation(typeFactory, hint);
} catch (T... | 3.26 |
flink_TemplateUtils_asFunctionTemplates_rdh | /**
* Converts {@link FunctionHint}s to {@link FunctionTemplate}.
*/
static Set<FunctionTemplate> asFunctionTemplates(DataTypeFactory typeFactory, Set<FunctionHint> hints) {
return hints.stream().map(hint -> {
try {
return FunctionTemplate.fromAnnotation(typeFactory, hint);
} catch (T... | 3.26 |
flink_TemplateUtils_extractLocalFunctionTemplates_rdh | /**
* Retrieve local templates from function method.
*/
static Set<FunctionTemplate> extractLocalFunctionTemplates(DataTypeFactory typeFactory, Method method) {return
asFunctionTemplates(typeFactory, collectAnnotationsOfMethod(FunctionHint.class, method));
} | 3.26 |
flink_TemplateUtils_extractProcedureGlobalFunctionTemplates_rdh | /**
* Retrieve global templates from procedure class.
*/
static Set<FunctionTemplate> extractProcedureGlobalFunctionTemplates(DataTypeFactory typeFactory, Class<? extends Procedure> procedure) {
return m0(typeFactory, collectAnnotationsOfClass(ProcedureHint.class, procedure));
} | 3.26 |
flink_TemplateUtils_findResultMappingTemplates_rdh | /**
* Hints that map a signature to a result.
*/
static Set<FunctionTemplate> findResultMappingTemplates(Set<FunctionTemplate> globalTemplates, Set<FunctionTemplate> localTemplates, Function<FunctionTemplate, FunctionResultTemplate> accessor) {
return Stream.concat(globalTemplates.stream(), localTemplates.str... | 3.26 |
flink_TemplateUtils_findResultOnlyTemplate_rdh | /**
* Hints that only declare a result (either accumulator or output).
*/
@Nullable
static FunctionResultTemplate findResultOnlyTemplate(Set<FunctionResultTemplate> globalResultOnly, Set<FunctionResultTemplate> localResultOnly, Set<FunctionTemplate> explicitMappings, Function<FunctionTemplate, FunctionResultTemplat... | 3.26 |
flink_ResolvedExpression_asSerializableString_rdh | /**
* Returns a string that fully serializes this instance. The serialized string can be used for
* storing the query in, for example, a {@link org.apache.flink.table.catalog.Catalog} as a
* view.
*
* @return detailed string for persisting in a catalog
*/
default String asSerializableString() {
throw new Tabl... | 3.26 |
flink_PythonStreamGroupWindowAggregateOperator_createTumblingGroupWindowAggregateOperator_rdh | // The below static create methods are reflected from the planner
public static <K, W extends Window> PythonStreamGroupWindowAggregateOperator<K, W> createTumblingGroupWindowAggregateOperator(Configuration config, RowType inputType, RowType outputType,
PythonAggregateFunctionInfo[] aggregateFunctions, DataViewSpec[][... | 3.26 |
flink_CheckpointedInputGate_processPriorityEvents_rdh | /**
* Eagerly pulls and processes all priority events. Must be called from task thread.
*
* <p>Basic assumption is that no priority event needs to be handled by the {@link StreamTaskNetworkInput}.
*/
private void processPriorityEvents() throws IOException, InterruptedException {
// check if the priority event i... | 3.26 |
flink_CheckpointedInputGate_getNumberOfInputChannels_rdh | /**
*
* @return number of underlying input channels.
*/
public int getNumberOfInputChannels() {
return inputGate.getNumberOfInputChannels();
} | 3.26 |
flink_CheckpointedInputGate_toString_rdh | // ------------------------------------------------------------------------
// Utilities
// ------------------------------------------------------------------------
@Override
public String toString() {
return barrierHandler.toString();
} | 3.26 |
flink_StreamingFileSink_initializeState_rdh | // --------------------------- Sink Methods -----------------------------
@Override
public void initializeState(FunctionInitializationContext context) throws
Exception {
this.helper = new StreamingFileSinkHelper<>(bucketsBuilder.createBuckets(getRuntimeContext().getIndexOfThisSubtask()), context.isRestored(), context... | 3.26 |
flink_StreamingFileSink_build_rdh | /**
* Creates the actual sink.
*/
public StreamingFileSink<IN> build() {
return new StreamingFileSink<>(this, bucketCheckInterval);
} | 3.26 |
flink_TimerGauge_getAccumulatedCount_rdh | /**
*
* @return the accumulated period by the given * TimerGauge.
*/
public synchronized long getAccumulatedCount() {
return accumulatedCount;
} | 3.26 |
flink_ReusingKeyGroupedIterator_next_rdh | /**
* Prior to call this method, call hasNext() once!
*/
@Override
public E next() {
if (this.currentIsUnconsumed || hasNext()) {
this.currentIsUnconsumed = false;
return ReusingKeyGroupedIterator.this.current;
} else {
throw new NoSuchElementException();
}
} | 3.26 |
flink_ClusterEntryPointExceptionUtils_tryEnrichClusterEntryPointError_rdh | /**
* Tries to enrich the passed exception or its causes with additional information.
*
* <p>This method improves error messages for direct and metaspace {@link OutOfMemoryError}. It
* adds descriptions about possible causes and ways of resolution.
*
* @param root
* The Throwable of which the cause tree shall ... | 3.26 |
flink_StreamOperatorFactory_isInputTypeConfigurable_rdh | /**
* If the stream operator need to be configured with the data type they will operate on.
*/
default boolean isInputTypeConfigurable() {
return false;
} | 3.26 |
flink_StreamOperatorFactory_setOutputType_rdh | /**
* Is called by the {@link StreamGraph#addOperator} method when the {@link StreamGraph} is
* generated. The method is called with the output {@link TypeInformation} which is also used
* for the {@link StreamTask} output serializer.
*
* @param type
* Output type information of the {@link StreamTask}
* @param... | 3.26 |
flink_StreamOperatorFactory_isStreamSource_rdh | /**
* Is this factory for {@link StreamSource}.
*/
default boolean isStreamSource() {
return false;
} | 3.26 |
flink_StreamOperatorFactory_setInputType_rdh | /**
* Is called by the {@link StreamGraph#addOperator} method when the {@link StreamGraph} is
* generated.
*
* @param type
* The data type of the input.
* @param executionConfig
* The execution config for this parallel execution.
*/
default void setInputType(TypeInformation<?> type, ExecutionConfig executio... | 3.26 |
flink_RequestJobDetails_shouldIncludeFinished_rdh | // ------------------------------------------------------------------------
public boolean shouldIncludeFinished() {
return includeFinished;
} | 3.26 |
flink_RequestJobDetails_equals_rdh | // ------------------------------------------------------------------------
@Override
public boolean equals(Object o) {
if (this == o) {return true;
} else if (o instanceof RequestJobDetails) {
RequestJobDetails that = ((RequestJobDetails) (o));
return (this.includeFinished... | 3.26 |
flink_ConstraintEnforcer_build_rdh | /**
* If neither of NOT NULL or CHAR/VARCHAR length or BINARY/VARBINARY enforcers are
* configured, null is returned.
*/
public ConstraintEnforcer build() {
if (isConfigured) {
String
operatorName = ("ConstraintEnforcer[" + String.join(", ", operatorNames)) + "]";
return new ConstraintEn... | 3.26 |
flink_AbstractHeapState_getStateTable_rdh | /**
* This should only be used for testing.
*/
@VisibleForTesting
public StateTable<K, N, SV> getStateTable() {
return stateTable;
} | 3.26 |
flink_AbstractHeapState_clear_rdh | // ------------------------------------------------------------------------
@Override public final void clear() {
stateTable.remove(currentNamespace);} | 3.26 |
flink_KeyGroupsStateHandle_getDelegateStateHandle_rdh | /**
*
* @return The handle to the actual states
*/
public StreamStateHandle getDelegateStateHandle() {
return stateHandle;
} | 3.26 |
flink_KeyGroupsStateHandle_getGroupRangeOffsets_rdh | /**
*
* @return the internal key-group range to offsets metadata
*/
public KeyGroupRangeOffsets getGroupRangeOffsets() {
return groupRangeOffsets; } | 3.26 |
flink_JobManagerCheckpointStorage_createFromConfig_rdh | /**
* Creates a new {@link JobManagerCheckpointStorage} using the given configuration.
*
* @param config
* The Flink configuration (loaded by the TaskManager).
* @param classLoader
* The clsas loader that should be used to load the checkpoint storage.
* @return The created checkpoint storage.
* @throws Ille... | 3.26 |
flink_JobManagerCheckpointStorage_configure_rdh | // ------------------------------------------------------------------------
// Reconfiguration
// ------------------------------------------------------------------------
/**
* Creates a copy of this checkpoint storage that uses the values defined in the configuration
* for fields where that were not specified in thi... | 3.26 |
flink_JobManagerCheckpointStorage_getCheckpointPath_rdh | /**
*
* @return The location where checkpoints will be externalized if set.
*/
@Nullable
public Path getCheckpointPath() {
return location.getBaseCheckpointPath();
} | 3.26 |
flink_JobManagerCheckpointStorage_getMaxStateSize_rdh | // ------------------------------------------------------------------------
// Properties
// ------------------------------------------------------------------------
/**
* Gets the maximum size that an individual state can have, as configured in the constructor (by
* default {@value #DEFAULT_MAX_STATE_SIZE}).
*
* @... | 3.26 |
flink_JobManagerCheckpointStorage_resolveCheckpoint_rdh | // ------------------------------------------------------------------------
// checkpoint state persistence
// ------------------------------------------------------------------------
@Override
public CompletedCheckpointStorageLocation resolveCheckpoint(String pointer) throws IOException {
return AbstractFsCheckpoi... | 3.26 |
flink_JobManagerCheckpointStorage_toString_rdh | // ------------------------------------------------------------------------
// utilities
// ------------------------------------------------------------------------
@Override
public String toString() {
return (("JobManagerCheckpointStorage (checkpoints to JobManager) " + "( maxStateSize: ") + maxStateSize) + "... | 3.26 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.