name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_PythonOperatorChainingOptimizer_buildOutputMap_rdh | /**
* Construct the key-value pairs where the value is the output transformations of the key
* transformation.
*/
private static Map<Transformation<?>, Set<Transformation<?>>> buildOutputMap(List<Transformation<?>> transformations) {
final Map<Transformation<?>, Set<Transformation<?>>> outputMap = new HashMap<>(... | 3.26 |
flink_PythonOperatorChainingOptimizer_replaceInput_rdh | // ----------------------- Utility Methods -----------------------
private static void replaceInput(Transformation<?> transformation, Transformation<?> oldInput, Transformation<?> newInput) {
try {
if (((((((transformation instanceof OneInputTransformation) || (transformation instanceof FeedbackTransformati... | 3.26 |
flink_PythonOperatorChainingOptimizer_optimize_rdh | /**
* Perform chaining optimization. It will returns the chained transformations and the
* transformation after chaining optimization for the given transformation.
*/
public static Tuple2<List<Transformation<?>>, Transformation<?>> optimize(List<Transformation<?>> transformations, Transformation<... | 3.26 |
flink_HybridShuffleConfiguration_getFullStrategyReleaseBufferRatio_rdh | /**
* The proportion of buffers to be released. Used by {@link HsFullSpillingStrategy}.
*/
public float getFullStrategyReleaseBufferRatio() {
return fullStrategyReleaseBufferRatio;
} | 3.26 |
flink_HybridShuffleConfiguration_getSpillingStrategyType_rdh | /**
* Get {@link SpillingStrategyType} for hybrid shuffle mode.
*/
public SpillingStrategyType getSpillingStrategyType() {
return spillingStrategyType;
} | 3.26 |
flink_HybridShuffleConfiguration_getSelectiveStrategySpillBufferRatio_rdh | /**
* The proportion of buffers to be spilled. Used by {@link HsSelectiveSpillingStrategy}.
*/
public float getSelectiveStrategySpillBufferRatio() {
return selectiveStrategySpillBufferRatio;
} | 3.26 |
flink_HybridShuffleConfiguration_getMaxBuffersReadAhead_rdh | /**
* Determine how many buffers to read ahead at most for each subpartition to prevent other
* consumers from starving.
*/
public int getMaxBuffersReadAhead() {
return maxBuffersReadAhead;
} | 3.26 |
flink_HybridShuffleConfiguration_getRegionGroupSizeInBytes_rdh | /**
* Segment size of hybrid spilled file data index.
*/
public int getRegionGroupSizeInBytes() {
return regionGroupSizeInBytes;
} | 3.26 |
flink_HybridShuffleConfiguration_getBufferRequestTimeout_rdh | /**
* Maximum time to wait when requesting read buffers from the buffer pool before throwing an
* exception.
*/
public Duration getBufferRequestTimeout() {
return bufferRequestTimeout;} | 3.26 |
flink_HybridShuffleConfiguration_getFullStrategyReleaseThreshold_rdh | /**
* When the number of buffers that have been requested exceeds this threshold, trigger the
* release operation. Used by {@link HsFullSpillingStrategy}.
*/
public float getFullStrategyReleaseThreshold() {
return fullStrategyReleaseThreshold;
} | 3.26 |
flink_HybridShuffleConfiguration_getBufferPoolSizeCheckIntervalMs_rdh | /**
* Check interval of buffer pool's size.
*/
public long getBufferPoolSizeCheckIntervalMs() {
return bufferPoolSizeCheckIntervalMs;
} | 3.26 |
flink_HybridShuffleConfiguration_getNumRetainedInMemoryRegionsMax_rdh | /**
* Max number of hybrid retained regions in memory.
*/
public long getNumRetainedInMemoryRegionsMax() {
return numRetainedInMemoryRegionsMax;
} | 3.26 |
flink_FileSource_forRecordStreamFormat_rdh | // ------------------------------------------------------------------------
// Entry-point Factory Methods
// ------------------------------------------------------------------------
/**
* Builds a new {@code FileSource} using a {@link StreamFormat} to read record-by-record from a
* file stream.
*
* <p>When possibl... | 3.26 |
flink_FileSource_forBulkFileFormat_rdh | /**
* Builds a new {@code FileSource} using a {@link BulkFormat} to read batches of records from
* files.
*
* <p>Examples for bulk readers are compressed and vectorized formats such as ORC or Parquet.
*/
public static <T> FileSourceBuilder<T> forBulkFileFormat(final BulkFormat<T, FileSourceSplit> bulkFormat, final... | 3.26 |
flink_FileSource_forRecordFileFormat_rdh | /**
* Builds a new {@code FileSource} using a {@link FileRecordFormat} to read record-by-record
* from a a file path.
*
* <p>A {@code FileRecordFormat} is more general than the {@link StreamFormat}, but also
* requires often more careful parametrization.
*
* @deprecated Please use {@link #forRecordStreamFormat(S... | 3.26 |
flink_SqlJsonValueFunctionWrapper_explicitTypeSpec_rdh | /**
* Copied and modified from the original {@link SqlJsonValueFunction}.
*
* <p>Changes: Instead of returning {@link Optional} this method returns null directly.
*/
private static RelDataType explicitTypeSpec(SqlOperatorBinding opBinding) {
if (((opBinding.getOperandCount() > 2) && opBinding.isOperandLiteral(2... | 3.26 |
flink_StateWithExecutionGraph_updateTaskExecutionState_rdh | /**
* Updates the execution graph with the given task execution state transition.
*
* @param taskExecutionStateTransition
* taskExecutionStateTransition to update the ExecutionGraph
* with
* @param failureLabels
* the failure labels to attach to the task failure cause
* @return {@code true} if the update ... | 3.26 |
flink_AllGroupReduceDriver_setup_rdh | // ------------------------------------------------------------------------
@Override
public void setup(TaskContext<GroupReduceFunction<IT, OT>, OT> context) {this.taskContext = context;
} | 3.26 |
flink_AllGroupReduceDriver_prepare_rdh | // --------------------------------------------------------------------------------------------
@Override
public void prepare() throws Exception {
final TaskConfig config = this.taskContext.getTaskConfig();this.strategy = config.getDriverStrategy();
switch (this.strategy)... | 3.26 |
flink_S3Recoverable_m0_rdh | // ------------------------------------------------------------------------
public String m0() {
return
uploadId;
} | 3.26 |
flink_LeaderRetrievalUtils_retrieveLeaderInformation_rdh | /**
* Retrieves the leader pekko url and the current leader session ID. The values are stored in a
* {@link LeaderInformation} instance.
*
* @param leaderRetrievalService
* Leader retrieval service to retrieve the leader connection
* information
* @param timeout
* Timeout when to give up looking for the l... | 3.26 |
flink_RecordAndPosition_getRecord_rdh | // ------------------------------------------------------------------------
public E getRecord() {
return record;
} | 3.26 |
flink_RecordAndPosition_toString_rdh | // ------------------------------------------------------------------------
@Override
public String toString() {
return String.format("%s @ %d + %d", record, offset, recordSkipCount);
} | 3.26 |
flink_HadoopOutputFormatBase_writeObject_rdh | // --------------------------------------------------------------------------------------------
// Custom serialization methods
// --------------------------------------------------------------------------------------------
private void writeObject(ObjectOutputStream out) throws IOException {
super.write(out);
... | 3.26 |
flink_HadoopOutputFormatBase_open_rdh | /**
* create the temporary output file for hadoop RecordWriter.
*
* @param taskNumber
* The number of the parallel instance.
* @param numTasks
* The number of parallel tasks.
* @throws java.io.IOException
*/
@Overridepublic void open(int taskNumber, int numTasks) throws IOException {
// enforce sequenti... | 3.26 |
flink_HadoopOutputFormatBase_configure_rdh | // --------------------------------------------------------------------------------------------
// OutputFormat
// --------------------------------------------------------------------------------------------
@Override
public void configure(Configuration parameters) {
// enforce sequential configure() calls
sync... | 3.26 |
flink_HadoopOutputFormatBase_close_rdh | /**
* commit the task by moving the output file out from the temporary directory.
*
* @throws java.io.IOException
*/
@Override
public void close() throws IOException {
// enforce sequential close() calls
synchronized(CLOSE_MUTEX) {
this.recordWriter.close(new HadoopDummyReporter());
if
... | 3.26 |
flink_SqlJsonUtils_createArrayNode_rdh | /**
* Returns a new {@link ArrayNode}.
*/
public static ArrayNode createArrayNode() {
return MAPPER.createArrayNode();} | 3.26 |
flink_SqlJsonUtils_getNodeFactory_rdh | /**
* Returns the {@link JsonNodeFactory} for creating nodes.
*/
public static JsonNodeFactory getNodeFactory() {
return MAPPER.getNodeFactory();
}
/**
* Returns a new {@link ObjectNode} | 3.26 |
flink_SqlJsonUtils_serializeJson_rdh | /**
* Serializes the given {@link JsonNode} to a JSON string.
*/
public static String serializeJson(JsonNode node) {
try {
// For JSON functions to have deterministic output, we need to sort the keys. However,
// Jackson's built-in features don't work on the tree representation, so we need to
... | 3.26 |
flink_StateMap_releaseSnapshot_rdh | /**
* Releases a snapshot for this {@link StateMap}. This method should be called once a snapshot
* is no more needed.
*
* @param snapshotToRelease
* the snapshot to release, which was previously created by this state
* map.
*/
public void releaseSnapshot(StateMapSnapshot<K, N, S, ? extends StateMap<K, N, S>... | 3.26 |
flink_SerializableHadoopConfigWrapper_writeObject_rdh | // ------------------------------------------------------------------------
private void writeObject(ObjectOutputStream out) throws IOException {
out.defaultWriteObject();
// we write the Hadoop config through a separate serializer to avoid cryptic exceptions when
// it
// corrupts the s... | 3.26 |
flink_IterationTailTask_initialize_rdh | // --------------------------------------------------------------------------------------------
@Override
protected void initialize() throws Exception {
super.initialize();
// sanity check: the tail has to update either the workset or the solution set
if ((!isWorksetUpdate) && (!isSolutionSetUpdate)) {
... | 3.26 |
flink_ColumnStats_copy_rdh | /**
* Create a deep copy of "this" instance.
*
* @return a deep copy
*/
public ColumnStats copy() {
if ((maxValue != null) || (f0 != null)) {
return new ColumnStats(this.ndv, this.nullCount, this.avgLen, this.maxLen, this.maxValue, this.f0);
} else {
return new ColumnStats(this.ndv, this.nu... | 3.26 |
flink_ColumnStats_merge_rdh | /**
* Merges two column stats. When the stats are unknown, whatever the other are, we need return
* unknown stats. The unknown definition for column stats is null.
*
* @param other
* The other column stats to merge.
* @return The merged column stats.
*/
public ColumnStats merge(ColumnStats other, boolean isPar... | 3.26 |
flink_ColumnStats_getMaxValue_rdh | /**
* Deprecated because Number type max/min is not well supported comparable type, e.g. {@link java.util.Date}, {@link java.sql.Timestamp}.
*
* <p>Returns null if this instance is constructed by {@link ColumnStats.Builder}.
*/
@Deprecated
public Number getMaxValue() {
return maxValue;
} | 3.26 |
flink_ColumnStats_getMax_rdh | /**
* Returns null if this instance is constructed by {@link ColumnStats#ColumnStats(Long, Long,
* Double, Integer, Number, Number)}.
*/
public Comparable<?> getMax() {return max;
}
/**
* Deprecated because Number type max/min is not well supported comparable type, e.g. {@link java.util.Date}, {@link java.sql.Time... | 3.26 |
flink_ColumnStats_getMin_rdh | /**
* Returns null if this instance is constructed by {@link ColumnStats#ColumnStats(Long, Long,
* Double, Integer, Number, Number)}.
*/
public Comparable<?> getMin() {
return min;
} | 3.26 |
flink_HttpHeader_m0_rdh | /**
* Returns the name of this HTTP header.
*
* @return the name of this HTTP header
*/
public String m0() {
return name;
} | 3.26 |
flink_HttpHeader_getValue_rdh | /**
* Returns the value of this HTTP header.
*
* @return the value of this HTTP header
*/public String getValue() {
return value;
} | 3.26 |
flink_ConnectedStreams_process_rdh | /**
* Applies the given {@link KeyedCoProcessFunction} on the connected input streams, thereby
* creating a transformed output stream.
*
* <p>The function will be called for every element in the input streams and can produce zero or
* more output elements. Contrary to the {@link #flatMap(CoFlatMapFunction)} functi... | 3.26 |
flink_ConnectedStreams_flatMap_rdh | /**
* Applies a CoFlatMap transformation on a {@link ConnectedStreams} and maps the output to a
* common type. The transformation calls a {@link CoFlatMapFunction#flatMap1} for each element
* of the first input and {@link CoFlatMapFunction#flatMap2} for each element of the second
* input. Each CoFlatMapFunction cal... | 3.26 |
flink_ConnectedStreams_getType2_rdh | /**
* Gets the type of the second input.
*
* @return The type of the second input
*/
public TypeInformation<IN2> getType2() {
return inputStream2.getType();
} | 3.26 |
flink_ConnectedStreams_map_rdh | /**
* Applies a CoMap transformation on a {@link ConnectedStreams} and maps the output to a common
* type. The transformation calls a {@link CoMapFunction#map1} for each element of the first
* input and {@link CoMapFunction#map2} for each element of the second input. Each CoMapFunction
* call returns exactly one el... | 3.26 |
flink_ConnectedStreams_keyBy_rdh | /**
* KeyBy operation for connected data stream. Assigns keys to the elements of input1 and input2
* using keySelector1 and keySelector2 with explicit type information for the common key type.
*
* @param keySelector1
* The {@link KeySelector} used for grouping the first input
* @param keySelector2
* The {@li... | 3.26 |
flink_ConnectedStreams_getFirstInput_rdh | /**
* Returns the first {@link DataStream}.
*
* @return The first DataStream.
*/
public DataStream<IN1> getFirstInput() {
return inputStream1;
} | 3.26 |
flink_ConnectedStreams_getType1_rdh | /**
* Gets the type of the first input.
*
* @return The type of the first input
*/
public TypeInformation<IN1> getType1() {
return inputStream1.getType();
} | 3.26 |
flink_SqlGatewayOptionsParser_getSqlGatewayOptions_rdh | // --------------------------------------------------------------------------------------------
private static Options getSqlGatewayOptions() {
Options options = new Options();
options.addOption(OPTION_HELP);
options.addOption(DYNAMIC_PROPERTY_OPTION);return options;
} | 3.26 |
flink_SqlGatewayOptionsParser_parseSqlGatewayOptions_rdh | // --------------------------------------------------------------------------------------------
// Line Parsing
// --------------------------------------------------------------------------------------------
public static SqlGatewayOptions parseSqlGatewayOptions(String[] args) {
try {
DefaultParser parser =... | 3.26 |
flink_SqlGatewayOptionsParser_printHelpSqlGateway_rdh | // --------------------------------------------------------------------------------------------
// Help
// --------------------------------------------------------------------------------------------
/**
* Prints the help for the client.
*/
public static void printHelpSqlGateway(PrintStream writer) {
writer.prin... | 3.26 |
flink_OperationTreeBuilder_addAliasToTheCallInAggregate_rdh | /**
* Add a default name to the call in the grouping expressions, e.g., groupBy(a % 5) to groupBy(a
* % 5 as TMP_0) or make aggregate a named aggregate.
*/
private List<Expressi... | 3.26 |
flink_OperationTreeBuilder_addColumns_rdh | /**
* Adds additional columns. Existing fields will be replaced if replaceIfExist is true.
*/
public QueryOperation addColumns(boolean replaceIfExist, List<Expression> fieldLists, QueryOperation child) {
final List<Expression> newColumns;
if (replaceIfExist) {
final List<String> fieldNames = child.g... | 3.26 |
flink_OperationTreeBuilder_aliasBackwardFields_rdh | /**
* Rename fields in the input {@link QueryOperation}.
*/
private QueryOperation aliasBackwardFields(QueryOperation inputOperation, List<String> alias, int aliasStartIndex) {
if (!alias.isEmpty()) {
List<String> namesBeforeAlias = inputOperation.getResolvedSchema().getColumnNames();
List<Str... | 3.26 |
flink_OperationTreeBuilder_getUniqueName_rdh | /**
* Return a unique name that does not exist in usedFieldNames according to the input name.
*/
private String getUniqueName(String inputName, Collection<String> usedFieldNames) {
int i = 0;
String resultName = inputName;
while (usedFieldNames.contains(resultName)) {resultName = (resultName + "_") + i;
... | 3.26 |
flink_FlinkSemiAntiJoinJoinTransposeRule_setJoinAdjustments_rdh | /**
* Sets an array to reflect how much each index corresponding to a field needs to be adjusted.
* The array corresponds to fields in a 3-way join between (X, Y, and Z). X remains unchanged,
* but Y and Z need to be adjusted by some fixed amount as determined by the input.
*
* @param adjustments
* array to be ... | 3.26 |
flink_ZookeeperModuleFactory_createModule_rdh | /**
* A {@link SecurityModuleFactory} for {@link ZooKeeperModule}.
*/public class ZookeeperModuleFactory implements SecurityModuleFactory {
@Overridepublic SecurityModule createModule(SecurityConfiguration securityConfig) {
return new ZooKeeperModule(securityConfig);
} | 3.26 |
flink_VertexFlameGraphFactory_createFullFlameGraphFrom_rdh | /**
* Converts {@link VertexThreadInfoStats} into a FlameGraph.
*
* @param sample
* Thread details sample containing stack traces.
* @return FlameGraph data structure
*/
public static VertexFlameGraph createFullFlameGraphFrom(VertexThreadInfoStats sample) {
EnumSet<Thread.State> included = EnumSet.allOf(Th... | 3.26 |
flink_VertexFlameGraphFactory_createOnCpuFlameGraph_rdh | /**
* Converts {@link VertexThreadInfoStats} into a FlameGraph representing actively running
* (On-CPU) threads.
*
* <p>Includes threads in states Thread.State.[RUNNABLE, NEW].
*
* @param sample
* Thread details sample containing stack traces.
* @return FlameGraph data structure
*/
public static VertexFlameG... | 3.26 |
flink_VertexFlameGraphFactory_createOffCpuFlameGraph_rdh | /**
* Converts {@link VertexThreadInfoStats} into a FlameGraph representing blocked (Off-CPU)
* threads.
*
* <p>Includes threads in states Thread.State.[TIMED_WAITING, BLOCKED, WAITING].
*
* @param sample
* Thread details sample containing stack traces.
* @return FlameGraph data structure.
*/
public static V... | 3.26 |
flink_VertexFlameGraphFactory_m0_rdh | // Note that Thread.getStackTrace() performs a similar logic - the stack trace returned
// by this method will not contain lambda references with it. But ThreadMXBean does collect
// lambdas, so we have to clean them up explicitly.
private static StackTraceElement[] m0(StackTraceElement[] stackTrace) {
StackTraceEl... | 3.26 |
flink_PartitionLoader_loadPartition_rdh | /**
* Load a single partition.
*
* @param partSpec
* the specification for the single partition
* @param srcPaths
* the paths for the files used to load to the single partition
* @param srcPathIsDir
* whether the every path in {@param srcPaths} is directory or not. If true,
* it will load the files und... | 3.26 |
flink_PartitionLoader_loadNonPartition_rdh | /**
* Load a non-partition files to output path.
*
* @param srcPaths
* the paths for the files used to load to the single partition
* @param srcPathIsDir
* whether the every path in {@param srcPaths} is directory or not. If true,
* it will load the files under the directory of the every path. If false, eve... | 3.26 |
flink_PartitionLoader_moveFiles_rdh | /**
* Moves files from srcDir to destDir.
*/
private void moveFiles(List<Path> srcPaths, Path destDir, boolean srcPathIsDir) throws Exception {
if (srcPathIsDir) {
// if the src path is still a directory, list the directory to get the files that needed
... | 3.26 |
flink_PartitionLoader_loadEmptyPartition_rdh | /**
* The flink job does not write data to the partition, but the corresponding partition needs to
* be created or updated.
*
* <p>The partition does not exist, create it.
*
* <p>The partition exists:
*
* <pre>
* if overwrite is true, delete the path, then create it;
* if overwrite is false, do noth... | 3.26 |
flink_PartitionLoader_commitPartition_rdh | /**
* Reuse of PartitionCommitPolicy mechanisms. The default in Batch mode is metastore and
* success-file.
*/
private void commitPartition(LinkedHashMap<String, String> partitionSpec, Path path) throws Exception {
PartitionCommitPolicy.Context context = new CommitPolicyContextImpl(partitionSpec, path);
fo... | 3.26 |
flink_SourceCoordinator_aggregate_rdh | /**
* Update the {@link Watermark} for the given {@code key)}.
*
* @return the new updated combined {@link Watermark} if the value has changed. {@code Optional.empty()} otherwise.
*/
public Optional<Watermark> aggregate(T key, Watermark watermark) {
Watermark oldAggregatedWatermark = getAggregatedWatermark();
Water... | 3.26 |
flink_SourceCoordinator_deserializeCheckpoint_rdh | /**
* Restore the state of this source coordinator from the state bytes.
*
* @param bytes
* The checkpoint bytes that was returned from {@link #toBytes(long)}
* @throws Exception
* When the deserialization failed.
*/
private EnumChkT
deserializeCheckpoint(byte[] bytes) throws Exception {
try (ByteArrayInputS... | 3.26 |
flink_SourceCoordinator_handleRequestSplitEvent_rdh | // --------------------- private methods -------------
private void handleRequestSplitEvent(int subtask, int attemptNumber, RequestSplitEvent event) {
LOG.info("Source {} received split request from parallel task {} (#{})", operatorName, subtask, attemptNumber);
// request splits from the enumerator only if the enumer... | 3.26 |
flink_SourceCoordinator_toBytes_rdh | // --------------------- Serde -----------------------
/**
* Serialize the coordinator state. The current implementation may not be super efficient, but
* it should not matter that much because most of the state should be rather small. Large states
* themselves may already be a problem regardless of how the serializ... | 3.26 |
flink_SortingThread_go_rdh | /**
* Entry point of the thread.
*/
@Override
public void go() throws InterruptedException {
boolean alive = true;
// loop as long as the thread is marked alive
while (isRunning() && alive) {
final CircularElement<E> element = this.dispatcher.take(SortStage.SORT... | 3.26 |
flink_FlinkConvertletTable_convertTryCast_rdh | // Slightly modified version of StandardConvertletTable::convertCast
private RexNode convertTryCast(SqlRexContext cx, final SqlCall call) {
RelDataTypeFactory typeFactory = cx.getTypeFactory();
final SqlNode leftNode = call.operand(0);
final SqlNode rightNode = call.operand(1);
final RexNode valueRex... | 3.26 |
flink_TemporalTableJoinUtil_isRowTimeTemporalTableJoinCondition_rdh | /**
* Check if the given rexCall is a rewrote join condition on event time.
*/
public static boolean isRowTimeTemporalTableJoinCondition(RexCall call) {
// (LEFT_TIME_ATTRIBUTE, RIGHT_TIME_ATTRIBUTE, LEFT_KEY, RIGHT_KEY, PRIMARY_KEY)
return (call.getOperator() == TemporalJoinUtil.TEMPORAL_JOIN_CONDITION())... | 3.26 |
flink_TemporalTableJoinUtil_isEventTimeTemporalJoin_rdh | /**
* Check if the given join condition is an initial temporal join condition or a rewrote join
* condition on event time.
*/
public static boolean isEventTimeTemporalJoin(@Nonnull
RexNode joinCondition) {
RexVisitor<Void> temporalConditionFinder = new RexVisitorImpl<Void>(true) {
... | 3.26 |
flink_ViewUpdater_notifyOfAddedView_rdh | /**
* Notifies this ViewUpdater of a new metric that should be regularly updated.
*
* @param view
* metric that should be regularly updated
*/
public void notifyOfAddedView(View view) {
synchronized(lock) {
toAdd.add(view);
}
} | 3.26 |
flink_ViewUpdater_notifyOfRemovedView_rdh | /**
* Notifies this ViewUpdater of a metric that should no longer be regularly updated.
*
* @param view
* metric that should no longer be regularly updated
*/
public void notifyOfRemovedView(View view) {
synchronized(lock) {
toRemove.add(view);
}
} | 3.26 |
flink_FeedbackTransformation_getFeedbackEdges_rdh | /**
* Returns the list of feedback {@code Transformations}.
*/
public List<Transformation<T>> getFeedbackEdges() {
return feedbackEdges;
} | 3.26 |
flink_FeedbackTransformation_addFeedbackEdge_rdh | /**
* Adds a feedback edge. The parallelism of the {@code Transformation} must match the
* parallelism of the input {@code Transformation} of this {@code FeedbackTransformation}
*
* @param transform
* The new feedback {@code Transformation}.
*/
public void
addFeedbackEdge(Transformation<T> transform) {
if (... | 3.26 |
flink_FeedbackTransformation_getWaitTime_rdh | /**
* Returns the wait time. This is the amount of time that the feedback operator keeps listening
* for feedback elements. Once the time expires the operation will close and will not receive
* further elements.
*/
public Long getWaitTime() {
return waitTime;
} | 3.26 |
flink_SerializedThrowable_printStackTrace_rdh | // ------------------------------------------------------------------------
// Override the behavior of Throwable
// ------------------------------------------------------------------------
@Override
public void printStackTrace(PrintStream s) {
s.print(fullStringifiedStackTrace);
s.flush();
} | 3.26 |
flink_SerializedThrowable_get_rdh | // ------------------------------------------------------------------------
// Static utilities
// ------------------------------------------------------------------------
public static Throwable get(Throwable serThrowable, ClassLoader loader) {
if (serThrowable instanceof SerializedThrowable) {
return ((S... | 3.26 |
flink_RegisteredOperatorStateBackendMetaInfo_deepCopy_rdh | /**
* Creates a deep copy of the itself.
*/
@Nonnull
public RegisteredOperatorStateBackendMetaInfo<S> deepCopy() {
return new RegisteredOperatorStateBackendMetaInfo<>(this);
} | 3.26 |
flink_SpillChannelManager_m0_rdh | /**
* Removes a channel reader/writer from the list of channels that are to be removed at shutdown.
*
* @param channel
* The channel reader/writer.
*/
synchronized void m0(FileIOChannel channel) {
openChannels.remove(channel);
} | 3.26 |
flink_SpillChannelManager_registerChannelToBeRemovedAtShutdown_rdh | /**
* Adds a channel to the list of channels that are to be removed at shutdown.
*
* @param channel
* The channel id.
*/
synchronized void registerChannelToBeRemovedAtShutdown(FileIOChannel.ID channel) {
channelsToDeleteAtShutdown.add(channel);
} | 3.26 |
flink_SpillChannelManager_registerOpenChannelToBeRemovedAtShutdown_rdh | /**
* Adds a channel reader/writer to the list of channels that are to be removed at shutdown.
*
* @param channel
* The channel reader/writer.
*/
synchronized void registerOpenChannelToBeRemovedAtShutdown(FileIOChannel channel) {openChannels.add(channel);
} | 3.26 |
flink_SpillChannelManager_unregisterChannelToBeRemovedAtShutdown_rdh | /**
* Removes a channel from the list of channels that are to be removed at shutdown.
*
* @param channel
* The channel id.
*/
synchronized void unregisterChannelToBeRemovedAtShutdown(FileIOChannel.ID channel) {
channelsToDeleteAtShutdown.remove(channel);
} | 3.26 |
flink_TieredStorageIdMappingUtils_convertId_rdh | /**
* Utils to convert the Ids to Tiered Storage Ids, or vice versa.
*/
| 3.26 |
flink_CepRuntimeContext_addAccumulator_rdh | // -----------------------------------------------------------------------------------
// Unsupported operations
// -----------------------------------------------------------------------------------
@Override
public <V, A extends Serializable> void addAccumulator(final String name, final Accumulator<V, A> accumulator)... | 3.26 |
flink_JoinSpec_getJoinKeySize_rdh | /**
* Gets number of keys in join key.
*/
@JsonIgnore
public int getJoinKeySize() {
return
leftKeys.length;
} | 3.26 |
flink_HivePartitionUtils_getAllPartitions_rdh | /**
* Returns all HiveTablePartitions of a hive table, returns single HiveTablePartition if the
* hive table is not partitioned.
*/
public static List<HiveTablePartition> getAllPartitions(JobConf jobConf, String hiveVersion, ObjectPath tablePath, List<String> partitionColNames, List<Map<String, String>> remainin... | 3.26 |
flink_HivePartitionUtils_createPartitionSpec_rdh | /**
* Creates a {@link CatalogPartitionSpec} from a Hive partition name string. Example of Hive
* partition name string - "name=bob/year=2019". If the partition name for the given partition
* column is equal to {@param defaultPartitionName}, the partition value in returned {@link CatalogPartitionSpec} will be null.
... | 3.26 |
flink_HivePartitionUtils_parsePartitionValues_rdh | /**
* Parse partition string specs into object values.
*/
public static Map<String, Object> parsePartitionValues(Map<String, String> partitionSpecs, String[] fieldNames, DataType[] fieldTypes, String defaultPartitionName, HiveShim shim) {
checkArgument(fieldNames.length == fieldTypes.length);
List<String> fi... | 3.26 |
flink_Tuple1_toString_rdh | // -------------------------------------------------------------------------------------------------
// standard utilities
// -------------------------------------------------------------------------------------------------
/**
* Creates a string representation of the tuple in the form (f0), where the individual field... | 3.26 |
flink_Tuple1_setFields_rdh | /**
* Sets new values to all fields of the tuple.
*
* @param f0
* The value for field 0
*/ public void setFields(T0 f0) {
this.f0 = f0;
} | 3.26 |
flink_Tuple1_equals_rdh | /**
* Deep equality for tuples by calling equals() on the tuple members.
*
* @param o
* the object checked for equality
* @return true if this is equal to o.
*/
@Override
public boolean equals(Object o) {
if (this == o) {
return true;
}
if (!(o instanceof Tuple1)) {
return false;
... | 3.26 |
flink_HiveTableSink_toStagingDir_rdh | // get a staging dir
private String toStagingDir(String stagingParentDir, Configuration conf) throws IOException {
if (!stagingParentDir.endsWith(Path.SEPARATOR)) {
stagingParentDir += Path.SEPARATOR;
}
// TODO: may append something more meaningful than a timestamp, like query ID
stagingParentDi... | 3.26 |
flink_JsonRowSchemaConverter_convert_rdh | /**
* Converts a JSON schema into Flink's type information. Throws an exception if the schema
* cannot converted because of loss of precision or too flexible schema.
*
* <p>The converter can resolve simple schema references to solve those cases where entities are
* defined at the beginning and then used throughout... | 3.26 |
flink_MaterializedCollectStreamResult_processRecord_rdh | // --------------------------------------------------------------------------------------------
@Override
protected void processRecord(RowData row) {synchronized(resultLock) {
boolean isInsertOp = (row.getRowKind() == RowKind.INSERT) || (row.getRowKind() == RowKind.UPDATE_AFTER);
// Always set the Row... | 3.26 |
flink_MaterializedCollectStreamResult_processInsert_rdh | // --------------------------------------------------------------------------------------------
private void processInsert(RowData row) {
// limit the materialized table
if
((materializedTable.size() - validRowPosition) >= maxRowCount) {
cleanUp();
}materializedTable.add(row);
f0.put(row, ma... | 3.26 |
flink_CustomHeadersDecorator_getCustomHeaders_rdh | /**
* Returns the custom headers added to the message.
*
* @return The custom headers as a collection of {@link HttpHeader}.
*/
@Override
public Collection<HttpHeader> getCustomHeaders() {
return customHeaders;
} | 3.26 |
flink_CustomHeadersDecorator_addCustomHeader_rdh | /**
* Adds a custom header to the message. Initializes the custom headers collection if it hasn't
* been initialized yet.
*
* @param httpHeader
* The header to add.
*/
public void addCustomHeader(HttpHeader httpHeader) {
if (customHeaders == null) {
customHeaders = new ArrayList<>();
}
cus... | 3.26 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.