name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
flink_PythonOperatorChainingOptimizer_buildOutputMap_rdh
/** * Construct the key-value pairs where the value is the output transformations of the key * transformation. */ private static Map<Transformation<?>, Set<Transformation<?>>> buildOutputMap(List<Transformation<?>> transformations) { final Map<Transformation<?>, Set<Transformation<?>>> outputMap = new HashMap<>(...
3.26
flink_PythonOperatorChainingOptimizer_replaceInput_rdh
// ----------------------- Utility Methods ----------------------- private static void replaceInput(Transformation<?> transformation, Transformation<?> oldInput, Transformation<?> newInput) { try { if (((((((transformation instanceof OneInputTransformation) || (transformation instanceof FeedbackTransformati...
3.26
flink_PythonOperatorChainingOptimizer_optimize_rdh
/** * Perform chaining optimization. It will returns the chained transformations and the * transformation after chaining optimization for the given transformation. */ public static Tuple2<List<Transformation<?>>, Transformation<?>> optimize(List<Transformation<?>> transformations, Transformation<...
3.26
flink_HybridShuffleConfiguration_getFullStrategyReleaseBufferRatio_rdh
/** * The proportion of buffers to be released. Used by {@link HsFullSpillingStrategy}. */ public float getFullStrategyReleaseBufferRatio() { return fullStrategyReleaseBufferRatio; }
3.26
flink_HybridShuffleConfiguration_getSpillingStrategyType_rdh
/** * Get {@link SpillingStrategyType} for hybrid shuffle mode. */ public SpillingStrategyType getSpillingStrategyType() { return spillingStrategyType; }
3.26
flink_HybridShuffleConfiguration_getSelectiveStrategySpillBufferRatio_rdh
/** * The proportion of buffers to be spilled. Used by {@link HsSelectiveSpillingStrategy}. */ public float getSelectiveStrategySpillBufferRatio() { return selectiveStrategySpillBufferRatio; }
3.26
flink_HybridShuffleConfiguration_getMaxBuffersReadAhead_rdh
/** * Determine how many buffers to read ahead at most for each subpartition to prevent other * consumers from starving. */ public int getMaxBuffersReadAhead() { return maxBuffersReadAhead; }
3.26
flink_HybridShuffleConfiguration_getRegionGroupSizeInBytes_rdh
/** * Segment size of hybrid spilled file data index. */ public int getRegionGroupSizeInBytes() { return regionGroupSizeInBytes; }
3.26
flink_HybridShuffleConfiguration_getBufferRequestTimeout_rdh
/** * Maximum time to wait when requesting read buffers from the buffer pool before throwing an * exception. */ public Duration getBufferRequestTimeout() { return bufferRequestTimeout;}
3.26
flink_HybridShuffleConfiguration_getFullStrategyReleaseThreshold_rdh
/** * When the number of buffers that have been requested exceeds this threshold, trigger the * release operation. Used by {@link HsFullSpillingStrategy}. */ public float getFullStrategyReleaseThreshold() { return fullStrategyReleaseThreshold; }
3.26
flink_HybridShuffleConfiguration_getBufferPoolSizeCheckIntervalMs_rdh
/** * Check interval of buffer pool's size. */ public long getBufferPoolSizeCheckIntervalMs() { return bufferPoolSizeCheckIntervalMs; }
3.26
flink_HybridShuffleConfiguration_getNumRetainedInMemoryRegionsMax_rdh
/** * Max number of hybrid retained regions in memory. */ public long getNumRetainedInMemoryRegionsMax() { return numRetainedInMemoryRegionsMax; }
3.26
flink_FileSource_forRecordStreamFormat_rdh
// ------------------------------------------------------------------------ // Entry-point Factory Methods // ------------------------------------------------------------------------ /** * Builds a new {@code FileSource} using a {@link StreamFormat} to read record-by-record from a * file stream. * * <p>When possibl...
3.26
flink_FileSource_forBulkFileFormat_rdh
/** * Builds a new {@code FileSource} using a {@link BulkFormat} to read batches of records from * files. * * <p>Examples for bulk readers are compressed and vectorized formats such as ORC or Parquet. */ public static <T> FileSourceBuilder<T> forBulkFileFormat(final BulkFormat<T, FileSourceSplit> bulkFormat, final...
3.26
flink_FileSource_forRecordFileFormat_rdh
/** * Builds a new {@code FileSource} using a {@link FileRecordFormat} to read record-by-record * from a a file path. * * <p>A {@code FileRecordFormat} is more general than the {@link StreamFormat}, but also * requires often more careful parametrization. * * @deprecated Please use {@link #forRecordStreamFormat(S...
3.26
flink_SqlJsonValueFunctionWrapper_explicitTypeSpec_rdh
/** * Copied and modified from the original {@link SqlJsonValueFunction}. * * <p>Changes: Instead of returning {@link Optional} this method returns null directly. */ private static RelDataType explicitTypeSpec(SqlOperatorBinding opBinding) { if (((opBinding.getOperandCount() > 2) && opBinding.isOperandLiteral(2...
3.26
flink_StateWithExecutionGraph_updateTaskExecutionState_rdh
/** * Updates the execution graph with the given task execution state transition. * * @param taskExecutionStateTransition * taskExecutionStateTransition to update the ExecutionGraph * with * @param failureLabels * the failure labels to attach to the task failure cause * @return {@code true} if the update ...
3.26
flink_AllGroupReduceDriver_setup_rdh
// ------------------------------------------------------------------------ @Override public void setup(TaskContext<GroupReduceFunction<IT, OT>, OT> context) {this.taskContext = context; }
3.26
flink_AllGroupReduceDriver_prepare_rdh
// -------------------------------------------------------------------------------------------- @Override public void prepare() throws Exception { final TaskConfig config = this.taskContext.getTaskConfig();this.strategy = config.getDriverStrategy(); switch (this.strategy)...
3.26
flink_S3Recoverable_m0_rdh
// ------------------------------------------------------------------------ public String m0() { return uploadId; }
3.26
flink_LeaderRetrievalUtils_retrieveLeaderInformation_rdh
/** * Retrieves the leader pekko url and the current leader session ID. The values are stored in a * {@link LeaderInformation} instance. * * @param leaderRetrievalService * Leader retrieval service to retrieve the leader connection * information * @param timeout * Timeout when to give up looking for the l...
3.26
flink_RecordAndPosition_getRecord_rdh
// ------------------------------------------------------------------------ public E getRecord() { return record; }
3.26
flink_RecordAndPosition_toString_rdh
// ------------------------------------------------------------------------ @Override public String toString() { return String.format("%s @ %d + %d", record, offset, recordSkipCount); }
3.26
flink_HadoopOutputFormatBase_writeObject_rdh
// -------------------------------------------------------------------------------------------- // Custom serialization methods // -------------------------------------------------------------------------------------------- private void writeObject(ObjectOutputStream out) throws IOException { super.write(out); ...
3.26
flink_HadoopOutputFormatBase_open_rdh
/** * create the temporary output file for hadoop RecordWriter. * * @param taskNumber * The number of the parallel instance. * @param numTasks * The number of parallel tasks. * @throws java.io.IOException */ @Overridepublic void open(int taskNumber, int numTasks) throws IOException { // enforce sequenti...
3.26
flink_HadoopOutputFormatBase_configure_rdh
// -------------------------------------------------------------------------------------------- // OutputFormat // -------------------------------------------------------------------------------------------- @Override public void configure(Configuration parameters) { // enforce sequential configure() calls sync...
3.26
flink_HadoopOutputFormatBase_close_rdh
/** * commit the task by moving the output file out from the temporary directory. * * @throws java.io.IOException */ @Override public void close() throws IOException { // enforce sequential close() calls synchronized(CLOSE_MUTEX) { this.recordWriter.close(new HadoopDummyReporter()); if ...
3.26
flink_SqlJsonUtils_createArrayNode_rdh
/** * Returns a new {@link ArrayNode}. */ public static ArrayNode createArrayNode() { return MAPPER.createArrayNode();}
3.26
flink_SqlJsonUtils_getNodeFactory_rdh
/** * Returns the {@link JsonNodeFactory} for creating nodes. */ public static JsonNodeFactory getNodeFactory() { return MAPPER.getNodeFactory(); } /** * Returns a new {@link ObjectNode}
3.26
flink_SqlJsonUtils_serializeJson_rdh
/** * Serializes the given {@link JsonNode} to a JSON string. */ public static String serializeJson(JsonNode node) { try { // For JSON functions to have deterministic output, we need to sort the keys. However, // Jackson's built-in features don't work on the tree representation, so we need to ...
3.26
flink_StateMap_releaseSnapshot_rdh
/** * Releases a snapshot for this {@link StateMap}. This method should be called once a snapshot * is no more needed. * * @param snapshotToRelease * the snapshot to release, which was previously created by this state * map. */ public void releaseSnapshot(StateMapSnapshot<K, N, S, ? extends StateMap<K, N, S>...
3.26
flink_SerializableHadoopConfigWrapper_writeObject_rdh
// ------------------------------------------------------------------------ private void writeObject(ObjectOutputStream out) throws IOException { out.defaultWriteObject(); // we write the Hadoop config through a separate serializer to avoid cryptic exceptions when // it // corrupts the s...
3.26
flink_IterationTailTask_initialize_rdh
// -------------------------------------------------------------------------------------------- @Override protected void initialize() throws Exception { super.initialize(); // sanity check: the tail has to update either the workset or the solution set if ((!isWorksetUpdate) && (!isSolutionSetUpdate)) { ...
3.26
flink_ColumnStats_copy_rdh
/** * Create a deep copy of "this" instance. * * @return a deep copy */ public ColumnStats copy() { if ((maxValue != null) || (f0 != null)) { return new ColumnStats(this.ndv, this.nullCount, this.avgLen, this.maxLen, this.maxValue, this.f0); } else { return new ColumnStats(this.ndv, this.nu...
3.26
flink_ColumnStats_merge_rdh
/** * Merges two column stats. When the stats are unknown, whatever the other are, we need return * unknown stats. The unknown definition for column stats is null. * * @param other * The other column stats to merge. * @return The merged column stats. */ public ColumnStats merge(ColumnStats other, boolean isPar...
3.26
flink_ColumnStats_getMaxValue_rdh
/** * Deprecated because Number type max/min is not well supported comparable type, e.g. {@link java.util.Date}, {@link java.sql.Timestamp}. * * <p>Returns null if this instance is constructed by {@link ColumnStats.Builder}. */ @Deprecated public Number getMaxValue() { return maxValue; }
3.26
flink_ColumnStats_getMax_rdh
/** * Returns null if this instance is constructed by {@link ColumnStats#ColumnStats(Long, Long, * Double, Integer, Number, Number)}. */ public Comparable<?> getMax() {return max; } /** * Deprecated because Number type max/min is not well supported comparable type, e.g. {@link java.util.Date}, {@link java.sql.Time...
3.26
flink_ColumnStats_getMin_rdh
/** * Returns null if this instance is constructed by {@link ColumnStats#ColumnStats(Long, Long, * Double, Integer, Number, Number)}. */ public Comparable<?> getMin() { return min; }
3.26
flink_HttpHeader_m0_rdh
/** * Returns the name of this HTTP header. * * @return the name of this HTTP header */ public String m0() { return name; }
3.26
flink_HttpHeader_getValue_rdh
/** * Returns the value of this HTTP header. * * @return the value of this HTTP header */public String getValue() { return value; }
3.26
flink_ConnectedStreams_process_rdh
/** * Applies the given {@link KeyedCoProcessFunction} on the connected input streams, thereby * creating a transformed output stream. * * <p>The function will be called for every element in the input streams and can produce zero or * more output elements. Contrary to the {@link #flatMap(CoFlatMapFunction)} functi...
3.26
flink_ConnectedStreams_flatMap_rdh
/** * Applies a CoFlatMap transformation on a {@link ConnectedStreams} and maps the output to a * common type. The transformation calls a {@link CoFlatMapFunction#flatMap1} for each element * of the first input and {@link CoFlatMapFunction#flatMap2} for each element of the second * input. Each CoFlatMapFunction cal...
3.26
flink_ConnectedStreams_getType2_rdh
/** * Gets the type of the second input. * * @return The type of the second input */ public TypeInformation<IN2> getType2() { return inputStream2.getType(); }
3.26
flink_ConnectedStreams_map_rdh
/** * Applies a CoMap transformation on a {@link ConnectedStreams} and maps the output to a common * type. The transformation calls a {@link CoMapFunction#map1} for each element of the first * input and {@link CoMapFunction#map2} for each element of the second input. Each CoMapFunction * call returns exactly one el...
3.26
flink_ConnectedStreams_keyBy_rdh
/** * KeyBy operation for connected data stream. Assigns keys to the elements of input1 and input2 * using keySelector1 and keySelector2 with explicit type information for the common key type. * * @param keySelector1 * The {@link KeySelector} used for grouping the first input * @param keySelector2 * The {@li...
3.26
flink_ConnectedStreams_getFirstInput_rdh
/** * Returns the first {@link DataStream}. * * @return The first DataStream. */ public DataStream<IN1> getFirstInput() { return inputStream1; }
3.26
flink_ConnectedStreams_getType1_rdh
/** * Gets the type of the first input. * * @return The type of the first input */ public TypeInformation<IN1> getType1() { return inputStream1.getType(); }
3.26
flink_SqlGatewayOptionsParser_getSqlGatewayOptions_rdh
// -------------------------------------------------------------------------------------------- private static Options getSqlGatewayOptions() { Options options = new Options(); options.addOption(OPTION_HELP); options.addOption(DYNAMIC_PROPERTY_OPTION);return options; }
3.26
flink_SqlGatewayOptionsParser_parseSqlGatewayOptions_rdh
// -------------------------------------------------------------------------------------------- // Line Parsing // -------------------------------------------------------------------------------------------- public static SqlGatewayOptions parseSqlGatewayOptions(String[] args) { try { DefaultParser parser =...
3.26
flink_SqlGatewayOptionsParser_printHelpSqlGateway_rdh
// -------------------------------------------------------------------------------------------- // Help // -------------------------------------------------------------------------------------------- /** * Prints the help for the client. */ public static void printHelpSqlGateway(PrintStream writer) { writer.prin...
3.26
flink_OperationTreeBuilder_addAliasToTheCallInAggregate_rdh
/** * Add a default name to the call in the grouping expressions, e.g., groupBy(a % 5) to groupBy(a * % 5 as TMP_0) or make aggregate a named aggregate. */ private List<Expressi...
3.26
flink_OperationTreeBuilder_addColumns_rdh
/** * Adds additional columns. Existing fields will be replaced if replaceIfExist is true. */ public QueryOperation addColumns(boolean replaceIfExist, List<Expression> fieldLists, QueryOperation child) { final List<Expression> newColumns; if (replaceIfExist) { final List<String> fieldNames = child.g...
3.26
flink_OperationTreeBuilder_aliasBackwardFields_rdh
/** * Rename fields in the input {@link QueryOperation}. */ private QueryOperation aliasBackwardFields(QueryOperation inputOperation, List<String> alias, int aliasStartIndex) { if (!alias.isEmpty()) { List<String> namesBeforeAlias = inputOperation.getResolvedSchema().getColumnNames(); List<Str...
3.26
flink_OperationTreeBuilder_getUniqueName_rdh
/** * Return a unique name that does not exist in usedFieldNames according to the input name. */ private String getUniqueName(String inputName, Collection<String> usedFieldNames) { int i = 0; String resultName = inputName; while (usedFieldNames.contains(resultName)) {resultName = (resultName + "_") + i; ...
3.26
flink_FlinkSemiAntiJoinJoinTransposeRule_setJoinAdjustments_rdh
/** * Sets an array to reflect how much each index corresponding to a field needs to be adjusted. * The array corresponds to fields in a 3-way join between (X, Y, and Z). X remains unchanged, * but Y and Z need to be adjusted by some fixed amount as determined by the input. * * @param adjustments * array to be ...
3.26
flink_ZookeeperModuleFactory_createModule_rdh
/** * A {@link SecurityModuleFactory} for {@link ZooKeeperModule}. */public class ZookeeperModuleFactory implements SecurityModuleFactory { @Overridepublic SecurityModule createModule(SecurityConfiguration securityConfig) { return new ZooKeeperModule(securityConfig); }
3.26
flink_VertexFlameGraphFactory_createFullFlameGraphFrom_rdh
/** * Converts {@link VertexThreadInfoStats} into a FlameGraph. * * @param sample * Thread details sample containing stack traces. * @return FlameGraph data structure */ public static VertexFlameGraph createFullFlameGraphFrom(VertexThreadInfoStats sample) { EnumSet<Thread.State> included = EnumSet.allOf(Th...
3.26
flink_VertexFlameGraphFactory_createOnCpuFlameGraph_rdh
/** * Converts {@link VertexThreadInfoStats} into a FlameGraph representing actively running * (On-CPU) threads. * * <p>Includes threads in states Thread.State.[RUNNABLE, NEW]. * * @param sample * Thread details sample containing stack traces. * @return FlameGraph data structure */ public static VertexFlameG...
3.26
flink_VertexFlameGraphFactory_createOffCpuFlameGraph_rdh
/** * Converts {@link VertexThreadInfoStats} into a FlameGraph representing blocked (Off-CPU) * threads. * * <p>Includes threads in states Thread.State.[TIMED_WAITING, BLOCKED, WAITING]. * * @param sample * Thread details sample containing stack traces. * @return FlameGraph data structure. */ public static V...
3.26
flink_VertexFlameGraphFactory_m0_rdh
// Note that Thread.getStackTrace() performs a similar logic - the stack trace returned // by this method will not contain lambda references with it. But ThreadMXBean does collect // lambdas, so we have to clean them up explicitly. private static StackTraceElement[] m0(StackTraceElement[] stackTrace) { StackTraceEl...
3.26
flink_PartitionLoader_loadPartition_rdh
/** * Load a single partition. * * @param partSpec * the specification for the single partition * @param srcPaths * the paths for the files used to load to the single partition * @param srcPathIsDir * whether the every path in {@param srcPaths} is directory or not. If true, * it will load the files und...
3.26
flink_PartitionLoader_loadNonPartition_rdh
/** * Load a non-partition files to output path. * * @param srcPaths * the paths for the files used to load to the single partition * @param srcPathIsDir * whether the every path in {@param srcPaths} is directory or not. If true, * it will load the files under the directory of the every path. If false, eve...
3.26
flink_PartitionLoader_moveFiles_rdh
/** * Moves files from srcDir to destDir. */ private void moveFiles(List<Path> srcPaths, Path destDir, boolean srcPathIsDir) throws Exception { if (srcPathIsDir) { // if the src path is still a directory, list the directory to get the files that needed ...
3.26
flink_PartitionLoader_loadEmptyPartition_rdh
/** * The flink job does not write data to the partition, but the corresponding partition needs to * be created or updated. * * <p>The partition does not exist, create it. * * <p>The partition exists: * * <pre> * if overwrite is true, delete the path, then create it; * if overwrite is false, do noth...
3.26
flink_PartitionLoader_commitPartition_rdh
/** * Reuse of PartitionCommitPolicy mechanisms. The default in Batch mode is metastore and * success-file. */ private void commitPartition(LinkedHashMap<String, String> partitionSpec, Path path) throws Exception { PartitionCommitPolicy.Context context = new CommitPolicyContextImpl(partitionSpec, path); fo...
3.26
flink_SourceCoordinator_aggregate_rdh
/** * Update the {@link Watermark} for the given {@code key)}. * * @return the new updated combined {@link Watermark} if the value has changed. {@code Optional.empty()} otherwise. */ public Optional<Watermark> aggregate(T key, Watermark watermark) { Watermark oldAggregatedWatermark = getAggregatedWatermark(); Water...
3.26
flink_SourceCoordinator_deserializeCheckpoint_rdh
/** * Restore the state of this source coordinator from the state bytes. * * @param bytes * The checkpoint bytes that was returned from {@link #toBytes(long)} * @throws Exception * When the deserialization failed. */ private EnumChkT deserializeCheckpoint(byte[] bytes) throws Exception { try (ByteArrayInputS...
3.26
flink_SourceCoordinator_handleRequestSplitEvent_rdh
// --------------------- private methods ------------- private void handleRequestSplitEvent(int subtask, int attemptNumber, RequestSplitEvent event) { LOG.info("Source {} received split request from parallel task {} (#{})", operatorName, subtask, attemptNumber); // request splits from the enumerator only if the enumer...
3.26
flink_SourceCoordinator_toBytes_rdh
// --------------------- Serde ----------------------- /** * Serialize the coordinator state. The current implementation may not be super efficient, but * it should not matter that much because most of the state should be rather small. Large states * themselves may already be a problem regardless of how the serializ...
3.26
flink_SortingThread_go_rdh
/** * Entry point of the thread. */ @Override public void go() throws InterruptedException { boolean alive = true; // loop as long as the thread is marked alive while (isRunning() && alive) { final CircularElement<E> element = this.dispatcher.take(SortStage.SORT...
3.26
flink_FlinkConvertletTable_convertTryCast_rdh
// Slightly modified version of StandardConvertletTable::convertCast private RexNode convertTryCast(SqlRexContext cx, final SqlCall call) { RelDataTypeFactory typeFactory = cx.getTypeFactory(); final SqlNode leftNode = call.operand(0); final SqlNode rightNode = call.operand(1); final RexNode valueRex...
3.26
flink_TemporalTableJoinUtil_isRowTimeTemporalTableJoinCondition_rdh
/** * Check if the given rexCall is a rewrote join condition on event time. */ public static boolean isRowTimeTemporalTableJoinCondition(RexCall call) { // (LEFT_TIME_ATTRIBUTE, RIGHT_TIME_ATTRIBUTE, LEFT_KEY, RIGHT_KEY, PRIMARY_KEY) return (call.getOperator() == TemporalJoinUtil.TEMPORAL_JOIN_CONDITION())...
3.26
flink_TemporalTableJoinUtil_isEventTimeTemporalJoin_rdh
/** * Check if the given join condition is an initial temporal join condition or a rewrote join * condition on event time. */ public static boolean isEventTimeTemporalJoin(@Nonnull RexNode joinCondition) { RexVisitor<Void> temporalConditionFinder = new RexVisitorImpl<Void>(true) { ...
3.26
flink_ViewUpdater_notifyOfAddedView_rdh
/** * Notifies this ViewUpdater of a new metric that should be regularly updated. * * @param view * metric that should be regularly updated */ public void notifyOfAddedView(View view) { synchronized(lock) { toAdd.add(view); } }
3.26
flink_ViewUpdater_notifyOfRemovedView_rdh
/** * Notifies this ViewUpdater of a metric that should no longer be regularly updated. * * @param view * metric that should no longer be regularly updated */ public void notifyOfRemovedView(View view) { synchronized(lock) { toRemove.add(view); } }
3.26
flink_FeedbackTransformation_getFeedbackEdges_rdh
/** * Returns the list of feedback {@code Transformations}. */ public List<Transformation<T>> getFeedbackEdges() { return feedbackEdges; }
3.26
flink_FeedbackTransformation_addFeedbackEdge_rdh
/** * Adds a feedback edge. The parallelism of the {@code Transformation} must match the * parallelism of the input {@code Transformation} of this {@code FeedbackTransformation} * * @param transform * The new feedback {@code Transformation}. */ public void addFeedbackEdge(Transformation<T> transform) { if (...
3.26
flink_FeedbackTransformation_getWaitTime_rdh
/** * Returns the wait time. This is the amount of time that the feedback operator keeps listening * for feedback elements. Once the time expires the operation will close and will not receive * further elements. */ public Long getWaitTime() { return waitTime; }
3.26
flink_SerializedThrowable_printStackTrace_rdh
// ------------------------------------------------------------------------ // Override the behavior of Throwable // ------------------------------------------------------------------------ @Override public void printStackTrace(PrintStream s) { s.print(fullStringifiedStackTrace); s.flush(); }
3.26
flink_SerializedThrowable_get_rdh
// ------------------------------------------------------------------------ // Static utilities // ------------------------------------------------------------------------ public static Throwable get(Throwable serThrowable, ClassLoader loader) { if (serThrowable instanceof SerializedThrowable) { return ((S...
3.26
flink_RegisteredOperatorStateBackendMetaInfo_deepCopy_rdh
/** * Creates a deep copy of the itself. */ @Nonnull public RegisteredOperatorStateBackendMetaInfo<S> deepCopy() { return new RegisteredOperatorStateBackendMetaInfo<>(this); }
3.26
flink_SpillChannelManager_m0_rdh
/** * Removes a channel reader/writer from the list of channels that are to be removed at shutdown. * * @param channel * The channel reader/writer. */ synchronized void m0(FileIOChannel channel) { openChannels.remove(channel); }
3.26
flink_SpillChannelManager_registerChannelToBeRemovedAtShutdown_rdh
/** * Adds a channel to the list of channels that are to be removed at shutdown. * * @param channel * The channel id. */ synchronized void registerChannelToBeRemovedAtShutdown(FileIOChannel.ID channel) { channelsToDeleteAtShutdown.add(channel); }
3.26
flink_SpillChannelManager_registerOpenChannelToBeRemovedAtShutdown_rdh
/** * Adds a channel reader/writer to the list of channels that are to be removed at shutdown. * * @param channel * The channel reader/writer. */ synchronized void registerOpenChannelToBeRemovedAtShutdown(FileIOChannel channel) {openChannels.add(channel); }
3.26
flink_SpillChannelManager_unregisterChannelToBeRemovedAtShutdown_rdh
/** * Removes a channel from the list of channels that are to be removed at shutdown. * * @param channel * The channel id. */ synchronized void unregisterChannelToBeRemovedAtShutdown(FileIOChannel.ID channel) { channelsToDeleteAtShutdown.remove(channel); }
3.26
flink_TieredStorageIdMappingUtils_convertId_rdh
/** * Utils to convert the Ids to Tiered Storage Ids, or vice versa. */
3.26
flink_CepRuntimeContext_addAccumulator_rdh
// ----------------------------------------------------------------------------------- // Unsupported operations // ----------------------------------------------------------------------------------- @Override public <V, A extends Serializable> void addAccumulator(final String name, final Accumulator<V, A> accumulator)...
3.26
flink_JoinSpec_getJoinKeySize_rdh
/** * Gets number of keys in join key. */ @JsonIgnore public int getJoinKeySize() { return leftKeys.length; }
3.26
flink_HivePartitionUtils_getAllPartitions_rdh
/** * Returns all HiveTablePartitions of a hive table, returns single HiveTablePartition if the * hive table is not partitioned. */ public static List<HiveTablePartition> getAllPartitions(JobConf jobConf, String hiveVersion, ObjectPath tablePath, List<String> partitionColNames, List<Map<String, String>> remainin...
3.26
flink_HivePartitionUtils_createPartitionSpec_rdh
/** * Creates a {@link CatalogPartitionSpec} from a Hive partition name string. Example of Hive * partition name string - "name=bob/year=2019". If the partition name for the given partition * column is equal to {@param defaultPartitionName}, the partition value in returned {@link CatalogPartitionSpec} will be null. ...
3.26
flink_HivePartitionUtils_parsePartitionValues_rdh
/** * Parse partition string specs into object values. */ public static Map<String, Object> parsePartitionValues(Map<String, String> partitionSpecs, String[] fieldNames, DataType[] fieldTypes, String defaultPartitionName, HiveShim shim) { checkArgument(fieldNames.length == fieldTypes.length); List<String> fi...
3.26
flink_Tuple1_toString_rdh
// ------------------------------------------------------------------------------------------------- // standard utilities // ------------------------------------------------------------------------------------------------- /** * Creates a string representation of the tuple in the form (f0), where the individual field...
3.26
flink_Tuple1_setFields_rdh
/** * Sets new values to all fields of the tuple. * * @param f0 * The value for field 0 */ public void setFields(T0 f0) { this.f0 = f0; }
3.26
flink_Tuple1_equals_rdh
/** * Deep equality for tuples by calling equals() on the tuple members. * * @param o * the object checked for equality * @return true if this is equal to o. */ @Override public boolean equals(Object o) { if (this == o) { return true; } if (!(o instanceof Tuple1)) { return false; ...
3.26
flink_HiveTableSink_toStagingDir_rdh
// get a staging dir private String toStagingDir(String stagingParentDir, Configuration conf) throws IOException { if (!stagingParentDir.endsWith(Path.SEPARATOR)) { stagingParentDir += Path.SEPARATOR; } // TODO: may append something more meaningful than a timestamp, like query ID stagingParentDi...
3.26
flink_JsonRowSchemaConverter_convert_rdh
/** * Converts a JSON schema into Flink's type information. Throws an exception if the schema * cannot converted because of loss of precision or too flexible schema. * * <p>The converter can resolve simple schema references to solve those cases where entities are * defined at the beginning and then used throughout...
3.26
flink_MaterializedCollectStreamResult_processRecord_rdh
// -------------------------------------------------------------------------------------------- @Override protected void processRecord(RowData row) {synchronized(resultLock) { boolean isInsertOp = (row.getRowKind() == RowKind.INSERT) || (row.getRowKind() == RowKind.UPDATE_AFTER); // Always set the Row...
3.26
flink_MaterializedCollectStreamResult_processInsert_rdh
// -------------------------------------------------------------------------------------------- private void processInsert(RowData row) { // limit the materialized table if ((materializedTable.size() - validRowPosition) >= maxRowCount) { cleanUp(); }materializedTable.add(row); f0.put(row, ma...
3.26
flink_CustomHeadersDecorator_getCustomHeaders_rdh
/** * Returns the custom headers added to the message. * * @return The custom headers as a collection of {@link HttpHeader}. */ @Override public Collection<HttpHeader> getCustomHeaders() { return customHeaders; }
3.26
flink_CustomHeadersDecorator_addCustomHeader_rdh
/** * Adds a custom header to the message. Initializes the custom headers collection if it hasn't * been initialized yet. * * @param httpHeader * The header to add. */ public void addCustomHeader(HttpHeader httpHeader) { if (customHeaders == null) { customHeaders = new ArrayList<>(); } cus...
3.26