name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
flink_JobManagerCheckpointStorage_getSavepointPath_rdh
/** * * @return The default location where savepoints will be externalized if set. */ @Nullable public Path getSavepointPath() { return location.getBaseSavepointPath(); }
3.26
flink_ReduceNode_getOperator_rdh
// ------------------------------------------------------------------------ @Override public ReduceOperatorBase<?, ?> getOperator() { return ((ReduceOperatorBase<?, ?>) (super.getOperator())); }
3.26
flink_ValueDataTypeConverter_extractDataType_rdh
/** * Returns the clearly identifiable data type if possible. For example, {@code 12L} can be * expressed as {@code DataTypes.BIGINT().notNull()}. However, for example, {@code null} could * be any type and is not supported. * * <p>All types of the {@link LogicalTypeFamily#PREDEFINED} family, symbols, and arrays ar...
3.26
flink_MemorySize_add_rdh
// ------------------------------------------------------------------------ // Calculations // ------------------------------------------------------------------------ public MemorySize add(MemorySize that) { return new MemorySize(Math.addExact(this.bytes, that.bytes)); }
3.26
flink_MemorySize_parseBytes_rdh
/** * Parses the given string as bytes. The supported expressions are listed under {@link MemorySize}. * * @param text * The string to parse * @return The parsed size, in bytes. * @throws IllegalArgumentException * Thrown, if the expression cannot be parsed. */ public static long parseBytes(String text) thr...
3.26
flink_MemorySize_m0_rdh
// ------------------------------------------------------------------------ // Parsing // ------------------------------------------------------------------------ /** * Parses the given string as as MemorySize. * * @param text * The string to parse * @return The parsed MemorySize * @throws IllegalArgumentExcept...
3.26
flink_MemorySize_getMebiBytes_rdh
/** * Gets the memory size in Mebibytes (= 1024 Kibibytes). */ public int getMebiBytes() { return ((int) (bytes >> 20)); }
3.26
flink_MemorySize_getKibiBytes_rdh
/** * Gets the memory size in Kibibytes (= 1024 bytes). */public long getKibiBytes() { return bytes >> 10; }
3.26
flink_MemorySize_parse_rdh
/** * Parses the given string with a default unit. * * @param text * The string to parse. * @param defaultUnit * specify the default unit. * @return The parsed MemorySize. * @throws IllegalArgumentException * Thrown, if the expression cannot be parsed. */ public static MemorySize parse(String text, Memo...
3.26
flink_MemorySize_getTebiBytes_rdh
/** * Gets the memory size in Tebibytes (= 1024 Gibibytes). */ public long getTebiBytes() { return bytes >> 40; }
3.26
flink_MemorySize_getBytes_rdh
// ------------------------------------------------------------------------ /** * Gets the memory size in bytes. */ public long getBytes() { return bytes; }
3.26
flink_MemorySize_hashCode_rdh
// ------------------------------------------------------------------------ @Override public int hashCode() { return ((int) (bytes ^ (bytes >>> 32))); }
3.26
flink_MemorySize_getGibiBytes_rdh
/** * Gets the memory size in Gibibytes (= 1024 Mebibytes). */ public long getGibiBytes() { return bytes >> 30; }
3.26
flink_SliceAssigners_hopping_rdh
/** * Creates a hopping window {@link SliceAssigner} that assigns elements to slices of hopping * windows. * * @param rowtimeIndex * the index of rowtime field in the input row, {@code -1} if based on * * processing time. * @param shiftTimeZone * The shift timezone of the window, if the proctime or rowtim...
3.26
flink_SliceAssigners_withOffset_rdh
/** * Creates a new {@link CumulativeSliceAssigner} with a new specified offset. */ public CumulativeSliceAssigner withOffset(Duration offset) { return new CumulativeSliceAssigner(rowtimeIndex, shiftTimeZone, maxSize, step, offset.toMillis()); }
3.26
flink_SliceAssigners_sliced_rdh
/** * Creates a {@link SliceAssigner} that assigns elements which has been attached slice end * timestamp. * * @param sliceEndIndex * the index of slice end field in the input row, mustn't be a negative * value. * @param innerAssigner * the inner assigner which assigns ...
3.26
flink_SliceAssigners_windowed_rdh
/** * Creates a {@link SliceAssigner} that assigns elements which has been attached window start * and window end timestamp to slices. The assigned slice is equal to the given window. * * @param windowEndIndex * the index of window end field in the input row, mustn't be a negative * value. * @param innerAssi...
3.26
flink_SliceAssigners_tumbling_rdh
// ------—------—------—------—------—------—------—------—------—------—------—------—------— // Utilities // ------—------—------—------—------—------—------—------—------—------—------—------—------— /** * Creates a tumbling window {@link SliceAssigner} that assigns elements to slices of tumbling * windows. * * ...
3.26
flink_SliceAssigners_cumulative_rdh
/** * Creates a cumulative window {@link SliceAssigner} that assigns elements to slices of * cumulative windows. * * @param rowtimeIndex * the index of rowtime field in the input row, {@code -1} if based on * * processing time. * @param shiftTimeZone * The shift timezone of the window, if the proctime or ...
3.26
flink_DataStructureConverter_open_rdh
/** * Converter between internal and external data structure. * * <p>Converters are serializable and can be passed to runtime operators. However, converters are * not thread-safe. * * @param <I> * internal data structure (see {@link RowData}) * @param <E> * external data structure (see {@link DataType#getC...
3.26
flink_DataStructureConverter_toInternalOrNull_rdh
/** * Converts to internal data structure or {@code null}. * * <p>The nullability could be derived from the data type. However, this method reduces null * checks. */ default I toInternalOrNull(E external) { if (external == null) { return null; } return toInternal(ext...
3.26
flink_StreamingFileWriter_closePartFileForPartitions_rdh
/** * Close in-progress part file when partition is committable. */ private void closePartFileForPartitions() throws Exception {if (partitionCommitPredicate != null) { final Iterator<Map.Entry<String, Long>> iterator = inProgressPartitions.entrySet().iterator(); while (iterator.hasNext()) { Map.Ent...
3.26
flink_ParserUtils_parsePluginOutput_rdh
/** * Iterates over the given lines, identifying plugin execution blocks with the given pattern and * parses the plugin output with the given parser. * * <p>This method assumes that the given pattern matches at most once for each module. * * <p>The given pattern must include a {@code m...
3.26
flink_Channel_getLocalStrategyComparator_rdh
/** * Gets the local strategy comparator from this Channel. * * @return The local strategy comparator. */public TypeComparatorFactory<?> getLocalStrategyComparator() { return localStrategyComparator;}
3.26
flink_Channel_getRelativeTempMemory_rdh
/** * Gets the memory for materializing the channel's result from this Channel. * * @return The temp memory. */ public double getRelativeTempMemory() { return this.relativeTempMemory; }
3.26
flink_Channel_getRequiredGlobalProps_rdh
// -------------------------------------------------------------------------------------------- // Data Property Handling // -------------------------------------------------------------------------------------------- public RequestedGlobalProperties getRequiredGlobalProps() { return requiredGlobalProps; }
3.26
flink_Channel_getDataExchangeMode_rdh
/** * Gets the data exchange mode (batch / pipelined) to use for the data exchange of this channel. * * @return The data exchange mode of this channel. */ public DataExchangeMode getDataExchangeMode() { return dataExchangeMode; }
3.26
flink_Channel_getSerializer_rdh
/** * Gets the serializer from this Channel. * * @return The serializer. */ public TypeSerializerFactory<?> getSerializer() { return f0;}
3.26
flink_Channel_setLocalStrategyComparator_rdh
/** * Sets the local strategy comparator for this Channel. * * @param localStrategyComparator * The local strategy comparator to set. */public void setLocalStrategyComparator(TypeComparatorFactory<?> localStrategyComparator) { this.localStrategyComparator = localStrategyComparator; }
3.26
flink_Channel_toString_rdh
// -------------------------------------------------------------------------------------------- @Override public String toString() { return ((((((("Channel (" + this.source) + (this.target == null ? ')' : (") -> (" + this.target) + ')')) + '[') + this.shipStrategy) + "] [") + this.localStrategy) + "] ") + ((this.tempM...
3.26
flink_Channel_getTarget_rdh
/** * Gets the target of this Channel. * * @return The target. */ public PlanNode getTarget() { return this.target; }
3.26
flink_Channel_m3_rdh
// -------------------------------------------------------------------------------------------- /** * Utility method used while swapping binary union nodes for n-ary union nodes. */ public void m3(PlanNode newUnionNode) { if (!(this.source instanceof BinaryUnionPlanNode)) { throw new IllegalStateException(); } else...
3.26
flink_Channel_getSource_rdh
// -------------------------------------------------------------------------------------------- // Accessors // -------------------------------------------------------------------------------------------- /** * Gets the source of this Channel. * * @return The source. */ @Override public PlanNode getSource() { r...
3.26
flink_Channel_getReplicationFactor_rdh
/** * Returns the replication factor of the connection. * * @return The replication factor of the connection. */ public int getReplicationFactor() { return this.replicationFactor; }
3.26
flink_Channel_setTarget_rdh
/** * Sets the target of this Channel. * * @param target * The target. */ public void setTarget(PlanNode target) { this.target = target; }
3.26
flink_Channel_getEstimatedOutputSize_rdh
// -------------------------------------------------------------------------------------------- // Statistic Estimates // -------------------------------------------------------------------------------------------- @Override public long getEstimatedOutputSize() { long estimate = this.source.template.getEstimatedOutputS...
3.26
flink_Channel_setTempMode_rdh
/** * Sets the temp mode of the connection. * * @param tempMode * The temp mode of the connection. */ public void setTempMode(TempMode tempMode) { this.tempMode = tempMode; }
3.26
flink_Channel_setSerializer_rdh
/** * Sets the serializer for this Channel. * * @param serializer * The serializer to set. */ public void setSerializer(TypeSerializerFactory<?> serializer) { this.f0 = serializer; }
3.26
flink_Channel_setShipStrategyComparator_rdh
/** * Sets the ship strategy comparator for this Channel. * * @param shipStrategyComparator * The ship strategy comparator to set. */ public void setShipStrategyComparator(TypeComparatorFactory<?> shipStrategyComparator) { this.shipStrategyComparator = shipStrategyComparator; }
3.26
flink_Channel_m4_rdh
// -------------------------------------------------------------------------------------------- public int m4() { return this.source.getOptimizerNode().getMaxDepth() + 1; }
3.26
flink_Channel_setReplicationFactor_rdh
/** * Sets the replication factor of the connection. * * @param factor * The replication factor of the connection. */ public void setReplicationFactor(int factor) { this.replicationFactor = factor; }
3.26
flink_Channel_setRelativeTempMemory_rdh
/** * Sets the memory for materializing the channel's result from this Channel. * * @param relativeTempMemory * The memory for materialization. */ public void setRelativeTempMemory(double relativeTempMemory) { this.relativeTempMemory = relativeTempMemory; }
3.26
flink_AbstractColumnReader_readToVector_rdh
/** * Reads `total` values from this columnReader into column. */ @Override public final void readToVector(int readNumber, VECTOR vector) throws IOException { int rowId = 0; WritableIntVector dictionaryIds = null; if (dictionary != null) { dictionaryIds = vector.reserveDictionaryIds(readN...
3.26
flink_AbstractColumnReader_afterReadPage_rdh
/** * After read a page, we may need some initialization. */ protected void afterReadPage() { }
3.26
flink_TaskExecutor_tryLoadLocalAllocationSnapshots_rdh
/** * This method tries to repopulate the {@link JobTable} and {@link TaskSlotTable} from the local * filesystem in a best-effort manner. */ private void tryLoadLocalAllocationSnapshots() { Collection<SlotAllocationSnapshot> slotAllocationSnapshots = slotAllocationSnapshotPersistenceService.loadAllocationSnapsho...
3.26
flink_TaskExecutor_failTask_rdh
// ------------------------------------------------------------------------ // Internal task methods // ------------------------------------------------------------------------ private void failTask(final ExecutionAttemptID executionAttemptID, final Throwable cause) { final Task task = taskSlotTable.getTask(executi...
3.26
flink_TaskExecutor_onStart_rdh
// ------------------------------------------------------------------------ // Life cycle // ------------------------------------------------------------------------ @Override public void onStart() throws Exception { try { startTaskExecutorServices(); } catch (Throwable t) { fina...
3.26
flink_TaskExecutor_offerSlotsToJobManager_rdh
// ------------------------------------------------------------------------ // Internal job manager connection methods // ------------------------------------------------------------------------ private void offerSlotsToJobManager(final JobID jobId) { jobTable.getConnection(jobId).ifPresent(this::internalOfferSlot...
3.26
flink_TaskExecutor_submitTask_rdh
// ---------------------------------------------------------------------- // Task lifecycle RPCs // ---------------------------------------------------------------------- @Override public CompletableFuture<Acknowledge> submitTask(TaskDeploymentDescriptor tdd, JobMasterId jobMasterId, Time timeout) { try { f...
3.26
flink_TaskExecutor_triggerCheckpoint_rdh
// ---------------------------------------------------------------------- // Checkpointing RPCs // ---------------------------------------------------------------------- @Override public CompletableFuture<Acknowledge> triggerCheckpoint(ExecutionAttemptID executionAttemptID, long checkpointId, long checkpointTimestamp, ...
3.26
flink_TaskExecutor_m1_rdh
// ------------------------------------------------------------------------ // Internal resource manager connection methods // ------------------------------------------------------------------------ private void m1(String newLeaderAddress, ResourceManagerId newResourceManagerId) { resourceManagerAddress = create...
3.26
flink_TaskExecutor_heartbeatFromJobManager_rdh
// ---------------------------------------------------------------------- // Heartbeat RPC // ---------------------------------------------------------------------- @Override public CompletableFuture<Void> heartbeatFromJobManager(ResourceID resourceID, AllocatedSlotReport allocatedSlotReport) { return jobManagerHea...
3.26
flink_TaskExecutor_syncSlotsWithSnapshotFromJobMaster_rdh
/** * Syncs the TaskExecutor's view on its allocated slots with the JobMaster's view. Slots which * are no longer reported by the JobMaster are being freed. Slots which the JobMaster thinks it * still owns but which are no longer allocated to it will be failed via {@li...
3.26
flink_TaskExecutor_onFatalError_rdh
// ------------------------------------------------------------------------ // Error Handling // ------------------------------------------------------------------------ /** * Notifies the TaskExecutor that a fatal error has occurred and it cannot proceed. * * @param t * The except...
3.26
flink_TaskExecutor_getResourceManagerConnection_rdh
// ------------------------------------------------------------------------ // Access to fields for testing // ------------------------------------------------------------------------ @VisibleForTesting TaskExecutorToResourceManagerConnection getResourceManagerConnection() { return resourceManagerConnection; }
3.26
flink_TaskExecutor_updatePartitions_rdh
// ---------------------------------------------------------------------- // Partition lifecycle RPCs // ---------------------------------------------------------------------- @Override public CompletableFuture<Acknowledge> updatePartitions(final ExecutionAtte...
3.26
flink_TaskExecutor_sendOperatorEventToTask_rdh
// ---------------------------------------------------------------------- // Other RPCs // ---------------------------------------------------------------------- @Override public CompletableFuture<Acknowledge> sendOperatorEventToTask(ExecutionAttemptID executionAttemptID, OperatorID operatorId, SerializedValue<Operator...
3.26
flink_TaskExecutor_disconnectJobManager_rdh
// ---------------------------------------------------------------------- // Disconnection RPCs // ---------------------------------------------------------------------- @Override public void disconnectJobManager(JobID j...
3.26
flink_TaskExecutor_isConnectedToResourceManager_rdh
// ------------------------------------------------------------------------ // Internal utility methods // ------------------------------------------------------------------------ private boolean isConnectedToResourceManager() {return establishedResourceManagerConnection != null; }
3.26
flink_TaskExecutor_requestSlot_rdh
// ---------------------------------------------------------------------- // Slot allocation RPCs // ---------------------------------------------------------------------- @Override public CompletableFuture<Acknowledge> requestSlot(final SlotID slotId, final J...
3.26
flink_TaskExecutor_getResourceID_rdh
// ------------------------------------------------------------------------ // Properties // ------------------------------------------------------------------------ public ResourceID getResourceID() { return unresolvedTaskManagerLocation.getResourceID(); }
3.26
flink_TaskExecutor_onStop_rdh
/** * Called to shut down the TaskManager. The method closes all TaskManager services. */ @Override public CompletableFuture<Void> onStop() { log.info("Stopping TaskExecutor {}.", getAddress()); Throwable jobManagerDisconnectThrowable = null; FlinkExpectedException cause = new FlinkExpectedException...
3.26
flink_AbstractUdfOperator_asArray_rdh
// -------------------------------------------------------------------------------------------- /** * Generic utility function that wraps a single class object into an array of that class type. * * @param <U> * The type of the classes. * @param clazz * The class object to be wrapped. * @return An array wrapp...
3.26
flink_AbstractUdfOperator_getUserCodeWrapper_rdh
// -------------------------------------------------------------------------------------------- /** * Gets the function that is held by this operator. The function is the actual implementation of * the user code. * * <p>This throws an exception if the pact does not contain an object but a class for the user * code...
3.26
flink_AbstractUdfOperator_getBroadcastInputs_rdh
// -------------------------------------------------------------------------------------------- /** * Returns the input, or null, if none is set. * * @return The broadcast input root operator. */ public Map<String, Operator<?>> getBroadcastInputs() { return this.broadcastInputs; }
3.26
flink_AbstractUdfOperator_setBroadcastVariable_rdh
/** * Binds the result produced by a plan rooted at {@code root} to a variable used by the UDF * wrapped in this operator. * * @param root * The root of the plan producing this input. */ public void setBroadcastVariable(String name, Operator<?> root) { if (name == null) { throw new IllegalArgument...
3.26
flink_AbstractUdfOperator_setBroadcastVariables_rdh
/** * Clears all previous broadcast inputs and binds the given inputs as broadcast variables of * this operator. * * @param inputs * The {@code<name, root>} pairs to be set as broadcast inputs. */ public <T> void setBroadcastVariables(Map<String, Operator<T>> inputs) {this.broadcastInputs.clear(); this.br...
3.26
flink_CsvBulkWriter_forPojo_rdh
/** * Builds a writer based on a POJO class definition. * * @param pojoClass * The class of the POJO. * @param stream * The output stream. * @param <T> * The type of the elements accepted by this writer. */ static <T> CsvBulkWriter<T, T, Void> forPojo(Class<T> pojoClass, FSDataOutputStream stream) { ...
3.26
flink_CsvBulkWriter_forSchema_rdh
/** * Builds a writer with Jackson schema and a type converter. * * @param mapper * The specialized mapper for producing CSV. * @param schema * The schema that defined the mapping properties. * @param converter * The type converter that converts incoming elements of type {@code <T>} into * elements of ...
3.26
flink_CommittableMessageTypeInfo_of_rdh
/** * Returns the type information based on the serializer for a {@link CommittableMessage}. * * @param committableSerializerFactory * factory to create the serializer for a {@link CommittableMessage} * @param <CommT> * type of the committable * @return */ public static <CommT> TypeInformation<CommittableMe...
3.26
flink_CommittableMessageTypeInfo_noOutput_rdh
/** * Returns the type information for a {@link CommittableMessage} with no committable. * * @return {@link TypeInformation} with {@link CommittableMessage} */public static TypeInformation<CommittableMessage<Void>> noOutput() { return new CommittableMessageTypeInfo<>(NoOutputSerializer::new); }
3.26
flink_LogicalTypeJsonDeserializer_deserializeLengthFieldType_rdh
// -------------------------------------------------------------------------------------------- // Helper methods for some complex types // -------------------------------------------------------------------------------------------- private LogicalType deserializeLengthFieldType(LogicalTypeRoot typeRoot, JsonNode logic...
3.26
flink_LogicalTypeJsonDeserializer_deserializeInternal_rdh
/** * Deserialize json according to the original type root. It's reverse operation of {@code SerializerWIP#serializeinternal}. */ private LogicalType deserializeInternal(JsonNode logicalTypeNode) { LogicalTypeRoot typeRoot = LogicalTypeRoot.valueOf(logicalTypeNode.get(FIELD_NAME_TYPE_NAME).asText()); // the ...
3.26
flink_FlatMapNode_computeOperatorSpecificDefaultEstimates_rdh
/** * Computes the estimates for the FlatMap operator. Since it un-nests, we assume a cardinality * increase. To give the system a hint at data increase, we take a default magic number of a 5 * times increase. */ @Override protected void computeOperatorSpecificDefaultEstimates(DataStatistics statistics) { this....
3.26
flink_GroupCombineOperatorBase_executeOnCollections_rdh
// -------------------------------------------------------------------------------------------- @Override protected List<OUT> executeOnCollections(List<IN> inputData, RuntimeContext ctx, ExecutionConfig executionConfig) throws Exception { GroupCombineFunction<IN, OUT> function = this.userFunction.getUserCodeOb...
3.26
flink_GroupCombineOperatorBase_setGroupOrder_rdh
// -------------------------------------------------------------------------------------------- /** * Sets the order of the elements within a reduce group. * * @param order * The order for the elements in a reduce group. */ public void setGroupOrder(Ordering order) { this.groupOrder = order; }
3.26
flink_ExceptionHistoryEntry_fromTaskManagerLocation_rdh
/** * Creates a {@code ArchivedTaskManagerLocation} copy of the passed {@link TaskManagerLocation}. * * @param taskManagerLocation * The {@code TaskManagerLocation} that's going to be copied. * @return The corresponding {@code ArchivedTaskManagerLocation} or {@code null} if {@code null} was passed. */ @VisibleF...
3.26
flink_ExceptionHistoryEntry_createGlobal_rdh
/** * Creates an {@code ExceptionHistoryEntry} that is not based on an {@code Execution}. */ public static ExceptionHistoryEntry createGlobal(Throwable cause, CompletableFuture<Map<String, String>> failureLabels) { return new ExceptionHistoryEntry(cause, System.currentTimeMillis(), failureLabels, null, ((Archived...
3.26
flink_ExceptionHistoryEntry_create_rdh
/** * Creates an {@code ExceptionHistoryEntry} based on the provided {@code Execution}. * * @param failedExecution * the failed {@code Execution}. * @param taskName * the name of the task. * @param failureLabels * the labels associated with the failure. * @return The {@code ExceptionHistoryEntry}. * @th...
3.26
flink_BlobInputStream_throwEOFException_rdh
/** * Convenience method to throw an {@link EOFException}. * * @throws EOFException * thrown to indicate the underlying input stream did not provide as much * data as expected */ private void throwEOFException() throws EOFException { throw new EOFException(String.format("Expected to read %d more bytes fr...
3.26
flink_TypeStringUtils_containsDelimiter_rdh
// -------------------------------------------------------------------------------------------- private static boolean containsDelimiter(String string) { final char[] charArray = string.toCharArray(); for (char c : charArray) { if (isDelimiter(c)) { return true; }} return false;...
3.26
flink_CompositeType_hasDeterministicFieldOrder_rdh
/** * True if this type has an inherent ordering of the fields, such that a user can always be sure * in which order the fields will be in. This is true for Tuples and Case Classes. It is not * true for Regular Java Objects, since there, the ordering of the fields can be arbitrary. * * <p>This is used when transla...
3.26
flink_CompositeType_hasField_rdh
/** * Returns true when this type has a composite field with the given name. */ @PublicEvolving public boolean hasField(String fieldName) { return getFieldIndex(fieldName) >= 0; }
3.26
flink_InternalTimerServiceImpl_restoreTimersForKeyGroup_rdh
/** * Restore the timers (both processing and event time ones) for a given {@code keyGroupIdx}. * * @param restoredSnapshot * the restored snapshot containing the key-group's timers, and the * serializers that were used to write them * @param keyGroupIdx * the id of the key-group to be put in the snapshot....
3.26
flink_InternalTimerServiceImpl_startTimerService_rdh
/** * Starts the local {@link InternalTimerServiceImpl} by: * * <ol> * <li>Setting the {@code keySerialized} and {@code namespaceSerializer} for the timers it * will contain. * <li>Setting the {@code triggerTarget} which contains the action to be performed when a * ...
3.26
flink_HiveParserSubQueryDiagnostic_getRewrite_rdh
/** * Counterpart of hive's org.apache.hadoop.hive.ql.parse.SubQueryDiagnostic. */public class HiveParserSubQueryDiagnostic { static QBSubQueryRewrite getRewrite(HiveParserQBSubQuery subQuery, TokenRewriteStream stream, HiveParserContext ctx) { if (ctx.isExplainSkipExecution()) { return new QB...
3.26
flink_NoResourceAvailableException_equals_rdh
// -------------------------------------------------------------------------------------------- @Override public boolean equals(Object obj) {return (obj instanceof NoResourceAvailableException) && getMessage().equals(((NoResourceAvailableException) (obj)).getMessage()); }
3.26
flink_StreamTaskNetworkInput_getRecordDeserializers_rdh
// Initialize one deserializer per input channel private static Map<InputChannelInfo, SpillingAdaptiveSpanningRecordDeserializer<DeserializationDelegate<StreamElement>>> getRecordDeserializers(CheckpointedInputGate checkpointedInputGate, IOManager ioManager) { return checkpointedInputGate.getChannelInfos().stream(...
3.26
flink_SqlValidatorUtils_adjustTypeForMultisetConstructor_rdh
/** * When the element element does not equal with the component type, making explicit casting. * * @param evenType * derived type for element with even index * @param oddType * derived type for element with odd index * @param sqlCallBinding * description of call */ private static void adjustTypeForMulti...
3.26
flink_RowData_createFieldGetter_rdh
// ------------------------------------------------------------------------------------------ // Access Utilities // ------------------------------------------------------------------------------------------ /** * Creates an accessor for getting elements in an internal row data structure at the given * position. * ...
3.26
flink_JobMasterPartitionTracker_getAllTrackedNonClusterPartitions_rdh
/** * Gets all the non-cluster partitions under tracking. */ default Collection<ResultPartitionDeploymentDescriptor> getAllTrackedNonClusterPartitions() { return m0().stream().filter(descriptor -> !descriptor.getPartitionType().isPersistent()).collect(Collectors.toList()); }
3.26
flink_JobMasterPartitionTracker_getAllTrackedClusterPartitions_rdh
/** * Gets all the cluster partitions under tracking. */ default Collection<ResultPartitionDeploymentDescriptor> getAllTrackedClusterPartitions() { return m0().stream().filter(descriptor -> descriptor.getPartitionType().isPersistent()).collect(Collectors.toList()); }
3.26
flink_JobMasterPartitionTracker_stopTrackingAndReleasePartitions_rdh
/** * Releases the given partitions and stop the tracking of partitions that were released. */ default void stopTrackingAndReleasePartitions(Collection<ResultPartitionID> resultPartitionIds) { stopTrackingAndReleasePartitions(resultPartitionIds, true); }
3.26
flink_CallExpression_temporary_rdh
/** * Creates a {@link CallExpression} to a temporary function (potentially shadowing a {@link Catalog} function or providing a system function). */ public static CallExpression temporary(FunctionIdentifier functionIdentifier, FunctionDefinition functionDefinition, List<ResolvedExpression> args, DataType dataType) { ...
3.26
flink_CallExpression_permanent_rdh
/** * Creates a {@link CallExpression} to a resolved built-in function. It assumes that the {@link BuiltInFunctionDefinition} instance is provided by the framework (usually the core module). */ @Internal public static CallExpression permanent(BuiltInFunctionDefinition builtInFunctionDefinition, List<ResolvedExpressio...
3.26
flink_CallExpression_anonymous_rdh
/** * Creates a {@link CallExpression} to an anonymous function that has been declared inline * without a {@link FunctionIdentifier}. */ public static CallExpression anonymous(FunctionDefinition functionDefinition, List<ResolvedExpression> args, DataType dataType) { return new CallExpression(true, null, functio...
3.26
flink_CallExpression_getFunctionName_rdh
/** * Returns a string representation of the call's function for logging or printing to a console. */ public String getFunctionName() { if (functionIdentifier == null) { return functionDefinition.toString(); } else { return functionIdentifier.asSummaryString(); }}
3.26
flink_PartitionTempFileManager_m0_rdh
/** * Generate a new partition directory with partitions. */ public Path m0(String... partitions) { Path parentPath = taskTmpDir; for (String dir : partitions) { parentPath = new Path(parentPath, dir); } return new Path(parentPath, newFileName()); }
3.26
flink_PartitionTempFileManager_listTaskTemporaryPaths_rdh
/** * Returns task temporary paths in this checkpoint. */ public static List<Path> listTaskTemporaryPaths(FileSystem fs, Path basePath, BiPredicate<Integer, Integer> taskAttemptFilter) throws Exception { List<Path> taskTmpPaths = new ArrayList<>(); if (fs.exists(basePath)) {for (FileStatus taskStatus : ...
3.26
flink_PartitionTempFileManager_collectPartSpecToPaths_rdh
/** * Collect all partitioned paths, aggregate according to partition spec. */ public static Map<LinkedHashMap<String, String>, List<Path>> collectPartSpecToPaths(FileSystem fs, List<Path> taskPaths, int partColSize) { Map<LinkedHashMap<String, String>, List<Path>> specToPaths = new HashMap<>(); for (Path t...
3.26