name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_JobManagerCheckpointStorage_getSavepointPath_rdh | /**
*
* @return The default location where savepoints will be externalized if set.
*/
@Nullable
public Path getSavepointPath() {
return location.getBaseSavepointPath();
} | 3.26 |
flink_ReduceNode_getOperator_rdh | // ------------------------------------------------------------------------
@Override
public ReduceOperatorBase<?, ?> getOperator() {
return ((ReduceOperatorBase<?, ?>) (super.getOperator()));
} | 3.26 |
flink_ValueDataTypeConverter_extractDataType_rdh | /**
* Returns the clearly identifiable data type if possible. For example, {@code 12L} can be
* expressed as {@code DataTypes.BIGINT().notNull()}. However, for example, {@code null} could
* be any type and is not supported.
*
* <p>All types of the {@link LogicalTypeFamily#PREDEFINED} family, symbols, and arrays ar... | 3.26 |
flink_MemorySize_add_rdh | // ------------------------------------------------------------------------
// Calculations
// ------------------------------------------------------------------------
public MemorySize add(MemorySize that) {
return new MemorySize(Math.addExact(this.bytes, that.bytes));
} | 3.26 |
flink_MemorySize_parseBytes_rdh | /**
* Parses the given string as bytes. The supported expressions are listed under {@link MemorySize}.
*
* @param text
* The string to parse
* @return The parsed size, in bytes.
* @throws IllegalArgumentException
* Thrown, if the expression cannot be parsed.
*/
public static long parseBytes(String text) thr... | 3.26 |
flink_MemorySize_m0_rdh | // ------------------------------------------------------------------------
// Parsing
// ------------------------------------------------------------------------
/**
* Parses the given string as as MemorySize.
*
* @param text
* The string to parse
* @return The parsed MemorySize
* @throws IllegalArgumentExcept... | 3.26 |
flink_MemorySize_getMebiBytes_rdh | /**
* Gets the memory size in Mebibytes (= 1024 Kibibytes).
*/
public int getMebiBytes() {
return ((int) (bytes >> 20));
} | 3.26 |
flink_MemorySize_getKibiBytes_rdh | /**
* Gets the memory size in Kibibytes (= 1024 bytes).
*/public long getKibiBytes() {
return bytes >> 10;
} | 3.26 |
flink_MemorySize_parse_rdh | /**
* Parses the given string with a default unit.
*
* @param text
* The string to parse.
* @param defaultUnit
* specify the default unit.
* @return The parsed MemorySize.
* @throws IllegalArgumentException
* Thrown, if the expression cannot be parsed.
*/
public static MemorySize parse(String text, Memo... | 3.26 |
flink_MemorySize_getTebiBytes_rdh | /**
* Gets the memory size in Tebibytes (= 1024 Gibibytes).
*/
public long getTebiBytes() {
return bytes >> 40;
} | 3.26 |
flink_MemorySize_getBytes_rdh | // ------------------------------------------------------------------------
/**
* Gets the memory size in bytes.
*/
public long getBytes()
{
return bytes;
} | 3.26 |
flink_MemorySize_hashCode_rdh | // ------------------------------------------------------------------------
@Override
public int hashCode()
{
return ((int)
(bytes ^ (bytes >>> 32)));
} | 3.26 |
flink_MemorySize_getGibiBytes_rdh | /**
* Gets the memory size in Gibibytes (= 1024 Mebibytes).
*/
public long getGibiBytes() { return bytes >> 30;
} | 3.26 |
flink_SliceAssigners_hopping_rdh | /**
* Creates a hopping window {@link SliceAssigner} that assigns elements to slices of hopping
* windows.
*
* @param rowtimeIndex
* the index of rowtime field in the input row, {@code -1} if based on *
* processing time.
* @param shiftTimeZone
* The shift timezone of the window, if the proctime or rowtim... | 3.26 |
flink_SliceAssigners_withOffset_rdh | /**
* Creates a new {@link CumulativeSliceAssigner} with a new specified offset.
*/
public CumulativeSliceAssigner withOffset(Duration offset) {
return new CumulativeSliceAssigner(rowtimeIndex, shiftTimeZone, maxSize, step, offset.toMillis());
} | 3.26 |
flink_SliceAssigners_sliced_rdh | /**
* Creates a {@link SliceAssigner} that assigns elements which has been attached slice end
* timestamp.
*
* @param sliceEndIndex
* the index of slice end field in the input row, mustn't be a negative
* value.
* @param innerAssigner
* the inner assigner which assigns ... | 3.26 |
flink_SliceAssigners_windowed_rdh | /**
* Creates a {@link SliceAssigner} that assigns elements which has been attached window start
* and window end timestamp to slices. The assigned slice is equal to the given window.
*
* @param windowEndIndex
* the index of window end field in the input row, mustn't be a negative
* value.
* @param innerAssi... | 3.26 |
flink_SliceAssigners_tumbling_rdh | // ------—------—------—------—------—------—------—------—------—------—------—------—------—
// Utilities
// ------—------—------—------—------—------—------—------—------—------—------—------—------—
/**
* Creates a tumbling window {@link SliceAssigner} that assigns elements to slices of tumbling
* windows.
*
* ... | 3.26 |
flink_SliceAssigners_cumulative_rdh | /**
* Creates a cumulative window {@link SliceAssigner} that assigns elements to slices of
* cumulative windows.
*
* @param rowtimeIndex
* the index of rowtime field in the input row, {@code -1} if based on *
* processing time.
* @param shiftTimeZone
* The shift timezone of the window, if the proctime or ... | 3.26 |
flink_DataStructureConverter_open_rdh | /**
* Converter between internal and external data structure.
*
* <p>Converters are serializable and can be passed to runtime operators. However, converters are
* not thread-safe.
*
* @param <I>
* internal data structure (see {@link RowData})
* @param <E>
* external data structure (see {@link DataType#getC... | 3.26 |
flink_DataStructureConverter_toInternalOrNull_rdh | /**
* Converts to internal data structure or {@code null}.
*
* <p>The nullability could be derived from the data type. However, this method reduces null
* checks.
*/
default I toInternalOrNull(E external) {
if (external == null) {
return null;
}
return toInternal(ext... | 3.26 |
flink_StreamingFileWriter_closePartFileForPartitions_rdh | /**
* Close in-progress part file when partition is committable.
*/
private void closePartFileForPartitions() throws Exception {if (partitionCommitPredicate != null) {
final Iterator<Map.Entry<String, Long>> iterator = inProgressPartitions.entrySet().iterator();
while (iterator.hasNext()) {
Map.Ent... | 3.26 |
flink_ParserUtils_parsePluginOutput_rdh | /**
* Iterates over the given lines, identifying plugin execution blocks with the given pattern and
* parses the plugin output with the given parser.
*
* <p>This method assumes that the given pattern matches at most once for each module.
*
* <p>The given pattern must include a {@code m... | 3.26 |
flink_Channel_getLocalStrategyComparator_rdh | /**
* Gets the local strategy comparator from this Channel.
*
* @return The local strategy comparator.
*/public TypeComparatorFactory<?> getLocalStrategyComparator() {
return localStrategyComparator;} | 3.26 |
flink_Channel_getRelativeTempMemory_rdh | /**
* Gets the memory for materializing the channel's result from this Channel.
*
* @return The temp memory.
*/
public double
getRelativeTempMemory() {
return this.relativeTempMemory;
} | 3.26 |
flink_Channel_getRequiredGlobalProps_rdh | // --------------------------------------------------------------------------------------------
// Data Property Handling
// --------------------------------------------------------------------------------------------
public RequestedGlobalProperties getRequiredGlobalProps() {
return requiredGlobalProps;
} | 3.26 |
flink_Channel_getDataExchangeMode_rdh | /**
* Gets the data exchange mode (batch / pipelined) to use for the data exchange of this channel.
*
* @return The data exchange mode of this channel.
*/
public DataExchangeMode getDataExchangeMode() {
return dataExchangeMode;
} | 3.26 |
flink_Channel_getSerializer_rdh | /**
* Gets the serializer from this Channel.
*
* @return The serializer.
*/
public TypeSerializerFactory<?> getSerializer() {
return f0;} | 3.26 |
flink_Channel_setLocalStrategyComparator_rdh | /**
* Sets the local strategy comparator for this Channel.
*
* @param localStrategyComparator
* The local strategy comparator to set.
*/public void setLocalStrategyComparator(TypeComparatorFactory<?> localStrategyComparator) {
this.localStrategyComparator
= localStrategyComparator;
} | 3.26 |
flink_Channel_toString_rdh | // --------------------------------------------------------------------------------------------
@Override
public String toString() {
return ((((((("Channel (" + this.source) + (this.target == null ? ')' : (") -> ("
+ this.target) + ')')) + '[') + this.shipStrategy) + "] [") + this.localStrategy) + "] ") + ((this.tempM... | 3.26 |
flink_Channel_getTarget_rdh | /**
* Gets the target of this Channel.
*
* @return The target.
*/
public PlanNode getTarget() {
return this.target;
} | 3.26 |
flink_Channel_m3_rdh | // --------------------------------------------------------------------------------------------
/**
* Utility method used while swapping binary union nodes for n-ary union nodes.
*/
public void m3(PlanNode newUnionNode) {
if (!(this.source instanceof BinaryUnionPlanNode)) {
throw new IllegalStateException();
}
else... | 3.26 |
flink_Channel_getSource_rdh | // --------------------------------------------------------------------------------------------
// Accessors
// --------------------------------------------------------------------------------------------
/**
* Gets the source of this Channel.
*
* @return The source.
*/
@Override
public PlanNode getSource() {
r... | 3.26 |
flink_Channel_getReplicationFactor_rdh | /**
* Returns the replication factor of the connection.
*
* @return The replication factor of the connection.
*/
public int getReplicationFactor() {
return this.replicationFactor;
} | 3.26 |
flink_Channel_setTarget_rdh | /**
* Sets the target of this Channel.
*
* @param target
* The target.
*/
public void setTarget(PlanNode target) {
this.target = target;
} | 3.26 |
flink_Channel_getEstimatedOutputSize_rdh | // --------------------------------------------------------------------------------------------
// Statistic Estimates
// --------------------------------------------------------------------------------------------
@Override
public long getEstimatedOutputSize() {
long estimate = this.source.template.getEstimatedOutputS... | 3.26 |
flink_Channel_setTempMode_rdh | /**
* Sets the temp mode of the connection.
*
* @param tempMode
* The temp mode of the connection.
*/
public void setTempMode(TempMode tempMode)
{
this.tempMode = tempMode;
} | 3.26 |
flink_Channel_setSerializer_rdh | /**
* Sets the serializer for this Channel.
*
* @param serializer
* The serializer to set.
*/
public void setSerializer(TypeSerializerFactory<?> serializer) {
this.f0 = serializer;
} | 3.26 |
flink_Channel_setShipStrategyComparator_rdh | /**
* Sets the ship strategy comparator for this Channel.
*
* @param shipStrategyComparator
* The ship strategy comparator to set.
*/
public void setShipStrategyComparator(TypeComparatorFactory<?> shipStrategyComparator) {
this.shipStrategyComparator = shipStrategyComparator;
} | 3.26 |
flink_Channel_m4_rdh | // --------------------------------------------------------------------------------------------
public int m4() {
return this.source.getOptimizerNode().getMaxDepth() + 1;
} | 3.26 |
flink_Channel_setReplicationFactor_rdh | /**
* Sets the replication factor of the connection.
*
* @param factor
* The replication factor of the connection.
*/
public void setReplicationFactor(int factor) {
this.replicationFactor = factor;
} | 3.26 |
flink_Channel_setRelativeTempMemory_rdh | /**
* Sets the memory for materializing the channel's result from this Channel.
*
* @param relativeTempMemory
* The memory for materialization.
*/
public void setRelativeTempMemory(double relativeTempMemory) {
this.relativeTempMemory = relativeTempMemory;
} | 3.26 |
flink_AbstractColumnReader_readToVector_rdh | /**
* Reads `total` values from this columnReader into column.
*/
@Override
public final void readToVector(int readNumber, VECTOR vector) throws IOException {
int
rowId =
0;
WritableIntVector dictionaryIds = null;
if (dictionary != null) {
dictionaryIds = vector.reserveDictionaryIds(readN... | 3.26 |
flink_AbstractColumnReader_afterReadPage_rdh | /**
* After read a page, we may need some initialization.
*/
protected void afterReadPage() {
} | 3.26 |
flink_TaskExecutor_tryLoadLocalAllocationSnapshots_rdh | /**
* This method tries to repopulate the {@link JobTable} and {@link TaskSlotTable} from the local
* filesystem in a best-effort manner.
*/
private void tryLoadLocalAllocationSnapshots() {
Collection<SlotAllocationSnapshot> slotAllocationSnapshots = slotAllocationSnapshotPersistenceService.loadAllocationSnapsho... | 3.26 |
flink_TaskExecutor_failTask_rdh | // ------------------------------------------------------------------------
// Internal task methods
// ------------------------------------------------------------------------
private void failTask(final ExecutionAttemptID executionAttemptID, final Throwable
cause) {
final Task task = taskSlotTable.getTask(executi... | 3.26 |
flink_TaskExecutor_onStart_rdh | // ------------------------------------------------------------------------
// Life cycle
// ------------------------------------------------------------------------
@Override
public void onStart() throws Exception {
try {
startTaskExecutorServices();
} catch (Throwable t) {
fina... | 3.26 |
flink_TaskExecutor_offerSlotsToJobManager_rdh | // ------------------------------------------------------------------------
// Internal job manager connection methods
// ------------------------------------------------------------------------
private void offerSlotsToJobManager(final JobID jobId) {
jobTable.getConnection(jobId).ifPresent(this::internalOfferSlot... | 3.26 |
flink_TaskExecutor_submitTask_rdh | // ----------------------------------------------------------------------
// Task lifecycle RPCs
// ----------------------------------------------------------------------
@Override
public CompletableFuture<Acknowledge> submitTask(TaskDeploymentDescriptor tdd, JobMasterId jobMasterId, Time timeout)
{
try {
f... | 3.26 |
flink_TaskExecutor_triggerCheckpoint_rdh | // ----------------------------------------------------------------------
// Checkpointing RPCs
// ----------------------------------------------------------------------
@Override
public CompletableFuture<Acknowledge> triggerCheckpoint(ExecutionAttemptID executionAttemptID, long checkpointId, long checkpointTimestamp, ... | 3.26 |
flink_TaskExecutor_m1_rdh | // ------------------------------------------------------------------------
// Internal resource manager connection methods
// ------------------------------------------------------------------------
private void m1(String newLeaderAddress, ResourceManagerId newResourceManagerId) {
resourceManagerAddress = create... | 3.26 |
flink_TaskExecutor_heartbeatFromJobManager_rdh | // ----------------------------------------------------------------------
// Heartbeat RPC
// ----------------------------------------------------------------------
@Override
public CompletableFuture<Void> heartbeatFromJobManager(ResourceID resourceID, AllocatedSlotReport allocatedSlotReport) {
return jobManagerHea... | 3.26 |
flink_TaskExecutor_syncSlotsWithSnapshotFromJobMaster_rdh | /**
* Syncs the TaskExecutor's view on its allocated slots with the JobMaster's view. Slots which
* are no longer reported by the JobMaster are being freed. Slots which the JobMaster thinks it
* still owns but which are no longer allocated to it will be failed via {@li... | 3.26 |
flink_TaskExecutor_onFatalError_rdh | // ------------------------------------------------------------------------
// Error Handling
// ------------------------------------------------------------------------
/**
* Notifies the TaskExecutor that a fatal error has occurred and it cannot proceed.
*
* @param t
* The except... | 3.26 |
flink_TaskExecutor_getResourceManagerConnection_rdh | // ------------------------------------------------------------------------
// Access to fields for testing
// ------------------------------------------------------------------------
@VisibleForTesting
TaskExecutorToResourceManagerConnection getResourceManagerConnection() {
return resourceManagerConnection;
} | 3.26 |
flink_TaskExecutor_updatePartitions_rdh | // ----------------------------------------------------------------------
// Partition lifecycle RPCs
// ----------------------------------------------------------------------
@Override
public CompletableFuture<Acknowledge> updatePartitions(final ExecutionAtte... | 3.26 |
flink_TaskExecutor_sendOperatorEventToTask_rdh | // ----------------------------------------------------------------------
// Other RPCs
// ----------------------------------------------------------------------
@Override
public CompletableFuture<Acknowledge> sendOperatorEventToTask(ExecutionAttemptID executionAttemptID, OperatorID operatorId, SerializedValue<Operator... | 3.26 |
flink_TaskExecutor_disconnectJobManager_rdh | // ----------------------------------------------------------------------
// Disconnection RPCs
// ----------------------------------------------------------------------
@Override
public void disconnectJobManager(JobID j... | 3.26 |
flink_TaskExecutor_isConnectedToResourceManager_rdh | // ------------------------------------------------------------------------
// Internal utility methods
// ------------------------------------------------------------------------
private boolean isConnectedToResourceManager() {return establishedResourceManagerConnection != null;
} | 3.26 |
flink_TaskExecutor_requestSlot_rdh | // ----------------------------------------------------------------------
// Slot allocation RPCs
// ----------------------------------------------------------------------
@Override
public CompletableFuture<Acknowledge> requestSlot(final SlotID slotId, final J... | 3.26 |
flink_TaskExecutor_getResourceID_rdh | // ------------------------------------------------------------------------
// Properties
// ------------------------------------------------------------------------
public ResourceID getResourceID() {
return unresolvedTaskManagerLocation.getResourceID();
} | 3.26 |
flink_TaskExecutor_onStop_rdh | /**
* Called to shut down the TaskManager. The method closes all TaskManager services.
*/
@Override
public CompletableFuture<Void> onStop() {
log.info("Stopping TaskExecutor {}.", getAddress());
Throwable jobManagerDisconnectThrowable
= null;
FlinkExpectedException cause = new FlinkExpectedException... | 3.26 |
flink_AbstractUdfOperator_asArray_rdh | // --------------------------------------------------------------------------------------------
/**
* Generic utility function that wraps a single class object into an array of that class type.
*
* @param <U>
* The type of the classes.
* @param clazz
* The class object to be wrapped.
* @return An array wrapp... | 3.26 |
flink_AbstractUdfOperator_getUserCodeWrapper_rdh | // --------------------------------------------------------------------------------------------
/**
* Gets the function that is held by this operator. The function is the actual implementation of
* the user code.
*
* <p>This throws an exception if the pact does not contain an object but a class for the user
* code... | 3.26 |
flink_AbstractUdfOperator_getBroadcastInputs_rdh | // --------------------------------------------------------------------------------------------
/**
* Returns the input, or null, if none is set.
*
* @return The broadcast input root operator.
*/
public Map<String, Operator<?>> getBroadcastInputs() {
return this.broadcastInputs;
} | 3.26 |
flink_AbstractUdfOperator_setBroadcastVariable_rdh | /**
* Binds the result produced by a plan rooted at {@code root} to a variable used by the UDF
* wrapped in this operator.
*
* @param root
* The root of the plan producing this input.
*/
public void setBroadcastVariable(String name, Operator<?> root) {
if (name == null) {
throw new IllegalArgument... | 3.26 |
flink_AbstractUdfOperator_setBroadcastVariables_rdh | /**
* Clears all previous broadcast inputs and binds the given inputs as broadcast variables of
* this operator.
*
* @param inputs
* The {@code<name, root>} pairs to be set as broadcast inputs.
*/
public <T> void setBroadcastVariables(Map<String, Operator<T>> inputs) {this.broadcastInputs.clear();
this.br... | 3.26 |
flink_CsvBulkWriter_forPojo_rdh | /**
* Builds a writer based on a POJO class definition.
*
* @param pojoClass
* The class of the POJO.
* @param stream
* The output stream.
* @param <T>
* The type of the elements accepted by this writer.
*/
static <T> CsvBulkWriter<T, T, Void> forPojo(Class<T> pojoClass, FSDataOutputStream stream) {
... | 3.26 |
flink_CsvBulkWriter_forSchema_rdh | /**
* Builds a writer with Jackson schema and a type converter.
*
* @param mapper
* The specialized mapper for producing CSV.
* @param schema
* The schema that defined the mapping properties.
* @param converter
* The type converter that converts incoming elements of type {@code <T>} into
* elements of ... | 3.26 |
flink_CommittableMessageTypeInfo_of_rdh | /**
* Returns the type information based on the serializer for a {@link CommittableMessage}.
*
* @param committableSerializerFactory
* factory to create the serializer for a {@link CommittableMessage}
* @param <CommT>
* type of the committable
* @return */
public static <CommT> TypeInformation<CommittableMe... | 3.26 |
flink_CommittableMessageTypeInfo_noOutput_rdh | /**
* Returns the type information for a {@link CommittableMessage} with no committable.
*
* @return {@link TypeInformation} with {@link CommittableMessage}
*/public static TypeInformation<CommittableMessage<Void>> noOutput() {
return new CommittableMessageTypeInfo<>(NoOutputSerializer::new);
} | 3.26 |
flink_LogicalTypeJsonDeserializer_deserializeLengthFieldType_rdh | // --------------------------------------------------------------------------------------------
// Helper methods for some complex types
// --------------------------------------------------------------------------------------------
private LogicalType deserializeLengthFieldType(LogicalTypeRoot typeRoot, JsonNode logic... | 3.26 |
flink_LogicalTypeJsonDeserializer_deserializeInternal_rdh | /**
* Deserialize json according to the original type root. It's reverse operation of {@code SerializerWIP#serializeinternal}.
*/
private LogicalType deserializeInternal(JsonNode logicalTypeNode) {
LogicalTypeRoot typeRoot = LogicalTypeRoot.valueOf(logicalTypeNode.get(FIELD_NAME_TYPE_NAME).asText());
// the ... | 3.26 |
flink_FlatMapNode_computeOperatorSpecificDefaultEstimates_rdh | /**
* Computes the estimates for the FlatMap operator. Since it un-nests, we assume a cardinality
* increase. To give the system a hint at data increase, we take a default magic number of a 5
* times increase.
*/
@Override
protected void computeOperatorSpecificDefaultEstimates(DataStatistics statistics) {
this.... | 3.26 |
flink_GroupCombineOperatorBase_executeOnCollections_rdh | // --------------------------------------------------------------------------------------------
@Override
protected List<OUT> executeOnCollections(List<IN> inputData, RuntimeContext ctx, ExecutionConfig executionConfig) throws Exception {
GroupCombineFunction<IN, OUT>
function = this.userFunction.getUserCodeOb... | 3.26 |
flink_GroupCombineOperatorBase_setGroupOrder_rdh | // --------------------------------------------------------------------------------------------
/**
* Sets the order of the elements within a reduce group.
*
* @param order
* The order for the elements in a reduce group.
*/
public void setGroupOrder(Ordering order) {
this.groupOrder = order;
} | 3.26 |
flink_ExceptionHistoryEntry_fromTaskManagerLocation_rdh | /**
* Creates a {@code ArchivedTaskManagerLocation} copy of the passed {@link TaskManagerLocation}.
*
* @param taskManagerLocation
* The {@code TaskManagerLocation} that's going to be copied.
* @return The corresponding {@code ArchivedTaskManagerLocation} or {@code null} if {@code null} was passed.
*/
@VisibleF... | 3.26 |
flink_ExceptionHistoryEntry_createGlobal_rdh | /**
* Creates an {@code ExceptionHistoryEntry} that is not based on an {@code Execution}.
*/
public static ExceptionHistoryEntry createGlobal(Throwable cause, CompletableFuture<Map<String,
String>> failureLabels) {
return new ExceptionHistoryEntry(cause, System.currentTimeMillis(), failureLabels, null, ((Archived... | 3.26 |
flink_ExceptionHistoryEntry_create_rdh | /**
* Creates an {@code ExceptionHistoryEntry} based on the provided {@code Execution}.
*
* @param failedExecution
* the failed {@code Execution}.
* @param taskName
* the name of the task.
* @param failureLabels
* the labels associated with the failure.
* @return The {@code ExceptionHistoryEntry}.
* @th... | 3.26 |
flink_BlobInputStream_throwEOFException_rdh | /**
* Convenience method to throw an {@link EOFException}.
*
* @throws EOFException
* thrown to indicate the underlying input stream did not provide as much
* data as expected
*/
private void throwEOFException() throws EOFException {
throw new EOFException(String.format("Expected to read %d more bytes fr... | 3.26 |
flink_TypeStringUtils_containsDelimiter_rdh | // --------------------------------------------------------------------------------------------
private static boolean containsDelimiter(String string) {
final char[] charArray = string.toCharArray();
for (char c : charArray) {
if (isDelimiter(c)) {
return true;
}}
return false;... | 3.26 |
flink_CompositeType_hasDeterministicFieldOrder_rdh | /**
* True if this type has an inherent ordering of the fields, such that a user can always be sure
* in which order the fields will be in. This is true for Tuples and Case Classes. It is not
* true for Regular Java Objects, since there, the ordering of the fields can be arbitrary.
*
* <p>This is used when transla... | 3.26 |
flink_CompositeType_hasField_rdh | /**
* Returns true when this type has a composite field with the given name.
*/
@PublicEvolving
public boolean hasField(String fieldName) {
return getFieldIndex(fieldName) >= 0;
} | 3.26 |
flink_InternalTimerServiceImpl_restoreTimersForKeyGroup_rdh | /**
* Restore the timers (both processing and event time ones) for a given {@code keyGroupIdx}.
*
* @param restoredSnapshot
* the restored snapshot containing the key-group's timers, and the
* serializers that were used to write them
* @param keyGroupIdx
* the id of the key-group to be put in the snapshot.... | 3.26 |
flink_InternalTimerServiceImpl_startTimerService_rdh | /**
* Starts the local {@link InternalTimerServiceImpl} by:
*
* <ol>
* <li>Setting the {@code keySerialized} and {@code namespaceSerializer} for the timers it
* will contain.
* <li>Setting the {@code triggerTarget} which contains the action to be performed when a
* ... | 3.26 |
flink_HiveParserSubQueryDiagnostic_getRewrite_rdh | /**
* Counterpart of hive's org.apache.hadoop.hive.ql.parse.SubQueryDiagnostic.
*/public class HiveParserSubQueryDiagnostic {
static QBSubQueryRewrite getRewrite(HiveParserQBSubQuery subQuery, TokenRewriteStream stream, HiveParserContext ctx) {
if (ctx.isExplainSkipExecution()) {
return new QB... | 3.26 |
flink_NoResourceAvailableException_equals_rdh | // --------------------------------------------------------------------------------------------
@Override
public boolean equals(Object obj) {return (obj instanceof NoResourceAvailableException) && getMessage().equals(((NoResourceAvailableException) (obj)).getMessage());
} | 3.26 |
flink_StreamTaskNetworkInput_getRecordDeserializers_rdh | // Initialize one deserializer per input channel
private static Map<InputChannelInfo, SpillingAdaptiveSpanningRecordDeserializer<DeserializationDelegate<StreamElement>>> getRecordDeserializers(CheckpointedInputGate checkpointedInputGate, IOManager ioManager) {
return checkpointedInputGate.getChannelInfos().stream(... | 3.26 |
flink_SqlValidatorUtils_adjustTypeForMultisetConstructor_rdh | /**
* When the element element does not equal with the component type, making explicit casting.
*
* @param evenType
* derived type for element with even index
* @param oddType
* derived type for element with odd index
* @param sqlCallBinding
* description of call
*/
private static void adjustTypeForMulti... | 3.26 |
flink_RowData_createFieldGetter_rdh | // ------------------------------------------------------------------------------------------
// Access Utilities
// ------------------------------------------------------------------------------------------
/**
* Creates an accessor for getting elements in an internal row data structure at the given
* position.
*
... | 3.26 |
flink_JobMasterPartitionTracker_getAllTrackedNonClusterPartitions_rdh | /**
* Gets all the non-cluster partitions under tracking.
*/
default Collection<ResultPartitionDeploymentDescriptor> getAllTrackedNonClusterPartitions() {
return m0().stream().filter(descriptor -> !descriptor.getPartitionType().isPersistent()).collect(Collectors.toList());
} | 3.26 |
flink_JobMasterPartitionTracker_getAllTrackedClusterPartitions_rdh | /**
* Gets all the cluster partitions under tracking.
*/
default Collection<ResultPartitionDeploymentDescriptor>
getAllTrackedClusterPartitions()
{
return m0().stream().filter(descriptor -> descriptor.getPartitionType().isPersistent()).collect(Collectors.toList());
} | 3.26 |
flink_JobMasterPartitionTracker_stopTrackingAndReleasePartitions_rdh | /**
* Releases the given partitions and stop the tracking of partitions that were released.
*/
default void stopTrackingAndReleasePartitions(Collection<ResultPartitionID> resultPartitionIds) {
stopTrackingAndReleasePartitions(resultPartitionIds, true);
} | 3.26 |
flink_CallExpression_temporary_rdh | /**
* Creates a {@link CallExpression} to a temporary function (potentially shadowing a {@link Catalog} function or providing a system function).
*/
public static CallExpression temporary(FunctionIdentifier functionIdentifier, FunctionDefinition functionDefinition, List<ResolvedExpression> args, DataType dataType) {
... | 3.26 |
flink_CallExpression_permanent_rdh | /**
* Creates a {@link CallExpression} to a resolved built-in function. It assumes that the {@link BuiltInFunctionDefinition} instance is provided by the framework (usually the core module).
*/
@Internal
public static CallExpression permanent(BuiltInFunctionDefinition builtInFunctionDefinition, List<ResolvedExpressio... | 3.26 |
flink_CallExpression_anonymous_rdh | /**
* Creates a {@link CallExpression} to an anonymous function that has been declared inline
* without a {@link FunctionIdentifier}.
*/
public static CallExpression anonymous(FunctionDefinition functionDefinition, List<ResolvedExpression> args, DataType dataType) {
return new CallExpression(true, null, functio... | 3.26 |
flink_CallExpression_getFunctionName_rdh | /**
* Returns a string representation of the call's function for logging or printing to a console.
*/ public String
getFunctionName() {
if (functionIdentifier == null) {
return
functionDefinition.toString();
} else {
return functionIdentifier.asSummaryString();
}} | 3.26 |
flink_PartitionTempFileManager_m0_rdh | /**
* Generate a new partition directory with partitions.
*/
public Path m0(String... partitions) {
Path parentPath = taskTmpDir;
for (String dir : partitions) {
parentPath = new Path(parentPath, dir);
}
return new Path(parentPath, newFileName());
} | 3.26 |
flink_PartitionTempFileManager_listTaskTemporaryPaths_rdh | /**
* Returns task temporary paths in this checkpoint.
*/
public static List<Path> listTaskTemporaryPaths(FileSystem fs, Path basePath, BiPredicate<Integer, Integer> taskAttemptFilter) throws Exception
{
List<Path> taskTmpPaths = new ArrayList<>();
if (fs.exists(basePath)) {for (FileStatus taskStatus :
... | 3.26 |
flink_PartitionTempFileManager_collectPartSpecToPaths_rdh | /**
* Collect all partitioned paths, aggregate according to partition spec.
*/
public static Map<LinkedHashMap<String, String>, List<Path>> collectPartSpecToPaths(FileSystem fs,
List<Path> taskPaths, int partColSize) {
Map<LinkedHashMap<String, String>, List<Path>> specToPaths = new HashMap<>();
for (Path t... | 3.26 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.