name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_LocalTimeTypeInfo_instantiateComparator_rdh | // --------------------------------------------------------------------------------------------
private static <X> TypeComparator<X> instantiateComparator(Class<? extends TypeComparator<X>> comparatorClass, boolean ascendingOrder) {
try {
Constructor<? extends TypeComparator<X>> constructor = comparatorClas... | 3.26 |
flink_WrappingRuntimeException_unwrap_rdh | /**
* Recursively unwraps this WrappingRuntimeException and its causes, getting the first non
* wrapping exception.
*
* @return The first cause that is not a wrapping exception.
*/
public Throwable unwrap() {
Throwable cause = getCause();
return cause instanceof WrappingRuntimeException ? ((WrappingRunt... | 3.26 |
flink_GenericRowData_setField_rdh | /**
* Sets the field value at the given position.
*
* <p>Note: The given field value must be an internal data structures. Otherwise the {@link GenericRowData} is corrupted and may throw exception when processing. See {@link RowData} for
* more information about internal data structures.
*
... | 3.26 |
flink_GenericRowData_ofKind_rdh | /**
* Creates an instance of {@link GenericRowData} with given kind and field values.
*
* <p>Note: All fields of the row must be internal data structures.
*/
public static GenericRowData ofKind(RowKind kind, Object... values) {
GenericRowData row = new GenericRowData(kind, values.length);
for (int i = 0; i ... | 3.26 |
flink_GenericRowData_getField_rdh | /**
* Returns the field value at the given position.
*
* <p>Note: The returned value is in internal data structure. See {@link RowData} for more
* information about internal data structures.
*
* <p>The returned field value can be null for representing nullability.
*/
public Object getField(int pos) {
return ... | 3.26 |
flink_ContinuousFileMonitoringFunction_listEligibleFiles_rdh | /**
* Returns the paths of the files not yet processed.
*
* @param fileSystem
* The filesystem where the monitored directory resides.
*/
private Map<Path, FileStatus> listEligibleFiles(FileSystem fileSystem, Path path) {
final FileStatus[] statuses;
try {
statuses = fileSystem.listStatus(path);
... | 3.26 |
flink_ContinuousFileMonitoringFunction_shouldIgnore_rdh | /**
* Returns {@code true} if the file is NOT to be processed further. This happens if the
* modification time of the file is smaller than the {@link #globalModificationTime}.
*
* @param filePath
* the path of the file to check.
* @param modificationTime
* the modification time of the file.
*/
private boole... | 3.26 |
flink_ContinuousFileMonitoringFunction_getInputSplitsSortedByModTime_rdh | /**
* Creates the input splits to be forwarded to the downstream tasks of the {@link ContinuousFileReaderOperator}. Splits are sorted <b>by modification time</b> before being
* forwarded and only splits belonging to files in the {@code eligibleFiles} list will be
* processed.
*
* @param eligibleFiles
* The file... | 3.26 |
flink_ContinuousFileMonitoringFunction_snapshotState_rdh | // --------------------- Checkpointing --------------------------
@Override
public void snapshotState(FunctionSnapshotContext context) throws Exception {
Preconditions.checkState(this.checkpointedState != null, ("The " + getClass().getSimpleName()) + " state has not been properly initialized.");
this.check... | 3.26 |
flink_DecimalData_fromBigDecimal_rdh | // ------------------------------------------------------------------------------------------
// Constructor Utilities
// ------------------------------------------------------------------------------------------
/**
* Creates an instance of {@link DecimalData} from a {@link BigDecimal} and the given precision
* and ... | 3.26 |
flink_DecimalData_isCompact_rdh | // ------------------------------------------------------------------------------------------
// Utilities
// ------------------------------------------------------------------------------------------
/**
* Returns whether the decimal value is small enough to be stored in a long.
*/
public static boolean isCompact(in... | 3.26 |
flink_DecimalData_toBigDecimal_rdh | /**
* Converts this {@link DecimalData} into an instance of {@link BigDecimal}.
*/
public BigDecimal toBigDecimal() {
BigDecimal bd = decimalVal;
if (bd == null) {
decimalVal = bd = BigDecimal.valueOf(longVal, scale);
}
return bd;
} | 3.26 |
flink_DecimalData_fromUnscaledBytes_rdh | /**
* Creates an instance of {@link DecimalData} from an unscaled byte array value and the given
* precision and scale.
*/public static DecimalData fromUnscaledBytes(byte[] unscaledBytes, int precision, int scale) {
BigDecimal bd = new BigDecimal(new BigInteger(unscaledBytes), scale);return fromBigDecimal(bd, ... | 3.26 |
flink_DecimalData_scale_rdh | /**
* Returns the <i>scale</i> of this {@link DecimalData}.
*/
public int scale() {
return scale;
} | 3.26 |
flink_DecimalData_toUnscaledLong_rdh | /**
* Returns a long describing the <i>unscaled value</i> of this {@link DecimalData}.
*
* @throws ArithmeticException
* if this {@link DecimalData} does not exactly fit in a long.
*/
public long toUnscaledLong() {
if (isCompact()) {
return
longVal;
} else {
return toBigDecimal()... | 3.26 |
flink_DecimalData_zero_rdh | /**
* Creates an instance of {@link DecimalData} for a zero value with the given precision and
* scale.
*
* <p>The precision will be checked. If the precision overflows, null will be returned.
*/
@Nullable
public static DecimalData zero(int precision, int scale) {
if (precision <= MAX_COMPACT_PRECISION) {
... | 3.26 |
flink_DecimalData_fromUnscaledLong_rdh | /**
* Creates an instance of {@link DecimalData} from an unscaled long value and the given
* precision and scale.
*/
public static DecimalData fromUnscaledLong(long unscaledLong, int precision, int scale) {
checkArgument((precision > 0) && (precision <= MAX_LONG_DIGITS));
return new DecimalData(precision, ... | 3.26 |
flink_DecimalData_precision_rdh | // ------------------------------------------------------------------------------------------
// Public Interfaces
// ------------------------------------------------------------------------------------------
/**
* Returns the <i>precision</i> of this {@link DecimalData}.
*
* <p>The precision is the number of digits... | 3.26 |
flink_DecimalData_copy_rdh | /**
* Returns a copy of this {@link DecimalData} object.
*/
public DecimalData copy() {
return new DecimalData(precision, scale, longVal, decimalVal);
} | 3.26 |
flink_BinaryRowWriter_reset_rdh | /**
* First, reset.
*/
@Override
public void reset() {
this.cursor = fixedSize;
for (int i = 0; i < nullBitsSizeInBytes; i += 8)
{
segment.putLong(i, 0L);
}
} | 3.26 |
flink_BinaryRowWriter_setNullAt_rdh | /**
* Default not null.
*/@Override
public void setNullAt(int pos) {
setNullBit(pos);
segment.putLong(getFieldOffset(pos), 0L);
} | 3.26 |
flink_TransientBlobCleanupTask_run_rdh | /**
* Cleans up transient BLOBs whose TTL is up, tolerating that files do not exist (anymore).
*/
@Override
public void run() {
// let's cache the current time - we do not operate on a millisecond precision anyway
final long currentTimeMillis = System.currentTimeMillis();
// iterate through all entries ... | 3.26 |
flink_RetryingRegistration_getFuture_rdh | // ------------------------------------------------------------------------
// completion and cancellation
// ------------------------------------------------------------------------
public CompletableFuture<RetryingRegistrationResult<G,
S, R>> getFuture() {
return completionFuture;
} | 3.26 |
flink_RetryingRegistration_startRegistration_rdh | /**
* This method resolves the target address to a callable gateway and starts the registration
* after that.
*/
@SuppressWarnings("unchecked")
public void startRegistration() {
if (canceled) {
// we already got canceled
return;
}
try {
// trigger resolutio... | 3.26 |
flink_RetryingRegistration_cancel_rdh | /**
* Cancels the registration procedure.
*/
public void cancel() {
canceled = true;
completionFuture.cancel(false);
} | 3.26 |
flink_RetryingRegistration_register_rdh | /**
* This method performs a registration attempt and triggers either a success notification or a
* retry, depending on the result.
*/
@SuppressWarnings("unchecked")
private void register(final G g... | 3.26 |
flink_LogicalType_is_rdh | /**
* Returns whether the family type of the type equals to the {@code family} or not.
*
* @param family
* The family type to check against for equality
*/
public boolean is(LogicalTypeFamily family) {
return typeRoot.getFamilies().contains(family);
} | 3.26 |
flink_LogicalType_asSummaryString_rdh | /**
* Returns a string that summarizes this type for printing to a console. An implementation might
* shorten long names or skips very specific properties.
*
* <p>Use {@link #asSerializableString()} for a type string that fully serializes this instance.
*
* @return summary string of this type for debugging purpos... | 3.26 |
flink_LogicalType_isAnyOf_rdh | /**
* Returns whether the root of the type is part of at least one family of the {@code typeFamily}
* or not.
*
* @param typeFamilies
* The families to check against for equality
*/
public boolean isAnyOf(LogicalTypeFamily... typeFamilies) {
return Arrays.stream(typeFamilies).anyMatch(tf -> this.typeRoot.ge... | 3.26 |
flink_LogicalType_isNullable_rdh | /**
* Returns whether a value of this type can be {@code null}.
*/
public boolean isNullable() {
return isNullable;
} | 3.26 |
flink_FormatFactory_forwardOptions_rdh | /**
* Returns a set of {@link ConfigOption} that are directly forwarded to the runtime
* implementation but don't affect the final execution topology.
*
* <p>Options declared here can override options of the persisted plan during an enrichment
* phase. Since a restored topology is static, a... | 3.26 |
flink_AsynchronousJobOperationKey_getJobId_rdh | /**
* Get the job id for the given operation key.
*
* @return job id
*/public JobID getJobId()
{return jobId;
} | 3.26 |
flink_CopyableValueComparator_readObject_rdh | // --------------------------------------------------------------------------------------------
private void readObject(ObjectInputStream s)
throws IOException, ClassNotFoundException {
// read basic object and the type
s.defaultReadObject();
this.reference = InstantiationUtil.instantiate(type, CopyableValu... | 3.26 |
flink_CopyableValueComparator_supportsSerializationWithKeyNormalization_rdh | // --------------------------------------------------------------------------------------------
// unsupported normalization
// --------------------------------------------------------------------------------------------
@Override
public boolean supportsSerializationWithKeyNormalization() {
return false;
} | 3.26 |
flink_UserFacingListState_get_rdh | // ------------------------------------------------------------------------
@Override
public Iterable<T> get() throws Exception {
Iterable<T> original = originalState.get();
return original != null ? original : emptyState;
} | 3.26 |
flink_Tuple13_toString_rdh | // -------------------------------------------------------------------------------------------------
// standard utilities
// -------------------------------------------------------------------------------------------------
/**
* Creates a string representation of the tuple in the form (f0, f1, f2, f3, f4, f5, f6, f7,... | 3.26 |
flink_Tuple13_setFields_rdh | /**
* Sets new values to all fields of the tuple.
*
* @param f0
* The value for field 0
* @param f1
* The value for field 1
* @param f2
* The value for field 2
* @param f3
* The value for field 3
* @param f4
* The value for field 4
* @param f5
* The value for field 5
* @param f6
* The valu... | 3.26 |
flink_Tuple13_copy_rdh | /**
* Shallow tuple copy.
*
* @return A new Tuple with the same fields as this.
*/
@Override
@SuppressWarnings("unchecked")
public Tuple13<T0, T1, T2, T3, T4, T5, T6,
T7, T8, T9, T10, T11, T12> copy() {
return new Tuple13<>(this.f0,
this.f1, this.f2,
this.f3, this.f4, this.f5, this.f6, this.f7, this.f8... | 3.26 |
flink_Tuple13_of_rdh | /**
* Creates a new tuple and assigns the given values to the tuple's fields. This is more
* convenient than using the constructor, because the compiler can infer the generic type
* arguments implicitly. For example: {@code Tuple3.of(n, x, s)} instead of {@code new
* Tuple3<Integer, Double, String>(n, x, s)}
*/
pu... | 3.26 |
flink_TaskExecutorProcessUtils_newProcessSpecBuilder_rdh | // ------------------------------------------------------------------------
// Memory Configuration Calculations
// ------------------------------------------------------------------------
public static TaskExecutorProcessSpecBuilder newProcessSpecBuilder(final Configuration config) {
return TaskExecutorProcessSpec... | 3.26 |
flink_TaskExecutorProcessUtils_generateDynamicConfigsStr_rdh | // ------------------------------------------------------------------------
// Generating Dynamic Config Options
// ------------------------------------------------------------------------
public static String generateDynamicConfigsStr(final TaskExecutorProcessSpec taskExecutorProcessSpec) {
final Map<String, Strin... | 3.26 |
flink_AsynchronousBlockWriter_getReturnQueue_rdh | /**
* Gets the queue in which the memory segments are queued after the asynchronous write is
* completed.
*
* @return The queue with the written memory segments.
*/
@Override
public LinkedBlockingQueue<MemorySegment> getReturnQueue() {
return this.returnSegments;} | 3.26 |
flink_AsynchronousBlockWriter_getNextReturnedBlock_rdh | /**
* Gets the next memory segment that has been written and is available again. This method blocks
* until such a segment is available, or until an error occurs in the writer, or the writer is
* closed.
*
* <p>NOTE: If this method is invoked without any segment ever returning (for example, because
* the {@link #... | 3.26 |
flink_EmptyIterator_remove_rdh | /**
* Throws a {@link java.lang.UnsupportedOperationException}.
*
* @see java.util.Iterator#remove()
*/
@Override
public void remove() {
throw new UnsupportedOperationException();
} | 3.26 |
flink_EmptyIterator_m0_rdh | /**
* Always throws a {@link java.util.NoSuchElementException}.
*
* @see java.util.Iterator#next()
*/
@Override
public E m0() {
throw new NoSuchElementException();} | 3.26 |
flink_EmptyIterator_get_rdh | /**
* Gets a singleton instance of the empty iterator.
*
* @param <E>
* The type of the objects (not) returned by the iterator.
* @return An instance of the iterator.
*/
public static <E> EmptyIterator<E> get() {
@SuppressWarnings("unchecked")
EmptyIterator<E> iter = ((EmptyIterator<E>) (INSTANCE));
... | 3.26 |
flink_FlinkDatabaseMetaData_storesMixedCaseQuotedIdentifiers_rdh | /**
* Flink sql is mixed case as sensitive.
*/
@Override
public boolean storesMixedCaseQuotedIdentifiers() throws SQLException {return true;
} | 3.26 |
flink_FlinkDatabaseMetaData_m4_rdh | /**
* Flink sql is mixed case as sensitive.
*/
@Override
public boolean m4() throws SQLException {
return true;
} | 3.26 |
flink_FlinkDatabaseMetaData_supportsMixedCaseIdentifiers_rdh | /**
* Flink sql is mixed case as sensitive.
*/
@Override
public boolean supportsMixedCaseIdentifiers() throws SQLException {return true;
} | 3.26 |
flink_FlinkDatabaseMetaData_nullsAreSortedLow_rdh | /**
* In flink null value will be used as low value for sort.
*/@Override
public boolean nullsAreSortedLow() throws SQLException {
return true;
} | 3.26 |
flink_FlinkDatabaseMetaData_nullPlusNonNullIsNull_rdh | /**
* Null value plus non-null in flink will be null result.
*/
@Override
public boolean nullPlusNonNullIsNull()
throws SQLException {
return true;
} | 3.26 |
flink_FlinkDatabaseMetaData_supportsMixedCaseQuotedIdentifiers_rdh | /**
* Flink sql is mixed case as sensitive.
*/
@Override
public boolean supportsMixedCaseQuotedIdentifiers() throws SQLException {
return true;
} | 3.26 |
flink_FlinkDatabaseMetaData_isCatalogAtStart_rdh | /**
* Catalog name appears at the start of full name.
*/
@Override
public boolean isCatalogAtStart() throws SQLException {
return true;
} | 3.26 |
flink_FlinkDatabaseMetaData_getSchemas_rdh | // TODO Flink will support SHOW DATABASES LIKE statement in FLIP-297, this method will be
// supported after that issue.
@Override
public ResultSet
getSchemas(String catalog, String schemaPattern) throws SQLException {throw new UnsupportedOperationException();
} | 3.26 |
flink_DefaultCostEstimator_addCachedHybridHashCosts_rdh | /**
* Calculates the costs for the cached variant of the hybrid hash join. We are assuming by
* default that half of the cached hash table fit into memory.
*/
@Override
public void addCachedHybridHashCosts(EstimateProvider buildSideInput, EstimateProvider probeSideInput, Costs costs, int costWeight) {
if (costW... | 3.26 |
flink_DefaultCostEstimator_addFileInputCost_rdh | // --------------------------------------------------------------------------------------------
// Local Strategy Cost
// --------------------------------------------------------------------------------------------
@Override
public void addFileInputCost(long fileSizeInBytes, Costs costs) {
if (fileSizeInBytes >= ... | 3.26 |
flink_DefaultCostEstimator_addArtificialDamCost_rdh | // --------------------------------------------------------------------------------------------
// Damming Cost
// --------------------------------------------------------------------------------------------
@Override
public void addArtificialDamCost(EstimateProvider estimates, long bufferSize, Costs costs) {
final... | 3.26 |
flink_DefaultCostEstimator_addRandomPartitioningCost_rdh | // --------------------------------------------------------------------------------------------
@Override
public void addRandomPartitioningCost(EstimateProvider
estimates, Costs costs) {
// conservative estimate: we need ship the whole data over the network to establish the
// partitioning. no disk costs.
... | 3.26 |
flink_SlotStatus_getResourceProfile_rdh | /**
* Get the resource profile of this slot.
*
* @return The resource profile
*/
public ResourceProfile getResourceProfile() {
return f0;
} | 3.26 |
flink_SlotStatus_getAllocationID_rdh | /**
* Get the allocation id of this slot.
*
* @return The allocation id if this slot is allocated, otherwise null
*/
public AllocationID getAllocationID() {
return allocationID;
} | 3.26 |
flink_BlockResettableMutableObjectIterator_next_rdh | // --------------------------------------------------------------------------------------------
@Override
public T next(T target) throws IOException {
// check for the left over element
if (this.readPhase) {
return getNextRecord(target);} else // writing phase. check for leftover first
if (this.le... | 3.26 |
flink_BlockResettableMutableObjectIterator_hasFurtherInput_rdh | /**
* Checks, whether the input that is blocked by this iterator, has further elements available.
* This method may be used to forecast (for example at the point where a block is full) whether
* there will be more data (possibly in another block).
*
* @return True, if there will be more data, false otherwise.
*/
... | 3.26 |
flink_AdvancedFunctionsExample_executeLastDatedValueFunction_rdh | /**
* Aggregates data by name and returns the latest non-null {@code item_count} value with its
* corresponding {@code order_date}.
*/
private static void executeLastDatedValueFunction(TableEnvironment env) {
// create a table with example data
final Table customers = env.fromValues(DataTypes.of("ROW<name ST... | 3.26 |
flink_CatalogTableImpl_removeRedundant_rdh | /**
* Construct catalog table properties from {@link #toProperties()}.
*/
public static Map<String, String> removeRedundant(Map<String, String> properties, TableSchema schema, List<String> partitionKeys) {
Map<String, String> ret = new HashMap<>(properties);
DescriptorProperties descriptorProperties = new D... | 3.26 |
flink_CatalogTableImpl_fromProperties_rdh | /**
* Construct a {@link CatalogTableImpl} from complete properties that contains table schema.
*/
public static CatalogTableImpl fromProperties(Map<String, String> properties) {
DescriptorProperties descriptorProperties = new DescriptorProperties(false);
descriptorProperties.putProperties(properties);
Ta... | 3.26 |
flink_SourceOperatorFactory_instantiateSourceOperator_rdh | /**
* This is a utility method to conjure up a "SplitT" generics variable binding so that we can
* construct the SourceOperator without resorting to "all raw types". That way, this methods
* puts all "type non-safety" in one place and allows to maintain as much generics safety in the
... | 3.26 |
flink_HsSubpartitionMemoryDataManager_canBeCompressed_rdh | /**
* Whether the buffer can be compressed or not. Note that event is not compressed because it is
* usually small and the size can become even larger after compression.
*/
private boolean canBeCompressed(Buffer buffer) {
return ((bufferCompressor != null) && buffer.isBuffer()) && (buffer.readableBytes() > 0);
} | 3.26 |
flink_HsSubpartitionMemoryDataManager_getBuffersSatisfyStatus_rdh | /**
* Get buffers in {@link #allBuffers} that satisfy expected {@link SpillStatus} and {@link ConsumeStatus}.
*
* @param spillStatus
* the status of spilling expected.
* @param consumeStatusWithId
* the status and consumerId expected.
* @return buffers satisfy expected status in order.
*/
// Note that: call... | 3.26 |
flink_HsSubpartitionMemoryDataManager_writeEvent_rdh | // ------------------------------------------------------------------------
// Internal Methods
// ------------------------------------------------------------------------
private void writeEvent(ByteBuffer event, DataType dataType) {
checkArgument(dataType.isEvent());
// each Event must take an exclusive buf... | 3.26 |
flink_HsSubpartitionMemoryDataManager_releaseSubpartitionBuffers_rdh | /**
* Release this subpartition's buffers in a decision.
*
* @param toRelease
* All buffers that need to be released belong to this subpartition in a
* decision.
*/
// Note that: runWithLock ensure that code block guarded by resultPartitionReadLock and
// subpartitionLock.
@SuppressWarnings("FieldAccessNotGua... | 3.26 |
flink_HsSubpartitionMemoryDataManager_append_rdh | // ------------------------------------------------------------------------
// Called by MemoryDataManager
// ------------------------------------------------------------------------
/**
* Append record to {@link HsSubpartitionMemoryDataManager}.
*
* @param record
* to be managed by this class.
* @param dataType... | 3.26 |
flink_HsSubpartitionMemoryDataManager_trimHeadingReleasedBuffers_rdh | /**
* Remove all released buffer from head of queue until buffer queue is empty or meet un-released
* buffer.
*/
@GuardedBy("subpartitionLock")
private void trimHeadingReleasedBuffers(Deque<HsBufferContext> bufferQueue) {
while ((!bufferQueue.isEmpty()) && buff... | 3.26 |
flink_HsSubpartitionMemoryDataManager_addFinishedBuffer_rdh | // Note that: callWithLock ensure that code block guarded by resultPartitionReadLock and
// subpartitionLock.
@SuppressWarnings("FieldAccessNotGuarded")
private void addFinishedBuffer(HsBufferContext bufferContext) {
finishedBufferIndex++;
... | 3.26 |
flink_HsSubpartitionMemoryDataManager_spillSubpartitionBuffers_rdh | /**
* Spill this subpartition's buffers in a decision.
*
* @param toSpill
* All buffers that need to be spilled belong to this subpartition in a decision.
* @param spillDoneFuture
* completed when spill is finished.
* @return {@link BufferWithIdentity}s about these spill buffers.
*/
// Note that: callWithLo... | 3.26 |
flink_ListDelimitedSerializer_deserializeNextElement_rdh | /**
* Deserializes a single element from a serialized list.
*/
public static <T> T
deserializeNextElement(DataInputDeserializer in, TypeSerializer<T> elementSerializer) throws IOException {
if (in.available() > 0) {
T element = elementSerializer.deserialize(in);if (in.available() > 0) {
in.r... | 3.26 |
flink_Tuple22_equals_rdh | /**
* Deep equality for tuples by calling equals() on the tuple members.
*
* @param o
* the object checked for equality
* @return true if this is equal to o.
*/
@Override
public boolean equals(Object o) {
if (this == o) {
return true;
}
if (!(o instanceof Tuple22)) {return false;
}@Supp... | 3.26 |
flink_Tuple22_toString_rdh | // -------------------------------------------------------------------------------------------------
// standard utilities
// -------------------------------------------------------------------------------------------------
/**
* Creates a string representation of the tuple in the form (f0, f1, f2, f3, f4, f5, f6, f7,... | 3.26 |
flink_Tuple22_of_rdh | /**
* Creates a new tuple and assigns the given values to the tuple's fields. This is more
* convenient than using the constructor, because the compiler can infer the generic type
* arguments implicitly. For example: {@code Tuple3.of(n, x, s)} instead of {@code new
* Tuple3<Integer, Double, String>(n, x, s)}
*/
pu... | 3.26 |
flink_Tuple22_copy_rdh | /**
* Shallow tuple copy.
*
* @return A new Tuple with the same fields as this.
*/
@Override
@SuppressWarnings("unchecked")
public Tuple22<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15,
T16, T17, T18, T19, T20, T21> copy() {
return new Tuple22<>(this.f0, this.f1, this.f2, this.f3, this.f4, ... | 3.26 |
flink_Tuple22_setFields_rdh | /**
* Sets new values to all fields of the tuple.
*
* @param f0
* The value for field 0
* @param f1
* The value for field 1
* @param f2
* The value for field 2
* @param f3
* The value for field 3
* @param f4
* The value for field 4
* @param f5
* The value for field 5
* @param f6
* The valu... | 3.26 |
flink_RegisteredBroadcastStateBackendMetaInfo_deepCopy_rdh | /**
* Creates a deep copy of the itself.
*/
@Nonnull
public RegisteredBroadcastStateBackendMetaInfo<K, V> deepCopy() {
return new RegisteredBroadcastStateBackendMetaInfo<>(this);
} | 3.26 |
flink_AbstractBinaryExternalMerger_getMergingIterator_rdh | /**
* Returns an iterator that iterates over the merged result from all given channels.
*
* @param channelIDs
* The channels that are to be merged and returned.
* @return An iterator over the merged records of the input channels.
* @throws IOException
* Thrown, if the readers encounter an I/O problem.
*/
pu... | 3.26 |
flink_AbstractBinaryExternalMerger_mergeChannels_rdh | /**
* Merges the sorted runs described by the given Channel IDs into a single sorted run.
*
* @param channelIDs
* The IDs of the runs' channels.
* @return The ID and number of blocks of the channel that describes the merged run.
*/
private ChannelWithMeta mergeChannels(List<ChannelWithMeta> channelIDs) throws... | 3.26 |
flink_AbstractBinaryExternalMerger_mergeChannelList_rdh | /**
* Merges the given sorted runs to a smaller number of sorted runs.
*
* @param channelIDs
* The IDs of the sorted runs that need to be merged.
* @return A list of the IDs of the merged channels.
* @throws IOException
* Thrown, if the readers or writers encountered an I/O problem.
*/
public List<ChannelWi... | 3.26 |
flink_RocksDBConfigurableOptions_checkArgumentValid_rdh | /**
* Helper method to check whether the (key,value) is valid through given configuration and
* returns the formatted value.
*
* @param option
* The configuration key which is configurable in {@link RocksDBConfigurableOptions}.
* @param value
* The value within given configuration.
*/
static void checkArgum... | 3.26 |
flink_TypeSerializerSnapshotSerializationUtil_deserializeV2_rdh | /**
* Deserialization path for Flink versions 1.7+.
*/
@VisibleForTesting
static <T> TypeSerializerSnapshot<T> deserializeV2(DataInputView in, ClassLoader cl) throws IOException {
return TypeSerializerSnapshot.readVersionedSnapshot(in, cl);
} | 3.26 |
flink_TypeSerializerSnapshotSerializationUtil_writeSerializerSnapshot_rdh | /**
* Writes a {@link TypeSerializerSnapshot} to the provided data output view.
*
* <p>It is written with a format that can be later read again using {@link #readSerializerSnapshot(DataInputView, ClassLoader)}.
*
* @param out
* the data output view
* @param serializerSnapshot
* the serializer configuration ... | 3.26 |
flink_MessageSerializer_deserializeServerFailure_rdh | /**
* De-serializes the failure message sent to the {@link org.apache.flink.queryablestate.network.Client} in case of server related errors.
*
* <pre>
* <b>The buffer is expected to be at the correct position.</b>
* </pre>
*
* @param buf
* The {@link ByteBuf} containing the serialized failure message.
* @re... | 3.26 |
flink_MessageSerializer_deserializeRequest_rdh | /**
* De-serializes the request sent to the {@link org.apache.flink.queryablestate.network.AbstractServerBase}.
*
* <pre>
* <b>The buffer is expected to be at the request position.</b>
* </pre>
*
* @param buf
* The {@link ByteBuf} containing the serialized request.
* @return The request.
*/public REQ deser... | 3.26 |
flink_MessageSerializer_deserializeRequestFailure_rdh | /**
* De-serializes the {@link RequestFailure} sent to the {@link org.apache.flink.queryablestate.network.Client} in case of protocol related errors.
*
* <pre>
* <b>The buffer is expected to be at the correct position.</b>
* </pre>
*
* @param buf
* The {@link ByteBuf} containing the serialized failure messag... | 3.26 |
flink_MessageSerializer_writePayload_rdh | /**
* Helper for serializing the messages.
*
* @param alloc
* The {@link ByteBufAllocator} used to allocate the buffer to serialize the
* message into.
* @param requestId
* The id of the request to which the message refers to.
* @param messageType
* The {@link MessageType type of the message}.
* @para... | 3.26 |
flink_MessageSerializer_m1_rdh | /**
* Serializes the exception containing the failure message sent to the {@link org.apache.flink.queryablestate.network.Client} in case of protocol related errors.
*
* @param alloc
* The {@link ByteBufAllocator} used to allocate the buffer to serialize the
* message into.
* @param requestId
* The id of th... | 3.26 |
flink_MessageSerializer_serializeRequest_rdh | // ------------------------------------------------------------------------
// Serialization
// ------------------------------------------------------------------------
/**
* Serializes the request sent to the {@link org.apache.flink.queryablestate.network.AbstractServerBase}.
*
* @param alloc
* The {@link ByteBu... | 3.26 |
flink_MessageSerializer_m0_rdh | /**
* Serializes the response sent to the {@link org.apache.flink.queryablestate.network.Client}.
*
* @param alloc
* The {@link ByteBufAllocator} used to allocate the buffer to serialize the
* message into.
* @param requestId
* The id of the request to which the message refers to.
* @param response
* T... | 3.26 |
flink_MessageSerializer_getRequestId_rdh | /**
* De-serializes the header and returns the {@link MessageType}.
*
* <pre>
* <b>The buffer is expected to be at the request id position.</b>
* </pre>
*
* @param buf
* The {@link ByteBuf} containing the serialized request id.
* @return The request id.
*/
public static long getRequestId(final ByteBuf buf)... | 3.26 |
flink_MessageSerializer_deserializeResponse_rdh | /**
* De-serializes the response sent to the {@link org.apache.flink.queryablestate.network.Client}.
*
* <pre>
* <b>The buffer is expected to be at the response position.</b>
* </pre>
*
* @param buf
* The {@link ByteBuf} containing the serialized response.
* @return The response.
*/
public RESP
deserialize... | 3.26 |
flink_MessageSerializer_serializeServerFailure_rdh | /**
* Serializes the failure message sent to the {@link org.apache.flink.queryablestate.network.Client} in case of server related errors.
*
* @param alloc
* The {@link ByteBufAllocator} used to allocate the buffer to serialize the
* message into.
* @param cause
* The exception thrown at the server.
* @ret... | 3.26 |
flink_MessageSerializer_writeHeader_rdh | /**
* Helper for serializing the header.
*
* @param buf
* The {@link ByteBuf} to serialize the header into.
* @param messageType
* The {@link MessageType} of the message this header refers to.
*/
private static void writeHeader(final ByteBuf buf, final MessageType messageType) {
buf.writeInt(VERSION);
... | 3.26 |
flink_MessageSerializer_deserializeHeader_rdh | // ------------------------------------------------------------------------
// Deserialization
// ------------------------------------------------------------------------
/**
* De-serializes the header and returns the {@link MessageType}.
*
* <pre>
* <b>The buffer is expected to be at the header position.</b>
* <... | 3.26 |
flink_InternalServiceDecorator_getNamespacedInternalServiceName_rdh | /**
* Generate namespaced name of the internal Service.
*/
public static String getNamespacedInternalServiceName(String clusterId, String namespace) {
return (getInternalServiceName(clusterId) + ".") + namespace;
} | 3.26 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.