name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
flink_LocalTimeTypeInfo_instantiateComparator_rdh
// -------------------------------------------------------------------------------------------- private static <X> TypeComparator<X> instantiateComparator(Class<? extends TypeComparator<X>> comparatorClass, boolean ascendingOrder) { try { Constructor<? extends TypeComparator<X>> constructor = comparatorClas...
3.26
flink_WrappingRuntimeException_unwrap_rdh
/** * Recursively unwraps this WrappingRuntimeException and its causes, getting the first non * wrapping exception. * * @return The first cause that is not a wrapping exception. */ public Throwable unwrap() { Throwable cause = getCause(); return cause instanceof WrappingRuntimeException ? ((WrappingRunt...
3.26
flink_GenericRowData_setField_rdh
/** * Sets the field value at the given position. * * <p>Note: The given field value must be an internal data structures. Otherwise the {@link GenericRowData} is corrupted and may throw exception when processing. See {@link RowData} for * more information about internal data structures. * ...
3.26
flink_GenericRowData_ofKind_rdh
/** * Creates an instance of {@link GenericRowData} with given kind and field values. * * <p>Note: All fields of the row must be internal data structures. */ public static GenericRowData ofKind(RowKind kind, Object... values) { GenericRowData row = new GenericRowData(kind, values.length); for (int i = 0; i ...
3.26
flink_GenericRowData_getField_rdh
/** * Returns the field value at the given position. * * <p>Note: The returned value is in internal data structure. See {@link RowData} for more * information about internal data structures. * * <p>The returned field value can be null for representing nullability. */ public Object getField(int pos) { return ...
3.26
flink_ContinuousFileMonitoringFunction_listEligibleFiles_rdh
/** * Returns the paths of the files not yet processed. * * @param fileSystem * The filesystem where the monitored directory resides. */ private Map<Path, FileStatus> listEligibleFiles(FileSystem fileSystem, Path path) { final FileStatus[] statuses; try { statuses = fileSystem.listStatus(path); ...
3.26
flink_ContinuousFileMonitoringFunction_shouldIgnore_rdh
/** * Returns {@code true} if the file is NOT to be processed further. This happens if the * modification time of the file is smaller than the {@link #globalModificationTime}. * * @param filePath * the path of the file to check. * @param modificationTime * the modification time of the file. */ private boole...
3.26
flink_ContinuousFileMonitoringFunction_getInputSplitsSortedByModTime_rdh
/** * Creates the input splits to be forwarded to the downstream tasks of the {@link ContinuousFileReaderOperator}. Splits are sorted <b>by modification time</b> before being * forwarded and only splits belonging to files in the {@code eligibleFiles} list will be * processed. * * @param eligibleFiles * The file...
3.26
flink_ContinuousFileMonitoringFunction_snapshotState_rdh
// --------------------- Checkpointing -------------------------- @Override public void snapshotState(FunctionSnapshotContext context) throws Exception { Preconditions.checkState(this.checkpointedState != null, ("The " + getClass().getSimpleName()) + " state has not been properly initialized."); this.check...
3.26
flink_DecimalData_fromBigDecimal_rdh
// ------------------------------------------------------------------------------------------ // Constructor Utilities // ------------------------------------------------------------------------------------------ /** * Creates an instance of {@link DecimalData} from a {@link BigDecimal} and the given precision * and ...
3.26
flink_DecimalData_isCompact_rdh
// ------------------------------------------------------------------------------------------ // Utilities // ------------------------------------------------------------------------------------------ /** * Returns whether the decimal value is small enough to be stored in a long. */ public static boolean isCompact(in...
3.26
flink_DecimalData_toBigDecimal_rdh
/** * Converts this {@link DecimalData} into an instance of {@link BigDecimal}. */ public BigDecimal toBigDecimal() { BigDecimal bd = decimalVal; if (bd == null) { decimalVal = bd = BigDecimal.valueOf(longVal, scale); } return bd; }
3.26
flink_DecimalData_fromUnscaledBytes_rdh
/** * Creates an instance of {@link DecimalData} from an unscaled byte array value and the given * precision and scale. */public static DecimalData fromUnscaledBytes(byte[] unscaledBytes, int precision, int scale) { BigDecimal bd = new BigDecimal(new BigInteger(unscaledBytes), scale);return fromBigDecimal(bd, ...
3.26
flink_DecimalData_scale_rdh
/** * Returns the <i>scale</i> of this {@link DecimalData}. */ public int scale() { return scale; }
3.26
flink_DecimalData_toUnscaledLong_rdh
/** * Returns a long describing the <i>unscaled value</i> of this {@link DecimalData}. * * @throws ArithmeticException * if this {@link DecimalData} does not exactly fit in a long. */ public long toUnscaledLong() { if (isCompact()) { return longVal; } else { return toBigDecimal()...
3.26
flink_DecimalData_zero_rdh
/** * Creates an instance of {@link DecimalData} for a zero value with the given precision and * scale. * * <p>The precision will be checked. If the precision overflows, null will be returned. */ @Nullable public static DecimalData zero(int precision, int scale) { if (precision <= MAX_COMPACT_PRECISION) { ...
3.26
flink_DecimalData_fromUnscaledLong_rdh
/** * Creates an instance of {@link DecimalData} from an unscaled long value and the given * precision and scale. */ public static DecimalData fromUnscaledLong(long unscaledLong, int precision, int scale) { checkArgument((precision > 0) && (precision <= MAX_LONG_DIGITS)); return new DecimalData(precision, ...
3.26
flink_DecimalData_precision_rdh
// ------------------------------------------------------------------------------------------ // Public Interfaces // ------------------------------------------------------------------------------------------ /** * Returns the <i>precision</i> of this {@link DecimalData}. * * <p>The precision is the number of digits...
3.26
flink_DecimalData_copy_rdh
/** * Returns a copy of this {@link DecimalData} object. */ public DecimalData copy() { return new DecimalData(precision, scale, longVal, decimalVal); }
3.26
flink_BinaryRowWriter_reset_rdh
/** * First, reset. */ @Override public void reset() { this.cursor = fixedSize; for (int i = 0; i < nullBitsSizeInBytes; i += 8) { segment.putLong(i, 0L); } }
3.26
flink_BinaryRowWriter_setNullAt_rdh
/** * Default not null. */@Override public void setNullAt(int pos) { setNullBit(pos); segment.putLong(getFieldOffset(pos), 0L); }
3.26
flink_TransientBlobCleanupTask_run_rdh
/** * Cleans up transient BLOBs whose TTL is up, tolerating that files do not exist (anymore). */ @Override public void run() { // let's cache the current time - we do not operate on a millisecond precision anyway final long currentTimeMillis = System.currentTimeMillis(); // iterate through all entries ...
3.26
flink_RetryingRegistration_getFuture_rdh
// ------------------------------------------------------------------------ // completion and cancellation // ------------------------------------------------------------------------ public CompletableFuture<RetryingRegistrationResult<G, S, R>> getFuture() { return completionFuture; }
3.26
flink_RetryingRegistration_startRegistration_rdh
/** * This method resolves the target address to a callable gateway and starts the registration * after that. */ @SuppressWarnings("unchecked") public void startRegistration() { if (canceled) { // we already got canceled return; } try { // trigger resolutio...
3.26
flink_RetryingRegistration_cancel_rdh
/** * Cancels the registration procedure. */ public void cancel() { canceled = true; completionFuture.cancel(false); }
3.26
flink_RetryingRegistration_register_rdh
/** * This method performs a registration attempt and triggers either a success notification or a * retry, depending on the result. */ @SuppressWarnings("unchecked") private void register(final G g...
3.26
flink_LogicalType_is_rdh
/** * Returns whether the family type of the type equals to the {@code family} or not. * * @param family * The family type to check against for equality */ public boolean is(LogicalTypeFamily family) { return typeRoot.getFamilies().contains(family); }
3.26
flink_LogicalType_asSummaryString_rdh
/** * Returns a string that summarizes this type for printing to a console. An implementation might * shorten long names or skips very specific properties. * * <p>Use {@link #asSerializableString()} for a type string that fully serializes this instance. * * @return summary string of this type for debugging purpos...
3.26
flink_LogicalType_isAnyOf_rdh
/** * Returns whether the root of the type is part of at least one family of the {@code typeFamily} * or not. * * @param typeFamilies * The families to check against for equality */ public boolean isAnyOf(LogicalTypeFamily... typeFamilies) { return Arrays.stream(typeFamilies).anyMatch(tf -> this.typeRoot.ge...
3.26
flink_LogicalType_isNullable_rdh
/** * Returns whether a value of this type can be {@code null}. */ public boolean isNullable() { return isNullable; }
3.26
flink_FormatFactory_forwardOptions_rdh
/** * Returns a set of {@link ConfigOption} that are directly forwarded to the runtime * implementation but don't affect the final execution topology. * * <p>Options declared here can override options of the persisted plan during an enrichment * phase. Since a restored topology is static, a...
3.26
flink_AsynchronousJobOperationKey_getJobId_rdh
/** * Get the job id for the given operation key. * * @return job id */public JobID getJobId() {return jobId; }
3.26
flink_CopyableValueComparator_readObject_rdh
// -------------------------------------------------------------------------------------------- private void readObject(ObjectInputStream s) throws IOException, ClassNotFoundException { // read basic object and the type s.defaultReadObject(); this.reference = InstantiationUtil.instantiate(type, CopyableValu...
3.26
flink_CopyableValueComparator_supportsSerializationWithKeyNormalization_rdh
// -------------------------------------------------------------------------------------------- // unsupported normalization // -------------------------------------------------------------------------------------------- @Override public boolean supportsSerializationWithKeyNormalization() { return false; }
3.26
flink_UserFacingListState_get_rdh
// ------------------------------------------------------------------------ @Override public Iterable<T> get() throws Exception { Iterable<T> original = originalState.get(); return original != null ? original : emptyState; }
3.26
flink_Tuple13_toString_rdh
// ------------------------------------------------------------------------------------------------- // standard utilities // ------------------------------------------------------------------------------------------------- /** * Creates a string representation of the tuple in the form (f0, f1, f2, f3, f4, f5, f6, f7,...
3.26
flink_Tuple13_setFields_rdh
/** * Sets new values to all fields of the tuple. * * @param f0 * The value for field 0 * @param f1 * The value for field 1 * @param f2 * The value for field 2 * @param f3 * The value for field 3 * @param f4 * The value for field 4 * @param f5 * The value for field 5 * @param f6 * The valu...
3.26
flink_Tuple13_copy_rdh
/** * Shallow tuple copy. * * @return A new Tuple with the same fields as this. */ @Override @SuppressWarnings("unchecked") public Tuple13<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12> copy() { return new Tuple13<>(this.f0, this.f1, this.f2, this.f3, this.f4, this.f5, this.f6, this.f7, this.f8...
3.26
flink_Tuple13_of_rdh
/** * Creates a new tuple and assigns the given values to the tuple's fields. This is more * convenient than using the constructor, because the compiler can infer the generic type * arguments implicitly. For example: {@code Tuple3.of(n, x, s)} instead of {@code new * Tuple3<Integer, Double, String>(n, x, s)} */ pu...
3.26
flink_TaskExecutorProcessUtils_newProcessSpecBuilder_rdh
// ------------------------------------------------------------------------ // Memory Configuration Calculations // ------------------------------------------------------------------------ public static TaskExecutorProcessSpecBuilder newProcessSpecBuilder(final Configuration config) { return TaskExecutorProcessSpec...
3.26
flink_TaskExecutorProcessUtils_generateDynamicConfigsStr_rdh
// ------------------------------------------------------------------------ // Generating Dynamic Config Options // ------------------------------------------------------------------------ public static String generateDynamicConfigsStr(final TaskExecutorProcessSpec taskExecutorProcessSpec) { final Map<String, Strin...
3.26
flink_AsynchronousBlockWriter_getReturnQueue_rdh
/** * Gets the queue in which the memory segments are queued after the asynchronous write is * completed. * * @return The queue with the written memory segments. */ @Override public LinkedBlockingQueue<MemorySegment> getReturnQueue() { return this.returnSegments;}
3.26
flink_AsynchronousBlockWriter_getNextReturnedBlock_rdh
/** * Gets the next memory segment that has been written and is available again. This method blocks * until such a segment is available, or until an error occurs in the writer, or the writer is * closed. * * <p>NOTE: If this method is invoked without any segment ever returning (for example, because * the {@link #...
3.26
flink_EmptyIterator_remove_rdh
/** * Throws a {@link java.lang.UnsupportedOperationException}. * * @see java.util.Iterator#remove() */ @Override public void remove() { throw new UnsupportedOperationException(); }
3.26
flink_EmptyIterator_m0_rdh
/** * Always throws a {@link java.util.NoSuchElementException}. * * @see java.util.Iterator#next() */ @Override public E m0() { throw new NoSuchElementException();}
3.26
flink_EmptyIterator_get_rdh
/** * Gets a singleton instance of the empty iterator. * * @param <E> * The type of the objects (not) returned by the iterator. * @return An instance of the iterator. */ public static <E> EmptyIterator<E> get() { @SuppressWarnings("unchecked") EmptyIterator<E> iter = ((EmptyIterator<E>) (INSTANCE)); ...
3.26
flink_FlinkDatabaseMetaData_storesMixedCaseQuotedIdentifiers_rdh
/** * Flink sql is mixed case as sensitive. */ @Override public boolean storesMixedCaseQuotedIdentifiers() throws SQLException {return true; }
3.26
flink_FlinkDatabaseMetaData_m4_rdh
/** * Flink sql is mixed case as sensitive. */ @Override public boolean m4() throws SQLException { return true; }
3.26
flink_FlinkDatabaseMetaData_supportsMixedCaseIdentifiers_rdh
/** * Flink sql is mixed case as sensitive. */ @Override public boolean supportsMixedCaseIdentifiers() throws SQLException {return true; }
3.26
flink_FlinkDatabaseMetaData_nullsAreSortedLow_rdh
/** * In flink null value will be used as low value for sort. */@Override public boolean nullsAreSortedLow() throws SQLException { return true; }
3.26
flink_FlinkDatabaseMetaData_nullPlusNonNullIsNull_rdh
/** * Null value plus non-null in flink will be null result. */ @Override public boolean nullPlusNonNullIsNull() throws SQLException { return true; }
3.26
flink_FlinkDatabaseMetaData_supportsMixedCaseQuotedIdentifiers_rdh
/** * Flink sql is mixed case as sensitive. */ @Override public boolean supportsMixedCaseQuotedIdentifiers() throws SQLException { return true; }
3.26
flink_FlinkDatabaseMetaData_isCatalogAtStart_rdh
/** * Catalog name appears at the start of full name. */ @Override public boolean isCatalogAtStart() throws SQLException { return true; }
3.26
flink_FlinkDatabaseMetaData_getSchemas_rdh
// TODO Flink will support SHOW DATABASES LIKE statement in FLIP-297, this method will be // supported after that issue. @Override public ResultSet getSchemas(String catalog, String schemaPattern) throws SQLException {throw new UnsupportedOperationException(); }
3.26
flink_DefaultCostEstimator_addCachedHybridHashCosts_rdh
/** * Calculates the costs for the cached variant of the hybrid hash join. We are assuming by * default that half of the cached hash table fit into memory. */ @Override public void addCachedHybridHashCosts(EstimateProvider buildSideInput, EstimateProvider probeSideInput, Costs costs, int costWeight) { if (costW...
3.26
flink_DefaultCostEstimator_addFileInputCost_rdh
// -------------------------------------------------------------------------------------------- // Local Strategy Cost // -------------------------------------------------------------------------------------------- @Override public void addFileInputCost(long fileSizeInBytes, Costs costs) { if (fileSizeInBytes >= ...
3.26
flink_DefaultCostEstimator_addArtificialDamCost_rdh
// -------------------------------------------------------------------------------------------- // Damming Cost // -------------------------------------------------------------------------------------------- @Override public void addArtificialDamCost(EstimateProvider estimates, long bufferSize, Costs costs) { final...
3.26
flink_DefaultCostEstimator_addRandomPartitioningCost_rdh
// -------------------------------------------------------------------------------------------- @Override public void addRandomPartitioningCost(EstimateProvider estimates, Costs costs) { // conservative estimate: we need ship the whole data over the network to establish the // partitioning. no disk costs. ...
3.26
flink_SlotStatus_getResourceProfile_rdh
/** * Get the resource profile of this slot. * * @return The resource profile */ public ResourceProfile getResourceProfile() { return f0; }
3.26
flink_SlotStatus_getAllocationID_rdh
/** * Get the allocation id of this slot. * * @return The allocation id if this slot is allocated, otherwise null */ public AllocationID getAllocationID() { return allocationID; }
3.26
flink_BlockResettableMutableObjectIterator_next_rdh
// -------------------------------------------------------------------------------------------- @Override public T next(T target) throws IOException { // check for the left over element if (this.readPhase) { return getNextRecord(target);} else // writing phase. check for leftover first if (this.le...
3.26
flink_BlockResettableMutableObjectIterator_hasFurtherInput_rdh
/** * Checks, whether the input that is blocked by this iterator, has further elements available. * This method may be used to forecast (for example at the point where a block is full) whether * there will be more data (possibly in another block). * * @return True, if there will be more data, false otherwise. */ ...
3.26
flink_AdvancedFunctionsExample_executeLastDatedValueFunction_rdh
/** * Aggregates data by name and returns the latest non-null {@code item_count} value with its * corresponding {@code order_date}. */ private static void executeLastDatedValueFunction(TableEnvironment env) { // create a table with example data final Table customers = env.fromValues(DataTypes.of("ROW<name ST...
3.26
flink_CatalogTableImpl_removeRedundant_rdh
/** * Construct catalog table properties from {@link #toProperties()}. */ public static Map<String, String> removeRedundant(Map<String, String> properties, TableSchema schema, List<String> partitionKeys) { Map<String, String> ret = new HashMap<>(properties); DescriptorProperties descriptorProperties = new D...
3.26
flink_CatalogTableImpl_fromProperties_rdh
/** * Construct a {@link CatalogTableImpl} from complete properties that contains table schema. */ public static CatalogTableImpl fromProperties(Map<String, String> properties) { DescriptorProperties descriptorProperties = new DescriptorProperties(false); descriptorProperties.putProperties(properties); Ta...
3.26
flink_SourceOperatorFactory_instantiateSourceOperator_rdh
/** * This is a utility method to conjure up a "SplitT" generics variable binding so that we can * construct the SourceOperator without resorting to "all raw types". That way, this methods * puts all "type non-safety" in one place and allows to maintain as much generics safety in the ...
3.26
flink_HsSubpartitionMemoryDataManager_canBeCompressed_rdh
/** * Whether the buffer can be compressed or not. Note that event is not compressed because it is * usually small and the size can become even larger after compression. */ private boolean canBeCompressed(Buffer buffer) { return ((bufferCompressor != null) && buffer.isBuffer()) && (buffer.readableBytes() > 0); }
3.26
flink_HsSubpartitionMemoryDataManager_getBuffersSatisfyStatus_rdh
/** * Get buffers in {@link #allBuffers} that satisfy expected {@link SpillStatus} and {@link ConsumeStatus}. * * @param spillStatus * the status of spilling expected. * @param consumeStatusWithId * the status and consumerId expected. * @return buffers satisfy expected status in order. */ // Note that: call...
3.26
flink_HsSubpartitionMemoryDataManager_writeEvent_rdh
// ------------------------------------------------------------------------ // Internal Methods // ------------------------------------------------------------------------ private void writeEvent(ByteBuffer event, DataType dataType) { checkArgument(dataType.isEvent()); // each Event must take an exclusive buf...
3.26
flink_HsSubpartitionMemoryDataManager_releaseSubpartitionBuffers_rdh
/** * Release this subpartition's buffers in a decision. * * @param toRelease * All buffers that need to be released belong to this subpartition in a * decision. */ // Note that: runWithLock ensure that code block guarded by resultPartitionReadLock and // subpartitionLock. @SuppressWarnings("FieldAccessNotGua...
3.26
flink_HsSubpartitionMemoryDataManager_append_rdh
// ------------------------------------------------------------------------ // Called by MemoryDataManager // ------------------------------------------------------------------------ /** * Append record to {@link HsSubpartitionMemoryDataManager}. * * @param record * to be managed by this class. * @param dataType...
3.26
flink_HsSubpartitionMemoryDataManager_trimHeadingReleasedBuffers_rdh
/** * Remove all released buffer from head of queue until buffer queue is empty or meet un-released * buffer. */ @GuardedBy("subpartitionLock") private void trimHeadingReleasedBuffers(Deque<HsBufferContext> bufferQueue) { while ((!bufferQueue.isEmpty()) && buff...
3.26
flink_HsSubpartitionMemoryDataManager_addFinishedBuffer_rdh
// Note that: callWithLock ensure that code block guarded by resultPartitionReadLock and // subpartitionLock. @SuppressWarnings("FieldAccessNotGuarded") private void addFinishedBuffer(HsBufferContext bufferContext) { finishedBufferIndex++; ...
3.26
flink_HsSubpartitionMemoryDataManager_spillSubpartitionBuffers_rdh
/** * Spill this subpartition's buffers in a decision. * * @param toSpill * All buffers that need to be spilled belong to this subpartition in a decision. * @param spillDoneFuture * completed when spill is finished. * @return {@link BufferWithIdentity}s about these spill buffers. */ // Note that: callWithLo...
3.26
flink_ListDelimitedSerializer_deserializeNextElement_rdh
/** * Deserializes a single element from a serialized list. */ public static <T> T deserializeNextElement(DataInputDeserializer in, TypeSerializer<T> elementSerializer) throws IOException { if (in.available() > 0) { T element = elementSerializer.deserialize(in);if (in.available() > 0) { in.r...
3.26
flink_Tuple22_equals_rdh
/** * Deep equality for tuples by calling equals() on the tuple members. * * @param o * the object checked for equality * @return true if this is equal to o. */ @Override public boolean equals(Object o) { if (this == o) { return true; } if (!(o instanceof Tuple22)) {return false; }@Supp...
3.26
flink_Tuple22_toString_rdh
// ------------------------------------------------------------------------------------------------- // standard utilities // ------------------------------------------------------------------------------------------------- /** * Creates a string representation of the tuple in the form (f0, f1, f2, f3, f4, f5, f6, f7,...
3.26
flink_Tuple22_of_rdh
/** * Creates a new tuple and assigns the given values to the tuple's fields. This is more * convenient than using the constructor, because the compiler can infer the generic type * arguments implicitly. For example: {@code Tuple3.of(n, x, s)} instead of {@code new * Tuple3<Integer, Double, String>(n, x, s)} */ pu...
3.26
flink_Tuple22_copy_rdh
/** * Shallow tuple copy. * * @return A new Tuple with the same fields as this. */ @Override @SuppressWarnings("unchecked") public Tuple22<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21> copy() { return new Tuple22<>(this.f0, this.f1, this.f2, this.f3, this.f4, ...
3.26
flink_Tuple22_setFields_rdh
/** * Sets new values to all fields of the tuple. * * @param f0 * The value for field 0 * @param f1 * The value for field 1 * @param f2 * The value for field 2 * @param f3 * The value for field 3 * @param f4 * The value for field 4 * @param f5 * The value for field 5 * @param f6 * The valu...
3.26
flink_RegisteredBroadcastStateBackendMetaInfo_deepCopy_rdh
/** * Creates a deep copy of the itself. */ @Nonnull public RegisteredBroadcastStateBackendMetaInfo<K, V> deepCopy() { return new RegisteredBroadcastStateBackendMetaInfo<>(this); }
3.26
flink_AbstractBinaryExternalMerger_getMergingIterator_rdh
/** * Returns an iterator that iterates over the merged result from all given channels. * * @param channelIDs * The channels that are to be merged and returned. * @return An iterator over the merged records of the input channels. * @throws IOException * Thrown, if the readers encounter an I/O problem. */ pu...
3.26
flink_AbstractBinaryExternalMerger_mergeChannels_rdh
/** * Merges the sorted runs described by the given Channel IDs into a single sorted run. * * @param channelIDs * The IDs of the runs' channels. * @return The ID and number of blocks of the channel that describes the merged run. */ private ChannelWithMeta mergeChannels(List<ChannelWithMeta> channelIDs) throws...
3.26
flink_AbstractBinaryExternalMerger_mergeChannelList_rdh
/** * Merges the given sorted runs to a smaller number of sorted runs. * * @param channelIDs * The IDs of the sorted runs that need to be merged. * @return A list of the IDs of the merged channels. * @throws IOException * Thrown, if the readers or writers encountered an I/O problem. */ public List<ChannelWi...
3.26
flink_RocksDBConfigurableOptions_checkArgumentValid_rdh
/** * Helper method to check whether the (key,value) is valid through given configuration and * returns the formatted value. * * @param option * The configuration key which is configurable in {@link RocksDBConfigurableOptions}. * @param value * The value within given configuration. */ static void checkArgum...
3.26
flink_TypeSerializerSnapshotSerializationUtil_deserializeV2_rdh
/** * Deserialization path for Flink versions 1.7+. */ @VisibleForTesting static <T> TypeSerializerSnapshot<T> deserializeV2(DataInputView in, ClassLoader cl) throws IOException { return TypeSerializerSnapshot.readVersionedSnapshot(in, cl); }
3.26
flink_TypeSerializerSnapshotSerializationUtil_writeSerializerSnapshot_rdh
/** * Writes a {@link TypeSerializerSnapshot} to the provided data output view. * * <p>It is written with a format that can be later read again using {@link #readSerializerSnapshot(DataInputView, ClassLoader)}. * * @param out * the data output view * @param serializerSnapshot * the serializer configuration ...
3.26
flink_MessageSerializer_deserializeServerFailure_rdh
/** * De-serializes the failure message sent to the {@link org.apache.flink.queryablestate.network.Client} in case of server related errors. * * <pre> * <b>The buffer is expected to be at the correct position.</b> * </pre> * * @param buf * The {@link ByteBuf} containing the serialized failure message. * @re...
3.26
flink_MessageSerializer_deserializeRequest_rdh
/** * De-serializes the request sent to the {@link org.apache.flink.queryablestate.network.AbstractServerBase}. * * <pre> * <b>The buffer is expected to be at the request position.</b> * </pre> * * @param buf * The {@link ByteBuf} containing the serialized request. * @return The request. */public REQ deser...
3.26
flink_MessageSerializer_deserializeRequestFailure_rdh
/** * De-serializes the {@link RequestFailure} sent to the {@link org.apache.flink.queryablestate.network.Client} in case of protocol related errors. * * <pre> * <b>The buffer is expected to be at the correct position.</b> * </pre> * * @param buf * The {@link ByteBuf} containing the serialized failure messag...
3.26
flink_MessageSerializer_writePayload_rdh
/** * Helper for serializing the messages. * * @param alloc * The {@link ByteBufAllocator} used to allocate the buffer to serialize the * message into. * @param requestId * The id of the request to which the message refers to. * @param messageType * The {@link MessageType type of the message}. * @para...
3.26
flink_MessageSerializer_m1_rdh
/** * Serializes the exception containing the failure message sent to the {@link org.apache.flink.queryablestate.network.Client} in case of protocol related errors. * * @param alloc * The {@link ByteBufAllocator} used to allocate the buffer to serialize the * message into. * @param requestId * The id of th...
3.26
flink_MessageSerializer_serializeRequest_rdh
// ------------------------------------------------------------------------ // Serialization // ------------------------------------------------------------------------ /** * Serializes the request sent to the {@link org.apache.flink.queryablestate.network.AbstractServerBase}. * * @param alloc * The {@link ByteBu...
3.26
flink_MessageSerializer_m0_rdh
/** * Serializes the response sent to the {@link org.apache.flink.queryablestate.network.Client}. * * @param alloc * The {@link ByteBufAllocator} used to allocate the buffer to serialize the * message into. * @param requestId * The id of the request to which the message refers to. * @param response * T...
3.26
flink_MessageSerializer_getRequestId_rdh
/** * De-serializes the header and returns the {@link MessageType}. * * <pre> * <b>The buffer is expected to be at the request id position.</b> * </pre> * * @param buf * The {@link ByteBuf} containing the serialized request id. * @return The request id. */ public static long getRequestId(final ByteBuf buf)...
3.26
flink_MessageSerializer_deserializeResponse_rdh
/** * De-serializes the response sent to the {@link org.apache.flink.queryablestate.network.Client}. * * <pre> * <b>The buffer is expected to be at the response position.</b> * </pre> * * @param buf * The {@link ByteBuf} containing the serialized response. * @return The response. */ public RESP deserialize...
3.26
flink_MessageSerializer_serializeServerFailure_rdh
/** * Serializes the failure message sent to the {@link org.apache.flink.queryablestate.network.Client} in case of server related errors. * * @param alloc * The {@link ByteBufAllocator} used to allocate the buffer to serialize the * message into. * @param cause * The exception thrown at the server. * @ret...
3.26
flink_MessageSerializer_writeHeader_rdh
/** * Helper for serializing the header. * * @param buf * The {@link ByteBuf} to serialize the header into. * @param messageType * The {@link MessageType} of the message this header refers to. */ private static void writeHeader(final ByteBuf buf, final MessageType messageType) { buf.writeInt(VERSION); ...
3.26
flink_MessageSerializer_deserializeHeader_rdh
// ------------------------------------------------------------------------ // Deserialization // ------------------------------------------------------------------------ /** * De-serializes the header and returns the {@link MessageType}. * * <pre> * <b>The buffer is expected to be at the header position.</b> * <...
3.26
flink_InternalServiceDecorator_getNamespacedInternalServiceName_rdh
/** * Generate namespaced name of the internal Service. */ public static String getNamespacedInternalServiceName(String clusterId, String namespace) { return (getInternalServiceName(clusterId) + ".") + namespace; }
3.26