name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_CatalogSchemaTable_getContextResolvedTable_rdh | // ~ Methods ----------------------------------------------------------------
public ContextResolvedTable getContextResolvedTable() {
return contextResolvedTable;
} | 3.26 |
flink_DelegationTokenProvider_serviceConfigPrefix_rdh | /**
* Config prefix of the service.
*/
default String serviceConfigPrefix() {
return String.format("%s.%s", CONFIG_PREFIX, serviceName());
} | 3.26 |
flink_RuntimeRestAPIVersion_isStableVersion_rdh | /**
* Returns whether this version is considered stable.
*
* @return whether this version is stable
*/
@Override
public boolean isStableVersion() {
return isStable;
} | 3.26 |
flink_SinkModifyOperation_getTargetColumns_rdh | /**
* return null when no column list specified.
*/
@Nullable
public int[][] getTargetColumns() {
return targetColumns;
} | 3.26 |
flink_DefaultLookupCache_maximumSize_rdh | /**
* Specifies the maximum number of entries of the cache.
*/
public Builder maximumSize(long maximumSize) {
this.maximumSize = maximumSize;
return this;
} | 3.26 |
flink_DefaultLookupCache_newBuilder_rdh | /**
* Creates a builder for the cache.
*/
public static Builder newBuilder() {
return new Builder();
} | 3.26 |
flink_DefaultLookupCache_expireAfterAccess_rdh | /**
* Specifies the duration after an entry is last accessed that it should be automatically
* removed.
*/
public Builder expireAfterAccess(Duration duration) {
expireAfterAccessDuration = duration;
return this;
} | 3.26 |
flink_DefaultLookupCache_cacheMissingKey_rdh | /**
* Specifies whether to cache empty value into the cache.
*
* <p>Please note that "empty" means a collection without any rows in it instead of null.
* The cache will not accept any null key or value.
*/
public Builder cacheMissingKey(boolean cacheMissingKey) {
this.cacheMissingKey = cacheMissingKey;
return this... | 3.26 |
flink_DefaultLookupCache_expireAfterWrite_rdh | /**
* Specifies the duration after an entry is created that it should be automatically removed.
*/
public Builder expireAfterWrite(Duration duration) {
expireAfterWriteDuration = duration;
return this;
} | 3.26 |
flink_DefaultLookupCache_build_rdh | /**
* Creates the cache.
*/
public DefaultLookupCache build() {
return new DefaultLookupCache(expireAfterAccessDuration, expireAfterWriteDuration, maximumSize, cacheMissingKey);
} | 3.26 |
flink_BooleanColumnSummary_getNonNullCount_rdh | /**
* The number of non-null values in this column.
*/
@Override
public long getNonNullCount() {
return trueCount + falseCount;
} | 3.26 |
flink_RichAndCondition_getLeft_rdh | /**
*
* @return One of the {@link IterativeCondition conditions} combined in this condition.
*/
public IterativeCondition<T> getLeft() {
return getNestedConditions()[0];
} | 3.26 |
flink_RichAndCondition_getRight_rdh | /**
*
* @return One of the {@link IterativeCondition conditions} combined in this condition.
*/
public IterativeCondition<T> getRight() {
return getNestedConditions()[1];
} | 3.26 |
flink_SchemaValidator_deriveProctimeAttribute_rdh | /**
* Finds the proctime attribute if defined.
*/
public static Optional<String> deriveProctimeAttribute(DescriptorProperties properties) {
Map<String, String> names = properties.getIndexedProperty(SCHEMA, SCHEMA_NAME);
for (int i = 0; i < names.size(); i++) {Optional<Boolean> isProctime = properties.getOpti... | 3.26 |
flink_SchemaValidator_deriveFieldMapping_rdh | /**
* Finds a table source field mapping.
*
* @param properties
* The properties describing a schema.
* @param inputType
* The input type that a connector and/or format produces. This parameter can
* be used to resolve a rowtime field against an input field.
*/
public static Map<String, String> deriveFiel... | 3.26 |
flink_SchemaValidator_deriveRowtimeAttributes_rdh | /**
* Finds the rowtime attributes if defined.
*/
public static List<RowtimeAttributeDescriptor> deriveRowtimeAttributes(DescriptorProperties properties) {
Map<String, String> names = properties.getIndexedProperty(SCHEMA, SCHEMA_NAME);
List<RowtimeAttributeDescriptor> attributes = new ArrayList<>();
// ch... | 3.26 |
flink_CharValue_setValue_rdh | /**
* Sets the encapsulated char to the specified value.
*
* @param value
* the new value of the encapsulated char.
*/
public void setValue(char value) {
this.value = value;
} | 3.26 |
flink_CharValue_compareTo_rdh | // --------------------------------------------------------------------------------------------
@Override
public int compareTo(CharValue o) {
final int other = o.value;
return this.value < other ? -1 : this.value > other ? 1 : 0;
} | 3.26 |
flink_CharValue_read_rdh | // --------------------------------------------------------------------------------------------
@Override
public void read(DataInputView in) throws IOException {
this.value = in.readChar();
} | 3.26 |
flink_CharValue_getMaxNormalizedKeyLen_rdh | // --------------------------------------------------------------------------------------------
@Override
public int getMaxNormalizedKeyLen() {
return 2;
} | 3.26 |
flink_SlideWithSize_every_rdh | /**
* Specifies the window's slide as time or row-count interval.
*
* <p>The slide determines the interval in which windows are started. Hence, sliding windows can
* overlap if the slide is smaller than the size of the window.
*
* <p>For example, you could have windows of size 15 minutes that slide by 3 minutes. ... | 3.26 |
flink_ContinuousFileReaderOperator_prepareToProcessRecord_rdh | // the split was added and message to itself was enqueued to process it
public <T extends TimestampedInputSplit> boolean prepareToProcessRecord(ContinuousFileReaderOperator<?, T> op) throws IOException {if
(op.splits.isEmpty()) {
op.switchState(ReaderState.IDLE);
return false;
} else {
o... | 3.26 |
flink_KvStateClientProxyHandler_getKvStateLookupInfo_rdh | /**
* Lookup the {@link KvStateLocation} for the given job and queryable state name.
*
* <p>The job manager will be queried for the location only if forced or no cached location can
* be found. There are no guarantees about
*
* @param jobId
* JobID the state instance belongs to.
* @param queryableStateName
*... | 3.26 |
flink_AbstractKeyedStateBackend_getNumberOfKeyGroups_rdh | /**
*
* @see KeyedStateBackend
*/
public int getNumberOfKeyGroups() {
return numberOfKeyGroups;
} | 3.26 |
flink_AbstractKeyedStateBackend_getKeyGroupRange_rdh | /**
*
* @see KeyedStateBackend
*/
@Override
public KeyGroupRange getKeyGroupRange() {return keyGroupRange;
} | 3.26 |
flink_AbstractKeyedStateBackend_getCurrentKey_rdh | /**
*
* @see KeyedStateBackend
*/
@Override
public K getCurrentKey() {
return this.keyContext.getCurrentKey();
} | 3.26 |
flink_AbstractKeyedStateBackend_dispose_rdh | /**
* Closes the state backend, releasing all internal resources, but does not delete any
* persistent checkpoint data.
*/
@Override
public void dispose()
{
IOUtils.closeQuietly(cancelStreamRegistry);if (kvStateRegistry != null) {
kvStateRegistry.unregisterAll();
}
lastName = null;
lastSta... | 3.26 |
flink_AbstractKeyedStateBackend_getOrCreateKeyedState_rdh | /**
*
* @see KeyedStateBackend
*/
@Override
@SuppressWarnings("unchecked")
public <N, S extends State, V> S getOrCreateKeyedState(final TypeSerializer<N> namespaceSerializer, StateDescriptor<S, V> stateDescriptor) throws Exception {
checkNotNull(namespaceSerializer, "Namespace serializer");
checkNotNull(k... | 3.26 |
flink_AbstractKeyedStateBackend_getPartitionedState_rdh | /**
* TODO: NOTE: This method does a lot of work caching / retrieving states just to update the
* namespace. This method should be removed for the sake of namespaces being lazily fetched from
* the keyed state backend, or being set on the state directly.
*
* @see KeyedStateBackend
*/
@SuppressWarnings("unchecked"... | 3.26 |
flink_AbstractKeyedStateBackend_getCurrentKeyGroupIndex_rdh | /**
*
* @see KeyedStateBackend
*/
public int getCurrentKeyGroupIndex() {
return this.keyContext.getCurrentKeyGroupIndex();
} | 3.26 |
flink_AbstractKeyedStateBackend_applyToAllKeys_rdh | /**
*
* @see KeyedStateBackend
*/
@Override
public <N, S extends State, T> void applyToAllKeys(final N namespace, final TypeSerializer<N> namespaceSerializer, final StateDescriptor<S, T> stateDescriptor, final KeyedStateFunction<K, S> function) throws Exception {
applyToAllKeys(namespace, namespaceSerializer,... | 3.26 |
flink_AbstractKeyedStateBackend_setCurrentKey_rdh | /**
*
* @see KeyedStateBackend
*/
@Override
public void setCurrentKey(K newKey) {
notifyKeySelected(newKey);
this.keyContext.setCurrentKey(newKey);
this.keyContext.setCurrentKeyGroupIndex(KeyGroupRangeAssignment.assignToKeyGroup(newKey,
numberOfKeyGroups));
} | 3.26 |
flink_AbstractKeyedStateBackend_getKeySerializer_rdh | /**
*
* @see KeyedStateBackend
*/
@Override
public TypeSerializer<K> getKeySerializer() {
return keySerializer;
} | 3.26 |
flink_RestAPIVersion_getLatestVersion_rdh | /**
* Accept versions and one of them as a comparator, and get the latest one.
*
* @return latest version that implement RestAPIVersion interface>
*/
static <E
extends RestAPIVersion<E>> E getLatestVersion(Collection<E> versions) {
return Collections.max(versions);
} | 3.26 |
flink_ChainedStateHandle_isEmpty_rdh | /**
* Check if there are any states handles present. Notice that this can be true even if {@link #getLength()} is greater than zero, because state handles can be null.
*
* @return true if there are no state handles for any operator.
*/
public boolean isEmpty() {
for (T
state : operatorStateHandles) {
... | 3.26 |
flink_ChainedStateHandle_getLength_rdh | /**
* Returns the length of the operator chain. This can be different from the number of operator
* state handles, because the some operators in the chain can have no state and thus their state
* handle can be null.
*
* @return length of the operator chain
*/
public int getLength() {
return operatorStateHa... | 3.26 |
flink_FlinkRelUtil_isMergeable_rdh | /**
* Return two neighbouring {@link Calc} can merge into one {@link Calc} or not. If the two
* {@link Calc} can merge into one, each non-deterministic {@link RexNode} of bottom {@link Calc} should appear at most once in the project list of top {@link Calc}.
*/
public static boolean isMergeable(Calc topCalc, Calc bo... | 3.26 |
flink_FlinkRelUtil_merge_rdh | /**
* Merges the programs of two {@link Calc} instances and returns a new {@link Calc} instance
* with the merged program.
*/
public static Calc merge(Calc topCalc, Calc bottomCalc) {
RexProgram topProgram = topCalc.getProgram();
RexBuilder rexBuilder = topCalc.getCluster().getRe... | 3.26 |
flink_FlinkRelUtil_mergeable_rdh | /**
* The internal reusable method for filter, project nd calc.
*/
private static boolean mergeable(int[] topInputRefCounter, List<RexNode> topProjects, List<RexNode> bottomProjects) {
RexUtil.apply(new InputRefCounter(true, topInputRefCounter), topProjects, null);
boolean mergeable = true;
for (int idx ... | 3.26 |
flink_FlinkRelUtil_initializeArray_rdh | /**
* Returns an int array with given length and initial value.
*
* @param length
* array length
* @param initVal
* initial value
* @return initialized int array
*/
public static int[] initializeArray(int length, int initVal) {
final
int[] array = new int[length];
Arrays.fill(array, initVal);
... | 3.26 |
flink_InnerJoinOperatorBase_executeOnCollections_rdh | // --------------------------------------------------------------------------------------------
@SuppressWarnings("unchecked")
@Override
protected List<OUT> executeOnCollections(List<IN1> inputData1, List<IN2> inputData2, RuntimeContext runtimeContext, ExecutionConfig executionConfig) throws Exception {
FlatJoinFun... | 3.26 |
flink_SourcePredicates_isJavaClass_rdh | /**
* Checks whether the given {@link JavaClass} is actually a Java class, and not a Scala class.
*
* <p>ArchUnit does not yet fully support Scala. Rules should ensure that they restrict
* themselves to only Java classes for correct results.
*/
static boolean isJavaClass(JavaClass clazz) {
if (!clazz.getSource... | 3.26 |
flink_SourcePredicates_areJavaClasses_rdh | /**
* Tests that a given class is a Java class.
*
* <p>ArchUnit does not yet fully support Scala. Rules should ensure that they restrict
* themselves to only Java classes for correct results.
*/
public static DescribedPredicate<JavaClass> areJavaClasses() {
return new DescribedPredicate<JavaClass>("are Java cl... | 3.26 |
flink_FileWriterBucket_assembleNewPartPath_rdh | /**
* Constructor a new PartPath and increment the partCounter.
*/
private Path assembleNewPartPath() {
long v6 = partCounter++;
return new Path(bucketPath, ((((outputFileConfig.getPartPrefix() + '-') + uniqueId) + '-') + v6) + outputFileConfig.getPartSuffix());
} | 3.26 |
flink_FileWriterBucket_getUniqueId_rdh | // --------------------------- Testing Methods -----------------------------
@VisibleForTesting
public String getUniqueId() {
return uniqueId;
} | 3.26 |
flink_FileWriterBucket_getNew_rdh | // --------------------------- Static Factory Methods -----------------------------
/**
* Creates a new empty {@code Bucket}.
*
* @param bucketId
* the identifier of the bucket, as returned by the {@link BucketAssigner}.
* @param bucketPath
* the path to where the part files for the bucket will be written to.... | 3.26 |
flink_FileWriterBucket_restore_rdh | /**
* Restores a {@code Bucket} from the state included in the provided {@link FileWriterBucketState}.
*
* @param bucketWriter
* the {@link BucketWriter} used to write part files in the bucket.
* @param bucketState
* the initial state of the restored bucket.
* @param <IN>
* the type of input elements to t... | 3.26 |
flink_RemoteCacheManager_startSegment_rdh | // ------------------------------------------------------------------------
// Called by RemoteTierProducerAgent
// ------------------------------------------------------------------------
void startSegment(int subpartitionId, int segmentId) {
subpartitionCacheDataManagers[subpartitionId].startSegment(segmentId);
... | 3.26 |
flink_SqlGatewayRestAPIVersion_getStableVersions_rdh | /**
* Returns the supported stable versions.
*
* @return the list of the stable versions.
*/
public static List<SqlGatewayRestAPIVersion> getStableVersions() {
return Arrays.stream(SqlGatewayRestAPIVersion.values()).filter(SqlGatewayRestAPIVersion::isStableVersion).collect(Collectors.toList());
} | 3.26 |
flink_SqlGatewayRestAPIVersion_getURLVersionPrefix_rdh | /**
* Returns the URL version prefix (e.g. "v1") for this version.
*
* @return URL version prefix
*/
@Override
public String getURLVersionPrefix() {
return name().toLowerCase();
} | 3.26 |
flink_SqlGatewayRestAPIVersion_fromURIToVersion_rdh | /**
* Convert uri to SqlGatewayRestAPIVersion. If failed, return default version.
*
* @return SqlGatewayRestAPIVersion
*/
public static SqlGatewayRestAPIVersion fromURIToVersion(String uri) {
int slashIndex = uri.indexOf('/', 1);
if (slashIndex < 0) {
slashIndex = uri.length();
}
try {
... | 3.26 |
flink_KvStateLocation_getJobVertexId_rdh | /**
* Returns the JobVertexID the KvState instances belong to.
*
* @return JobVertexID the KvState instances belong to
*/
public JobVertexID getJobVertexId() {
return jobVertexId; } | 3.26 |
flink_KvStateLocation_getKvStateID_rdh | /**
* Returns the registered KvStateID for the key group index or <code>null</code> if none is
* registered yet.
*
* @param keyGroupIndex
* Key group index to get ID for.
* @return KvStateID for the key group index or <code>null</code> if none is registered yet
* @throws IndexOutOfBoundsException
* If key g... | 3.26 |
flink_KvStateLocation_unregisterKvState_rdh | /**
* Registers a KvState instance for the given key group index.
*
* @param keyGroupRange
* Key group range to unregister.
* @throws IndexOutOfBoundsException
* If key group range start < 0 or key group range end >=
* Number of key groups
* @throws IllegalArgumentException
* If no location information... | 3.26 |
flink_KvStateLocation_registerKvState_rdh | /**
* Registers a KvState instance for the given key group index.
*
* @param keyGroupRange
* Key group range to register
* @param kvStateId
* ID of the KvState instance at the key group index.
* @param kvStateAddress
* Server address of the KvState instance at the key group index.
* @throws IndexOutOfBou... | 3.26 |
flink_KvStateLocation_getRegistrationName_rdh | /**
* Returns the name under which the KvState instances have been registered.
*
* @return Name under which the KvState instances have been registered.
*/
public String getRegistrationName() {
return f0;
} | 3.26 |
flink_BlobOutputStream_receiveAndCheckPutResponse_rdh | /**
* Reads the response from the input stream and throws in case of errors.
*
* @param is
* stream to read from
* @param md
* message digest to check the response against
* @param blobType
* whether the BLOB should be permanent or transient
* @throws IOException
* if the response is an error, the mes... | 3.26 |
flink_BlobOutputStream_sendPutHeader_rdh | /**
* Constructs and writes the header data for a PUT request to the given output stream.
*
* @param outputStream
* the output stream to write the PUT header data to
* @param jobId
* the ID of job the BLOB belongs to (or <tt>null</tt> if job-unrelated)
* @param blobType
* whether the BLOB should become pe... | 3.26 |
flink_JoinedStreams_evictor_rdh | /**
* Sets the {@code Evictor} that should be used to evict elements from a window before
* emission.
*
* <p>Note: When using an evictor window performance will degrade significantly, since
* pre-aggregation of window results cannot be used.
*/
@PublicEvolving
public WithWindow<T1, T2, KEY, W> evictor(Evictor<? s... | 3.26 |
flink_JoinedStreams_window_rdh | /**
* Specifies the window on which the join operation works.
*/
@PublicEvolving
public <W extends Window> WithWindow<T1, T2, KEY, W> window(WindowAssigner<? super TaggedUnion<T1, T2>, W> assigner) {
return new WithWindow<>(input1, input2, keySelector1, keySelector2, keyType, assigner, null, null, null);
} | 3.26 |
flink_JoinedStreams_where_rdh | /**
* Specifies a {@link KeySelector} for elements from the first input with explicit type
* information for the key type.
*
* @param keySelector
* The KeySelector to be used for extracting the first input's key for
* partitioning.
* @param keyType
* The type information describing the key type.
*/
publi... | 3.26 |
flink_JoinedStreams_allowedLateness_rdh | /**
* Sets the time by which elements are allowed to be late.
*
* @see WindowedStream#allowedLateness(Time)
*/
@PublicEvolving
public WithWindow<T1, T2,
KEY, W> allowedLateness(Time newLateness) {
return
new WithWindow<>(input1, input2, keySelector1, keySelector2, keyType, windowAssigner, trigger, evictor, ... | 3.26 |
flink_JoinedStreams_equalTo_rdh | /**
* Specifies a {@link KeySelector} for elements from the second input with explicit type
* information for the key type.
*
* @param keySelector
* The KeySelector to be used for extracting the second input's key for
* partitioning.
* @param keyType
* The type information describing the key type.
*/
pub... | 3.26 |
flink_JoinedStreams_trigger_rdh | /**
* Sets the {@code Trigger} that should be used to trigger window emission.
*/
@PublicEvolving
public WithWindow<T1, T2, KEY, W> trigger(Trigger<? super TaggedUnion<T1, T2>, ? super W> newTrigger) {
return new WithWindow<>(input1, input2, keySelector1, keySelector2, keyType, windowAssigner, newTrigger, evictor... | 3.26 |
flink_JoinedStreams_apply_rdh | /**
* Completes the join operation with the user function that is executed for each combination
* of elements with the same key in a window.
*
* <p>Note: This method's return type does not support setting an operator-specific
* parallelism. Due to binary backwards compatibility, this cannot be altered. Use the
* ... | 3.26 |
flink_MapValue_isEmpty_rdh | /* (non-Javadoc)
@see java.util.Map#isEmpty()
*/
@Override
public boolean isEmpty() {
return this.map.isEmpty();
} | 3.26 |
flink_MapValue_get_rdh | /* (non-Javadoc)
@see java.util.Map#get(java.lang.Object)
*/@Override
public V get(final Object key) {
return this.map.get(key);
} | 3.26 |
flink_MapValue_containsValue_rdh | /* (non-Javadoc)
@see java.util.Map#containsValue(java.lang.Object)
*/
@Override
public boolean containsValue(final Object value) { return this.map.containsValue(value);
} | 3.26 |
flink_MapValue_keySet_rdh | /* (non-Javadoc)
@see java.util.Map#keySet()
*/
@Override
public Set<K> keySet() {
return this.map.keySet();
} | 3.26 |
flink_MapValue_put_rdh | /* (non-Javadoc)
@see java.util.Map#put(java.lang.Object, java.lang.Object)
*/
@Override
public V put(final K key, final
V value) {
return this.map.put(key, value);
} | 3.26 |
flink_MapValue_clear_rdh | /* (non-Javadoc)
@see java.util.Map#clear()
*/
@Override
public void clear() {
this.map.clear();
} | 3.26 |
flink_MapValue_m0_rdh | /* (non-Javadoc)
@see java.util.Map#putAll(java.util.Map)
*/
@Override
public void m0(final Map<? extends K, ? extends V> m) {
this.map.putAll(m);
} | 3.26 |
flink_MapValue_containsKey_rdh | /* (non-Javadoc)
@see java.util.Map#containsKey(java.lang.Object)
*/
@Override
public boolean containsKey(final Object key) {
return this.map.containsKey(key);
} | 3.26 |
flink_MapValue_toString_rdh | /* (non-Javadoc)
@see java.lang.Object#toString()
*/
@Override
public String toString()
{
return this.map.toString();
} | 3.26 |
flink_MapValue_equals_rdh | /* (non-Javadoc)
@see java.lang.Object#equals(java.lang.Object)
*/
@Override
public boolean equals(final Object obj) {
if (this == obj) {
return true;
}
if (obj == null) {
return false;
}
if (this.getClass() != obj.getClass()) {
return false;
}
final MapValue<?, ?> ... | 3.26 |
flink_MapValue_entrySet_rdh | /* (non-Javadoc)
@see java.util.Map#entrySet()
*/
@Override
public Set<Entry<K, V>> entrySet() {
return this.map.entrySet();
} | 3.26 |
flink_MapValue_remove_rdh | /* (non-Javadoc)
@see java.util.Map#remove(java.lang.Object)
*/
@Override
public V remove(final Object key) {
return this.map.remove(key);
} | 3.26 |
flink_MapValue_size_rdh | /* (non-Javadoc)
@see java.util.Map#size()
*/
@Override
public int size() {
return this.map.size();
} | 3.26 |
flink_MapValue_hashCode_rdh | /* (non-Javadoc)
@see java.lang.Object#hashCode()
*/
@Override
public int hashCode() {
final int prime = 47;
int result = 1;
result = (prime * result) + this.map.hashCode();
return result;
} | 3.26 |
flink_Tuple9_setFields_rdh | /**
* Sets new values to all fields of the tuple.
*
* @param f0
* The value for field 0
* @param f1
* The value for field 1
* @param f2
* The value for field 2
* @param f3
* The value for field 3
* @param f4
* The value for field 4
* @param f5
* The value for field 5
* @param f6
* The valu... | 3.26 |
flink_Tuple9_toString_rdh | // -------------------------------------------------------------------------------------------------
// standard utilities
// -------------------------------------------------------------------------------------------------
/**
* Creates a string representation of the tuple in the form (f0, f1, f2, f3, f4, f5, f6, f7,... | 3.26 |
flink_Tuple9_copy_rdh | /**
* Shallow tuple copy.
*
* @return A new Tuple with the same fields as this.
*/
@Override
@SuppressWarnings("unchecked")
public Tuple9<T0, T1, T2, T3, T4, T5, T6, T7, T8> copy() {
return new Tuple9<>(this.f0, this.f1, this.f2, this.f3, this.f4, this.f5, this.f6, this.f7, this.f8);
} | 3.26 |
flink_Tuple9_m0_rdh | /**
* Deep equality for tuples by calling equals() on the tuple members.
*
* @param o
* the object checked for equality
* @return true if this is equal to o.
*/
@Override
public boolean m0(Object o) {
if (this == o) {
return true;
}
if (!(o instanceof Tuple9)) {
return false;
} @Suppress... | 3.26 |
flink_Tuple9_of_rdh | /**
* Creates a new tuple and assigns the given values to the tuple's fields. This is more
* convenient than using the constructor, because the compiler can infer the generic type
* arguments implicitly. For example: {@code Tuple3.of(n, x, s)} instead of {@code new
* Tuple3<Integer, Double, String>(n, x, s)}
*/
pu... | 3.26 |
flink_DoubleSumAggregator_aggregate_rdh | /**
* Adds the given value to the current aggregate.
*
* @param value
* The value to add to the aggregate.
*/
public void aggregate(double value) {
sum += value;
} | 3.26 |
flink_TpcdsResultComparator_compareLine_rdh | // ------------------------------------------------------------------------
private static boolean compareLine(String expectedLine, String actualLine) {
return compareLineInternal(expectedLine, actualLine, -1, false);
} | 3.26 |
flink_TpcdsResultComparator_isEqualCol_rdh | // ------------------------------------------------------------------------
private static boolean isEqualCol(String expected, String actual) {
return (isEqualNull(expected, actual) ||
isEqualNumber(expected, actual)) || expected.equals(actual);
} | 3.26 |
flink_TpcdsResultComparator_compareQuery34_rdh | // ------------------------------------------------------------------------
private static boolean compareQuery34(String[] expectedLines, String[] actualLines) {
// take the first two lines and move them back to lines 7 and 8
final String expected1 = expectedLines[0];
final String expected2 = expectedLines[... | 3.26 |
flink_OptionalFailure_get_rdh | /**
*
* @return stored value or throw a {@link FlinkException} with {@code failureCause}.
*/
public T get() throws FlinkException {
if (value != null) {
return value;
}
checkNotNull(failureCause);
throw new FlinkException(failureCause);
}
/**
*
* @return same as {@link #get()} but throws a... | 3.26 |
flink_PushFilterIntoSourceScanRuleBase_resolveFiltersAndCreateTableSourceTable_rdh | /**
* Resolves filters using the underlying sources {@link SupportsFilterPushDown} and creates a
* new {@link TableSourceTable} with the supplied predicates.
*
* @param convertiblePredicates
* Predicates to resolve
* @param oldTableSourceTable
* TableSourceTable to copy
* @param scan
* Underlying table s... | 3.26 |
flink_EdgeManagerBuildUtil_computeMaxEdgesToTargetExecutionVertex_rdh | /**
* Given parallelisms of two job vertices, compute the max number of edges connected to a target
* execution vertex from the source execution vertices. Note that edge is considered undirected
* here. It can be an edge connected from an upstream job vertex to a downstream job vertex, or
* in a reversed way.
*
*... | 3.26 |
flink_EdgeManagerBuildUtil_connectInternal_rdh | /**
* Connect all execution vertices to all partitions.
*/
private static void connectInternal(List<ExecutionVertex> taskVertices,
List<IntermediateResultPartition> partitions, ResultPartitionType
resultPartitionType, EdgeManager edgeManager) {
checkState(!taskVertices.isEmpty());
checkState(!partitions.isEm... | 3.26 |
flink_EdgeManagerBuildUtil_connectVertexToResult_rdh | /**
* Calculate the connections between {@link ExecutionJobVertex} and {@link IntermediateResult} *
* based on the {@link DistributionPattern}.
*
* @param vertex
* the downstream consumer {@link ExecutionJobVertex}
* @param intermediateResult
* the upstream consumed {@link IntermediateResult}
*/
static void... | 3.26 |
flink_TumblingWindowAssigner_of_rdh | // ------------------------------------------------------------------------
// Utilities
// ------------------------------------------------------------------------
/**
* Creates a new {@code TumblingWindowAssigner} {@link WindowAssigner} that assigns elements to
* time windows based on the element timestamp.
*
* @... | 3.26 |
flink_TumblingWindowAssigner_withOffset_rdh | /**
* Creates a new {@code TumblingWindowAssigner} {@link WindowAssigner} that assigns elements to
* time windows based on the element timestamp and offset.
*
* <p>For example, if you want window a stream by hour,but window begins at the 15th minutes of
* each hour, you can use {@code of(Time.hours(1),Time.minutes... | 3.26 |
flink_CliRowView_init_rdh | // --------------------------------------------------------------------------------------------
@Override
protected void init() {
// nothing to do
} | 3.26 |
flink_Tuple16_equals_rdh | /**
* Deep equality for tuples by calling equals() on the tuple members.
*
* @param o
* the object checked for equality
* @return true if this is equal to o.
*/
@Override
public boolean equals(Object o) {
if (this == o) {
return true;
}
if
(!(o instanceof Tuple16)) {
return false;
}
@SuppressWarnings("rawtype... | 3.26 |
flink_Tuple16_of_rdh | /**
* Creates a new tuple and assigns the given values to the tuple's fields. This is more
* convenient than using the constructor, because the compiler can infer the generic type
* arguments implicitly. For example: {@code Tuple3.of(n, x, s)} instead of {@code new
* Tuple3<Integer, Double, String>(n, x, s)}
*/
pu... | 3.26 |
flink_Tuple16_copy_rdh | /**
* Shallow tuple copy.
*
* @return A new Tuple with the same fields as this.
*/
@Override
@SuppressWarnings("unchecked")
public Tuple16<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15> copy() {
return new Tuple16<>(this.f0, this.f1, this.f2,
this.f3, this.f4, this.f5, this.f6, this.f7, this... | 3.26 |
flink_Tuple16_toString_rdh | // -------------------------------------------------------------------------------------------------
// standard utilities
// -------------------------------------------------------------------------------------------------
/**
* Creates a string representation of the tuple in the form (f0, f1, f2, f3, f4, f5, f6, f7,... | 3.26 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.