name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
flink_CatalogSchemaTable_getContextResolvedTable_rdh
// ~ Methods ---------------------------------------------------------------- public ContextResolvedTable getContextResolvedTable() { return contextResolvedTable; }
3.26
flink_DelegationTokenProvider_serviceConfigPrefix_rdh
/** * Config prefix of the service. */ default String serviceConfigPrefix() { return String.format("%s.%s", CONFIG_PREFIX, serviceName()); }
3.26
flink_RuntimeRestAPIVersion_isStableVersion_rdh
/** * Returns whether this version is considered stable. * * @return whether this version is stable */ @Override public boolean isStableVersion() { return isStable; }
3.26
flink_SinkModifyOperation_getTargetColumns_rdh
/** * return null when no column list specified. */ @Nullable public int[][] getTargetColumns() { return targetColumns; }
3.26
flink_DefaultLookupCache_maximumSize_rdh
/** * Specifies the maximum number of entries of the cache. */ public Builder maximumSize(long maximumSize) { this.maximumSize = maximumSize; return this; }
3.26
flink_DefaultLookupCache_newBuilder_rdh
/** * Creates a builder for the cache. */ public static Builder newBuilder() { return new Builder(); }
3.26
flink_DefaultLookupCache_expireAfterAccess_rdh
/** * Specifies the duration after an entry is last accessed that it should be automatically * removed. */ public Builder expireAfterAccess(Duration duration) { expireAfterAccessDuration = duration; return this; }
3.26
flink_DefaultLookupCache_cacheMissingKey_rdh
/** * Specifies whether to cache empty value into the cache. * * <p>Please note that "empty" means a collection without any rows in it instead of null. * The cache will not accept any null key or value. */ public Builder cacheMissingKey(boolean cacheMissingKey) { this.cacheMissingKey = cacheMissingKey; return this...
3.26
flink_DefaultLookupCache_expireAfterWrite_rdh
/** * Specifies the duration after an entry is created that it should be automatically removed. */ public Builder expireAfterWrite(Duration duration) { expireAfterWriteDuration = duration; return this; }
3.26
flink_DefaultLookupCache_build_rdh
/** * Creates the cache. */ public DefaultLookupCache build() { return new DefaultLookupCache(expireAfterAccessDuration, expireAfterWriteDuration, maximumSize, cacheMissingKey); }
3.26
flink_BooleanColumnSummary_getNonNullCount_rdh
/** * The number of non-null values in this column. */ @Override public long getNonNullCount() { return trueCount + falseCount; }
3.26
flink_RichAndCondition_getLeft_rdh
/** * * @return One of the {@link IterativeCondition conditions} combined in this condition. */ public IterativeCondition<T> getLeft() { return getNestedConditions()[0]; }
3.26
flink_RichAndCondition_getRight_rdh
/** * * @return One of the {@link IterativeCondition conditions} combined in this condition. */ public IterativeCondition<T> getRight() { return getNestedConditions()[1]; }
3.26
flink_SchemaValidator_deriveProctimeAttribute_rdh
/** * Finds the proctime attribute if defined. */ public static Optional<String> deriveProctimeAttribute(DescriptorProperties properties) { Map<String, String> names = properties.getIndexedProperty(SCHEMA, SCHEMA_NAME); for (int i = 0; i < names.size(); i++) {Optional<Boolean> isProctime = properties.getOpti...
3.26
flink_SchemaValidator_deriveFieldMapping_rdh
/** * Finds a table source field mapping. * * @param properties * The properties describing a schema. * @param inputType * The input type that a connector and/or format produces. This parameter can * be used to resolve a rowtime field against an input field. */ public static Map<String, String> deriveFiel...
3.26
flink_SchemaValidator_deriveRowtimeAttributes_rdh
/** * Finds the rowtime attributes if defined. */ public static List<RowtimeAttributeDescriptor> deriveRowtimeAttributes(DescriptorProperties properties) { Map<String, String> names = properties.getIndexedProperty(SCHEMA, SCHEMA_NAME); List<RowtimeAttributeDescriptor> attributes = new ArrayList<>(); // ch...
3.26
flink_CharValue_setValue_rdh
/** * Sets the encapsulated char to the specified value. * * @param value * the new value of the encapsulated char. */ public void setValue(char value) { this.value = value; }
3.26
flink_CharValue_compareTo_rdh
// -------------------------------------------------------------------------------------------- @Override public int compareTo(CharValue o) { final int other = o.value; return this.value < other ? -1 : this.value > other ? 1 : 0; }
3.26
flink_CharValue_read_rdh
// -------------------------------------------------------------------------------------------- @Override public void read(DataInputView in) throws IOException { this.value = in.readChar(); }
3.26
flink_CharValue_getMaxNormalizedKeyLen_rdh
// -------------------------------------------------------------------------------------------- @Override public int getMaxNormalizedKeyLen() { return 2; }
3.26
flink_SlideWithSize_every_rdh
/** * Specifies the window's slide as time or row-count interval. * * <p>The slide determines the interval in which windows are started. Hence, sliding windows can * overlap if the slide is smaller than the size of the window. * * <p>For example, you could have windows of size 15 minutes that slide by 3 minutes. ...
3.26
flink_ContinuousFileReaderOperator_prepareToProcessRecord_rdh
// the split was added and message to itself was enqueued to process it public <T extends TimestampedInputSplit> boolean prepareToProcessRecord(ContinuousFileReaderOperator<?, T> op) throws IOException {if (op.splits.isEmpty()) { op.switchState(ReaderState.IDLE); return false; } else { o...
3.26
flink_KvStateClientProxyHandler_getKvStateLookupInfo_rdh
/** * Lookup the {@link KvStateLocation} for the given job and queryable state name. * * <p>The job manager will be queried for the location only if forced or no cached location can * be found. There are no guarantees about * * @param jobId * JobID the state instance belongs to. * @param queryableStateName *...
3.26
flink_AbstractKeyedStateBackend_getNumberOfKeyGroups_rdh
/** * * @see KeyedStateBackend */ public int getNumberOfKeyGroups() { return numberOfKeyGroups; }
3.26
flink_AbstractKeyedStateBackend_getKeyGroupRange_rdh
/** * * @see KeyedStateBackend */ @Override public KeyGroupRange getKeyGroupRange() {return keyGroupRange; }
3.26
flink_AbstractKeyedStateBackend_getCurrentKey_rdh
/** * * @see KeyedStateBackend */ @Override public K getCurrentKey() { return this.keyContext.getCurrentKey(); }
3.26
flink_AbstractKeyedStateBackend_dispose_rdh
/** * Closes the state backend, releasing all internal resources, but does not delete any * persistent checkpoint data. */ @Override public void dispose() { IOUtils.closeQuietly(cancelStreamRegistry);if (kvStateRegistry != null) { kvStateRegistry.unregisterAll(); } lastName = null; lastSta...
3.26
flink_AbstractKeyedStateBackend_getOrCreateKeyedState_rdh
/** * * @see KeyedStateBackend */ @Override @SuppressWarnings("unchecked") public <N, S extends State, V> S getOrCreateKeyedState(final TypeSerializer<N> namespaceSerializer, StateDescriptor<S, V> stateDescriptor) throws Exception { checkNotNull(namespaceSerializer, "Namespace serializer"); checkNotNull(k...
3.26
flink_AbstractKeyedStateBackend_getPartitionedState_rdh
/** * TODO: NOTE: This method does a lot of work caching / retrieving states just to update the * namespace. This method should be removed for the sake of namespaces being lazily fetched from * the keyed state backend, or being set on the state directly. * * @see KeyedStateBackend */ @SuppressWarnings("unchecked"...
3.26
flink_AbstractKeyedStateBackend_getCurrentKeyGroupIndex_rdh
/** * * @see KeyedStateBackend */ public int getCurrentKeyGroupIndex() { return this.keyContext.getCurrentKeyGroupIndex(); }
3.26
flink_AbstractKeyedStateBackend_applyToAllKeys_rdh
/** * * @see KeyedStateBackend */ @Override public <N, S extends State, T> void applyToAllKeys(final N namespace, final TypeSerializer<N> namespaceSerializer, final StateDescriptor<S, T> stateDescriptor, final KeyedStateFunction<K, S> function) throws Exception { applyToAllKeys(namespace, namespaceSerializer,...
3.26
flink_AbstractKeyedStateBackend_setCurrentKey_rdh
/** * * @see KeyedStateBackend */ @Override public void setCurrentKey(K newKey) { notifyKeySelected(newKey); this.keyContext.setCurrentKey(newKey); this.keyContext.setCurrentKeyGroupIndex(KeyGroupRangeAssignment.assignToKeyGroup(newKey, numberOfKeyGroups)); }
3.26
flink_AbstractKeyedStateBackend_getKeySerializer_rdh
/** * * @see KeyedStateBackend */ @Override public TypeSerializer<K> getKeySerializer() { return keySerializer; }
3.26
flink_RestAPIVersion_getLatestVersion_rdh
/** * Accept versions and one of them as a comparator, and get the latest one. * * @return latest version that implement RestAPIVersion interface> */ static <E extends RestAPIVersion<E>> E getLatestVersion(Collection<E> versions) { return Collections.max(versions); }
3.26
flink_ChainedStateHandle_isEmpty_rdh
/** * Check if there are any states handles present. Notice that this can be true even if {@link #getLength()} is greater than zero, because state handles can be null. * * @return true if there are no state handles for any operator. */ public boolean isEmpty() { for (T state : operatorStateHandles) { ...
3.26
flink_ChainedStateHandle_getLength_rdh
/** * Returns the length of the operator chain. This can be different from the number of operator * state handles, because the some operators in the chain can have no state and thus their state * handle can be null. * * @return length of the operator chain */ public int getLength() { return operatorStateHa...
3.26
flink_FlinkRelUtil_isMergeable_rdh
/** * Return two neighbouring {@link Calc} can merge into one {@link Calc} or not. If the two * {@link Calc} can merge into one, each non-deterministic {@link RexNode} of bottom {@link Calc} should appear at most once in the project list of top {@link Calc}. */ public static boolean isMergeable(Calc topCalc, Calc bo...
3.26
flink_FlinkRelUtil_merge_rdh
/** * Merges the programs of two {@link Calc} instances and returns a new {@link Calc} instance * with the merged program. */ public static Calc merge(Calc topCalc, Calc bottomCalc) { RexProgram topProgram = topCalc.getProgram(); RexBuilder rexBuilder = topCalc.getCluster().getRe...
3.26
flink_FlinkRelUtil_mergeable_rdh
/** * The internal reusable method for filter, project nd calc. */ private static boolean mergeable(int[] topInputRefCounter, List<RexNode> topProjects, List<RexNode> bottomProjects) { RexUtil.apply(new InputRefCounter(true, topInputRefCounter), topProjects, null); boolean mergeable = true; for (int idx ...
3.26
flink_FlinkRelUtil_initializeArray_rdh
/** * Returns an int array with given length and initial value. * * @param length * array length * @param initVal * initial value * @return initialized int array */ public static int[] initializeArray(int length, int initVal) { final int[] array = new int[length]; Arrays.fill(array, initVal); ...
3.26
flink_InnerJoinOperatorBase_executeOnCollections_rdh
// -------------------------------------------------------------------------------------------- @SuppressWarnings("unchecked") @Override protected List<OUT> executeOnCollections(List<IN1> inputData1, List<IN2> inputData2, RuntimeContext runtimeContext, ExecutionConfig executionConfig) throws Exception { FlatJoinFun...
3.26
flink_SourcePredicates_isJavaClass_rdh
/** * Checks whether the given {@link JavaClass} is actually a Java class, and not a Scala class. * * <p>ArchUnit does not yet fully support Scala. Rules should ensure that they restrict * themselves to only Java classes for correct results. */ static boolean isJavaClass(JavaClass clazz) { if (!clazz.getSource...
3.26
flink_SourcePredicates_areJavaClasses_rdh
/** * Tests that a given class is a Java class. * * <p>ArchUnit does not yet fully support Scala. Rules should ensure that they restrict * themselves to only Java classes for correct results. */ public static DescribedPredicate<JavaClass> areJavaClasses() { return new DescribedPredicate<JavaClass>("are Java cl...
3.26
flink_FileWriterBucket_assembleNewPartPath_rdh
/** * Constructor a new PartPath and increment the partCounter. */ private Path assembleNewPartPath() { long v6 = partCounter++; return new Path(bucketPath, ((((outputFileConfig.getPartPrefix() + '-') + uniqueId) + '-') + v6) + outputFileConfig.getPartSuffix()); }
3.26
flink_FileWriterBucket_getUniqueId_rdh
// --------------------------- Testing Methods ----------------------------- @VisibleForTesting public String getUniqueId() { return uniqueId; }
3.26
flink_FileWriterBucket_getNew_rdh
// --------------------------- Static Factory Methods ----------------------------- /** * Creates a new empty {@code Bucket}. * * @param bucketId * the identifier of the bucket, as returned by the {@link BucketAssigner}. * @param bucketPath * the path to where the part files for the bucket will be written to....
3.26
flink_FileWriterBucket_restore_rdh
/** * Restores a {@code Bucket} from the state included in the provided {@link FileWriterBucketState}. * * @param bucketWriter * the {@link BucketWriter} used to write part files in the bucket. * @param bucketState * the initial state of the restored bucket. * @param <IN> * the type of input elements to t...
3.26
flink_RemoteCacheManager_startSegment_rdh
// ------------------------------------------------------------------------ // Called by RemoteTierProducerAgent // ------------------------------------------------------------------------ void startSegment(int subpartitionId, int segmentId) { subpartitionCacheDataManagers[subpartitionId].startSegment(segmentId); ...
3.26
flink_SqlGatewayRestAPIVersion_getStableVersions_rdh
/** * Returns the supported stable versions. * * @return the list of the stable versions. */ public static List<SqlGatewayRestAPIVersion> getStableVersions() { return Arrays.stream(SqlGatewayRestAPIVersion.values()).filter(SqlGatewayRestAPIVersion::isStableVersion).collect(Collectors.toList()); }
3.26
flink_SqlGatewayRestAPIVersion_getURLVersionPrefix_rdh
/** * Returns the URL version prefix (e.g. "v1") for this version. * * @return URL version prefix */ @Override public String getURLVersionPrefix() { return name().toLowerCase(); }
3.26
flink_SqlGatewayRestAPIVersion_fromURIToVersion_rdh
/** * Convert uri to SqlGatewayRestAPIVersion. If failed, return default version. * * @return SqlGatewayRestAPIVersion */ public static SqlGatewayRestAPIVersion fromURIToVersion(String uri) { int slashIndex = uri.indexOf('/', 1); if (slashIndex < 0) { slashIndex = uri.length(); } try { ...
3.26
flink_KvStateLocation_getJobVertexId_rdh
/** * Returns the JobVertexID the KvState instances belong to. * * @return JobVertexID the KvState instances belong to */ public JobVertexID getJobVertexId() { return jobVertexId; }
3.26
flink_KvStateLocation_getKvStateID_rdh
/** * Returns the registered KvStateID for the key group index or <code>null</code> if none is * registered yet. * * @param keyGroupIndex * Key group index to get ID for. * @return KvStateID for the key group index or <code>null</code> if none is registered yet * @throws IndexOutOfBoundsException * If key g...
3.26
flink_KvStateLocation_unregisterKvState_rdh
/** * Registers a KvState instance for the given key group index. * * @param keyGroupRange * Key group range to unregister. * @throws IndexOutOfBoundsException * If key group range start < 0 or key group range end >= * Number of key groups * @throws IllegalArgumentException * If no location information...
3.26
flink_KvStateLocation_registerKvState_rdh
/** * Registers a KvState instance for the given key group index. * * @param keyGroupRange * Key group range to register * @param kvStateId * ID of the KvState instance at the key group index. * @param kvStateAddress * Server address of the KvState instance at the key group index. * @throws IndexOutOfBou...
3.26
flink_KvStateLocation_getRegistrationName_rdh
/** * Returns the name under which the KvState instances have been registered. * * @return Name under which the KvState instances have been registered. */ public String getRegistrationName() { return f0; }
3.26
flink_BlobOutputStream_receiveAndCheckPutResponse_rdh
/** * Reads the response from the input stream and throws in case of errors. * * @param is * stream to read from * @param md * message digest to check the response against * @param blobType * whether the BLOB should be permanent or transient * @throws IOException * if the response is an error, the mes...
3.26
flink_BlobOutputStream_sendPutHeader_rdh
/** * Constructs and writes the header data for a PUT request to the given output stream. * * @param outputStream * the output stream to write the PUT header data to * @param jobId * the ID of job the BLOB belongs to (or <tt>null</tt> if job-unrelated) * @param blobType * whether the BLOB should become pe...
3.26
flink_JoinedStreams_evictor_rdh
/** * Sets the {@code Evictor} that should be used to evict elements from a window before * emission. * * <p>Note: When using an evictor window performance will degrade significantly, since * pre-aggregation of window results cannot be used. */ @PublicEvolving public WithWindow<T1, T2, KEY, W> evictor(Evictor<? s...
3.26
flink_JoinedStreams_window_rdh
/** * Specifies the window on which the join operation works. */ @PublicEvolving public <W extends Window> WithWindow<T1, T2, KEY, W> window(WindowAssigner<? super TaggedUnion<T1, T2>, W> assigner) { return new WithWindow<>(input1, input2, keySelector1, keySelector2, keyType, assigner, null, null, null); }
3.26
flink_JoinedStreams_where_rdh
/** * Specifies a {@link KeySelector} for elements from the first input with explicit type * information for the key type. * * @param keySelector * The KeySelector to be used for extracting the first input's key for * partitioning. * @param keyType * The type information describing the key type. */ publi...
3.26
flink_JoinedStreams_allowedLateness_rdh
/** * Sets the time by which elements are allowed to be late. * * @see WindowedStream#allowedLateness(Time) */ @PublicEvolving public WithWindow<T1, T2, KEY, W> allowedLateness(Time newLateness) { return new WithWindow<>(input1, input2, keySelector1, keySelector2, keyType, windowAssigner, trigger, evictor, ...
3.26
flink_JoinedStreams_equalTo_rdh
/** * Specifies a {@link KeySelector} for elements from the second input with explicit type * information for the key type. * * @param keySelector * The KeySelector to be used for extracting the second input's key for * partitioning. * @param keyType * The type information describing the key type. */ pub...
3.26
flink_JoinedStreams_trigger_rdh
/** * Sets the {@code Trigger} that should be used to trigger window emission. */ @PublicEvolving public WithWindow<T1, T2, KEY, W> trigger(Trigger<? super TaggedUnion<T1, T2>, ? super W> newTrigger) { return new WithWindow<>(input1, input2, keySelector1, keySelector2, keyType, windowAssigner, newTrigger, evictor...
3.26
flink_JoinedStreams_apply_rdh
/** * Completes the join operation with the user function that is executed for each combination * of elements with the same key in a window. * * <p>Note: This method's return type does not support setting an operator-specific * parallelism. Due to binary backwards compatibility, this cannot be altered. Use the * ...
3.26
flink_MapValue_isEmpty_rdh
/* (non-Javadoc) @see java.util.Map#isEmpty() */ @Override public boolean isEmpty() { return this.map.isEmpty(); }
3.26
flink_MapValue_get_rdh
/* (non-Javadoc) @see java.util.Map#get(java.lang.Object) */@Override public V get(final Object key) { return this.map.get(key); }
3.26
flink_MapValue_containsValue_rdh
/* (non-Javadoc) @see java.util.Map#containsValue(java.lang.Object) */ @Override public boolean containsValue(final Object value) { return this.map.containsValue(value); }
3.26
flink_MapValue_keySet_rdh
/* (non-Javadoc) @see java.util.Map#keySet() */ @Override public Set<K> keySet() { return this.map.keySet(); }
3.26
flink_MapValue_put_rdh
/* (non-Javadoc) @see java.util.Map#put(java.lang.Object, java.lang.Object) */ @Override public V put(final K key, final V value) { return this.map.put(key, value); }
3.26
flink_MapValue_clear_rdh
/* (non-Javadoc) @see java.util.Map#clear() */ @Override public void clear() { this.map.clear(); }
3.26
flink_MapValue_m0_rdh
/* (non-Javadoc) @see java.util.Map#putAll(java.util.Map) */ @Override public void m0(final Map<? extends K, ? extends V> m) { this.map.putAll(m); }
3.26
flink_MapValue_containsKey_rdh
/* (non-Javadoc) @see java.util.Map#containsKey(java.lang.Object) */ @Override public boolean containsKey(final Object key) { return this.map.containsKey(key); }
3.26
flink_MapValue_toString_rdh
/* (non-Javadoc) @see java.lang.Object#toString() */ @Override public String toString() { return this.map.toString(); }
3.26
flink_MapValue_equals_rdh
/* (non-Javadoc) @see java.lang.Object#equals(java.lang.Object) */ @Override public boolean equals(final Object obj) { if (this == obj) { return true; } if (obj == null) { return false; } if (this.getClass() != obj.getClass()) { return false; } final MapValue<?, ?> ...
3.26
flink_MapValue_entrySet_rdh
/* (non-Javadoc) @see java.util.Map#entrySet() */ @Override public Set<Entry<K, V>> entrySet() { return this.map.entrySet(); }
3.26
flink_MapValue_remove_rdh
/* (non-Javadoc) @see java.util.Map#remove(java.lang.Object) */ @Override public V remove(final Object key) { return this.map.remove(key); }
3.26
flink_MapValue_size_rdh
/* (non-Javadoc) @see java.util.Map#size() */ @Override public int size() { return this.map.size(); }
3.26
flink_MapValue_hashCode_rdh
/* (non-Javadoc) @see java.lang.Object#hashCode() */ @Override public int hashCode() { final int prime = 47; int result = 1; result = (prime * result) + this.map.hashCode(); return result; }
3.26
flink_Tuple9_setFields_rdh
/** * Sets new values to all fields of the tuple. * * @param f0 * The value for field 0 * @param f1 * The value for field 1 * @param f2 * The value for field 2 * @param f3 * The value for field 3 * @param f4 * The value for field 4 * @param f5 * The value for field 5 * @param f6 * The valu...
3.26
flink_Tuple9_toString_rdh
// ------------------------------------------------------------------------------------------------- // standard utilities // ------------------------------------------------------------------------------------------------- /** * Creates a string representation of the tuple in the form (f0, f1, f2, f3, f4, f5, f6, f7,...
3.26
flink_Tuple9_copy_rdh
/** * Shallow tuple copy. * * @return A new Tuple with the same fields as this. */ @Override @SuppressWarnings("unchecked") public Tuple9<T0, T1, T2, T3, T4, T5, T6, T7, T8> copy() { return new Tuple9<>(this.f0, this.f1, this.f2, this.f3, this.f4, this.f5, this.f6, this.f7, this.f8); }
3.26
flink_Tuple9_m0_rdh
/** * Deep equality for tuples by calling equals() on the tuple members. * * @param o * the object checked for equality * @return true if this is equal to o. */ @Override public boolean m0(Object o) { if (this == o) { return true; } if (!(o instanceof Tuple9)) { return false; } @Suppress...
3.26
flink_Tuple9_of_rdh
/** * Creates a new tuple and assigns the given values to the tuple's fields. This is more * convenient than using the constructor, because the compiler can infer the generic type * arguments implicitly. For example: {@code Tuple3.of(n, x, s)} instead of {@code new * Tuple3<Integer, Double, String>(n, x, s)} */ pu...
3.26
flink_DoubleSumAggregator_aggregate_rdh
/** * Adds the given value to the current aggregate. * * @param value * The value to add to the aggregate. */ public void aggregate(double value) { sum += value; }
3.26
flink_TpcdsResultComparator_compareLine_rdh
// ------------------------------------------------------------------------ private static boolean compareLine(String expectedLine, String actualLine) { return compareLineInternal(expectedLine, actualLine, -1, false); }
3.26
flink_TpcdsResultComparator_isEqualCol_rdh
// ------------------------------------------------------------------------ private static boolean isEqualCol(String expected, String actual) { return (isEqualNull(expected, actual) || isEqualNumber(expected, actual)) || expected.equals(actual); }
3.26
flink_TpcdsResultComparator_compareQuery34_rdh
// ------------------------------------------------------------------------ private static boolean compareQuery34(String[] expectedLines, String[] actualLines) { // take the first two lines and move them back to lines 7 and 8 final String expected1 = expectedLines[0]; final String expected2 = expectedLines[...
3.26
flink_OptionalFailure_get_rdh
/** * * @return stored value or throw a {@link FlinkException} with {@code failureCause}. */ public T get() throws FlinkException { if (value != null) { return value; } checkNotNull(failureCause); throw new FlinkException(failureCause); } /** * * @return same as {@link #get()} but throws a...
3.26
flink_PushFilterIntoSourceScanRuleBase_resolveFiltersAndCreateTableSourceTable_rdh
/** * Resolves filters using the underlying sources {@link SupportsFilterPushDown} and creates a * new {@link TableSourceTable} with the supplied predicates. * * @param convertiblePredicates * Predicates to resolve * @param oldTableSourceTable * TableSourceTable to copy * @param scan * Underlying table s...
3.26
flink_EdgeManagerBuildUtil_computeMaxEdgesToTargetExecutionVertex_rdh
/** * Given parallelisms of two job vertices, compute the max number of edges connected to a target * execution vertex from the source execution vertices. Note that edge is considered undirected * here. It can be an edge connected from an upstream job vertex to a downstream job vertex, or * in a reversed way. * *...
3.26
flink_EdgeManagerBuildUtil_connectInternal_rdh
/** * Connect all execution vertices to all partitions. */ private static void connectInternal(List<ExecutionVertex> taskVertices, List<IntermediateResultPartition> partitions, ResultPartitionType resultPartitionType, EdgeManager edgeManager) { checkState(!taskVertices.isEmpty()); checkState(!partitions.isEm...
3.26
flink_EdgeManagerBuildUtil_connectVertexToResult_rdh
/** * Calculate the connections between {@link ExecutionJobVertex} and {@link IntermediateResult} * * based on the {@link DistributionPattern}. * * @param vertex * the downstream consumer {@link ExecutionJobVertex} * @param intermediateResult * the upstream consumed {@link IntermediateResult} */ static void...
3.26
flink_TumblingWindowAssigner_of_rdh
// ------------------------------------------------------------------------ // Utilities // ------------------------------------------------------------------------ /** * Creates a new {@code TumblingWindowAssigner} {@link WindowAssigner} that assigns elements to * time windows based on the element timestamp. * * @...
3.26
flink_TumblingWindowAssigner_withOffset_rdh
/** * Creates a new {@code TumblingWindowAssigner} {@link WindowAssigner} that assigns elements to * time windows based on the element timestamp and offset. * * <p>For example, if you want window a stream by hour,but window begins at the 15th minutes of * each hour, you can use {@code of(Time.hours(1),Time.minutes...
3.26
flink_CliRowView_init_rdh
// -------------------------------------------------------------------------------------------- @Override protected void init() { // nothing to do }
3.26
flink_Tuple16_equals_rdh
/** * Deep equality for tuples by calling equals() on the tuple members. * * @param o * the object checked for equality * @return true if this is equal to o. */ @Override public boolean equals(Object o) { if (this == o) { return true; } if (!(o instanceof Tuple16)) { return false; } @SuppressWarnings("rawtype...
3.26
flink_Tuple16_of_rdh
/** * Creates a new tuple and assigns the given values to the tuple's fields. This is more * convenient than using the constructor, because the compiler can infer the generic type * arguments implicitly. For example: {@code Tuple3.of(n, x, s)} instead of {@code new * Tuple3<Integer, Double, String>(n, x, s)} */ pu...
3.26
flink_Tuple16_copy_rdh
/** * Shallow tuple copy. * * @return A new Tuple with the same fields as this. */ @Override @SuppressWarnings("unchecked") public Tuple16<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15> copy() { return new Tuple16<>(this.f0, this.f1, this.f2, this.f3, this.f4, this.f5, this.f6, this.f7, this...
3.26
flink_Tuple16_toString_rdh
// ------------------------------------------------------------------------------------------------- // standard utilities // ------------------------------------------------------------------------------------------------- /** * Creates a string representation of the tuple in the form (f0, f1, f2, f3, f4, f5, f6, f7,...
3.26