name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_StructuredOptionsSplitter_escapeWithSingleQuote_rdh | /**
* Escapes the given string with single quotes, if the input string contains a double quote or
* any of the given {@code charsToEscape}. Any single quotes in the input string will be escaped
* by doubling.
*
* <p>Given that the escapeChar is (;)
*
* <p>Examples:
*
* <ul>
* <li>A,B,C,D => A,B,C,D
* <li... | 3.26 |
flink_BucketWriter_openNewCompactingFile_rdh | /**
* Used to create a new {@link CompactingFileWriter} of the requesting type. Requesting a writer
* of an unsupported type will result in UnsupportedOperationException. By default, only
* RECORD_WISE type is supported, for which a {@link InProgressFileWriter} will be created.
*
* @param type
* the type of thi... | 3.26 |
flink_JsonRowSerializationSchema_build_rdh | /**
* Finalizes the configuration and checks validity.
*
* @return Configured {@link JsonRowSerializationSchema}
*/
public JsonRowSerializationSchema build() {
checkArgument(typeInfo != null, "typeInfo should be set.");return
new JsonRowSeria... | 3.26 |
flink_JsonRowSerializationSchema_withTypeInfo_rdh | /**
* Sets type information for JSON serialization schema.
*
* @param typeInfo
* Type information describing the result type. The field names of {@link Row} are used to parse the JSON properties.
*/
public Builder withTypeInfo(TypeInformation<Row> typeInfo) {
checkArgument(typeInfo instanceof RowTypeInfo, "O... | 3.26 |
flink_SlidingWindowCheckMapper_verifyPreviousOccurences_rdh | /**
* Verifies if all values from previous windows appear in the new one. Returns union of all
* events seen so far that were not seen <b>slideFactor</b> number of times yet.
*/
private List<Tuple2<Event, Integer>> verifyPreviousOccurences(List<Tuple2<Event, Integer>> previousWindowValues, List<Event> newValues, Lon... | 3.26 |
flink_BuildSideIterator_setBucket_rdh | // update current bucket status.
private void setBucket(MemorySegment bucket, MemorySegment[] overflowSegments, int bucketInSegmentOffset) {
this.bucketSegment = bucket;
this.overflowSegments = overflowSegments;
this.bucketInSegmentOffset = bucketInSegmentOffset;
this.pointerOffset = bucketInSegmentOffs... | 3.26 |
flink_BuildSideIterator_m0_rdh | /**
* Move to next bucket, return true while move to a on heap bucket, return false while move
* to a spilled bucket or there is no more bucket.
*/
private boolean m0() {
scanCount++;
if (scanCount >= area.numBuck... | 3.26 |
flink_IntermediateDataSet_toString_rdh | // --------------------------------------------------------------------------------------------
@Override
public String toString() {return ("Intermediate Data Set (" + id) + ")";
} | 3.26 |
flink_IntermediateDataSet_getId_rdh | // --------------------------------------------------------------------------------------------
public IntermediateDataSetID getId() {
return id;
} | 3.26 |
flink_IntermediateDataSet_addConsumer_rdh | // --------------------------------------------------------------------------------------------
public void addConsumer(JobEdge edge) {
// sanity check
checkState(id.equals(edge.getSourceId()), "Incompatible dataset id.");
if (consumers.isEmpty()) {
distributionPattern = edge.getDistributionPattern();
isBroadc... | 3.26 |
flink_SourceCoordinatorContext_getAndRemoveUncheckpointedAssignment_rdh | /**
* Get the split to put back. This only happens when a source reader subtask has failed.
*
* @param subtaskId
* the failed subtask id.
* @param restoredCheckpointId
* the checkpoint that the task is recovered to.
* @return A list of splits that needs to be added back to the {@link SplitEnumerator}.
*/
Li... | 3.26 |
flink_SourceCoordinatorContext_schedulePeriodTask_rdh | /**
* To avoid period task lost, we should handle the potential exception throw by task.
*/
ScheduledFuture<?> schedulePeriodTask(Runnable command, long initDelay, long period, TimeUnit unit) {
return coordinatorExecutor.scheduleAtFixedRate(() -> {
try {
command.run();
} catch (Throwab... | 3.26 |
flink_SourceCoordinatorContext_m3_rdh | /**
* Unregister a source reader.
*
* @param subtaskId
* the subtask id of the source reader.
* @param attemptNumber
* the attempt number of the source reader.
*/
void m3(int subtaskId, int attemptNumber) {
final Map<Integer, ReaderInfo> attemptReaders = registeredReaders.get(subtaskId);
if
(atte... | 3.26 |
flink_SourceCoordinatorContext_checkSubtaskIndex_rdh | // ---------------- private helper methods -----------------
private void checkSubtaskIndex(int subtaskIndex)
{
if ((subtaskIndex < 0) || (subtaskIndex >= getCoordinatorContext().currentParallelism())) {throw new IllegalArgumentException(String.format("Subtask index %d is out of bounds [0, %s)", subtaskIndex, getCo... | 3.26 |
flink_SourceCoordinatorContext_submitTask_rdh | // ---------------- Executor methods to avoid use coordinatorExecutor directly -----------------
Future<?> submitTask(Runnable task) {
return coordinatorExecutor.submit(task);
} | 3.26 |
flink_SourceCoordinatorContext_onCheckpoint_rdh | /**
* Behavior of SourceCoordinatorContext on checkpoint.
*
* @param checkpointId
* The id of the ongoing checkpoint.
*/
void onCheckpoint(long checkpointId) throws Exception {
assignmentTracker.onCheckpoint(checkpointId);
} | 3.26 |
flink_SourceCoordinatorContext_attemptReady_rdh | // --------- Package private additional methods for the SourceCoordinator ------------
void attemptReady(OperatorCoordinator.SubtaskGateway gateway) {
checkState(coordinatorThreadFactory.isCurrentThreadCoordinatorThread());
subtaskGateways.registerSubtaskGateway(gateway);
} | 3.26 |
flink_SourceCoordinatorContext_callInCoordinatorThread_rdh | /**
* A helper method that delegates the callable to the coordinator thread if the current thread
* is not the coordinator thread, otherwise call the callable right away.
*
* @param callable
* the callable to delegate.
*/
private <V> V callInCoordinatorThread(Callable<V> callable, String errorMessage) {
// ... | 3.26 |
flink_SourceCoordinatorContext_failJob_rdh | /**
* Fail the job with the given cause.
*
* @param cause
* the cause of the job failure.
*/
void failJob(Throwable cause) {
operatorCoordinatorContext.failJob(cause);
} | 3.26 |
flink_SourceCoordinatorContext_registerSourceReader_rdh | /**
* Register a source reader.
*
* @param subtaskId
* the subtask id of the source reader.
* @param attemptNumber
* the attempt number of the source reader.
* @param location
* the location of the source reader.
*/
void registerSourceReader(int subtaskId,
int attemptNumber, String location) {
final ... | 3.26 |
flink_SourceCoordinatorContext_runInCoordinatorThread_rdh | /**
* {@inheritDoc } If the runnable throws an Exception, the corresponding job is failed.
*/
@Override
public void runInCoordinatorThread(Runnable runnable) {
// when using a ScheduledThreadPool, uncaught exception handler catches only
// exceptions thrown by the threadPool, so manually call it when the exc... | 3.26 |
flink_SourceCoordinatorContext_onCheckpointComplete_rdh | /**
* Invoked when a successful checkpoint has been taken.
*
* @param checkpointId
* the id of the successful checkpoint.
*/
void onCheckpointComplete(long checkpointId) {
assignmentTracker.onCheckpointComplete(checkpointId);
} | 3.26 |
flink_SegmentPartitionFile_getTieredStoragePath_rdh | // ------------------------------------------------------------------------
// File-related utilities
// ------------------------------------------------------------------------
public static String getTieredStoragePath(String basePath) {
return String.format("%s/%s", basePath, TIERED_STORAGE_DIR);
} | 3.26 |
flink_MapView_contains_rdh | /**
* Checks if the map view contains a value for a given key.
*
* @param key
* The key to check.
* @return True if there exists a value for the given key, false otherwise.
* @throws Exception
* Thrown if the system cannot access the map.
*/public
boolean contains(K key) throws Exception {
return map.co... | 3.26 |
flink_MapView_isEmpty_rdh | /**
* Returns true if the map view contains no key-value mappings, otherwise false.
*
* @return True if the map view contains no key-value mappings, otherwise false.
* @throws Exception
* Thrown if the system cannot access the state.
*/
public boolean isEmpty() throws Exception {
return map.isEmpty();
} | 3.26 |
flink_MapView_entries_rdh | /**
* Returns all entries of the map view.
*
* @return An iterable of all the key-value pairs in the map view.
* @throws Exception
* Thrown if the system cannot access the map.
*/
public Iterable<Map.Entry<K, V>> entries() throws
Exception {
return map.entrySet();
} | 3.26 |
flink_MapView_newMapViewDataType_rdh | // --------------------------------------------------------------------------------------------
// Utilities
// --------------------------------------------------------------------------------------------
/**
* Utility method for creating a {@link DataType} of {@link MapView} explicitly.
*/
public static DataType new... | 3.26 |
flink_MapView_put_rdh | /**
* Inserts a value for the given key into the map view. If the map view already contains a value
* for the key, the existing value is overwritten.
*
* @param key
* The key for which the value is inserted.
* @param value
* The value that is inserted for the key.
* @throws Exception
* Thrown if the syst... | 3.26 |
flink_MapView_remove_rdh | /**
* Deletes the value for the given key.
*
* @param key
* The key for which the value is deleted.
* @throws Exception
* Thrown if the system cannot access the map.
*/
public void remove(K key) throws Exception {
map.remove(key);
} | 3.26 |
flink_MapView_putAll_rdh | /**
* Inserts all mappings from the specified map to this map view.
*
* @param map
* The map whose entries are inserted into this map view.
* @throws Exception
* Thrown if the system cannot access the map.
*/
public void putAll(Map<K, V> map) throws Exception {
this.map.putAll(map);
} | 3.26 |
flink_MapView_values_rdh | /**
* Returns all the values in the map view.
*
* @return An iterable of all the values in the map.
* @throws Exception
* Thrown if the system cannot access the map.
*/
public Iterable<V> values() throws Exception {
return map.values();
} | 3.26 |
flink_MapView_iterator_rdh | /**
* Returns an iterator over all entries of the map view.
*
* @return An iterator over all the mappings in the map.
* @throws Exception
* Thrown if the system cannot access the map.
*/
public Iterator<Map.Entry<K, V>> iterator() throws Exception {
return map.entrySet().iterator(); } | 3.26 |
flink_MapView_setMap_rdh | /**
* Replaces the entire view's content with the content of the given {@link Map}.
*/
public void setMap(Map<K, V> map) {
this.map = map;
} | 3.26 |
flink_MapView_clear_rdh | /**
* Removes all entries of this map.
*/
@Override
public void clear() {
map.clear();
} | 3.26 |
flink_MapView_keys_rdh | /**
* Returns all the keys in the map view.
*
* @return An iterable of all the keys in the map.
* @throws Exception
* Thrown if the system cannot access the map.
*/
public Iterable<K> keys() throws Exception {
return
map.keySet();
} | 3.26 |
flink_MapView_get_rdh | /**
* Return the value for the specified key or {@code null} if the key is not in the map view.
*
* @param key
* The look up key.
* @return The value for the specified key.
* @throws Exception
* Thrown if the system cannot get data.
*/
public V get(K key)
throws Exception {
return map.get(key);
} | 3.26 |
flink_JoinedRowData_replace_rdh | /**
* Replaces the {@link RowData} backing this {@link JoinedRowData}.
*
* <p>This method replaces the backing rows in place and does not return a new object. This is
* done for performance reasons.
*/
public JoinedRowData replace(RowData row1,
RowData row2) {
this.row1 = row1;
this.row2 = row2;
return... | 3.26 |
flink_MutableHashTable_ensureNumBuffersReturned_rdh | /**
* This method makes sure that at least a certain number of memory segments is in the list of
* free segments. Free memory can be in the list of free segments, or in the return-queue where
* segments used to write behind are put. The number of segments that are in that return-queue,
* but are actually reclaimabl... | 3.26 |
flink_MutableHashTable_buildBloomFilterForBucket_rdh | /**
* Set all the bucket memory except bucket header as the bit set of bloom filter, and use hash
* code of build records to build bloom filter.
*/
final void buildBloomFilterForBucket(int bucketInSegmentPos, MemorySegment bucket, HashPartition<BT, PT> p) {
final int count = bucket.getShort(bucketInSegmentPos + HEAD... | 3.26 |
flink_MutableHashTable_getPartitioningFanOutNoEstimates_rdh | /**
* Gets the number of partitions to be used for an initial hash-table, when no estimates are
* available.
*
* <p>The current logic makes sure that there are always between 10 and 127 partitions, and
* close to 0.1 of the number of buffers.
*
* @param numBuffers
* The number of buffers available.
* @return... | 3.26 |
flink_MutableHashTable_releaseTable_rdh | /**
* Releases the table (the array of buckets) and returns the occupied memory segments to the
* list of free segments.
*/
protected void releaseTable() {
// set the counters back
this.numBuckets = 0;
if (this.buckets != null) {
for (MemorySegment bucket : this.buckets) {
this.availableMemory.add(bucket);
}
this.... | 3.26 |
flink_MutableHashTable_hash_rdh | /**
* The level parameter is needed so that we can have different hash functions when we
* recursively apply the partitioning, so that the working set eventually fits into memory.
*/
public static int hash(int code, int level) {
final int rotation = level * 11;
code = Integer.rotateLeft(code, rotation);
return Mat... | 3.26 |
flink_MutableHashTable_getNumWriteBehindBuffers_rdh | // --------------------------------------------------------------------------------------------
// Utility Computational Functions
// --------------------------------------------------------------------------------------------
/**
* Determines the number of buffers to be used for asynchronous write behind. It is curre... | 3.26 |
flink_MutableHashTable_getNewInMemoryPartition_rdh | // --------------------------------------------------------------------------------------------
// Setup and Tear Down of Structures
// --------------------------------------------------------------------------------------------
/**
* Returns a new inMemoryPartition object. This is required as a plug for
* ReOpenable... | 3.26 |
flink_MutableHashTable_moveToNextBucket_rdh | /**
* Move to next bucket, return true while move to a on heap bucket, return false while move
* to a spilled bucket or there is no more bucket.
*/
private boolean moveToNextBucket() {
scanCount++;
if (scanCount > (f3 - 1)) {
return false;
}
// move to next bucket, update all the current bucket status with new bucke... | 3.26 |
flink_MutableHashTable_m0_rdh | /**
* Gets the next buffer to be used with the hash-table, either for an in-memory partition, or
* for the table buckets. This method returns <tt>null</tt>, if no more buffer is available.
* Spilling a partition may free new buffers then.
*
* @return The next buffer to be used by the hash-table, or null, if no buf... | 3.26 |
flink_MutableHashTable_spillPartition_rdh | // --------------------------------------------------------------------------------------------
// Memory Handling
// --------------------------------------------------------------------------------------------
/**
* Selects a partition and spills it. The number of the spilled partition is returned.
*
* @return The ... | 3.26 |
flink_MutableHashTable_open_rdh | /**
* Opens the hash join. This method reads the build-side input and constructs the initial hash
* table, gradually spilling partitions that do not fit into memory.
*
* @param buildSide
* Build side input.
* @param probeSide
* Probe side input.
* @param buildOuterJoin
* Whether outer join on build side.... | 3.26 |
flink_MutableHashTable_close_rdh | /**
* Closes the hash table. This effectively releases all internal structures and closes all open
* files and removes them. The call to this method is valid both as a cleanup after the complete
* inputs were properly processed, and as an cancellation call, which cleans up all resources
* that are currently held by... | 3.26 |
flink_MutableHashTable_assignPartition_rdh | /**
* Assigns a partition to a bucket.
*
* @param bucket
* The bucket to get the partition for.
* @param numPartitions
* The number of partitions.
* @return The partition for the bucket.
*/
public static byte assignPartition(int bucket, byte numPartitions) {
return ((byte) (bucket % numPartitions));
} | 3.26 |
flink_MutableHashTable_setBucket_rdh | // update current bucket status.
private void setBucket(MemorySegment bucket, MemorySegment[] overflowSegments, HashPartition<BT, PT> partition, int bucketInSegmentOffset) {
this.bucketSegment = bucket;
this.overflowSegments = overflowSegments;
this.partition = partition;
this.bucketInSegmentOffset = bucketInSegmentOf... | 3.26 |
flink_MutableHashTable_buildInitialTable_rdh | // ------------------------------------------------------------------------
// Hash Table Building
// ------------------------------------------------------------------------
/**
* Creates the initial hash table. This method sets up partitions, hash index, and inserts the
* data from the given iterator.
*
* @param ... | 3.26 |
flink_MutableHashTable_clearPartitions_rdh | /**
* This method clears all partitions currently residing (partially) in memory. It releases all
* memory and deletes all spilled partitions.
*
* <p>This method is intended for a hard cleanup in the case that the join is aborted.
*/
protected void clearPartitions() {
for (int v104 = this.partitionsBeingBuilt.size... | 3.26 |
flink_MutableHashTable_nextSegment_rdh | /**
* This is the method called by the partitions to request memory to serialize records. It
* automatically spills partitions, if memory runs out.
*
* @return The next available memory segment.
*/
@Override
public MemorySegment nextSegment() {
final MemorySegment seg = m0();
if (seg != null) {
return seg;
} els... | 3.26 |
flink_BaseVectorizedColumnReader_nextInt_rdh | /**
* Reading zero always.
*/protected static final class NullIntIterator implements IntIterator {
@Override
public int nextInt() {
return 0;
} | 3.26 |
flink_FreeingBufferRecycler_recycle_rdh | /**
* Frees the given memory segment.
*
* @param memorySegment
* The memory segment to be recycled.
*/
@Override
public void recycle(MemorySegment memorySegment) {
memorySegment.free();
} | 3.26 |
flink_FlinkSemiAntiJoinFilterTransposeRule_onMatch_rdh | // implement RelOptRule
public void onMatch(RelOptRuleCall call) {
LogicalJoin join = call.rel(0);
LogicalFilter filter = call.rel(1);
RelNode newJoin = LogicalJoin.create(filter.getInput(), join.getRight(), join.getHints(), join.getCondition(), join.getVariablesSet(), join.getJoinType());final RelFactori... | 3.26 |
flink_FlinkSemiAntiJoinFilterTransposeRule_matches_rdh | // ~ Methods ----------------------------------------------------------------
@Override
public boolean matches(RelOptRuleCall call) {
LogicalJoin join = call.rel(0);
return (join.getJoinType() == JoinRelType.SEMI) || (join.getJoinType() == JoinRelType.ANTI);
} | 3.26 |
flink_ExternalResourceUtils_getExternalResourceAmountMap_rdh | /**
* Get the map of resource name and amount of all of enabled external resources.
*/
@VisibleForTesting
static Map<String, Long> getExternalResourceAmountMap(Configuration config) {
final Set<String> resourceSet = getExternalResourceSet(con... | 3.26 |
flink_ExternalResourceUtils_createStaticExternalResourceInfoProvider_rdh | /**
* Instantiate {@link StaticExternalResourceInfoProvider} for all of enabled external resources.
*/
@VisibleForTesting
static ExternalResourceInfoProvider createStaticExternalResourceInfoProvider(Map<String, Long> externalResourceAmountMap, Map<String, Ex... | 3.26 |
flink_ExternalResourceUtils_getExternalResourceSet_rdh | /**
* Get the enabled external resource list from configuration.
*/
private static Set<String> getExternalResourceSet(Configuration config) {
if (config.getValue(ExternalResourceOptions.EXTERNAL_RESOURCE_LIST).equals(ExternalResourceOptions.NONE)) {
return Collections.emptySet();
}
return new ... | 3.26 |
flink_ExternalResourceUtils_externalResourceDriversFromConfig_rdh | /**
* Instantiate the {@link ExternalResourceDriver ExternalResourceDrivers} for all of enabled
* external resources. {@link ExternalResourceDriver ExternalResourceDrivers} are mapped to its
* resource name.
*/
@VisibleF... | 3.26 |
flink_ExternalResourceUtils_createStaticExternalResourceInfoProviderFromConfig_rdh | /**
* Instantiate {@link StaticExternalResourceInfoProvider} for all of enabled external resources.
*/public static ExternalResourceInfoProvider createStaticExternalResourceInfoProviderFromConfig(Configuration configuration, PluginManager pluginManager) {
final Map<String, Long> externalResourceAmountMap = ge... | 3.26 |
flink_ExternalResourceUtils_m0_rdh | /**
* Get the collection of all enabled external resources.
*/
public static Collection<ExternalResource> m0(Configuration config) {
return getExternalResourceAmountMap(config).entrySet().stream().map(entry -> new ExternalResource(entry.getKey(), entry.getValue())).collect(Collectors.toList());
} | 3.26 |
flink_ExternalResourceUtils_generateExternalResourcesString_rdh | /**
* Generate the string expression of the given external resources.
*/
public static String generateExternalResourcesString(Collection<ExternalResource> extendedResources) {
return extendedResources.stream().map(resource -> (resource.getName() + "=") + resource.getValue()).collect(Collectors.joining(", "));
} | 3.26 |
flink_ExternalSorterBuilder_build_rdh | /**
* Creates a push-based {@link PushSorter}. The {@link PushSorter#getIterator()} will return
* when the {@link PushSorter#finishReading()} is called. Will spawn two threads: sort, spill.
*/
public PushSorter<T> build() throws MemoryAllocationException {
PushFactory<T> pushFactory = new PushFactory<>();
Ex... | 3.26 |
flink_FileSystemBlobStore_put_rdh | // - Put ------------------------------------------------------------------
@Override
public boolean put(File localFile, JobID jobId, BlobKey blobKey) throws IOException {
createBasePathIfNeeded();
String toBlobPath = BlobUtils.getStorageLocationPath(basePath, jobId, blobKey);
try (F... | 3.26 |
flink_FileSystemBlobStore_delete_rdh | // - Delete ---------------------------------------------------------------
@Override
public boolean delete(JobID jobId, BlobKey blobKey) {
return delete(BlobUtils.getStorageLocationPath(basePath, jobId, blobKey));
} | 3.26 |
flink_FileSystemBlobStore_get_rdh | // - Get ------------------------------------------------------------------
@Override
public boolean get(JobID jobId, BlobKey blobKey, File localFile) throws IOException {return get(BlobUtils.getStorageLocationPath(basePath, jobId, blobKey), localFile, blobKey);} | 3.26 |
flink_KryoSerializer_getKryoInstance_rdh | // --------------------------------------------------------------------------------------------
/**
* Returns the Chill Kryo Serializer which is implicitly added to the classpath via
* flink-runtime. Falls back to the default Kryo serializer if it can't be found.
*
* @return The Kryo serializer instance.
*/
privat... | 3.26 |
flink_KryoSerializer_readObject_rdh | // --------------------------------------------------------------------------------------------
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException {
in.defaultReadObject();
// kryoRegistrations may be null if this Kryo serializer is deserialized from an old version
if (kryoRegist... | 3.26 |
flink_KryoSerializer_isImmutableType_rdh | // ------------------------------------------------------------------------
@Override
public boolean isImmutableType() {
return false;
} | 3.26 |
flink_KryoSerializer_buildKryoRegistrations_rdh | // --------------------------------------------------------------------------------------------
// Utilities
// --------------------------------------------------------------------------------------------
/**
* Utility method that takes lists of registered types and their serializers, and resolve them
* into a single... | 3.26 |
flink_KryoSerializer_m1_rdh | // --------------------------------------------------------------------------------------------
@Override
public int m1() {
int result = type.hashCode();
result = (31 * result) + kryoRegistrations.hashCode();
result = (31 * result) + defaultSerializers.hashCode();
result = (31 * result) + defaultSerializerClasses.ha... | 3.26 |
flink_KryoSerializer_enterExclusiveThread_rdh | // --------------------------------------------------------------------------------------------
// For testing
// --------------------------------------------------------------------------------------------
private void enterExclusiveThread() {
// we use simple get, check, set here, rather than CAS
// we don't need loc... | 3.26 |
flink_KryoSerializer_snapshotConfiguration_rdh | // --------------------------------------------------------------------------------------------
// Serializer configuration snapshotting & compatibility
// --------------------------------------------------------------------------------------------
@Override
public TypeSerializerSnapshot<T> snapshotConfiguration() {
... | 3.26 |
flink_TtlUtils_expired_rdh | /**
* Common functions related to State TTL.
*/ public class TtlUtils {
static <V> boolean expired(@Nullable
TtlValue<V> ttlValue, long ttl, TtlTimeProvider timeProvider) {
return expired(ttlValue, ttl, timeProvider.currentTimestamp());
} | 3.26 |
flink_ProjectOperator_projectTuple14_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13> ProjectOperator<T, Tuple14<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13>> projectTu... | 3.26 |
flink_ProjectOperator_projectTuple6_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5> ProjectOperator<T, Tuple6<T0, T1, T2, T3, T4, T5>> projectTuple6() {
TypeInformation<?>[] fTypes = extractFieldTypes(fieldIndexe... | 3.26 |
flink_ProjectOperator_projectTuple25_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21, T22, T23, T24> ProjectOperator<T, Tuple25<T0, T1, T2, T3,
... | 3.26 |
flink_ProjectOperator_projectTuple21_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15,
T16, T17, T18, T19, T20> ProjectOperator<T, Tuple21<T0, T1, T2, T3, T4, T5, T6, T7, T8, T... | 3.26 |
flink_ProjectOperator_projectTuple2_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1> ProjectOperator<T, Tuple2<T0, T1>> projectTuple2() {
TypeInformation<?>[] fTypes = extractFieldTypes(fieldIndexes, ds.getType());
TupleTypeIn... | 3.26 |
flink_ProjectOperator_projectTuple18_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17> ProjectOperator<T, Tuple18<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11,... | 3.26 |
flink_ProjectOperator_projectTuple19_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18> ProjectOperator<T, Tuple19<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T... | 3.26 |
flink_ProjectOperator_projectTuple15_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14> ProjectOperator<T, Tuple15<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T... | 3.26 |
flink_ProjectOperator_projectTuple24_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21, T22, T23> ProjectOperator<T, Tuple24<T0, T1, T2, T3, T4, T... | 3.26 |
flink_ProjectOperator_projectTuple7_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5, T6> ProjectOperator<T, Tuple7<T0, T1, T2, T3, T4, T5, T6>> projectTuple7() {
TypeInformation<?>[] fTypes = extractFieldTypes(fi... | 3.26 |
flink_ProjectOperator_projectTuple3_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2> ProjectOperator<T, Tuple3<T0, T1, T2>> projectTuple3() {
TypeInformation<?>[] fTypes
= extractFieldTypes(fieldIndexes, ds.getType());
... | 3.26 |
flink_ProjectOperator_projectTuple20_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4,
T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19> ProjectOperator<T, Tuple20<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, ... | 3.26 |
flink_ProjectOperator_projectTuple11_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10> ProjectOperator<T, Tuple11<T0, T1, T2,
T3, T4, T5, T6, T7, T8, T9, T10>> projectTuple11() {
TypeInformation<?... | 3.26 |
flink_ProjectOperator_projectTupleX_rdh | // --------------------------------------------------------------------------------------------
// The following lines are generated.
// --------------------------------------------------------------------------------------------
// BEGIN_OF_TUPLE_DEPENDENT_CODE
// GENERATED FROM org.apache.flink.api.java.tuple.TupleGe... | 3.26 |
flink_ProjectOperator_projectTuple9_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8> ProjectOperator<T, Tuple9<T0, T1, T2, T3, T4, T5, T6, T7, T8>> projectTuple9() {
TypeInformation<?>[] fTypes = extract... | 3.26 |
flink_ProjectOperator_types_rdh | /**
*
* @deprecated Deprecated method only kept for compatibility.
*/
@SuppressWarnings("unchecked")
@Deprecated
@PublicEvolving
public <R extends Tuple> ProjectOperator<IN, R> types(Class<?>... types) {
TupleTypeInfo<R> typeInfo = ((TupleTypeInfo<R>) (this.getResultType()));
if (types.length != typeInfo.get... | 3.26 |
flink_ProjectOperator_projectTuple17_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16> ProjectOperator<T, Tuple17<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, ... | 3.26 |
flink_ProjectOperator_projectTuple5_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4> ProjectOperator<T, Tuple5<T0, T1,
T2, T3, T4>> projectTuple5() {
TypeInformation<?>[] fTypes = extractFieldTypes(fieldIndexes, ds.get... | 3.26 |
flink_ProjectOperator_projectTuple22_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21> ProjectOperator<T, Tuple22<T0, T1, T2, T3, T4, T5, T6, T... | 3.26 |
flink_ProjectOperator_projectTuple13_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2,
T3, T4, T5, T6, T7, T8, T9, T10, T11, T12> ProjectOperator<T, Tuple13<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12>> projectTuple13() {
... | 3.26 |
flink_ProjectOperator_projectTuple16_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15> ProjectOperator<T, Tuple16<T0, T1, T2,
T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T... | 3.26 |
flink_ProjectOperator_projectTuple4_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3> ProjectOperator<T, Tuple4<T0, T1, T2, T3>> projectTuple4() {
TypeInformation<?>[] fTypes = extractFieldTypes(fieldIndexes, ds.getType());
... | 3.26 |
flink_ProjectOperator_projectTuple23_rdh | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21, T22> ProjectOperator<T, Tuple23<T0, T1, T2, T3, T4, T5, T6... | 3.26 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.