name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
flink_StructuredOptionsSplitter_escapeWithSingleQuote_rdh
/** * Escapes the given string with single quotes, if the input string contains a double quote or * any of the given {@code charsToEscape}. Any single quotes in the input string will be escaped * by doubling. * * <p>Given that the escapeChar is (;) * * <p>Examples: * * <ul> * <li>A,B,C,D => A,B,C,D * <li...
3.26
flink_BucketWriter_openNewCompactingFile_rdh
/** * Used to create a new {@link CompactingFileWriter} of the requesting type. Requesting a writer * of an unsupported type will result in UnsupportedOperationException. By default, only * RECORD_WISE type is supported, for which a {@link InProgressFileWriter} will be created. * * @param type * the type of thi...
3.26
flink_JsonRowSerializationSchema_build_rdh
/** * Finalizes the configuration and checks validity. * * @return Configured {@link JsonRowSerializationSchema} */ public JsonRowSerializationSchema build() { checkArgument(typeInfo != null, "typeInfo should be set.");return new JsonRowSeria...
3.26
flink_JsonRowSerializationSchema_withTypeInfo_rdh
/** * Sets type information for JSON serialization schema. * * @param typeInfo * Type information describing the result type. The field names of {@link Row} are used to parse the JSON properties. */ public Builder withTypeInfo(TypeInformation<Row> typeInfo) { checkArgument(typeInfo instanceof RowTypeInfo, "O...
3.26
flink_SlidingWindowCheckMapper_verifyPreviousOccurences_rdh
/** * Verifies if all values from previous windows appear in the new one. Returns union of all * events seen so far that were not seen <b>slideFactor</b> number of times yet. */ private List<Tuple2<Event, Integer>> verifyPreviousOccurences(List<Tuple2<Event, Integer>> previousWindowValues, List<Event> newValues, Lon...
3.26
flink_BuildSideIterator_setBucket_rdh
// update current bucket status. private void setBucket(MemorySegment bucket, MemorySegment[] overflowSegments, int bucketInSegmentOffset) { this.bucketSegment = bucket; this.overflowSegments = overflowSegments; this.bucketInSegmentOffset = bucketInSegmentOffset; this.pointerOffset = bucketInSegmentOffs...
3.26
flink_BuildSideIterator_m0_rdh
/** * Move to next bucket, return true while move to a on heap bucket, return false while move * to a spilled bucket or there is no more bucket. */ private boolean m0() { scanCount++; if (scanCount >= area.numBuck...
3.26
flink_IntermediateDataSet_toString_rdh
// -------------------------------------------------------------------------------------------- @Override public String toString() {return ("Intermediate Data Set (" + id) + ")"; }
3.26
flink_IntermediateDataSet_getId_rdh
// -------------------------------------------------------------------------------------------- public IntermediateDataSetID getId() { return id; }
3.26
flink_IntermediateDataSet_addConsumer_rdh
// -------------------------------------------------------------------------------------------- public void addConsumer(JobEdge edge) { // sanity check checkState(id.equals(edge.getSourceId()), "Incompatible dataset id."); if (consumers.isEmpty()) { distributionPattern = edge.getDistributionPattern(); isBroadc...
3.26
flink_SourceCoordinatorContext_getAndRemoveUncheckpointedAssignment_rdh
/** * Get the split to put back. This only happens when a source reader subtask has failed. * * @param subtaskId * the failed subtask id. * @param restoredCheckpointId * the checkpoint that the task is recovered to. * @return A list of splits that needs to be added back to the {@link SplitEnumerator}. */ Li...
3.26
flink_SourceCoordinatorContext_schedulePeriodTask_rdh
/** * To avoid period task lost, we should handle the potential exception throw by task. */ ScheduledFuture<?> schedulePeriodTask(Runnable command, long initDelay, long period, TimeUnit unit) { return coordinatorExecutor.scheduleAtFixedRate(() -> { try { command.run(); } catch (Throwab...
3.26
flink_SourceCoordinatorContext_m3_rdh
/** * Unregister a source reader. * * @param subtaskId * the subtask id of the source reader. * @param attemptNumber * the attempt number of the source reader. */ void m3(int subtaskId, int attemptNumber) { final Map<Integer, ReaderInfo> attemptReaders = registeredReaders.get(subtaskId); if (atte...
3.26
flink_SourceCoordinatorContext_checkSubtaskIndex_rdh
// ---------------- private helper methods ----------------- private void checkSubtaskIndex(int subtaskIndex) { if ((subtaskIndex < 0) || (subtaskIndex >= getCoordinatorContext().currentParallelism())) {throw new IllegalArgumentException(String.format("Subtask index %d is out of bounds [0, %s)", subtaskIndex, getCo...
3.26
flink_SourceCoordinatorContext_submitTask_rdh
// ---------------- Executor methods to avoid use coordinatorExecutor directly ----------------- Future<?> submitTask(Runnable task) { return coordinatorExecutor.submit(task); }
3.26
flink_SourceCoordinatorContext_onCheckpoint_rdh
/** * Behavior of SourceCoordinatorContext on checkpoint. * * @param checkpointId * The id of the ongoing checkpoint. */ void onCheckpoint(long checkpointId) throws Exception { assignmentTracker.onCheckpoint(checkpointId); }
3.26
flink_SourceCoordinatorContext_attemptReady_rdh
// --------- Package private additional methods for the SourceCoordinator ------------ void attemptReady(OperatorCoordinator.SubtaskGateway gateway) { checkState(coordinatorThreadFactory.isCurrentThreadCoordinatorThread()); subtaskGateways.registerSubtaskGateway(gateway); }
3.26
flink_SourceCoordinatorContext_callInCoordinatorThread_rdh
/** * A helper method that delegates the callable to the coordinator thread if the current thread * is not the coordinator thread, otherwise call the callable right away. * * @param callable * the callable to delegate. */ private <V> V callInCoordinatorThread(Callable<V> callable, String errorMessage) { // ...
3.26
flink_SourceCoordinatorContext_failJob_rdh
/** * Fail the job with the given cause. * * @param cause * the cause of the job failure. */ void failJob(Throwable cause) { operatorCoordinatorContext.failJob(cause); }
3.26
flink_SourceCoordinatorContext_registerSourceReader_rdh
/** * Register a source reader. * * @param subtaskId * the subtask id of the source reader. * @param attemptNumber * the attempt number of the source reader. * @param location * the location of the source reader. */ void registerSourceReader(int subtaskId, int attemptNumber, String location) { final ...
3.26
flink_SourceCoordinatorContext_runInCoordinatorThread_rdh
/** * {@inheritDoc } If the runnable throws an Exception, the corresponding job is failed. */ @Override public void runInCoordinatorThread(Runnable runnable) { // when using a ScheduledThreadPool, uncaught exception handler catches only // exceptions thrown by the threadPool, so manually call it when the exc...
3.26
flink_SourceCoordinatorContext_onCheckpointComplete_rdh
/** * Invoked when a successful checkpoint has been taken. * * @param checkpointId * the id of the successful checkpoint. */ void onCheckpointComplete(long checkpointId) { assignmentTracker.onCheckpointComplete(checkpointId); }
3.26
flink_SegmentPartitionFile_getTieredStoragePath_rdh
// ------------------------------------------------------------------------ // File-related utilities // ------------------------------------------------------------------------ public static String getTieredStoragePath(String basePath) { return String.format("%s/%s", basePath, TIERED_STORAGE_DIR); }
3.26
flink_MapView_contains_rdh
/** * Checks if the map view contains a value for a given key. * * @param key * The key to check. * @return True if there exists a value for the given key, false otherwise. * @throws Exception * Thrown if the system cannot access the map. */public boolean contains(K key) throws Exception { return map.co...
3.26
flink_MapView_isEmpty_rdh
/** * Returns true if the map view contains no key-value mappings, otherwise false. * * @return True if the map view contains no key-value mappings, otherwise false. * @throws Exception * Thrown if the system cannot access the state. */ public boolean isEmpty() throws Exception { return map.isEmpty(); }
3.26
flink_MapView_entries_rdh
/** * Returns all entries of the map view. * * @return An iterable of all the key-value pairs in the map view. * @throws Exception * Thrown if the system cannot access the map. */ public Iterable<Map.Entry<K, V>> entries() throws Exception { return map.entrySet(); }
3.26
flink_MapView_newMapViewDataType_rdh
// -------------------------------------------------------------------------------------------- // Utilities // -------------------------------------------------------------------------------------------- /** * Utility method for creating a {@link DataType} of {@link MapView} explicitly. */ public static DataType new...
3.26
flink_MapView_put_rdh
/** * Inserts a value for the given key into the map view. If the map view already contains a value * for the key, the existing value is overwritten. * * @param key * The key for which the value is inserted. * @param value * The value that is inserted for the key. * @throws Exception * Thrown if the syst...
3.26
flink_MapView_remove_rdh
/** * Deletes the value for the given key. * * @param key * The key for which the value is deleted. * @throws Exception * Thrown if the system cannot access the map. */ public void remove(K key) throws Exception { map.remove(key); }
3.26
flink_MapView_putAll_rdh
/** * Inserts all mappings from the specified map to this map view. * * @param map * The map whose entries are inserted into this map view. * @throws Exception * Thrown if the system cannot access the map. */ public void putAll(Map<K, V> map) throws Exception { this.map.putAll(map); }
3.26
flink_MapView_values_rdh
/** * Returns all the values in the map view. * * @return An iterable of all the values in the map. * @throws Exception * Thrown if the system cannot access the map. */ public Iterable<V> values() throws Exception { return map.values(); }
3.26
flink_MapView_iterator_rdh
/** * Returns an iterator over all entries of the map view. * * @return An iterator over all the mappings in the map. * @throws Exception * Thrown if the system cannot access the map. */ public Iterator<Map.Entry<K, V>> iterator() throws Exception { return map.entrySet().iterator(); }
3.26
flink_MapView_setMap_rdh
/** * Replaces the entire view's content with the content of the given {@link Map}. */ public void setMap(Map<K, V> map) { this.map = map; }
3.26
flink_MapView_clear_rdh
/** * Removes all entries of this map. */ @Override public void clear() { map.clear(); }
3.26
flink_MapView_keys_rdh
/** * Returns all the keys in the map view. * * @return An iterable of all the keys in the map. * @throws Exception * Thrown if the system cannot access the map. */ public Iterable<K> keys() throws Exception { return map.keySet(); }
3.26
flink_MapView_get_rdh
/** * Return the value for the specified key or {@code null} if the key is not in the map view. * * @param key * The look up key. * @return The value for the specified key. * @throws Exception * Thrown if the system cannot get data. */ public V get(K key) throws Exception { return map.get(key); }
3.26
flink_JoinedRowData_replace_rdh
/** * Replaces the {@link RowData} backing this {@link JoinedRowData}. * * <p>This method replaces the backing rows in place and does not return a new object. This is * done for performance reasons. */ public JoinedRowData replace(RowData row1, RowData row2) { this.row1 = row1; this.row2 = row2; return...
3.26
flink_MutableHashTable_ensureNumBuffersReturned_rdh
/** * This method makes sure that at least a certain number of memory segments is in the list of * free segments. Free memory can be in the list of free segments, or in the return-queue where * segments used to write behind are put. The number of segments that are in that return-queue, * but are actually reclaimabl...
3.26
flink_MutableHashTable_buildBloomFilterForBucket_rdh
/** * Set all the bucket memory except bucket header as the bit set of bloom filter, and use hash * code of build records to build bloom filter. */ final void buildBloomFilterForBucket(int bucketInSegmentPos, MemorySegment bucket, HashPartition<BT, PT> p) { final int count = bucket.getShort(bucketInSegmentPos + HEAD...
3.26
flink_MutableHashTable_getPartitioningFanOutNoEstimates_rdh
/** * Gets the number of partitions to be used for an initial hash-table, when no estimates are * available. * * <p>The current logic makes sure that there are always between 10 and 127 partitions, and * close to 0.1 of the number of buffers. * * @param numBuffers * The number of buffers available. * @return...
3.26
flink_MutableHashTable_releaseTable_rdh
/** * Releases the table (the array of buckets) and returns the occupied memory segments to the * list of free segments. */ protected void releaseTable() { // set the counters back this.numBuckets = 0; if (this.buckets != null) { for (MemorySegment bucket : this.buckets) { this.availableMemory.add(bucket); } this....
3.26
flink_MutableHashTable_hash_rdh
/** * The level parameter is needed so that we can have different hash functions when we * recursively apply the partitioning, so that the working set eventually fits into memory. */ public static int hash(int code, int level) { final int rotation = level * 11; code = Integer.rotateLeft(code, rotation); return Mat...
3.26
flink_MutableHashTable_getNumWriteBehindBuffers_rdh
// -------------------------------------------------------------------------------------------- // Utility Computational Functions // -------------------------------------------------------------------------------------------- /** * Determines the number of buffers to be used for asynchronous write behind. It is curre...
3.26
flink_MutableHashTable_getNewInMemoryPartition_rdh
// -------------------------------------------------------------------------------------------- // Setup and Tear Down of Structures // -------------------------------------------------------------------------------------------- /** * Returns a new inMemoryPartition object. This is required as a plug for * ReOpenable...
3.26
flink_MutableHashTable_moveToNextBucket_rdh
/** * Move to next bucket, return true while move to a on heap bucket, return false while move * to a spilled bucket or there is no more bucket. */ private boolean moveToNextBucket() { scanCount++; if (scanCount > (f3 - 1)) { return false; } // move to next bucket, update all the current bucket status with new bucke...
3.26
flink_MutableHashTable_m0_rdh
/** * Gets the next buffer to be used with the hash-table, either for an in-memory partition, or * for the table buckets. This method returns <tt>null</tt>, if no more buffer is available. * Spilling a partition may free new buffers then. * * @return The next buffer to be used by the hash-table, or null, if no buf...
3.26
flink_MutableHashTable_spillPartition_rdh
// -------------------------------------------------------------------------------------------- // Memory Handling // -------------------------------------------------------------------------------------------- /** * Selects a partition and spills it. The number of the spilled partition is returned. * * @return The ...
3.26
flink_MutableHashTable_open_rdh
/** * Opens the hash join. This method reads the build-side input and constructs the initial hash * table, gradually spilling partitions that do not fit into memory. * * @param buildSide * Build side input. * @param probeSide * Probe side input. * @param buildOuterJoin * Whether outer join on build side....
3.26
flink_MutableHashTable_close_rdh
/** * Closes the hash table. This effectively releases all internal structures and closes all open * files and removes them. The call to this method is valid both as a cleanup after the complete * inputs were properly processed, and as an cancellation call, which cleans up all resources * that are currently held by...
3.26
flink_MutableHashTable_assignPartition_rdh
/** * Assigns a partition to a bucket. * * @param bucket * The bucket to get the partition for. * @param numPartitions * The number of partitions. * @return The partition for the bucket. */ public static byte assignPartition(int bucket, byte numPartitions) { return ((byte) (bucket % numPartitions)); }
3.26
flink_MutableHashTable_setBucket_rdh
// update current bucket status. private void setBucket(MemorySegment bucket, MemorySegment[] overflowSegments, HashPartition<BT, PT> partition, int bucketInSegmentOffset) { this.bucketSegment = bucket; this.overflowSegments = overflowSegments; this.partition = partition; this.bucketInSegmentOffset = bucketInSegmentOf...
3.26
flink_MutableHashTable_buildInitialTable_rdh
// ------------------------------------------------------------------------ // Hash Table Building // ------------------------------------------------------------------------ /** * Creates the initial hash table. This method sets up partitions, hash index, and inserts the * data from the given iterator. * * @param ...
3.26
flink_MutableHashTable_clearPartitions_rdh
/** * This method clears all partitions currently residing (partially) in memory. It releases all * memory and deletes all spilled partitions. * * <p>This method is intended for a hard cleanup in the case that the join is aborted. */ protected void clearPartitions() { for (int v104 = this.partitionsBeingBuilt.size...
3.26
flink_MutableHashTable_nextSegment_rdh
/** * This is the method called by the partitions to request memory to serialize records. It * automatically spills partitions, if memory runs out. * * @return The next available memory segment. */ @Override public MemorySegment nextSegment() { final MemorySegment seg = m0(); if (seg != null) { return seg; } els...
3.26
flink_BaseVectorizedColumnReader_nextInt_rdh
/** * Reading zero always. */protected static final class NullIntIterator implements IntIterator { @Override public int nextInt() { return 0; }
3.26
flink_FreeingBufferRecycler_recycle_rdh
/** * Frees the given memory segment. * * @param memorySegment * The memory segment to be recycled. */ @Override public void recycle(MemorySegment memorySegment) { memorySegment.free(); }
3.26
flink_FlinkSemiAntiJoinFilterTransposeRule_onMatch_rdh
// implement RelOptRule public void onMatch(RelOptRuleCall call) { LogicalJoin join = call.rel(0); LogicalFilter filter = call.rel(1); RelNode newJoin = LogicalJoin.create(filter.getInput(), join.getRight(), join.getHints(), join.getCondition(), join.getVariablesSet(), join.getJoinType());final RelFactori...
3.26
flink_FlinkSemiAntiJoinFilterTransposeRule_matches_rdh
// ~ Methods ---------------------------------------------------------------- @Override public boolean matches(RelOptRuleCall call) { LogicalJoin join = call.rel(0); return (join.getJoinType() == JoinRelType.SEMI) || (join.getJoinType() == JoinRelType.ANTI); }
3.26
flink_ExternalResourceUtils_getExternalResourceAmountMap_rdh
/** * Get the map of resource name and amount of all of enabled external resources. */ @VisibleForTesting static Map<String, Long> getExternalResourceAmountMap(Configuration config) { final Set<String> resourceSet = getExternalResourceSet(con...
3.26
flink_ExternalResourceUtils_createStaticExternalResourceInfoProvider_rdh
/** * Instantiate {@link StaticExternalResourceInfoProvider} for all of enabled external resources. */ @VisibleForTesting static ExternalResourceInfoProvider createStaticExternalResourceInfoProvider(Map<String, Long> externalResourceAmountMap, Map<String, Ex...
3.26
flink_ExternalResourceUtils_getExternalResourceSet_rdh
/** * Get the enabled external resource list from configuration. */ private static Set<String> getExternalResourceSet(Configuration config) { if (config.getValue(ExternalResourceOptions.EXTERNAL_RESOURCE_LIST).equals(ExternalResourceOptions.NONE)) { return Collections.emptySet(); } return new ...
3.26
flink_ExternalResourceUtils_externalResourceDriversFromConfig_rdh
/** * Instantiate the {@link ExternalResourceDriver ExternalResourceDrivers} for all of enabled * external resources. {@link ExternalResourceDriver ExternalResourceDrivers} are mapped to its * resource name. */ @VisibleF...
3.26
flink_ExternalResourceUtils_createStaticExternalResourceInfoProviderFromConfig_rdh
/** * Instantiate {@link StaticExternalResourceInfoProvider} for all of enabled external resources. */public static ExternalResourceInfoProvider createStaticExternalResourceInfoProviderFromConfig(Configuration configuration, PluginManager pluginManager) { final Map<String, Long> externalResourceAmountMap = ge...
3.26
flink_ExternalResourceUtils_m0_rdh
/** * Get the collection of all enabled external resources. */ public static Collection<ExternalResource> m0(Configuration config) { return getExternalResourceAmountMap(config).entrySet().stream().map(entry -> new ExternalResource(entry.getKey(), entry.getValue())).collect(Collectors.toList()); }
3.26
flink_ExternalResourceUtils_generateExternalResourcesString_rdh
/** * Generate the string expression of the given external resources. */ public static String generateExternalResourcesString(Collection<ExternalResource> extendedResources) { return extendedResources.stream().map(resource -> (resource.getName() + "=") + resource.getValue()).collect(Collectors.joining(", ")); }
3.26
flink_ExternalSorterBuilder_build_rdh
/** * Creates a push-based {@link PushSorter}. The {@link PushSorter#getIterator()} will return * when the {@link PushSorter#finishReading()} is called. Will spawn two threads: sort, spill. */ public PushSorter<T> build() throws MemoryAllocationException { PushFactory<T> pushFactory = new PushFactory<>(); Ex...
3.26
flink_FileSystemBlobStore_put_rdh
// - Put ------------------------------------------------------------------ @Override public boolean put(File localFile, JobID jobId, BlobKey blobKey) throws IOException { createBasePathIfNeeded(); String toBlobPath = BlobUtils.getStorageLocationPath(basePath, jobId, blobKey); try (F...
3.26
flink_FileSystemBlobStore_delete_rdh
// - Delete --------------------------------------------------------------- @Override public boolean delete(JobID jobId, BlobKey blobKey) { return delete(BlobUtils.getStorageLocationPath(basePath, jobId, blobKey)); }
3.26
flink_FileSystemBlobStore_get_rdh
// - Get ------------------------------------------------------------------ @Override public boolean get(JobID jobId, BlobKey blobKey, File localFile) throws IOException {return get(BlobUtils.getStorageLocationPath(basePath, jobId, blobKey), localFile, blobKey);}
3.26
flink_KryoSerializer_getKryoInstance_rdh
// -------------------------------------------------------------------------------------------- /** * Returns the Chill Kryo Serializer which is implicitly added to the classpath via * flink-runtime. Falls back to the default Kryo serializer if it can't be found. * * @return The Kryo serializer instance. */ privat...
3.26
flink_KryoSerializer_readObject_rdh
// -------------------------------------------------------------------------------------------- private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException { in.defaultReadObject(); // kryoRegistrations may be null if this Kryo serializer is deserialized from an old version if (kryoRegist...
3.26
flink_KryoSerializer_isImmutableType_rdh
// ------------------------------------------------------------------------ @Override public boolean isImmutableType() { return false; }
3.26
flink_KryoSerializer_buildKryoRegistrations_rdh
// -------------------------------------------------------------------------------------------- // Utilities // -------------------------------------------------------------------------------------------- /** * Utility method that takes lists of registered types and their serializers, and resolve them * into a single...
3.26
flink_KryoSerializer_m1_rdh
// -------------------------------------------------------------------------------------------- @Override public int m1() { int result = type.hashCode(); result = (31 * result) + kryoRegistrations.hashCode(); result = (31 * result) + defaultSerializers.hashCode(); result = (31 * result) + defaultSerializerClasses.ha...
3.26
flink_KryoSerializer_enterExclusiveThread_rdh
// -------------------------------------------------------------------------------------------- // For testing // -------------------------------------------------------------------------------------------- private void enterExclusiveThread() { // we use simple get, check, set here, rather than CAS // we don't need loc...
3.26
flink_KryoSerializer_snapshotConfiguration_rdh
// -------------------------------------------------------------------------------------------- // Serializer configuration snapshotting & compatibility // -------------------------------------------------------------------------------------------- @Override public TypeSerializerSnapshot<T> snapshotConfiguration() { ...
3.26
flink_TtlUtils_expired_rdh
/** * Common functions related to State TTL. */ public class TtlUtils { static <V> boolean expired(@Nullable TtlValue<V> ttlValue, long ttl, TtlTimeProvider timeProvider) { return expired(ttlValue, ttl, timeProvider.currentTimestamp()); }
3.26
flink_ProjectOperator_projectTuple14_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13> ProjectOperator<T, Tuple14<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13>> projectTu...
3.26
flink_ProjectOperator_projectTuple6_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5> ProjectOperator<T, Tuple6<T0, T1, T2, T3, T4, T5>> projectTuple6() { TypeInformation<?>[] fTypes = extractFieldTypes(fieldIndexe...
3.26
flink_ProjectOperator_projectTuple25_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21, T22, T23, T24> ProjectOperator<T, Tuple25<T0, T1, T2, T3, ...
3.26
flink_ProjectOperator_projectTuple21_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20> ProjectOperator<T, Tuple21<T0, T1, T2, T3, T4, T5, T6, T7, T8, T...
3.26
flink_ProjectOperator_projectTuple2_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1> ProjectOperator<T, Tuple2<T0, T1>> projectTuple2() { TypeInformation<?>[] fTypes = extractFieldTypes(fieldIndexes, ds.getType()); TupleTypeIn...
3.26
flink_ProjectOperator_projectTuple18_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17> ProjectOperator<T, Tuple18<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11,...
3.26
flink_ProjectOperator_projectTuple19_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18> ProjectOperator<T, Tuple19<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T...
3.26
flink_ProjectOperator_projectTuple15_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14> ProjectOperator<T, Tuple15<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T...
3.26
flink_ProjectOperator_projectTuple24_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21, T22, T23> ProjectOperator<T, Tuple24<T0, T1, T2, T3, T4, T...
3.26
flink_ProjectOperator_projectTuple7_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5, T6> ProjectOperator<T, Tuple7<T0, T1, T2, T3, T4, T5, T6>> projectTuple7() { TypeInformation<?>[] fTypes = extractFieldTypes(fi...
3.26
flink_ProjectOperator_projectTuple3_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1, T2> ProjectOperator<T, Tuple3<T0, T1, T2>> projectTuple3() { TypeInformation<?>[] fTypes = extractFieldTypes(fieldIndexes, ds.getType()); ...
3.26
flink_ProjectOperator_projectTuple20_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19> ProjectOperator<T, Tuple20<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, ...
3.26
flink_ProjectOperator_projectTuple11_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10> ProjectOperator<T, Tuple11<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10>> projectTuple11() { TypeInformation<?...
3.26
flink_ProjectOperator_projectTupleX_rdh
// -------------------------------------------------------------------------------------------- // The following lines are generated. // -------------------------------------------------------------------------------------------- // BEGIN_OF_TUPLE_DEPENDENT_CODE // GENERATED FROM org.apache.flink.api.java.tuple.TupleGe...
3.26
flink_ProjectOperator_projectTuple9_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8> ProjectOperator<T, Tuple9<T0, T1, T2, T3, T4, T5, T6, T7, T8>> projectTuple9() { TypeInformation<?>[] fTypes = extract...
3.26
flink_ProjectOperator_types_rdh
/** * * @deprecated Deprecated method only kept for compatibility. */ @SuppressWarnings("unchecked") @Deprecated @PublicEvolving public <R extends Tuple> ProjectOperator<IN, R> types(Class<?>... types) { TupleTypeInfo<R> typeInfo = ((TupleTypeInfo<R>) (this.getResultType())); if (types.length != typeInfo.get...
3.26
flink_ProjectOperator_projectTuple17_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16> ProjectOperator<T, Tuple17<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, ...
3.26
flink_ProjectOperator_projectTuple5_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4> ProjectOperator<T, Tuple5<T0, T1, T2, T3, T4>> projectTuple5() { TypeInformation<?>[] fTypes = extractFieldTypes(fieldIndexes, ds.get...
3.26
flink_ProjectOperator_projectTuple22_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21> ProjectOperator<T, Tuple22<T0, T1, T2, T3, T4, T5, T6, T...
3.26
flink_ProjectOperator_projectTuple13_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12> ProjectOperator<T, Tuple13<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12>> projectTuple13() { ...
3.26
flink_ProjectOperator_projectTuple16_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15> ProjectOperator<T, Tuple16<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T...
3.26
flink_ProjectOperator_projectTuple4_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3> ProjectOperator<T, Tuple4<T0, T1, T2, T3>> projectTuple4() { TypeInformation<?>[] fTypes = extractFieldTypes(fieldIndexes, ds.getType()); ...
3.26
flink_ProjectOperator_projectTuple23_rdh
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21, T22> ProjectOperator<T, Tuple23<T0, T1, T2, T3, T4, T5, T6...
3.26