name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
flink_ResultPartitionFactory_createBufferPoolFactory_rdh
/** * The minimum pool size should be <code>numberOfSubpartitions + 1</code> for two * considerations: * * <p>1. StreamTask can only process input if there is at-least one available buffer on output * side, so it might cause stuck problem if the minimum pool size is exactly equal to the number * of subpartitions,...
3.26
flink_SchedulerBase_startScheduling_rdh
// ------------------------------------------------------------------------ // SchedulerNG // ------------------------------------------------------------------------ @Override public final void startScheduling() { mainThreadExecutor.assertRunningInMainThread(); registerJobMetrics(jobManagerJobMetricGroup, executionGra...
3.26
flink_SchedulerBase_getExecutionGraph_rdh
/** * ExecutionGraph is exposed to make it easier to rework tests to be based on the new scheduler. * ExecutionGraph is expected to be used only for state check. Yet at the moment, before all the * actions are factored out from ExecutionGraph and its sub-components, some actions may still * be performed directly on...
3.26
flink_SchedulerBase_computeVertexParallelismStore_rdh
/** * Compute the {@link VertexParallelismStore} for all vertices of a given job graph, which will * set defaults and ensure that the returned store contains valid parallelisms. * * @param jobGraph * the job graph to retrieve vertices from * @return the computed parallelism store */ public static VertexParall...
3.26
flink_SchedulerBase_getJobId_rdh
// ------------------------------------------------------------------------ // access utils for testing // ------------------------------------------------------------------------ @VisibleForTesting JobID getJobId() { return f0.getJobID(); }
3.26
flink_SchedulerBase_getDefaultMaxParallelism_rdh
/** * Get a default value to use for a given vertex's max parallelism if none was specified. * * @param vertex * the vertex to compute a default max parallelism for * @return the computed max parallelism */ public static int getDefaultMaxParallelism(JobVertex vertex) { return KeyGroupRangeAssignment.compu...
3.26
flink_SchedulerBase_deliverOperatorEventToCoordinator_rdh
// Note: It may be worthwhile to move the OperatorCoordinators out // of the scheduler (have them owned by the JobMaster directly). // Then we could avoid routing these events through the scheduler and // doing this lazy initialization dance. However, this would require // that the Scheduler does not eagerly construct ...
3.26
flink_StreamElement_asWatermarkStatus_rdh
/** * Casts this element into a WatermarkStatus. * * @return This element as a WatermarkStatus. * @throws java.lang.ClassCastException * Thrown, if this element is actually not a Watermark * Status. */ public final WatermarkStatus asWatermarkStatus() {return ((WatermarkStatus) (this)); }
3.26
flink_StreamElement_asWatermark_rdh
/** * Casts this element into a Watermark. * * @return This element as a Watermark. * @throws java.lang.ClassCastException * Thrown, if this element is actually not a Watermark. */public final Watermark asWatermark() { return ((Watermark) (this)); }
3.26
flink_StreamElement_isLatencyMarker_rdh
/** * Checks whether this element is a latency marker. * * @return True, if this element is a latency marker, false otherwise. */ public final boolean isLatencyMarker() { return getClass() == LatencyMarker.class; }
3.26
flink_StreamElement_asRecord_rdh
/** * Casts this element into a StreamRecord. * * @return This element as a stream record. * @throws java.lang.ClassCastException * Thrown, if this element is actually not a stream record. */ @SuppressWarnings("unchecked") public final <E> StreamRecord<E> asRecord() { return ((StreamRecord<E>) (this)); }
3.26
flink_StreamElement_isWatermarkStatus_rdh
/** * Checks whether this element is a watermark status. * * @return True, if this element is a watermark status, false otherwise. */ public final boolean isWatermarkStatus() { return getClass() == WatermarkStatus.class; }
3.26
flink_StreamElement_asLatencyMarker_rdh
/** * Casts this element into a LatencyMarker. * * @return This element as a LatencyMarker. * @throws java.lang.ClassCastException * Thrown, if this element is actually not a LatencyMarker. */ public final LatencyMarker asLatencyMarker() { return ((LatencyMarker) (this)); }
3.26
flink_FileChannelManagerImpl_close_rdh
/** * Remove all the temp directories. */ @Override public void close() throws Exception { // Marks shutdown and exits if it has already shutdown. if (!isShutdown.compareAndSet(false, true)) { return; } IOUtils.closeAll(Arrays.stream(paths).filter(File::exists).map(FileChannelManagerImpl::getF...
3.26
flink_ClusterClient_reportHeartbeat_rdh
/** * The client reports the heartbeat to the dispatcher for aliveness. * * @param jobId * The jobId for the client and the job. * @return */ default CompletableFuture<Void> reportHeartbeat(JobID jobId, long expiredTimestamp) { return FutureUtils.completedVoidFuture(); }
3.26
flink_ClusterClient_getAccumulators_rdh
/** * Requests and returns the accumulators for the given job identifier. Accumulators can be * requested while a is running or after it has finished. The default class loader is used to * deserialize the incoming accumulator results. * * @param jobID * The job identifier of a job. * @return A Map containing t...
3.26
flink_ClusterClient_listCompletedClusterDatasetIds_rdh
/** * Return a set of ids of the completed cluster datasets. * * @return A set of ids of the completely cached intermediate dataset. */default CompletableFuture<Set<AbstractID>> listCompletedClusterDatasetIds() { return CompletableFuture.completedFuture(Collections.emptySet()); }
3.26
flink_NFA_extractCurrentMatches_rdh
/** * Extracts all the sequences of events from the start to the given computation state. An event * sequence is returned as a map which contains the events and the names of the states to which * the event...
3.26
flink_NFA_open_rdh
/** * Initialization method for the NFA. It is called before any element is passed and thus * suitable for one time setup work. * * @param cepRuntimeContext * runtime context of the enclosing operator * @param conf * The configuration containing the parameters attached to the contract. */ public void open(R...
3.26
flink_NFA_computeNextStates_rdh
/** * Computes the next computation states based on the given computation state, the current event, * its timestamp and the internal state machine. The algorithm is: * * <ol> * <li>Decide on valid transitions and number of branching paths. See {@link OutgoingEdges} * <li>Perform transitions: * <ol> * ...
3.26
flink_NFA_close_rdh
/** * Tear-down method for the NFA. */ public void close() throws Exception { for (State<T> v10 : getStates()) { for (StateTransition<T> transition : v10.getStateTransitions()) { IterativeCondition condition = transition.getCondition(); FunctionUtils.closeFunction(...
3.26
flink_NFA_advanceTime_rdh
/** * Prunes states assuming there will be no events with timestamp <b>lower</b> than the given * one. It clears the sharedBuffer and also emits all timed out partial matches. * * @param sharedBufferAccessor * the accessor to SharedBuffer object that we need to work upon * while processing * @param nfaState ...
3.26
flink_JoinOperationFactory_create_rdh
/** * Creates a valid {@link JoinQueryOperation} operation. * * <p>It performs validations such as: * * <ul> * <li>condition returns boolean * <li>the condition is either always true or contains equi join * <li>left and right side of the join do not contain ambiguous column names * <li>that correlated ...
3.26
flink_AnotherDummyFSFileSystem_getUri_rdh
// ------------------------------------------------------------------------ @Override public URI getUri() { return FS_URI; }
3.26
flink_RocksDBHandle_restoreInstanceDirectoryFromPath_rdh
/** * This recreates the new working directory of the recovered RocksDB instance and links/copies * the contents from a local state. */ private void restoreInstanceDirectoryFromPath(Path source) throws IOException { final Path instanceRocksDBDirectory = Paths.get(f0); final Path[] files = FileUtils.listDir...
3.26
flink_BridgingSqlFunction_of_rdh
/** * Creates an instance of a scalar or table built-in function during translation. */ public static BridgingSqlFunction of(RelOptCluster cluster, BuiltInFunctionDefinition functionDefinition) { return m0(cluster, ContextResolvedFunction.permanent(FunctionIdentifier.of(functionDefinition.getName()), functionDef...
3.26
flink_BridgingSqlFunction_getRowTypeInference_rdh
/** * The conversion to a row type is handled on the caller side. This allows us to perform it * SQL/Table API-specific. This is in particular important to set the aliases of fields * correctly (see {@link FlinkRelBuilder#pushFunctionScan(RelBuilder, SqlOperator, int, * Iterable, List)}). */@Override public SqlRet...
3.26
flink_FileStateHandle_discardState_rdh
/** * Discard the state by deleting the file that stores the state. If the parent directory of the * state is empty after deleting the state file, it is also deleted. * * @throws Exception * Thrown, if the file deletion (not the directory deletion) fails. */ @Override public void discardState() throws Exception...
3.26
flink_FileStateHandle_equals_rdh
// ------------------------------------------------------------------------ @Override public boolean equals(Object o) { if (this == o) { return true; } if (!(o instanceof FileStateHandle)) { return false; }FileStateHandle that = ((FileStateHandle) (o)); return filePath.equals(that.filePath); }
3.26
flink_FileStateHandle_getFilePath_rdh
/** * Gets the path where this handle's state is stored. * * @return The path where this handle's state is stored. */ public Path getFilePath() {return filePath; }
3.26
flink_FileStateHandle_getStateSize_rdh
/** * Returns the file size in bytes. * * @return The file size in bytes. */ @Override public long getStateSize() { return stateSize; }
3.26
flink_FileStateHandle_getFileSystem_rdh
/** * Gets the file system that stores the file state. * * @return The file system that stores the file state. * @throws IOException * Thrown if the file system cannot be accessed. */ private FileSystem getFileSystem() throws IOException { return FileSystem.get(filePath.toUri()); }
3.26
flink_ThreadInfoRequestCoordinator_triggerThreadInfoRequest_rdh
/** * Triggers collection of thread info stats of a job vertex by combining thread info responses * from given subtasks. A thread info response of a subtask in turn consists of {@code numSamples}, collected with {@code delayBetweenSamples} milliseconds delay between them. * * @param executionsWithGateways * Exec...
3.26
flink_ThreadInfoRequestCoordinator_requestThreadInfo_rdh
/** * Requests thread infos from given subtasks. The response would be ignored if it does not * return within timeout. */ private void requestThreadInfo(Map<ImmutableSet<ExecutionAttemptID>, CompletableFuture<TaskExecutorThreadInfoGateway>> executionWithGateways, ThreadInfoSamplesRequest requestParams, Time timeou...
3.26
flink_SortPartitionOperatorBase_executeOnCollections_rdh
// -------------------------------------------------------------------------------------------- @Override protected List<IN> executeOnCollections(List<IN> inputData, RuntimeContext runtimeContext, ExecutionConfig executionConfig) { TypeInformation<IN> inputType ...
3.26
flink_BroadcastPartitioner_selectChannel_rdh
/** * Note: Broadcast mode could be handled directly for all the output channels in record writer, * so it is no need to select channels via this method. */ @Override public int selectChannel(SerializationDelegate<StreamRecord<T>> record) { throw new UnsupportedOperationException("Broadcast partitioner does not ...
3.26
flink_TaskExecutorMemoryConfiguration_getTaskHeap_rdh
/** * Returns the configured heap size used by the tasks. */ public Long getTaskHeap() { return taskHeap; }
3.26
flink_TaskExecutorMemoryConfiguration_getFrameworkHeap_rdh
/** * Returns the configured heap size used by the framework. */ public Long getFrameworkHeap() { return frameworkHeap; }
3.26
flink_TaskExecutorMemoryConfiguration_getFrameworkOffHeap_rdh
/** * Returns the configured off-heap size used by the framework. */ public Long getFrameworkOffHeap() { return frameworkOffHeap; }
3.26
flink_TaskExecutorMemoryConfiguration_getNetworkMemory_rdh
/** * Returns the configured maximum network memory. */ public Long getNetworkMemory() { return networkMemory; }
3.26
flink_TaskExecutorMemoryConfiguration_m0_rdh
/** * Factory method for initializing a TaskExecutorMemoryConfiguration based on the passed * Configuration. * * @param config * The Configuration used for initializing the TaskExecutorMemoryConfiguration. * @return The newly instantiated TaskExecutorMemoryConfiguration. */ public static TaskExecutorMemoryConf...
3.26
flink_TaskExecutorMemoryConfiguration_getJvmMetaspace_rdh
/** * Returns the maximum Metaspace size allowed for the task manager. */ public Long getJvmMetaspace() { return jvmMetaspace; }
3.26
flink_TaskExecutorMemoryConfiguration_getTaskOffHeap_rdh
/** * Returns the configured off-heap size used by the tasks. */ public Long getTaskOffHeap() { return taskOffHeap; }
3.26
flink_TaskExecutorMemoryConfiguration_getManagedMemoryTotal_rdh
/** * Returns the total amount of memory reserved for by the MemoryManager. */ public Long getManagedMemoryTotal() { return managedMemoryTotal; }
3.26
flink_SubQueryInputTypeStrategy_getArgumentCount_rdh
/** * {@link InputTypeStrategy} for {@link BuiltInFunctionDefinitions#IN}. */ @Internalpublic class SubQueryInputTypeStrategy implements InputTypeStrategy { @Override public ArgumentCount getArgumentCount() {return ConstantArgumentCount.from(2); }
3.26
flink_SourceCoordinatorSerdeUtils_readAndVerifyCoordinatorSerdeVersion_rdh
/** * Read and verify the serde version. */ static int readAndVerifyCoordinatorSerdeVersion(DataInputStream in) throws IOException { int version = in.readInt(); if (version > CURRENT_VERSION) { throw new IOException("Unsupported source coordinator serde version " + version); } return version; ...
3.26
flink_SourceCoordinatorSerdeUtils_writeCoordinatorSerdeVersion_rdh
/** * Write the current serde version. */ static void writeCoordinatorSerdeVersion(DataOutputStream out) throws IOException {out.writeInt(CURRENT_VERSION); }
3.26
flink_DataSinkTask_getLogString_rdh
// ------------------------------------------------------------------------ // Utilities // ------------------------------------------------------------------------ /** * Utility function that composes a string for logging purposes. The string includes the given * message and the index of the task in its task group t...
3.26
flink_DataSinkTask_initOutputFormat_rdh
/** * Initializes the OutputFormat implementation and configuration. * * @throws RuntimeException * Throws if instance of OutputFormat implementation can not be * obtained. */ private void initOutputFormat() { ClassLoader userCodeClassLoader = getUserCodeClassLoader(); // obtain task configuration (in...
3.26
flink_DataSinkTask_initInputReaders_rdh
/** * Initializes the input readers of the DataSinkTask. * * @throws RuntimeException * Thrown in case of invalid task input configuration. */ @SuppressWarnings("unchecked") private void initInputReaders() throws Exception { int numGates = 0; // ---------------- create the input readers -----------...
3.26
flink_TransientBlobCache_m0_rdh
/** * Returns the blob expiry times - for testing purposes only! * * @return blob expiry times (internal state!) */ @VisibleForTesting ConcurrentMap<Tuple2<JobID, TransientBlobKey>, Long> m0() { return blobExpiryTimes; }
3.26
flink_TransientBlobCache_getStorageLocation_rdh
/** * Returns a file handle to the file associated with the given blob key on the blob server. * * @param jobId * ID of the job this blob belongs to (or <tt>null</tt> if job-unrelated) * @param key * identifying the file * @return file handle to the file * @throws IOException * if creating the directory ...
3.26
flink_DataOutputEncoder_writeFixed_rdh
// -------------------------------------------------------------------------------------------- // bytes // -------------------------------------------------------------------------------------------- @Override public void writeFixed(byte[] bytes, int start, int len) throws IOException { out.write(bytes, start, len...
3.26
flink_DataOutputEncoder_writeString_rdh
// -------------------------------------------------------------------------------------------- // strings // -------------------------------------------------------------------------------------------- @Override public void writeString(String str) throws IOException { byte[] bytes = Utf8.getBytesFor(str); writ...
3.26
flink_DataOutputEncoder_writeNull_rdh
// -------------------------------------------------------------------------------------------- // primitives // -------------------------------------------------------------------------------------------- @Override public void writeNull() { }
3.26
flink_DataOutputEncoder_writeVarLongCount_rdh
// -------------------------------------------------------------------------------------------- // utils // -------------------------------------------------------------------------------------------- public static void writeVarLongCount(DataOutput out, long val) throws IOException { if (val < 0) { throw ne...
3.26
flink_DataOutputEncoder_writeIndex_rdh
// -------------------------------------------------------------------------------------------- // union // -------------------------------------------------------------------------------------------- @Override public void writeIndex(int unionIndex) throws IOException { out.writeInt(unionIndex); }
3.26
flink_DataOutputEncoder_writeArrayStart_rdh
// -------------------------------------------------------------------------------------------- // collection types // -------------------------------------------------------------------------------------------- @Override public void writeArrayStart() { }
3.26
flink_AvroRowSerializationSchema_convertRowToAvroRecord_rdh
// -------------------------------------------------------------------------------------------- private GenericRecord convertRowToAvroRecord(Schema schema, Row row) { final List<Schema.Field> fields = schema.getFields(); final int length = fields.size(); final GenericRecord record = new GenericData.Record(s...
3.26
flink_KeyedStateTransformation_window_rdh
/** * Windows this transformation into a {@code WindowedOperatorTransformation}, which bootstraps * state that can be restored by a {@code WindowOperator}. Elements are put into windows by a * {@link WindowAssigner}. The grouping of elements is done both by key and by window. * * <p>A {@link org.apache.flink.strea...
3.26
flink_KeyedStateTransformation_transform_rdh
/** * Method for passing user defined operators along with the type information that will transform * the OperatorTransformation. * * <p><b>IMPORTANT:</b> Any output from this operator will be discarded. * * @param factory * A factory returning transformation logic type of the return stream * @return An {@lin...
3.26
flink_StreamProjection_projectTuple17_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16> SingleOutputStreamOperator<Tuple17<T0, T1, T2, T3, T4, T5, T6, T7, T8, T...
3.26
flink_StreamProjection_projectTuple4_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */public <T0, T1, T2, T3> SingleOutputStreamOperator<Tuple4<T0, T1, T2, T3>> projectTuple4() { TypeInformation<?>[] fTypes = m1(fieldIndexes, dataStream.get...
3.26
flink_StreamProjection_projectTuple13_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12> SingleOutputStreamOperator<Tuple13<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12>> p...
3.26
flink_StreamProjection_projectTuple22_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21> SingleOutputStreamOperator<Tuple22<T0, T1, T2, ...
3.26
flink_StreamProjection_projectTuple8_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public <T0, T1, T2, T3, T4, T5, T6, T7> SingleOutputStreamOperator<Tuple8<T0, T1, T2, T3, T4, T5, T6, T7>> projectTuple8() { TypeInformation<?>[] fTypes...
3.26
flink_StreamProjection_projectTupleX_rdh
/** * Chooses a projectTupleX according to the length of {@link org.apache.flink.streaming.api.datastream.StreamProjection#fieldIndexes}. * * @return The projected DataStream. * @see org.apache.flink.api.java.operators.ProjectOperator.Projection */ @SuppressWarnings("unchecked") public <OUT extends Tuple> SingleOu...
3.26
flink_StreamProjection_projectTuple9_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8> SingleOutputStreamOperator<Tuple9<T0, T1, T2, T3, T4, T5, T6, T7, T8>> projectTuple9() { TypeInformation<?>[]...
3.26
flink_StreamProjection_projectTuple5_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public <T0, T1, T2, T3, T4> SingleOutputStreamOperator<Tuple5<T0, T1, T2, T3, T4>> projectTuple5() { TypeInformation<?>[] fTypes = m1(fieldIndexes, data...
3.26
flink_StreamProjection_projectTuple16_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15> SingleOutputStreamOperator<Tuple16<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T...
3.26
flink_StreamProjection_projectTuple12_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11> SingleOutputStreamOperator<Tuple12<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11>> projectT...
3.26
flink_StreamProjection_projectTuple23_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21, T22> SingleOutputStreamOperator<Tuple23<T0, T1, T...
3.26
flink_StreamProjection_projectTuple18_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17> SingleOutputStreamOperator<Tuple18<T0, T1, T2, T3, T4, T5, T6, T7, T8...
3.26
flink_StreamProjection_projectTuple25_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21, T22, T23, T24> SingleOutputStreamOperator<Tuple2...
3.26
flink_StreamProjection_projectTuple14_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13> SingleOutputStreamOperator<Tuple14<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12,...
3.26
flink_StreamProjection_projectTuple7_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public <T0, T1, T2, T3, T4, T5, T6> SingleOutputStreamOperator<Tuple7<T0, T1, T2, T3, T4, T5, T6>> projectTuple7() { TypeInformation<?>[] fTypes = m1(fie...
3.26
flink_StreamProjection_projectTuple10_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9> SingleOutputStreamOperator<Tuple10<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9>> projectTuple10() { TypeInform...
3.26
flink_StreamProjection_projectTuple3_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public <T0, T1, T2> SingleOutputStreamOperator<Tuple3<T0, T1, T2>> projectTuple3() { TypeInformation<?>[] fTypes = m1(fieldIndexes, dataStream.getTyp...
3.26
flink_StreamProjection_projectTuple21_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20> SingleOutputStreamOperator<Tuple21<T0, T1, T2, T3...
3.26
flink_StreamProjection_projectTuple24_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21, T22, T23> SingleOutputStreamOperator<Tuple24<...
3.26
flink_StreamProjection_projectTuple6_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public <T0, T1, T2, T3, T4, T5> SingleOutputStreamOperator<Tuple6<T0, T1, T2, T3, T4, T5>> projectTuple6() { TypeInformation<?>[] fTypes = m1(fieldIndexe...
3.26
flink_StreamProjection_projectTuple15_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14> SingleOutputStreamOperator<Tuple15<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11...
3.26
flink_StreamProjection_projectTuple2_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public <T0, T1> SingleOutputStreamOperator<Tuple2<T0, T1>> projectTuple2() { TypeInformation<?>[] fTypes = m1(fieldIndexes, dataStream.getType()); Tu...
3.26
flink_StreamProjection_projectTuple20_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19> SingleOutputStreamOperator<Tuple20<T0, T1, T2, T3, T4, T5,...
3.26
flink_StreamProjection_projectTuple19_rdh
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18> SingleOutputStreamOperator<Tuple19<T0, T1, T2, T3, T4, T5, T6, T...
3.26
flink_OptimizedPlan_getJobName_rdh
/** * Returns the name of the program. * * @return The name of the program. */ public String getJobName() { return this.jobName; }
3.26
flink_OptimizedPlan_accept_rdh
// ------------------------------------------------------------------------ /** * Applies the given visitor top down to all nodes, starting at the sinks. * * @param visitor * The visitor to apply to the nodes in this plan. * @see org.apache.flink.util.Visitable#accept(org.apache.flink.util.Visitor) */ @Override...
3.26
flink_OptimizedPlan_getAllNodes_rdh
/** * Gets all the nodes from this OptimizedPlan. * * @return All nodes. */ public Collection<PlanNode> getAllNodes() { return allNodes; }
3.26
flink_CoGroupedStreams_equalTo_rdh
/** * Specifies a {@link KeySelector} for elements from the second input. * * @param keySelector * The KeySelector to be used for extracting the second input's key for * partitioning. */ public EqualTo equalTo(KeySelector<T2, KEY> keySelector) { Preconditions.checkNotNull(keySelector); final TypeInfor...
3.26
flink_CoGroupedStreams_where_rdh
/** * Specifies a {@link KeySelector} for elements from the first input with explicit type * information. * * @param keySelector * The KeySelector to be used for extracting the first input's key for * partitioning. * @param keyType * The type information describing the key type. */ public <KEY> Where<KEY...
3.26
flink_CoGroupedStreams_m0_rdh
/** * Sets the {@code Trigger} that should be used to trigger window emission. */ @PublicEvolving public WithWindow<T1, T2, KEY, W> m0(Trigger<? super TaggedUnion<T1, T2>, ? super W> newTrigger) { return new WithWindow<>(input1, input2, keySelector1, keySelector2, keyType, windowAssigner, newTrigger, evictor, a...
3.26
flink_CoGroupedStreams_apply_rdh
/** * Completes the co-group operation with the user function that is executed for windowed * groups. * * <p>Note: This method's return type does not support setting an operator-specific * parallelism. Due to binary backwards compatibility, this cannot be altered. Use the * {@link #with(CoGroupFunction, TypeInfor...
3.26
flink_CoGroupedStreams_allowedLateness_rdh
/** * Sets the time by which elements are allowed to be late. * * @see WindowedStream#allowedLateness(Time) */ @PublicEvolving public WithWindow<T1, T2, KEY, W> allowedLateness(Time newLateness) { return new WithWindow<>(input1, input2, keySelector1, keySelector2, keyType, windowAssigner, trigger, evictor, ne...
3.26
flink_MessageParameter_getKey_rdh
/** * Returns the key of this parameter, e.g. "jobid". * * @return key of this parameter */ public final String getKey() { return f0; }
3.26
flink_MessageParameter_resolve_rdh
/** * Resolves this parameter for the given value. * * @param value * value to resolve this parameter with */ public final void resolve(X value) { Preconditions.checkState(!resolved, "This parameter was already resolved."); this.value = Preconditions.checkNotNull(value);this.resolved = true; }
3.26
flink_MessageParameter_m0_rdh
/** * Returns the resolved value of this parameter as a string, or {@code null} if it isn't * resolved yet. * * @return resolved value, or null if it wasn't resolved yet */ final String m0() { return value == null ? null : convertToString(value); }
3.26
flink_MessageParameter_getValue_rdh
/** * Returns the resolved value of this parameter, or {@code null} if it isn't resolved yet. * * @return resolved value, or null if it wasn't resolved yet */ public final X getValue() { return value; }
3.26
flink_NotDuplicatingCheckpointStateToolset_canFastDuplicate_rdh
/** * An empty implementation of {@link CheckpointStateToolset}. */public final class NotDuplicatingCheckpointStateToolset implements CheckpointStateToolset { @Override public boolean canFastDuplicate(StreamStateHandle stateHandle) throws IOException { return false; }
3.26
flink_LegacySinkTransformation_setStateKeySelector_rdh
/** * Sets the {@link KeySelector} that must be used for partitioning keyed state of this Sink. * * @param stateKeySelector * The {@code KeySelector} to set */ public void setStateKeySelector(KeySelector<T, ?> stateKeySelector) { this.stateKeySelector = stateKeySelector; updateManagedMemoryStateBackendUs...
3.26
flink_LegacySinkTransformation_getOperatorFactory_rdh
/** * Returns the {@link StreamOperatorFactory} of this {@code LegacySinkTransformation}. */ public StreamOperatorFactory<Object> getOperatorFactory() { return operatorFactory; }
3.26