name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_ResultPartitionFactory_createBufferPoolFactory_rdh | /**
* The minimum pool size should be <code>numberOfSubpartitions + 1</code> for two
* considerations:
*
* <p>1. StreamTask can only process input if there is at-least one available buffer on output
* side, so it might cause stuck problem if the minimum pool size is exactly equal to the number
* of subpartitions,... | 3.26 |
flink_SchedulerBase_startScheduling_rdh | // ------------------------------------------------------------------------
// SchedulerNG
// ------------------------------------------------------------------------
@Override
public final void startScheduling() {
mainThreadExecutor.assertRunningInMainThread();
registerJobMetrics(jobManagerJobMetricGroup, executionGra... | 3.26 |
flink_SchedulerBase_getExecutionGraph_rdh | /**
* ExecutionGraph is exposed to make it easier to rework tests to be based on the new scheduler.
* ExecutionGraph is expected to be used only for state check. Yet at the moment, before all the
* actions are factored out from ExecutionGraph and its sub-components, some actions may still
* be performed directly on... | 3.26 |
flink_SchedulerBase_computeVertexParallelismStore_rdh | /**
* Compute the {@link VertexParallelismStore} for all vertices of a given job graph, which will
* set defaults and ensure that the returned store contains valid parallelisms.
*
* @param jobGraph
* the job graph to retrieve vertices from
* @return the computed parallelism store
*/
public static VertexParall... | 3.26 |
flink_SchedulerBase_getJobId_rdh | // ------------------------------------------------------------------------
// access utils for testing
// ------------------------------------------------------------------------
@VisibleForTesting
JobID getJobId() {
return f0.getJobID();
} | 3.26 |
flink_SchedulerBase_getDefaultMaxParallelism_rdh | /**
* Get a default value to use for a given vertex's max parallelism if none was specified.
*
* @param vertex
* the vertex to compute a default max parallelism for
* @return the computed max parallelism
*/
public static int getDefaultMaxParallelism(JobVertex vertex)
{
return KeyGroupRangeAssignment.compu... | 3.26 |
flink_SchedulerBase_deliverOperatorEventToCoordinator_rdh | // Note: It may be worthwhile to move the OperatorCoordinators out
// of the scheduler (have them owned by the JobMaster directly).
// Then we could avoid routing these events through the scheduler and
// doing this lazy initialization dance. However, this would require
// that the Scheduler does not eagerly construct ... | 3.26 |
flink_StreamElement_asWatermarkStatus_rdh | /**
* Casts this element into a WatermarkStatus.
*
* @return This element as a WatermarkStatus.
* @throws java.lang.ClassCastException
* Thrown, if this element is actually not a Watermark
* Status.
*/
public final WatermarkStatus asWatermarkStatus() {return ((WatermarkStatus) (this));
} | 3.26 |
flink_StreamElement_asWatermark_rdh | /**
* Casts this element into a Watermark.
*
* @return This element as a Watermark.
* @throws java.lang.ClassCastException
* Thrown, if this element is actually not a Watermark.
*/public final Watermark asWatermark() {
return ((Watermark) (this));
} | 3.26 |
flink_StreamElement_isLatencyMarker_rdh | /**
* Checks whether this element is a latency marker.
*
* @return True, if this element is a latency marker, false otherwise.
*/
public final boolean isLatencyMarker() {
return getClass() == LatencyMarker.class;
} | 3.26 |
flink_StreamElement_asRecord_rdh | /**
* Casts this element into a StreamRecord.
*
* @return This element as a stream record.
* @throws java.lang.ClassCastException
* Thrown, if this element is actually not a stream record.
*/
@SuppressWarnings("unchecked")
public final <E> StreamRecord<E> asRecord() {
return ((StreamRecord<E>) (this));
} | 3.26 |
flink_StreamElement_isWatermarkStatus_rdh | /**
* Checks whether this element is a watermark status.
*
* @return True, if this element is a watermark status, false otherwise.
*/
public final boolean
isWatermarkStatus() {
return getClass() == WatermarkStatus.class;
} | 3.26 |
flink_StreamElement_asLatencyMarker_rdh | /**
* Casts this element into a LatencyMarker.
*
* @return This element as a LatencyMarker.
* @throws java.lang.ClassCastException
* Thrown, if this element is actually not a LatencyMarker.
*/
public final LatencyMarker asLatencyMarker() {
return
((LatencyMarker) (this));
} | 3.26 |
flink_FileChannelManagerImpl_close_rdh | /**
* Remove all the temp directories.
*/
@Override
public void close() throws Exception {
// Marks shutdown and exits if it has already shutdown.
if (!isShutdown.compareAndSet(false, true)) {
return;
}
IOUtils.closeAll(Arrays.stream(paths).filter(File::exists).map(FileChannelManagerImpl::getF... | 3.26 |
flink_ClusterClient_reportHeartbeat_rdh | /**
* The client reports the heartbeat to the dispatcher for aliveness.
*
* @param jobId
* The jobId for the client and the job.
* @return */
default CompletableFuture<Void> reportHeartbeat(JobID jobId, long expiredTimestamp) {
return FutureUtils.completedVoidFuture();
} | 3.26 |
flink_ClusterClient_getAccumulators_rdh | /**
* Requests and returns the accumulators for the given job identifier. Accumulators can be
* requested while a is running or after it has finished. The default class loader is used to
* deserialize the incoming accumulator results.
*
* @param jobID
* The job identifier of a job.
* @return A Map containing t... | 3.26 |
flink_ClusterClient_listCompletedClusterDatasetIds_rdh | /**
* Return a set of ids of the completed cluster datasets.
*
* @return A set of ids of the completely cached intermediate dataset.
*/default CompletableFuture<Set<AbstractID>> listCompletedClusterDatasetIds() {
return CompletableFuture.completedFuture(Collections.emptySet());
} | 3.26 |
flink_NFA_extractCurrentMatches_rdh | /**
* Extracts all the sequences of events from the start to the given computation state. An event
* sequence is returned as a map which contains the events and the names of the states to which
* the event... | 3.26 |
flink_NFA_open_rdh | /**
* Initialization method for the NFA. It is called before any element is passed and thus
* suitable for one time setup work.
*
* @param cepRuntimeContext
* runtime context of the enclosing operator
* @param conf
* The configuration containing the parameters attached to the contract.
*/
public void open(R... | 3.26 |
flink_NFA_computeNextStates_rdh | /**
* Computes the next computation states based on the given computation state, the current event,
* its timestamp and the internal state machine. The algorithm is:
*
* <ol>
* <li>Decide on valid transitions and number of branching paths. See {@link OutgoingEdges}
* <li>Perform transitions:
* <ol>
* ... | 3.26 |
flink_NFA_close_rdh | /**
* Tear-down method for the NFA.
*/
public void close() throws Exception {
for (State<T> v10 : getStates()) {
for (StateTransition<T> transition : v10.getStateTransitions()) {
IterativeCondition condition = transition.getCondition();
FunctionUtils.closeFunction(... | 3.26 |
flink_NFA_advanceTime_rdh | /**
* Prunes states assuming there will be no events with timestamp <b>lower</b> than the given
* one. It clears the sharedBuffer and also emits all timed out partial matches.
*
* @param sharedBufferAccessor
* the accessor to SharedBuffer object that we need to work upon
* while processing
* @param nfaState
... | 3.26 |
flink_JoinOperationFactory_create_rdh | /**
* Creates a valid {@link JoinQueryOperation} operation.
*
* <p>It performs validations such as:
*
* <ul>
* <li>condition returns boolean
* <li>the condition is either always true or contains equi join
* <li>left and right side of the join do not contain ambiguous column names
* <li>that correlated ... | 3.26 |
flink_AnotherDummyFSFileSystem_getUri_rdh | // ------------------------------------------------------------------------
@Override
public URI getUri() {
return FS_URI;
} | 3.26 |
flink_RocksDBHandle_restoreInstanceDirectoryFromPath_rdh | /**
* This recreates the new working directory of the recovered RocksDB instance and links/copies
* the contents from a local state.
*/
private void restoreInstanceDirectoryFromPath(Path source) throws IOException {
final Path instanceRocksDBDirectory = Paths.get(f0);
final Path[] files = FileUtils.listDir... | 3.26 |
flink_BridgingSqlFunction_of_rdh | /**
* Creates an instance of a scalar or table built-in function during translation.
*/
public static BridgingSqlFunction of(RelOptCluster cluster, BuiltInFunctionDefinition functionDefinition) {
return m0(cluster, ContextResolvedFunction.permanent(FunctionIdentifier.of(functionDefinition.getName()), functionDef... | 3.26 |
flink_BridgingSqlFunction_getRowTypeInference_rdh | /**
* The conversion to a row type is handled on the caller side. This allows us to perform it
* SQL/Table API-specific. This is in particular important to set the aliases of fields
* correctly (see {@link FlinkRelBuilder#pushFunctionScan(RelBuilder, SqlOperator, int,
* Iterable, List)}).
*/@Override
public SqlRet... | 3.26 |
flink_FileStateHandle_discardState_rdh | /**
* Discard the state by deleting the file that stores the state. If the parent directory of the
* state is empty after deleting the state file, it is also deleted.
*
* @throws Exception
* Thrown, if the file deletion (not the directory deletion) fails.
*/
@Override
public void discardState() throws Exception... | 3.26 |
flink_FileStateHandle_equals_rdh | // ------------------------------------------------------------------------
@Override
public boolean equals(Object o) {
if (this == o) {
return true;
}
if (!(o instanceof FileStateHandle)) {
return false;
}FileStateHandle that = ((FileStateHandle) (o));
return filePath.equals(that.filePath);
} | 3.26 |
flink_FileStateHandle_getFilePath_rdh | /**
* Gets the path where this handle's state is stored.
*
* @return The path where this handle's state is stored.
*/
public Path getFilePath() {return filePath;
} | 3.26 |
flink_FileStateHandle_getStateSize_rdh | /**
* Returns the file size in bytes.
*
* @return The file size in bytes.
*/
@Override
public long getStateSize() {
return stateSize;
} | 3.26 |
flink_FileStateHandle_getFileSystem_rdh | /**
* Gets the file system that stores the file state.
*
* @return The file system that stores the file state.
* @throws IOException
* Thrown if the file system cannot be accessed.
*/
private FileSystem getFileSystem() throws IOException {
return FileSystem.get(filePath.toUri());
} | 3.26 |
flink_ThreadInfoRequestCoordinator_triggerThreadInfoRequest_rdh | /**
* Triggers collection of thread info stats of a job vertex by combining thread info responses
* from given subtasks. A thread info response of a subtask in turn consists of {@code numSamples}, collected with {@code delayBetweenSamples} milliseconds delay between them.
*
* @param executionsWithGateways
* Exec... | 3.26 |
flink_ThreadInfoRequestCoordinator_requestThreadInfo_rdh | /**
* Requests thread infos from given subtasks. The response would be ignored if it does not
* return within timeout.
*/
private void requestThreadInfo(Map<ImmutableSet<ExecutionAttemptID>, CompletableFuture<TaskExecutorThreadInfoGateway>>
executionWithGateways, ThreadInfoSamplesRequest requestParams, Time timeou... | 3.26 |
flink_SortPartitionOperatorBase_executeOnCollections_rdh | // --------------------------------------------------------------------------------------------
@Override
protected List<IN> executeOnCollections(List<IN> inputData, RuntimeContext runtimeContext, ExecutionConfig executionConfig) {
TypeInformation<IN> inputType
... | 3.26 |
flink_BroadcastPartitioner_selectChannel_rdh | /**
* Note: Broadcast mode could be handled directly for all the output channels in record writer,
* so it is no need to select channels via this method.
*/
@Override
public int selectChannel(SerializationDelegate<StreamRecord<T>> record) {
throw new UnsupportedOperationException("Broadcast partitioner does not ... | 3.26 |
flink_TaskExecutorMemoryConfiguration_getTaskHeap_rdh | /**
* Returns the configured heap size used by the tasks.
*/
public Long getTaskHeap() {
return taskHeap;
} | 3.26 |
flink_TaskExecutorMemoryConfiguration_getFrameworkHeap_rdh | /**
* Returns the configured heap size used by the framework.
*/
public Long getFrameworkHeap() {
return frameworkHeap;
} | 3.26 |
flink_TaskExecutorMemoryConfiguration_getFrameworkOffHeap_rdh | /**
* Returns the configured off-heap size used by the framework.
*/
public Long getFrameworkOffHeap() {
return frameworkOffHeap;
} | 3.26 |
flink_TaskExecutorMemoryConfiguration_getNetworkMemory_rdh | /**
* Returns the configured maximum network memory.
*/
public Long getNetworkMemory() {
return networkMemory;
} | 3.26 |
flink_TaskExecutorMemoryConfiguration_m0_rdh | /**
* Factory method for initializing a TaskExecutorMemoryConfiguration based on the passed
* Configuration.
*
* @param config
* The Configuration used for initializing the TaskExecutorMemoryConfiguration.
* @return The newly instantiated TaskExecutorMemoryConfiguration.
*/
public static TaskExecutorMemoryConf... | 3.26 |
flink_TaskExecutorMemoryConfiguration_getJvmMetaspace_rdh | /**
* Returns the maximum Metaspace size allowed for the task manager.
*/
public Long getJvmMetaspace() {
return jvmMetaspace;
} | 3.26 |
flink_TaskExecutorMemoryConfiguration_getTaskOffHeap_rdh | /**
* Returns the configured off-heap size used by the tasks.
*/
public Long getTaskOffHeap() {
return taskOffHeap;
} | 3.26 |
flink_TaskExecutorMemoryConfiguration_getManagedMemoryTotal_rdh | /**
* Returns the total amount of memory reserved for by the MemoryManager.
*/
public Long getManagedMemoryTotal() {
return managedMemoryTotal;
} | 3.26 |
flink_SubQueryInputTypeStrategy_getArgumentCount_rdh | /**
* {@link InputTypeStrategy} for {@link BuiltInFunctionDefinitions#IN}.
*/
@Internalpublic class SubQueryInputTypeStrategy implements InputTypeStrategy {
@Override
public ArgumentCount getArgumentCount() {return ConstantArgumentCount.from(2);
} | 3.26 |
flink_SourceCoordinatorSerdeUtils_readAndVerifyCoordinatorSerdeVersion_rdh | /**
* Read and verify the serde version.
*/
static int readAndVerifyCoordinatorSerdeVersion(DataInputStream in) throws IOException {
int version = in.readInt();
if (version > CURRENT_VERSION) {
throw new IOException("Unsupported source coordinator serde version " + version);
}
return version;
... | 3.26 |
flink_SourceCoordinatorSerdeUtils_writeCoordinatorSerdeVersion_rdh | /**
* Write the current serde version.
*/
static void writeCoordinatorSerdeVersion(DataOutputStream out) throws IOException {out.writeInt(CURRENT_VERSION);
} | 3.26 |
flink_DataSinkTask_getLogString_rdh | // ------------------------------------------------------------------------
// Utilities
// ------------------------------------------------------------------------
/**
* Utility function that composes a string for logging purposes. The string includes the given
* message and the index of the task in its task group t... | 3.26 |
flink_DataSinkTask_initOutputFormat_rdh | /**
* Initializes the OutputFormat implementation and configuration.
*
* @throws RuntimeException
* Throws if instance of OutputFormat implementation can not be
* obtained.
*/
private void initOutputFormat() {
ClassLoader userCodeClassLoader = getUserCodeClassLoader();
// obtain task configuration (in... | 3.26 |
flink_DataSinkTask_initInputReaders_rdh | /**
* Initializes the input readers of the DataSinkTask.
*
* @throws RuntimeException
* Thrown in case of invalid task input configuration.
*/
@SuppressWarnings("unchecked")
private void initInputReaders() throws Exception {
int numGates =
0;
// ---------------- create the input readers -----------... | 3.26 |
flink_TransientBlobCache_m0_rdh | /**
* Returns the blob expiry times - for testing purposes only!
*
* @return blob expiry times (internal state!)
*/
@VisibleForTesting
ConcurrentMap<Tuple2<JobID, TransientBlobKey>, Long> m0() {
return blobExpiryTimes;
} | 3.26 |
flink_TransientBlobCache_getStorageLocation_rdh | /**
* Returns a file handle to the file associated with the given blob key on the blob server.
*
* @param jobId
* ID of the job this blob belongs to (or <tt>null</tt> if job-unrelated)
* @param key
* identifying the file
* @return file handle to the file
* @throws IOException
* if creating the directory ... | 3.26 |
flink_DataOutputEncoder_writeFixed_rdh | // --------------------------------------------------------------------------------------------
// bytes
// --------------------------------------------------------------------------------------------
@Override
public void writeFixed(byte[] bytes, int start, int len) throws IOException {
out.write(bytes, start, len... | 3.26 |
flink_DataOutputEncoder_writeString_rdh | // --------------------------------------------------------------------------------------------
// strings
// --------------------------------------------------------------------------------------------
@Override
public void writeString(String str) throws IOException {
byte[] bytes = Utf8.getBytesFor(str);
writ... | 3.26 |
flink_DataOutputEncoder_writeNull_rdh | // --------------------------------------------------------------------------------------------
// primitives
// --------------------------------------------------------------------------------------------
@Override
public void writeNull() {
} | 3.26 |
flink_DataOutputEncoder_writeVarLongCount_rdh | // --------------------------------------------------------------------------------------------
// utils
// --------------------------------------------------------------------------------------------
public static void writeVarLongCount(DataOutput out, long val) throws IOException {
if (val < 0) {
throw ne... | 3.26 |
flink_DataOutputEncoder_writeIndex_rdh | // --------------------------------------------------------------------------------------------
// union
// --------------------------------------------------------------------------------------------
@Override
public void writeIndex(int unionIndex) throws IOException {
out.writeInt(unionIndex);
} | 3.26 |
flink_DataOutputEncoder_writeArrayStart_rdh | // --------------------------------------------------------------------------------------------
// collection types
// --------------------------------------------------------------------------------------------
@Override
public void writeArrayStart() {
} | 3.26 |
flink_AvroRowSerializationSchema_convertRowToAvroRecord_rdh | // --------------------------------------------------------------------------------------------
private GenericRecord
convertRowToAvroRecord(Schema schema, Row row) {
final List<Schema.Field> fields = schema.getFields();
final int length = fields.size();
final GenericRecord record = new GenericData.Record(s... | 3.26 |
flink_KeyedStateTransformation_window_rdh | /**
* Windows this transformation into a {@code WindowedOperatorTransformation}, which bootstraps
* state that can be restored by a {@code WindowOperator}. Elements are put into windows by a
* {@link WindowAssigner}. The grouping of elements is done both by key and by window.
*
* <p>A {@link org.apache.flink.strea... | 3.26 |
flink_KeyedStateTransformation_transform_rdh | /**
* Method for passing user defined operators along with the type information that will transform
* the OperatorTransformation.
*
* <p><b>IMPORTANT:</b> Any output from this operator will be discarded.
*
* @param factory
* A factory returning transformation logic type of the return stream
* @return An {@lin... | 3.26 |
flink_StreamProjection_projectTuple17_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16> SingleOutputStreamOperator<Tuple17<T0, T1, T2, T3, T4, T5, T6, T7, T8, T... | 3.26 |
flink_StreamProjection_projectTuple4_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/public
<T0, T1, T2, T3> SingleOutputStreamOperator<Tuple4<T0, T1, T2, T3>> projectTuple4() {
TypeInformation<?>[] fTypes = m1(fieldIndexes, dataStream.get... | 3.26 |
flink_StreamProjection_projectTuple13_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12> SingleOutputStreamOperator<Tuple13<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12>> p... | 3.26 |
flink_StreamProjection_projectTuple22_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4, T5, T6,
T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21> SingleOutputStreamOperator<Tuple22<T0, T1, T2, ... | 3.26 |
flink_StreamProjection_projectTuple8_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4, T5, T6, T7> SingleOutputStreamOperator<Tuple8<T0, T1, T2, T3, T4, T5, T6, T7>> projectTuple8() {
TypeInformation<?>[] fTypes... | 3.26 |
flink_StreamProjection_projectTupleX_rdh | /**
* Chooses a projectTupleX according to the length of {@link org.apache.flink.streaming.api.datastream.StreamProjection#fieldIndexes}.
*
* @return The projected DataStream.
* @see org.apache.flink.api.java.operators.ProjectOperator.Projection
*/
@SuppressWarnings("unchecked")
public <OUT extends Tuple> SingleOu... | 3.26 |
flink_StreamProjection_projectTuple9_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8> SingleOutputStreamOperator<Tuple9<T0, T1, T2, T3, T4, T5,
T6, T7, T8>> projectTuple9() {
TypeInformation<?>[]... | 3.26 |
flink_StreamProjection_projectTuple5_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4> SingleOutputStreamOperator<Tuple5<T0, T1, T2, T3, T4>> projectTuple5() {
TypeInformation<?>[] fTypes = m1(fieldIndexes, data... | 3.26 |
flink_StreamProjection_projectTuple16_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/public <T0, T1,
T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15> SingleOutputStreamOperator<Tuple16<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T... | 3.26 |
flink_StreamProjection_projectTuple12_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11> SingleOutputStreamOperator<Tuple12<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11>> projectT... | 3.26 |
flink_StreamProjection_projectTuple23_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21, T22> SingleOutputStreamOperator<Tuple23<T0, T1, T... | 3.26 |
flink_StreamProjection_projectTuple18_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17> SingleOutputStreamOperator<Tuple18<T0, T1,
T2, T3, T4, T5, T6, T7,
T8... | 3.26 |
flink_StreamProjection_projectTuple25_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21, T22, T23, T24> SingleOutputStreamOperator<Tuple2... | 3.26 |
flink_StreamProjection_projectTuple14_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13> SingleOutputStreamOperator<Tuple14<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12,... | 3.26 |
flink_StreamProjection_projectTuple7_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4, T5, T6> SingleOutputStreamOperator<Tuple7<T0, T1, T2, T3, T4, T5, T6>> projectTuple7() {
TypeInformation<?>[] fTypes = m1(fie... | 3.26 |
flink_StreamProjection_projectTuple10_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9> SingleOutputStreamOperator<Tuple10<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9>> projectTuple10() {
TypeInform... | 3.26 |
flink_StreamProjection_projectTuple3_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2> SingleOutputStreamOperator<Tuple3<T0, T1, T2>> projectTuple3() {
TypeInformation<?>[]
fTypes = m1(fieldIndexes, dataStream.getTyp... | 3.26 |
flink_StreamProjection_projectTuple21_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18,
T19, T20> SingleOutputStreamOperator<Tuple21<T0, T1, T2, T3... | 3.26 |
flink_StreamProjection_projectTuple24_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21, T22, T23> SingleOutputStreamOperator<Tuple24<... | 3.26 |
flink_StreamProjection_projectTuple6_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4, T5> SingleOutputStreamOperator<Tuple6<T0, T1, T2, T3, T4, T5>> projectTuple6() {
TypeInformation<?>[] fTypes = m1(fieldIndexe... | 3.26 |
flink_StreamProjection_projectTuple15_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14> SingleOutputStreamOperator<Tuple15<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11... | 3.26 |
flink_StreamProjection_projectTuple2_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1> SingleOutputStreamOperator<Tuple2<T0, T1>> projectTuple2() {
TypeInformation<?>[] fTypes = m1(fieldIndexes, dataStream.getType());
Tu... | 3.26 |
flink_StreamProjection_projectTuple20_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8,
T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19> SingleOutputStreamOperator<Tuple20<T0, T1, T2, T3, T4, T5,... | 3.26 |
flink_StreamProjection_projectTuple19_rdh | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18> SingleOutputStreamOperator<Tuple19<T0, T1, T2, T3, T4, T5, T6, T... | 3.26 |
flink_OptimizedPlan_getJobName_rdh | /**
* Returns the name of the program.
*
* @return The name of the program.
*/
public String getJobName() {
return this.jobName;
} | 3.26 |
flink_OptimizedPlan_accept_rdh | // ------------------------------------------------------------------------
/**
* Applies the given visitor top down to all nodes, starting at the sinks.
*
* @param visitor
* The visitor to apply to the nodes in this plan.
* @see org.apache.flink.util.Visitable#accept(org.apache.flink.util.Visitor)
*/
@Override... | 3.26 |
flink_OptimizedPlan_getAllNodes_rdh | /**
* Gets all the nodes from this OptimizedPlan.
*
* @return All nodes.
*/
public Collection<PlanNode> getAllNodes() {
return
allNodes;
} | 3.26 |
flink_CoGroupedStreams_equalTo_rdh | /**
* Specifies a {@link KeySelector} for elements from the second input.
*
* @param keySelector
* The KeySelector to be used for extracting the second input's key for
* partitioning.
*/
public EqualTo equalTo(KeySelector<T2, KEY> keySelector) {
Preconditions.checkNotNull(keySelector);
final TypeInfor... | 3.26 |
flink_CoGroupedStreams_where_rdh | /**
* Specifies a {@link KeySelector} for elements from the first input with explicit type
* information.
*
* @param keySelector
* The KeySelector to be used for extracting the first input's key for
* partitioning.
* @param keyType
* The type information describing the key type.
*/
public <KEY> Where<KEY... | 3.26 |
flink_CoGroupedStreams_m0_rdh | /**
* Sets the {@code Trigger} that should be used to trigger window emission.
*/
@PublicEvolving
public WithWindow<T1, T2, KEY, W> m0(Trigger<? super TaggedUnion<T1, T2>, ? super W> newTrigger) {
return new WithWindow<>(input1, input2, keySelector1, keySelector2, keyType, windowAssigner, newTrigger, evictor, a... | 3.26 |
flink_CoGroupedStreams_apply_rdh | /**
* Completes the co-group operation with the user function that is executed for windowed
* groups.
*
* <p>Note: This method's return type does not support setting an operator-specific
* parallelism. Due to binary backwards compatibility, this cannot be altered. Use the
* {@link #with(CoGroupFunction, TypeInfor... | 3.26 |
flink_CoGroupedStreams_allowedLateness_rdh | /**
* Sets the time by which elements are allowed to be late.
*
* @see WindowedStream#allowedLateness(Time)
*/
@PublicEvolving
public WithWindow<T1,
T2, KEY, W> allowedLateness(Time newLateness) {
return new WithWindow<>(input1, input2, keySelector1, keySelector2, keyType, windowAssigner, trigger, evictor, ne... | 3.26 |
flink_MessageParameter_getKey_rdh | /**
* Returns the key of this parameter, e.g. "jobid".
*
* @return key of this parameter
*/
public final String getKey() {
return f0;
} | 3.26 |
flink_MessageParameter_resolve_rdh | /**
* Resolves this parameter for the given value.
*
* @param value
* value to resolve this parameter with
*/
public final void resolve(X value) {
Preconditions.checkState(!resolved, "This parameter was already resolved.");
this.value = Preconditions.checkNotNull(value);this.resolved = true;
} | 3.26 |
flink_MessageParameter_m0_rdh | /**
* Returns the resolved value of this parameter as a string, or {@code null} if it isn't
* resolved yet.
*
* @return resolved value, or null if it wasn't resolved yet
*/
final String m0() {
return value == null ? null : convertToString(value);
} | 3.26 |
flink_MessageParameter_getValue_rdh | /**
* Returns the resolved value of this parameter, or {@code null} if it isn't resolved yet.
*
* @return resolved value, or null if it wasn't resolved yet
*/
public final X getValue() {
return value;
} | 3.26 |
flink_NotDuplicatingCheckpointStateToolset_canFastDuplicate_rdh | /**
* An empty implementation of {@link CheckpointStateToolset}.
*/public final class NotDuplicatingCheckpointStateToolset implements CheckpointStateToolset {
@Override
public boolean canFastDuplicate(StreamStateHandle stateHandle) throws IOException {
return false;
} | 3.26 |
flink_LegacySinkTransformation_setStateKeySelector_rdh | /**
* Sets the {@link KeySelector} that must be used for partitioning keyed state of this Sink.
*
* @param stateKeySelector
* The {@code KeySelector} to set
*/
public void setStateKeySelector(KeySelector<T, ?> stateKeySelector) {
this.stateKeySelector = stateKeySelector;
updateManagedMemoryStateBackendUs... | 3.26 |
flink_LegacySinkTransformation_getOperatorFactory_rdh | /**
* Returns the {@link StreamOperatorFactory} of this {@code LegacySinkTransformation}.
*/
public StreamOperatorFactory<Object> getOperatorFactory() {
return operatorFactory;
} | 3.26 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.