name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_Transformation_getBufferTimeout_rdh | /**
* Returns the buffer timeout of this {@code Transformation}.
*
* @see #setBufferTimeout(long)
*/
public long getBufferTimeout() {
return bufferTimeout;
} | 3.26 |
flink_Transformation_getOutputType_rdh | /**
* Returns the output type of this {@code Transformation} as a {@link TypeInformation}. Once
* this is used once the output type cannot be changed anymore using {@link #setOutputType}.
*
* @return The output type of this {@code Transformation}
*/
public TypeInformation<T> getOutputType() {
if (outputTyp... | 3.26 |
flink_Transformation_getManagedMemoryOperatorScopeUseCaseWeights_rdh | /**
* Get operator scope use cases that this transformation needs managed memory for, and the
* use-case-specific weights for this transformation. The weights are used for sharing managed
* memory across transformations for the use cases. Check the individual {@link ManagedMemoryUseCase} for the specific weight defi... | 3.26 |
flink_Transformation_getUserProvidedNodeHash_rdh | /**
* Gets the user provided hash.
*
* @return The user provided hash.
*/
public String getUserProvidedNodeHash() {
return userProvidedNodeHash;
} | 3.26 |
flink_Transformation_setDescription_rdh | /**
* Changes the description of this {@code Transformation}.
*/
public void setDescription(String description) {
this.description = Preconditions.checkNotNull(description);
} | 3.26 |
flink_Transformation_setUid_rdh | /**
* Sets an ID for this {@link Transformation}. This is will later be hashed to a uidHash which
* is then used to create the JobVertexID (that is shown in logs and the web ui).
*
* <p>The specified ID is used to assign the same operator ID across job submissions (for
* example when starting a job from a savepoin... | 3.26 |
flink_Transformation_setParallelism_rdh | /**
* Sets the parallelism of this {@code Transformation}.
*
* @param parallelism
* The new parallelism to set on this {@code Transformation}.
*/
public void setParallelism(int parallelism) {
setParallelism(parallelism, true);
} | 3.26 |
flink_Transformation_setMaxParallelism_rdh | /**
* Sets the maximum parallelism for this stream transformation.
*
* @param maxParallelism
* Maximum parallelism for this stream transformation.
*/
public void setMaxParallelism(int maxParallelism) {
OperatorValidationUtils.validateMaxParallelism(maxParallelism, UPPER_BOUND_MAX_PARALLELISM);
this.maxPa... | 3.26 |
flink_Transformation_declareManagedMemoryUseCaseAtOperatorScope_rdh | /**
* Declares that this transformation contains certain operator scope managed memory use case.
*
* @param managedMemoryUseCase
* The use case that this transformation declares needing managed
* memory for.
* @param weight
* Use-case-specific weights for this transformation. Used for sharing managed
* ... | 3.26 |
flink_Transformation_setSlotSharingGroup_rdh | /**
* Sets the slot sharing group of this transformation. Parallel instances of operations that are
* in the same slot sharing group will be co-located in the same TaskManager slot, if possible.
*
* <p>Initially, an operation is in the default slot sharing group. This can be explicitly set
* with constructing a {@... | 3.26 |
flink_Transformation_getSlotSharingGroup_rdh | /**
* Returns the slot sharing group of this transformation if present.
*
* @see #setSlotSharingGroup(SlotSharingGroup)
*/public Optional<SlotSharingGroup> getSlotSharingGroup() {
return f1;
} | 3.26 |
flink_Transformation_setBufferTimeout_rdh | /**
* Set the buffer timeout of this {@code Transformation}. The timeout defines how long data may
* linger in a partially full buffer before being sent over the network.
*
* <p>Lower timeouts lead to lower tail latencies, but may affect throughput. For Flink 1.5+,
* timeouts of 1ms are feasible for jobs with high... | 3.26 |
flink_Transformation_getParallelism_rdh | /**
* Returns the parallelism of this {@code Transformation}.
*/
public int getParallelism() {
return parallelism;
} | 3.26 |
flink_Transformation_getDescription_rdh | /**
* Returns the description of this {@code Transformation}.
*/
public String getDescription() { return description;
} | 3.26 |
flink_Transformation_setOutputType_rdh | /**
* Tries to fill in the type information. Type information can be filled in later when the
* program uses a type hint. This method checks whether the type information has ever been
* accessed before and does not allow modifications if the type was accessed already. This
* ensures consistency by making sure diffe... | 3.26 |
flink_Transformation_setResources_rdh | /**
* Sets the minimum and preferred resources for this stream transformation.
*
* @param minResources
* The minimum resource of this transformation.
* @param preferredResources
* The preferred resource of this transformation.
*/
public void setResources(ResourceSpec minResources, ResourceSpec preferredReso... | 3.26 |
flink_MultiShotLatch_await_rdh | /**
* Waits until {@link #trigger()} is called.
*/
public void await() throws InterruptedException {
synchronized(lock) {
while (!triggered) {
lock.wait();
}
triggered = false;
}
} | 3.26 |
flink_MultiShotLatch_trigger_rdh | /**
* Fires the latch. Code that is blocked on {@link #await()} will now return.
*/
public void trigger() {
synchronized(lock) {
triggered = true;
lock.notifyAll();
}
} | 3.26 |
flink_ColumnSummary_getTotalCount_rdh | /**
* The number of all rows in this column including both nulls and non-nulls.
*/
public long getTotalCount() {
return getNullCount() + getNonNullCount();
} | 3.26 |
flink_ColumnSummary_containsNonNull_rdh | /**
* True if this column contains any non-null values.
*/
public boolean containsNonNull() {
return getNonNullCount() > 0L;
} | 3.26 |
flink_FileSourceSplit_getReaderPosition_rdh | /**
* Gets the (checkpointed) position of the reader, if set. This value is typically absent for
* splits when assigned from the enumerator to the readers, and present when the splits are
* recovered from a checkpoint.
*/
public Optional<CheckpointedPosition> getReaderPosition() {
return Optional.ofNullable(f1);
} | 3.26 |
flink_FileSourceSplit_length_rdh | /**
* Returns the number of bytes in the file region described by this source split.
*/
public long length() {
return length;
} | 3.26 |
flink_FileSourceSplit_fileSize_rdh | /**
* Returns the full file size in bytes, from {@link FileStatus#getLen()}.
*/
public long fileSize() {
return fileSize;
} | 3.26 |
flink_FileSourceSplit_fileModificationTime_rdh | /**
* Returns the modification time of the file, from {@link FileStatus#getModificationTime()}.
*/
public long fileModificationTime() {
return fileModificationTime;
} | 3.26 |
flink_FileSourceSplit_updateWithCheckpointedPosition_rdh | /**
* Creates a copy of this split where the checkpointed position is replaced by the given new
* position.
*
* <p><b>IMPORTANT:</b> Subclasses that add additional information to the split must override
* this method to return that subclass type. This contract is enforced by checks in the file
* source implementa... | 3.26 |
flink_FileSourceSplit_splitId_rdh | // ------------------------------------------------------------------------
// split properties
// ------------------------------------------------------------------------
@Override
public String splitId() {
return id;
} | 3.26 |
flink_FileSourceSplit_m0_rdh | // ------------------------------------------------------------------------
// utils
// ------------------------------------------------------------------------
@Override
public String m0() { final String hosts = (hostnames.length == 0) ? "(no host info)" : " hosts=" + Arrays.toString(hostnames);
return String.format... | 3.26 |
flink_FileSourceSplit_offset_rdh | /**
* Returns the start of the file region referenced by this source split. The position is
* inclusive, the value indicates the first byte that is part of the split.
*/
public long offset() {
return f0;
} | 3.26 |
flink_FlinkContainersSettings_zookeeperHostname_rdh | /**
* Sets the {@code zookeeperHostname} and returns a reference to this Builder enabling
* method chaining.
*
* @param zookeeperHostname
* The Zookeeper hostname.
* @return A reference to this Builder.
*/
public Builder zookeeperHostname(String zookeeperHostname) {
this.zookeeperHostname = zookeeperHostna... | 3.26 |
flink_FlinkContainersSettings_defaultConfig_rdh | /**
* {@code FlinkContainersConfig} based on defaults.
*
* @return The Flink containers config.
*/
public static FlinkContainersSettings defaultConfig() {return builder().build();} | 3.26 |
flink_FlinkContainersSettings_jobManagerHostname_rdh | /**
* Sets the job manager hostname and returns a reference to this Builder enabling method
* chaining.
*
* @param jobManagerHostname
* The job manager hostname to set.
* @return A reference to this Builder.
*/
public Builder jobManagerHostname(String jobManagerHostname) {
return setConfigOption(JobManage... | 3.26 |
flink_FlinkContainersSettings_checkpointPath_rdh | /**
* Sets the {@code checkpointPath} and returns a reference to this Builder enabling method
* chaining.
*
* @param checkpointPath
* The checkpoint path to set.
* @return A reference to this Builder.
*/
public Builder
checkpointPath(String checkpointPath) {
this.checkpointPath = checkpointPath;
return... | 3.26 |
flink_FlinkContainersSettings_enableZookeeperHA_rdh | /**
* Enables Zookeeper HA. NOTE: this option uses default HA configuration. If you want to use
* non-default configuration, you should provide all settings, including the HA_MODE
* directly via the {@code basedOn()} method instead.
*
* @return A reference to this Builder.
*/
public Builder enableZookeeperHA() {
... | 3.26 |
flink_FlinkContainersSettings_getDefaultCheckpointPath_rdh | /**
* Gets default checkpoint path.
*
* @return The default checkpoint path.
*/
public static String getDefaultCheckpointPath() {
return DEFAULT_CHECKPOINT_PATH;
} | 3.26 |
flink_FlinkContainersSettings_getDefaultFlinkHome_rdh | /**
* Gets default flink home.
*
* @return The default flink home path.
*/
public static String getDefaultFlinkHome() {
return DEFAULT_FLINK_HOME;
} | 3.26 |
flink_FlinkContainersSettings_jarPaths_rdh | /**
* Sets the {@code jarPaths} and returns a reference to this Builder enabling method
* chaining.
*
* @param jarPaths
* The {@code jarPaths} to set.
* @return A reference to this Builder.
*/
public Builder jarPaths(Collection<String> jarPaths) {
this.jarPaths = jarPaths;
return this;
} | 3.26 |
flink_FlinkContainersSettings_getNumSlotsPerTaskManager_rdh | /**
* Gets number slots per task manager.
*
* @return The number slots per task manager.
*/
public int getNumSlotsPerTaskManager() {
return numSlotsPerTaskManager;
} | 3.26 |
flink_FlinkContainersSettings_numTaskManagers_rdh | /**
* Sets the {@code numTaskManagers} and returns a reference to this Builder enabling method
* chaining.
*
* @param numTaskManagers
* The {@code numTaskManagers} to set.
* @return A reference to this Builder.
*/
public Builder numTaskManagers(int numTaskManagers) {this.numTaskManagers = numTaskManagers;
... | 3.26 |
flink_FlinkContainersSettings_build_rdh | /**
* Returns a {@code FlinkContainersConfig} built from the parameters previously set.
*
* @return A {@code FlinkContainersConfig} built with parameters of this {@code FlinkContainersConfig.Builder}.
*/
public FlinkContainersSettings build() {
return new FlinkContainersSettings(this);
} | 3.26 |
flink_FlinkContainersSettings_isBuildFromFlinkDist_rdh | /**
* Returns whether to build from flink-dist or from an existing base container. Also see the
* {@code baseImage} property.
*/
public Boolean isBuildFromFlinkDist() {
return buildFromFlinkDist;} | 3.26 |
flink_FlinkContainersSettings_getHaStoragePath_rdh | /**
* Gets HA storage path.
*
* @return The ha storage path.
*/
public String getHaStoragePath() {
return
haStoragePath;
} | 3.26 |
flink_FlinkContainersSettings_setConfigOption_rdh | /**
* Sets a single Flink configuration parameter (the options for flink-conf.yaml) and returns
* a reference to this Builder enabling method chaining.
*
* @param <T>
* The type parameter.
* @param option
* The option.
* @param value
* The value.
* @return A reference to this Builder.
*/
public <T> Bui... | 3.26 |
flink_FlinkContainersSettings_m0_rdh | /**
* Gets jar paths.
*
* @return The jar paths.
*/public Collection<String> m0() {
return jarPaths;
} | 3.26 |
flink_FlinkContainersSettings_fullConfiguration_rdh | /**
* Sets the {@code flinkConfiguration} value to {@code config} and returns a reference to
* this Builder enabling method chaining.
*
* @param <T>
* the type parameter
* @param config
* The {@code config} to set.
* @return A reference to this Builder.
*/
public <T> Builder fullConfiguration(Configuration... | 3.26 |
flink_FlinkContainersSettings_flinkDistLocation_rdh | /**
* Sets the {@code flinkDistLocation} and returns a reference to this Builder enabling
* method chaining.
*
* @param flinkDistLocation
* The {@code flinkDistLocation} to set.
* @return A reference to this Builder.
*/
public Builder flinkDistLocation(String flinkDistLocation) {
this.flinkDistLocation = ... | 3.26 |
flink_FlinkContainersSettings_flinkHome_rdh | /**
* Sets the path of the Flink distribution inside the container. Returns a reference to this
* Builder enabling method chaining.
*
* @param flinkHome
* The {@code flinkHome} to set.
* @return A reference to this Builder.
*/
public Builder flinkHome(String
flinkHome) {
this.flinkHome = flinkHome;
r... | 3.26 |
flink_FlinkContainersSettings_numSlotsPerTaskManager_rdh | /**
* Sets the {@code numSlotsPerTaskManager} and returns a reference to this Builder enabling
* method chaining. It also adds this property into the {@code flinkConfiguration} field.
*
* @param numSlotsPerTaskManager
* The {@code numSlotsPerTaskManager} to set.
* @return A reference to this Builder.
*/public ... | 3.26 |
flink_FlinkContainersSettings_taskManagerHostnamePrefix_rdh | /**
* Sets the {@code taskManagerHostnamePrefix} and returns a reference to this Builder
* enabling method chaining.
*
* @param taskManagerHostnamePrefix
* The {@code taskManagerHostnamePrefix} to set.
* @return A reference to this Builder.
*/
public Builder taskManagerHostnamePrefix(String taskManagerHostname... | 3.26 |
flink_FlinkContainersSettings_haStoragePath_rdh | /**
* Sets the {@code haStoragePath} and returns a reference to this Builder enabling method
* chaining.
*
* @param haStoragePath
* The path for storing HA data.
* @return A reference to this Builder.
*/
public Builder haStoragePath(String haStoragePath) {
this.haStoragePath = haStoragePath;
return set... | 3.26 |
flink_FlinkContainersSettings_baseImage_rdh | /**
* Sets the {@code baseImage} and returns a reference to this Builder enabling method
* chaining.
*
* @param baseImage
* The {@code baseImage} to set.
* @return A reference to this Builder.
*/
public Builder baseImage(String baseImage) {
this.baseImage = baseImage;this.buildFromFlinkDist = false;
r... | 3.26 |
flink_FlinkContainersSettings_builder_rdh | /**
* A new builder for {@code FlinkContainersConfig}.
*
* @return The builder.
*/
public static Builder builder() { return new Builder();
} | 3.26 |
flink_FlinkContainersSettings_getBaseImage_rdh | /**
* Gets base image.
*
* @return The base image.
*/
public String getBaseImage() {
return baseImage;
} | 3.26 |
flink_FlinkContainersSettings_basedOn_rdh | /**
* Merges the provided {@code config} with the default config, potentially overwriting the
* defaults in case of collisions. Returns a reference to this Builder enabling method
* chaining.
*
* @param <T>
* the type parameter
* @param config
* The {@code config} to add.
* @return A reference to this Buil... | 3.26 |
flink_FlinkContainersSettings_isZookeeperHA_rdh | /**
* Is zookeeper HA boolean.
*/
public Boolean isZookeeperHA() {
return zookeeperHA;
} | 3.26 |
flink_FlinkContainersSettings_setLogProperty_rdh | /**
* Sets a single Flink logging configuration property in the log4j format and returns a
* reference to this Builder enabling method chaining.
*
* @param key
* The property key.
* @param value
* The property value.
* @return A reference to this Builder.
*/
public Builder setLogProperty(String key, String... | 3.26 |
flink_FlinkContainersSettings_getLogProperties_rdh | /**
* Gets logging properties.
*
* @return The logging properties.
*/public Properties getLogProperties() {
return logProperties;
} | 3.26 |
flink_FlinkContainersSettings_getCheckpointPath_rdh | /**
* Gets checkpoint path.
*
* @return The checkpoint path.
*/
public String getCheckpointPath() {
return f0;
} | 3.26 |
flink_PartitionRequestClientFactory_createPartitionRequestClient_rdh | /**
* Atomically establishes a TCP connection to the given remote address and creates a {@link NettyPartitionRequestClient} instance for this connection.
*/
NettyPartitionRequestClient createPartitionRequestClient(ConnectionID connectionId) throws IOException, InterruptedException {
// We map the input Connection... | 3.26 |
flink_PartitionRequestClientFactory_destroyPartitionRequestClient_rdh | /**
* Removes the client for the given {@link ConnectionID}.
*/
void destroyPartitionRequestClient(ConnectionID connectionId, PartitionRequestClient client) {
final CompletableFuture<NettyPartitionRequestClient> future = clients.get(connectionId);
if ((future != null) && future.isDone()) {
future.the... | 3.26 |
flink_AllGroupCombineDriver_setup_rdh | // ------------------------------------------------------------------------
@Override
public void setup(TaskContext<GroupCombineFunction<IN, OUT>, OUT> context) {
this.taskContext = context;
} | 3.26 |
flink_StreamPhysicalPythonCorrelateRule_findTableFunction_rdh | // find only calc and table function
private boolean findTableFunction(FlinkLogicalCalc calc) {
RelNode child = ((RelSubset) (calc.getInput())).getOriginal();
if (child instanceof FlinkLogicalTableFunctionScan) {
FlinkLogicalTableFunctionScan scan = ((FlinkLogicalTableFunctionScan) (child));
ret... | 3.26 |
flink_BridgingSqlAggFunction_of_rdh | /**
* Creates an instance of a aggregate function during translation.
*/
public static BridgingSqlAggFunction of(FlinkContext context, FlinkTypeFactory typeFactory, ContextResolvedFunction resolvedFunction) {
final DataTypeFactory dataTypeFactory = context.getCatalogManager().getDataTypeFactory();
final Type... | 3.26 |
flink_BridgingSqlAggFunction_createOrderRequirement_rdh | // --------------------------------------------------------------------------------------------
private static boolean createOrderRequirement() {
return false;
} | 3.26 |
flink_HadoopInputs_createHadoopInput_rdh | /**
* Creates a Flink {@link InputFormat} that wraps the given Hadoop {@link org.apache.hadoop.mapreduce.InputFormat}.
*
* @return A Flink InputFormat that wraps the Hadoop InputFormat.
*/
public static <K, V> HadoopInputFormat<K, V> createHadoopInput(InputFormat<K,
V> mapreduceInputFormat, Class<K> key, Class<V> v... | 3.26 |
flink_HadoopInputs_readHadoopFile_rdh | /**
* Creates a Flink {@link InputFormat} that wraps the given Hadoop {@link org.apache.hadoop.mapreduce.lib.input.FileInputFormat}.
*
* @return A Flink InputFormat that wraps the Hadoop FileInputFormat.
*/
public static <K, V> HadoopInputFormat<K, V> readHadoopFile(FileInputFormat<K, V> mapreduceInputFormat, Class... | 3.26 |
flink_FlinkResultSetMetaData_getType_rdh | /**
* Get column type name according type in {@link Types}.
*
* @param type
* the type in {@link Types}
* @return type class name
*/
static String getType(int type) throws SQLException {// see javax.sql.rowset.RowSetMetaDataImpl
switch (type) {
case Types.NUMERIC :
case Types.DECIMAL :
return Bi... | 3.26 |
flink_LogicalPipelinedRegionComputeUtil_computePipelinedRegions_rdh | /**
* Utils for computing {@link LogicalPipelinedRegion}s.
*/public final class LogicalPipelinedRegionComputeUtil {
public static Set<Set<LogicalVertex>> computePipelinedRegions(final Iterable<? extends LogicalVertex> topologicallySortedVertices) {
final Map<LogicalVertex, Set<LogicalVertex>> vertexToRegion... | 3.26 |
flink_GroupReduceOperator_translateToDataFlow_rdh | // --------------------------------------------------------------------------------------------
@Override
@SuppressWarnings("unchecked")
protected GroupReduceOperatorBase<?, OUT, ?> translateToDataFlow(Operator<IN> input) {
String name = (getName() != null) ? getName() : "GroupReduce at " + defaultName;
// wrap... | 3.26 |
flink_GroupReduceOperator_translateSelectorFunctionReducer_rdh | // --------------------------------------------------------------------------------------------
@SuppressWarnings("unchecked")
private static <IN, OUT, K> PlanUnwrappingReduceGroupOperator<IN, OUT, K> translateSelectorFunctionReducer(SelectorFunctionKeys<IN, ?> rawKeys, GroupReduceFunction<IN, OUT> function, TypeInform... | 3.26 |
flink_GroupReduceOperator_isCombinable_rdh | // --------------------------------------------------------------------------------------------
// Properties
// --------------------------------------------------------------------------------------------
@Internal
public boolean isCombinable() {
return combinable;
} | 3.26 |
flink_SlotSharingSlotAllocator_determineSlotsPerSharingGroup_rdh | /**
* Distributes free slots across the slot-sharing groups of the job. Slots are distributed as
* evenly as possible. If a group requires less than an even share of slots the remainder is
* distributed over the remaining groups.
*/
private static Map<SlotSharingGroupId, Integer> determineSlotsPerSharingGroup(JobIn... | 3.26 |
flink_JoinNode_computeOperatorSpecificDefaultEstimates_rdh | /**
* The default estimates build on the principle of inclusion: The smaller input key domain is
* included in the larger input key domain. We also assume that every key from the larger input
* has one join partner in the smaller input. The result cardinality is hence the larger one.
*/
@Override
protected void com... | 3.26 |
flink_FileMergingSnapshotManagerBuilder_build_rdh | /**
* Create file-merging snapshot manager based on configuration.
*
* <p>TODO (FLINK-32074): Support another type of FileMergingSnapshotManager that merges files
* across different checkpoints.
*
* @return the created manager.
*/
public FileMergingSnapshotManager build() {
re... | 3.26 |
flink_FileMergingSnapshotManagerBuilder_setIOExecutor_rdh | /**
* Set the executor for io operation in manager. If null(default), all io operation will be
* executed synchronously.
*/
public FileMergingSnapshotManagerBuilder setIOExecutor(@Nullable
Executor ioExecutor) {
this.ioExecutor = ioExecutor;
return this;
} | 3.26 |
flink_ProducerMergedPartitionFileIndex_getSize_rdh | /**
* Get the total size in bytes of this region, including the fields and the buffers.
*/
@Override
public int getSize() {
return f0 + numBuffers;
} | 3.26 |
flink_ProducerMergedPartitionFileIndex_getRegion_rdh | /**
* Get the subpartition's {@link FixedSizeRegion} containing the specific buffer index.
*
* @param subpartitionId
* the subpartition id
* @param bufferIndex
* the buffer index
* @return the region containing the buffer index, or return emtpy if the region is not found.
*/
Optional<FixedSizeRegion> getReg... | 3.26 |
flink_ProducerMergedPartitionFileIndex_addBuffers_rdh | /**
* Add buffers to the index.
*
* @param buffers
* to be added. Note, the provided buffers are required to be physically
* consecutive and in the same order as in the file.
*/
void addBuffers(List<FlushedBuffer> buffers) {
if (buffers.isEmpty()) {
return;
}
Map<Integer, List<FixedSizeReg... | 3.26 |
flink_ProducerMergedPartitionFileIndex_convertToRegions_rdh | // ------------------------------------------------------------------------
// Internal Methods
// ------------------------------------------------------------------------
private static Map<Integer, List<FixedSizeRegion>> convertToRegions(List<FlushedBuffer> buffers) {
Map<Integer, List<FixedSizeRegion>> subparti... | 3.26 |
flink_ExtendedParser_getCompletionHints_rdh | /**
* Returns completion hints for the given statement at the given cursor position. The completion
* happens case insensitively.
*
* @param statement
* Partial or slightly incorrect SQL statement
* @param cursor
* cursor position
* @return completion hints that fit at the current cursor position
*/
public... | 3.26 |
flink_ExtendedParser_parse_rdh | /**
* Parse the input statement to the {@link Operation}.
*
* @param statement
* the command to evaluate
* @return parsed operation that represents the command
*/
public Optional<Operation> parse(String statement) {
for (ExtendedParseStrategy strategy : PARSE_STRATEGIES) {
... | 3.26 |
flink_ConfigUtils_m0_rdh | /**
* Puts an array of values of type {@code IN} in a {@link WritableConfig} as a {@link ConfigOption} of type {@link List} of type {@code OUT}. If the {@code values} is {@code null}
* or empty, then nothing is put in the configuration.
*
* @param configuration
* the configuration object to put the list in
* @p... | 3.26 |
flink_ConfigUtils_encodeCollectionToConfig_rdh | /**
* Puts a {@link Collection} of values of type {@code IN} in a {@link WritableConfig} as a
* {@link ConfigOption} of type {@link List} of type {@code OUT}. If the {@code values} is
* {@code null} or empty, then nothing is put in the configuration.
*
* @param configuration
* the configuration object to put th... | 3.26 |
flink_SortOperationFactory_createLimitWithOffset_rdh | /**
* Creates a valid {@link SortQueryOperation} with offset (possibly merged into a preceding
* {@link SortQueryOperation}).
*
* @param offset
* offset to start from
* @param child
* relational expression on top of which to apply the sort operation
* @param postResolverFactory
* factory for creating res... | 3.26 |
flink_SortOperationFactory_createSort_rdh | /**
* Creates a valid {@link SortQueryOperation}.
*
* <p><b>NOTE:</b> If the collation is not explicitly specified for an expression, the
* expression is wrapped in a default ascending order. If no expression is specified, the result
* is not sorted but only limited.
*
* @param orders
* expressions describing... | 3.26 |
flink_SortOperationFactory_createLimitWithFetch_rdh | /**
* Creates a valid {@link SortQueryOperation} with fetch (possibly merged into a preceding
* {@link SortQueryOperation}).
*
* @param fetch
* fetch to limit
* @param child
* relational expression on top of which to apply the sort operation
* @param postResolverFactory
* factory for creating resolved ex... | 3.26 |
flink_SourceReaderContext_currentParallelism_rdh | /**
* Get the current parallelism of this Source.
*
* @return the parallelism of the Source.
*/default int currentParallelism() {
throw new UnsupportedOperationException();
} | 3.26 |
flink_ResolvedCatalogView_getOptions_rdh | // --------------------------------------------------------------------------------------------
// Delegations to original CatalogView
// --------------------------------------------------------------------------------------------
@Override
public Map<String, String> getOptions() {
return origin.getOptions();
} | 3.26 |
flink_HiveFunctionArguments_create_rdh | // create from a CallContext
public static HiveFunctionArguments create(CallContext callContext) {
DataType[] argTypes = callContext.getArgumentDataTypes().toArray(new DataType[0]);
Object[]
args = new Object[argTypes.length];
BitSet literalIndices = new BitSet(args.length);
... | 3.26 |
flink_GroupReduceCombineDriver_setup_rdh | // ------------------------------------------------------------------------
@Override
public void setup(TaskContext<GroupCombineFunction<IN, OUT>, OUT> context) {
this.taskContext = context;
this.running = true;
} | 3.26 |
flink_FileBasedOneShotLatch_await_rdh | /**
* Waits until the latch file is created.
*
* <p>When this method returns, subsequent invocations will not block even after the latch file
* is deleted. Note that this method may not return if the latch file is deleted before this
* method returns.
*
* @throws InterruptedException
* if interrupted while wa... | 3.26 |
flink_PythonTableFunctionOperator_isFinishResult_rdh | /**
* The received udtf execution result is a finish message when it is a byte with value 0x00.
*/
private boolean isFinishResult(byte[] rawUdtfResult, int length) {
return (length == 1) && (rawUdtfResult[0] == 0x0);
} | 3.26 |
flink_SingleOutputStreamOperator_m1_rdh | /**
* Sets the buffering timeout for data produced by this operation. The timeout defines how long
* data may linger in a partially full buffer before being sent over the network.
*
* <p>Lower timeouts lead to lower tail latencies, but may affect throughput. Timeouts of 1 ms
* still sustain high throughput, even f... | 3.26 |
flink_SingleOutputStreamOperator_setChainingStrategy_rdh | /**
* Sets the {@link ChainingStrategy} for the given operator affecting the way operators will
* possibly be co-located on the same thread for increased performance.
*
* @param strategy
* The selected {@link ChainingStrategy}
* @return The operator with the modified chaining strategy
*/
@PublicEvolving
privat... | 3.26 |
flink_SingleOutputStreamOperator_m0_rdh | /**
* Sets the parallelism for this operator.
*
* @param parallelism
* The parallelism for this operator.
* @return The operator with set parallelism.
*/
public SingleOutputStreamOperator<T> m0(int parallelism) {
OperatorValidationUtils.validateParallelism(parallelism, canBeParallel());
transformation.s... | 3.26 |
flink_SingleOutputStreamOperator_setDescription_rdh | /**
* Sets the description for this operation.
*
* <p>Description is used in json plan and web ui, but not in logging and metrics where only
* name is available. Description is expected to provide detailed information about the sink,
* while name is expected to be more simple, providing summary information only, s... | 3.26 |
flink_SingleOutputStreamOperator_slotSharingGroup_rdh | /**
* Sets the slot sharing group of this operation. Parallel instances of operations that are in
* the same slot sharing group will be co-located in the same TaskManager slot, if possible.
*
* <p>Operations inherit the slot sharing group of input operations if all input operations are
* in the same slot sharing g... | 3.26 |
flink_SingleOutputStreamOperator_setUidHash_rdh | /**
* Sets an user provided hash for this operator. This will be used AS IS the create the
* JobVertexID.
*
* <p>The user provided hash is an alternative to the generated hashes, that is considered when
* identifying an operator through the default hash mechanics fails (e.g. because of changes
* between Flink ver... | 3.26 |
flink_SingleOutputStreamOperator_disableChaining_rdh | /**
* Turns off chaining for this operator so thread co-location will not be used as an
* optimization.
*
* <p>Chaining can be turned off for the whole job by {@link StreamExecutionEnvironment#disableOperatorChaining()} however it is not advised for
* performance considerations.
*
* @return The operator with cha... | 3.26 |
flink_SingleOutputStreamOperator_forceNonParallel_rdh | /**
* Sets the parallelism and maximum parallelism of this operator to one. And mark this operator
* cannot set a non-1 degree of parallelism.
*
* @return The operator with only one parallelism.
*/@PublicEvolving
public SingleOutputStreamOperator<T> forceNonParallel() {
transformation.setParallelism(1);
tr... | 3.26 |
flink_SingleOutputStreamOperator_cache_rdh | /**
* Cache the intermediate result of the transformation. Only support bounded streams and
* currently only block mode is supported. The cache is generated lazily at the first time the
* intermediate result is computed. The cache will be clear when {@link CachedDataStream#invalidate()} called or the {@link StreamEx... | 3.26 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.