name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
flink_Transformation_getBufferTimeout_rdh
/** * Returns the buffer timeout of this {@code Transformation}. * * @see #setBufferTimeout(long) */ public long getBufferTimeout() { return bufferTimeout; }
3.26
flink_Transformation_getOutputType_rdh
/** * Returns the output type of this {@code Transformation} as a {@link TypeInformation}. Once * this is used once the output type cannot be changed anymore using {@link #setOutputType}. * * @return The output type of this {@code Transformation} */ public TypeInformation<T> getOutputType() { if (outputTyp...
3.26
flink_Transformation_getManagedMemoryOperatorScopeUseCaseWeights_rdh
/** * Get operator scope use cases that this transformation needs managed memory for, and the * use-case-specific weights for this transformation. The weights are used for sharing managed * memory across transformations for the use cases. Check the individual {@link ManagedMemoryUseCase} for the specific weight defi...
3.26
flink_Transformation_getUserProvidedNodeHash_rdh
/** * Gets the user provided hash. * * @return The user provided hash. */ public String getUserProvidedNodeHash() { return userProvidedNodeHash; }
3.26
flink_Transformation_setDescription_rdh
/** * Changes the description of this {@code Transformation}. */ public void setDescription(String description) { this.description = Preconditions.checkNotNull(description); }
3.26
flink_Transformation_setUid_rdh
/** * Sets an ID for this {@link Transformation}. This is will later be hashed to a uidHash which * is then used to create the JobVertexID (that is shown in logs and the web ui). * * <p>The specified ID is used to assign the same operator ID across job submissions (for * example when starting a job from a savepoin...
3.26
flink_Transformation_setParallelism_rdh
/** * Sets the parallelism of this {@code Transformation}. * * @param parallelism * The new parallelism to set on this {@code Transformation}. */ public void setParallelism(int parallelism) { setParallelism(parallelism, true); }
3.26
flink_Transformation_setMaxParallelism_rdh
/** * Sets the maximum parallelism for this stream transformation. * * @param maxParallelism * Maximum parallelism for this stream transformation. */ public void setMaxParallelism(int maxParallelism) { OperatorValidationUtils.validateMaxParallelism(maxParallelism, UPPER_BOUND_MAX_PARALLELISM); this.maxPa...
3.26
flink_Transformation_declareManagedMemoryUseCaseAtOperatorScope_rdh
/** * Declares that this transformation contains certain operator scope managed memory use case. * * @param managedMemoryUseCase * The use case that this transformation declares needing managed * memory for. * @param weight * Use-case-specific weights for this transformation. Used for sharing managed * ...
3.26
flink_Transformation_setSlotSharingGroup_rdh
/** * Sets the slot sharing group of this transformation. Parallel instances of operations that are * in the same slot sharing group will be co-located in the same TaskManager slot, if possible. * * <p>Initially, an operation is in the default slot sharing group. This can be explicitly set * with constructing a {@...
3.26
flink_Transformation_getSlotSharingGroup_rdh
/** * Returns the slot sharing group of this transformation if present. * * @see #setSlotSharingGroup(SlotSharingGroup) */public Optional<SlotSharingGroup> getSlotSharingGroup() { return f1; }
3.26
flink_Transformation_setBufferTimeout_rdh
/** * Set the buffer timeout of this {@code Transformation}. The timeout defines how long data may * linger in a partially full buffer before being sent over the network. * * <p>Lower timeouts lead to lower tail latencies, but may affect throughput. For Flink 1.5+, * timeouts of 1ms are feasible for jobs with high...
3.26
flink_Transformation_getParallelism_rdh
/** * Returns the parallelism of this {@code Transformation}. */ public int getParallelism() { return parallelism; }
3.26
flink_Transformation_getDescription_rdh
/** * Returns the description of this {@code Transformation}. */ public String getDescription() { return description; }
3.26
flink_Transformation_setOutputType_rdh
/** * Tries to fill in the type information. Type information can be filled in later when the * program uses a type hint. This method checks whether the type information has ever been * accessed before and does not allow modifications if the type was accessed already. This * ensures consistency by making sure diffe...
3.26
flink_Transformation_setResources_rdh
/** * Sets the minimum and preferred resources for this stream transformation. * * @param minResources * The minimum resource of this transformation. * @param preferredResources * The preferred resource of this transformation. */ public void setResources(ResourceSpec minResources, ResourceSpec preferredReso...
3.26
flink_MultiShotLatch_await_rdh
/** * Waits until {@link #trigger()} is called. */ public void await() throws InterruptedException { synchronized(lock) { while (!triggered) { lock.wait(); } triggered = false; } }
3.26
flink_MultiShotLatch_trigger_rdh
/** * Fires the latch. Code that is blocked on {@link #await()} will now return. */ public void trigger() { synchronized(lock) { triggered = true; lock.notifyAll(); } }
3.26
flink_ColumnSummary_getTotalCount_rdh
/** * The number of all rows in this column including both nulls and non-nulls. */ public long getTotalCount() { return getNullCount() + getNonNullCount(); }
3.26
flink_ColumnSummary_containsNonNull_rdh
/** * True if this column contains any non-null values. */ public boolean containsNonNull() { return getNonNullCount() > 0L; }
3.26
flink_FileSourceSplit_getReaderPosition_rdh
/** * Gets the (checkpointed) position of the reader, if set. This value is typically absent for * splits when assigned from the enumerator to the readers, and present when the splits are * recovered from a checkpoint. */ public Optional<CheckpointedPosition> getReaderPosition() { return Optional.ofNullable(f1); }
3.26
flink_FileSourceSplit_length_rdh
/** * Returns the number of bytes in the file region described by this source split. */ public long length() { return length; }
3.26
flink_FileSourceSplit_fileSize_rdh
/** * Returns the full file size in bytes, from {@link FileStatus#getLen()}. */ public long fileSize() { return fileSize; }
3.26
flink_FileSourceSplit_fileModificationTime_rdh
/** * Returns the modification time of the file, from {@link FileStatus#getModificationTime()}. */ public long fileModificationTime() { return fileModificationTime; }
3.26
flink_FileSourceSplit_updateWithCheckpointedPosition_rdh
/** * Creates a copy of this split where the checkpointed position is replaced by the given new * position. * * <p><b>IMPORTANT:</b> Subclasses that add additional information to the split must override * this method to return that subclass type. This contract is enforced by checks in the file * source implementa...
3.26
flink_FileSourceSplit_splitId_rdh
// ------------------------------------------------------------------------ // split properties // ------------------------------------------------------------------------ @Override public String splitId() { return id; }
3.26
flink_FileSourceSplit_m0_rdh
// ------------------------------------------------------------------------ // utils // ------------------------------------------------------------------------ @Override public String m0() { final String hosts = (hostnames.length == 0) ? "(no host info)" : " hosts=" + Arrays.toString(hostnames); return String.format...
3.26
flink_FileSourceSplit_offset_rdh
/** * Returns the start of the file region referenced by this source split. The position is * inclusive, the value indicates the first byte that is part of the split. */ public long offset() { return f0; }
3.26
flink_FlinkContainersSettings_zookeeperHostname_rdh
/** * Sets the {@code zookeeperHostname} and returns a reference to this Builder enabling * method chaining. * * @param zookeeperHostname * The Zookeeper hostname. * @return A reference to this Builder. */ public Builder zookeeperHostname(String zookeeperHostname) { this.zookeeperHostname = zookeeperHostna...
3.26
flink_FlinkContainersSettings_defaultConfig_rdh
/** * {@code FlinkContainersConfig} based on defaults. * * @return The Flink containers config. */ public static FlinkContainersSettings defaultConfig() {return builder().build();}
3.26
flink_FlinkContainersSettings_jobManagerHostname_rdh
/** * Sets the job manager hostname and returns a reference to this Builder enabling method * chaining. * * @param jobManagerHostname * The job manager hostname to set. * @return A reference to this Builder. */ public Builder jobManagerHostname(String jobManagerHostname) { return setConfigOption(JobManage...
3.26
flink_FlinkContainersSettings_checkpointPath_rdh
/** * Sets the {@code checkpointPath} and returns a reference to this Builder enabling method * chaining. * * @param checkpointPath * The checkpoint path to set. * @return A reference to this Builder. */ public Builder checkpointPath(String checkpointPath) { this.checkpointPath = checkpointPath; return...
3.26
flink_FlinkContainersSettings_enableZookeeperHA_rdh
/** * Enables Zookeeper HA. NOTE: this option uses default HA configuration. If you want to use * non-default configuration, you should provide all settings, including the HA_MODE * directly via the {@code basedOn()} method instead. * * @return A reference to this Builder. */ public Builder enableZookeeperHA() { ...
3.26
flink_FlinkContainersSettings_getDefaultCheckpointPath_rdh
/** * Gets default checkpoint path. * * @return The default checkpoint path. */ public static String getDefaultCheckpointPath() { return DEFAULT_CHECKPOINT_PATH; }
3.26
flink_FlinkContainersSettings_getDefaultFlinkHome_rdh
/** * Gets default flink home. * * @return The default flink home path. */ public static String getDefaultFlinkHome() { return DEFAULT_FLINK_HOME; }
3.26
flink_FlinkContainersSettings_jarPaths_rdh
/** * Sets the {@code jarPaths} and returns a reference to this Builder enabling method * chaining. * * @param jarPaths * The {@code jarPaths} to set. * @return A reference to this Builder. */ public Builder jarPaths(Collection<String> jarPaths) { this.jarPaths = jarPaths; return this; }
3.26
flink_FlinkContainersSettings_getNumSlotsPerTaskManager_rdh
/** * Gets number slots per task manager. * * @return The number slots per task manager. */ public int getNumSlotsPerTaskManager() { return numSlotsPerTaskManager; }
3.26
flink_FlinkContainersSettings_numTaskManagers_rdh
/** * Sets the {@code numTaskManagers} and returns a reference to this Builder enabling method * chaining. * * @param numTaskManagers * The {@code numTaskManagers} to set. * @return A reference to this Builder. */ public Builder numTaskManagers(int numTaskManagers) {this.numTaskManagers = numTaskManagers; ...
3.26
flink_FlinkContainersSettings_build_rdh
/** * Returns a {@code FlinkContainersConfig} built from the parameters previously set. * * @return A {@code FlinkContainersConfig} built with parameters of this {@code FlinkContainersConfig.Builder}. */ public FlinkContainersSettings build() { return new FlinkContainersSettings(this); }
3.26
flink_FlinkContainersSettings_isBuildFromFlinkDist_rdh
/** * Returns whether to build from flink-dist or from an existing base container. Also see the * {@code baseImage} property. */ public Boolean isBuildFromFlinkDist() { return buildFromFlinkDist;}
3.26
flink_FlinkContainersSettings_getHaStoragePath_rdh
/** * Gets HA storage path. * * @return The ha storage path. */ public String getHaStoragePath() { return haStoragePath; }
3.26
flink_FlinkContainersSettings_setConfigOption_rdh
/** * Sets a single Flink configuration parameter (the options for flink-conf.yaml) and returns * a reference to this Builder enabling method chaining. * * @param <T> * The type parameter. * @param option * The option. * @param value * The value. * @return A reference to this Builder. */ public <T> Bui...
3.26
flink_FlinkContainersSettings_m0_rdh
/** * Gets jar paths. * * @return The jar paths. */public Collection<String> m0() { return jarPaths; }
3.26
flink_FlinkContainersSettings_fullConfiguration_rdh
/** * Sets the {@code flinkConfiguration} value to {@code config} and returns a reference to * this Builder enabling method chaining. * * @param <T> * the type parameter * @param config * The {@code config} to set. * @return A reference to this Builder. */ public <T> Builder fullConfiguration(Configuration...
3.26
flink_FlinkContainersSettings_flinkDistLocation_rdh
/** * Sets the {@code flinkDistLocation} and returns a reference to this Builder enabling * method chaining. * * @param flinkDistLocation * The {@code flinkDistLocation} to set. * @return A reference to this Builder. */ public Builder flinkDistLocation(String flinkDistLocation) { this.flinkDistLocation = ...
3.26
flink_FlinkContainersSettings_flinkHome_rdh
/** * Sets the path of the Flink distribution inside the container. Returns a reference to this * Builder enabling method chaining. * * @param flinkHome * The {@code flinkHome} to set. * @return A reference to this Builder. */ public Builder flinkHome(String flinkHome) { this.flinkHome = flinkHome; r...
3.26
flink_FlinkContainersSettings_numSlotsPerTaskManager_rdh
/** * Sets the {@code numSlotsPerTaskManager} and returns a reference to this Builder enabling * method chaining. It also adds this property into the {@code flinkConfiguration} field. * * @param numSlotsPerTaskManager * The {@code numSlotsPerTaskManager} to set. * @return A reference to this Builder. */public ...
3.26
flink_FlinkContainersSettings_taskManagerHostnamePrefix_rdh
/** * Sets the {@code taskManagerHostnamePrefix} and returns a reference to this Builder * enabling method chaining. * * @param taskManagerHostnamePrefix * The {@code taskManagerHostnamePrefix} to set. * @return A reference to this Builder. */ public Builder taskManagerHostnamePrefix(String taskManagerHostname...
3.26
flink_FlinkContainersSettings_haStoragePath_rdh
/** * Sets the {@code haStoragePath} and returns a reference to this Builder enabling method * chaining. * * @param haStoragePath * The path for storing HA data. * @return A reference to this Builder. */ public Builder haStoragePath(String haStoragePath) { this.haStoragePath = haStoragePath; return set...
3.26
flink_FlinkContainersSettings_baseImage_rdh
/** * Sets the {@code baseImage} and returns a reference to this Builder enabling method * chaining. * * @param baseImage * The {@code baseImage} to set. * @return A reference to this Builder. */ public Builder baseImage(String baseImage) { this.baseImage = baseImage;this.buildFromFlinkDist = false; r...
3.26
flink_FlinkContainersSettings_builder_rdh
/** * A new builder for {@code FlinkContainersConfig}. * * @return The builder. */ public static Builder builder() { return new Builder(); }
3.26
flink_FlinkContainersSettings_getBaseImage_rdh
/** * Gets base image. * * @return The base image. */ public String getBaseImage() { return baseImage; }
3.26
flink_FlinkContainersSettings_basedOn_rdh
/** * Merges the provided {@code config} with the default config, potentially overwriting the * defaults in case of collisions. Returns a reference to this Builder enabling method * chaining. * * @param <T> * the type parameter * @param config * The {@code config} to add. * @return A reference to this Buil...
3.26
flink_FlinkContainersSettings_isZookeeperHA_rdh
/** * Is zookeeper HA boolean. */ public Boolean isZookeeperHA() { return zookeeperHA; }
3.26
flink_FlinkContainersSettings_setLogProperty_rdh
/** * Sets a single Flink logging configuration property in the log4j format and returns a * reference to this Builder enabling method chaining. * * @param key * The property key. * @param value * The property value. * @return A reference to this Builder. */ public Builder setLogProperty(String key, String...
3.26
flink_FlinkContainersSettings_getLogProperties_rdh
/** * Gets logging properties. * * @return The logging properties. */public Properties getLogProperties() { return logProperties; }
3.26
flink_FlinkContainersSettings_getCheckpointPath_rdh
/** * Gets checkpoint path. * * @return The checkpoint path. */ public String getCheckpointPath() { return f0; }
3.26
flink_PartitionRequestClientFactory_createPartitionRequestClient_rdh
/** * Atomically establishes a TCP connection to the given remote address and creates a {@link NettyPartitionRequestClient} instance for this connection. */ NettyPartitionRequestClient createPartitionRequestClient(ConnectionID connectionId) throws IOException, InterruptedException { // We map the input Connection...
3.26
flink_PartitionRequestClientFactory_destroyPartitionRequestClient_rdh
/** * Removes the client for the given {@link ConnectionID}. */ void destroyPartitionRequestClient(ConnectionID connectionId, PartitionRequestClient client) { final CompletableFuture<NettyPartitionRequestClient> future = clients.get(connectionId); if ((future != null) && future.isDone()) { future.the...
3.26
flink_AllGroupCombineDriver_setup_rdh
// ------------------------------------------------------------------------ @Override public void setup(TaskContext<GroupCombineFunction<IN, OUT>, OUT> context) { this.taskContext = context; }
3.26
flink_StreamPhysicalPythonCorrelateRule_findTableFunction_rdh
// find only calc and table function private boolean findTableFunction(FlinkLogicalCalc calc) { RelNode child = ((RelSubset) (calc.getInput())).getOriginal(); if (child instanceof FlinkLogicalTableFunctionScan) { FlinkLogicalTableFunctionScan scan = ((FlinkLogicalTableFunctionScan) (child)); ret...
3.26
flink_BridgingSqlAggFunction_of_rdh
/** * Creates an instance of a aggregate function during translation. */ public static BridgingSqlAggFunction of(FlinkContext context, FlinkTypeFactory typeFactory, ContextResolvedFunction resolvedFunction) { final DataTypeFactory dataTypeFactory = context.getCatalogManager().getDataTypeFactory(); final Type...
3.26
flink_BridgingSqlAggFunction_createOrderRequirement_rdh
// -------------------------------------------------------------------------------------------- private static boolean createOrderRequirement() { return false; }
3.26
flink_HadoopInputs_createHadoopInput_rdh
/** * Creates a Flink {@link InputFormat} that wraps the given Hadoop {@link org.apache.hadoop.mapreduce.InputFormat}. * * @return A Flink InputFormat that wraps the Hadoop InputFormat. */ public static <K, V> HadoopInputFormat<K, V> createHadoopInput(InputFormat<K, V> mapreduceInputFormat, Class<K> key, Class<V> v...
3.26
flink_HadoopInputs_readHadoopFile_rdh
/** * Creates a Flink {@link InputFormat} that wraps the given Hadoop {@link org.apache.hadoop.mapreduce.lib.input.FileInputFormat}. * * @return A Flink InputFormat that wraps the Hadoop FileInputFormat. */ public static <K, V> HadoopInputFormat<K, V> readHadoopFile(FileInputFormat<K, V> mapreduceInputFormat, Class...
3.26
flink_FlinkResultSetMetaData_getType_rdh
/** * Get column type name according type in {@link Types}. * * @param type * the type in {@link Types} * @return type class name */ static String getType(int type) throws SQLException {// see javax.sql.rowset.RowSetMetaDataImpl switch (type) { case Types.NUMERIC : case Types.DECIMAL : return Bi...
3.26
flink_LogicalPipelinedRegionComputeUtil_computePipelinedRegions_rdh
/** * Utils for computing {@link LogicalPipelinedRegion}s. */public final class LogicalPipelinedRegionComputeUtil { public static Set<Set<LogicalVertex>> computePipelinedRegions(final Iterable<? extends LogicalVertex> topologicallySortedVertices) { final Map<LogicalVertex, Set<LogicalVertex>> vertexToRegion...
3.26
flink_GroupReduceOperator_translateToDataFlow_rdh
// -------------------------------------------------------------------------------------------- @Override @SuppressWarnings("unchecked") protected GroupReduceOperatorBase<?, OUT, ?> translateToDataFlow(Operator<IN> input) { String name = (getName() != null) ? getName() : "GroupReduce at " + defaultName; // wrap...
3.26
flink_GroupReduceOperator_translateSelectorFunctionReducer_rdh
// -------------------------------------------------------------------------------------------- @SuppressWarnings("unchecked") private static <IN, OUT, K> PlanUnwrappingReduceGroupOperator<IN, OUT, K> translateSelectorFunctionReducer(SelectorFunctionKeys<IN, ?> rawKeys, GroupReduceFunction<IN, OUT> function, TypeInform...
3.26
flink_GroupReduceOperator_isCombinable_rdh
// -------------------------------------------------------------------------------------------- // Properties // -------------------------------------------------------------------------------------------- @Internal public boolean isCombinable() { return combinable; }
3.26
flink_SlotSharingSlotAllocator_determineSlotsPerSharingGroup_rdh
/** * Distributes free slots across the slot-sharing groups of the job. Slots are distributed as * evenly as possible. If a group requires less than an even share of slots the remainder is * distributed over the remaining groups. */ private static Map<SlotSharingGroupId, Integer> determineSlotsPerSharingGroup(JobIn...
3.26
flink_JoinNode_computeOperatorSpecificDefaultEstimates_rdh
/** * The default estimates build on the principle of inclusion: The smaller input key domain is * included in the larger input key domain. We also assume that every key from the larger input * has one join partner in the smaller input. The result cardinality is hence the larger one. */ @Override protected void com...
3.26
flink_FileMergingSnapshotManagerBuilder_build_rdh
/** * Create file-merging snapshot manager based on configuration. * * <p>TODO (FLINK-32074): Support another type of FileMergingSnapshotManager that merges files * across different checkpoints. * * @return the created manager. */ public FileMergingSnapshotManager build() { re...
3.26
flink_FileMergingSnapshotManagerBuilder_setIOExecutor_rdh
/** * Set the executor for io operation in manager. If null(default), all io operation will be * executed synchronously. */ public FileMergingSnapshotManagerBuilder setIOExecutor(@Nullable Executor ioExecutor) { this.ioExecutor = ioExecutor; return this; }
3.26
flink_ProducerMergedPartitionFileIndex_getSize_rdh
/** * Get the total size in bytes of this region, including the fields and the buffers. */ @Override public int getSize() { return f0 + numBuffers; }
3.26
flink_ProducerMergedPartitionFileIndex_getRegion_rdh
/** * Get the subpartition's {@link FixedSizeRegion} containing the specific buffer index. * * @param subpartitionId * the subpartition id * @param bufferIndex * the buffer index * @return the region containing the buffer index, or return emtpy if the region is not found. */ Optional<FixedSizeRegion> getReg...
3.26
flink_ProducerMergedPartitionFileIndex_addBuffers_rdh
/** * Add buffers to the index. * * @param buffers * to be added. Note, the provided buffers are required to be physically * consecutive and in the same order as in the file. */ void addBuffers(List<FlushedBuffer> buffers) { if (buffers.isEmpty()) { return; } Map<Integer, List<FixedSizeReg...
3.26
flink_ProducerMergedPartitionFileIndex_convertToRegions_rdh
// ------------------------------------------------------------------------ // Internal Methods // ------------------------------------------------------------------------ private static Map<Integer, List<FixedSizeRegion>> convertToRegions(List<FlushedBuffer> buffers) { Map<Integer, List<FixedSizeRegion>> subparti...
3.26
flink_ExtendedParser_getCompletionHints_rdh
/** * Returns completion hints for the given statement at the given cursor position. The completion * happens case insensitively. * * @param statement * Partial or slightly incorrect SQL statement * @param cursor * cursor position * @return completion hints that fit at the current cursor position */ public...
3.26
flink_ExtendedParser_parse_rdh
/** * Parse the input statement to the {@link Operation}. * * @param statement * the command to evaluate * @return parsed operation that represents the command */ public Optional<Operation> parse(String statement) { for (ExtendedParseStrategy strategy : PARSE_STRATEGIES) { ...
3.26
flink_ConfigUtils_m0_rdh
/** * Puts an array of values of type {@code IN} in a {@link WritableConfig} as a {@link ConfigOption} of type {@link List} of type {@code OUT}. If the {@code values} is {@code null} * or empty, then nothing is put in the configuration. * * @param configuration * the configuration object to put the list in * @p...
3.26
flink_ConfigUtils_encodeCollectionToConfig_rdh
/** * Puts a {@link Collection} of values of type {@code IN} in a {@link WritableConfig} as a * {@link ConfigOption} of type {@link List} of type {@code OUT}. If the {@code values} is * {@code null} or empty, then nothing is put in the configuration. * * @param configuration * the configuration object to put th...
3.26
flink_SortOperationFactory_createLimitWithOffset_rdh
/** * Creates a valid {@link SortQueryOperation} with offset (possibly merged into a preceding * {@link SortQueryOperation}). * * @param offset * offset to start from * @param child * relational expression on top of which to apply the sort operation * @param postResolverFactory * factory for creating res...
3.26
flink_SortOperationFactory_createSort_rdh
/** * Creates a valid {@link SortQueryOperation}. * * <p><b>NOTE:</b> If the collation is not explicitly specified for an expression, the * expression is wrapped in a default ascending order. If no expression is specified, the result * is not sorted but only limited. * * @param orders * expressions describing...
3.26
flink_SortOperationFactory_createLimitWithFetch_rdh
/** * Creates a valid {@link SortQueryOperation} with fetch (possibly merged into a preceding * {@link SortQueryOperation}). * * @param fetch * fetch to limit * @param child * relational expression on top of which to apply the sort operation * @param postResolverFactory * factory for creating resolved ex...
3.26
flink_SourceReaderContext_currentParallelism_rdh
/** * Get the current parallelism of this Source. * * @return the parallelism of the Source. */default int currentParallelism() { throw new UnsupportedOperationException(); }
3.26
flink_ResolvedCatalogView_getOptions_rdh
// -------------------------------------------------------------------------------------------- // Delegations to original CatalogView // -------------------------------------------------------------------------------------------- @Override public Map<String, String> getOptions() { return origin.getOptions(); }
3.26
flink_HiveFunctionArguments_create_rdh
// create from a CallContext public static HiveFunctionArguments create(CallContext callContext) { DataType[] argTypes = callContext.getArgumentDataTypes().toArray(new DataType[0]); Object[] args = new Object[argTypes.length]; BitSet literalIndices = new BitSet(args.length); ...
3.26
flink_GroupReduceCombineDriver_setup_rdh
// ------------------------------------------------------------------------ @Override public void setup(TaskContext<GroupCombineFunction<IN, OUT>, OUT> context) { this.taskContext = context; this.running = true; }
3.26
flink_FileBasedOneShotLatch_await_rdh
/** * Waits until the latch file is created. * * <p>When this method returns, subsequent invocations will not block even after the latch file * is deleted. Note that this method may not return if the latch file is deleted before this * method returns. * * @throws InterruptedException * if interrupted while wa...
3.26
flink_PythonTableFunctionOperator_isFinishResult_rdh
/** * The received udtf execution result is a finish message when it is a byte with value 0x00. */ private boolean isFinishResult(byte[] rawUdtfResult, int length) { return (length == 1) && (rawUdtfResult[0] == 0x0); }
3.26
flink_SingleOutputStreamOperator_m1_rdh
/** * Sets the buffering timeout for data produced by this operation. The timeout defines how long * data may linger in a partially full buffer before being sent over the network. * * <p>Lower timeouts lead to lower tail latencies, but may affect throughput. Timeouts of 1 ms * still sustain high throughput, even f...
3.26
flink_SingleOutputStreamOperator_setChainingStrategy_rdh
/** * Sets the {@link ChainingStrategy} for the given operator affecting the way operators will * possibly be co-located on the same thread for increased performance. * * @param strategy * The selected {@link ChainingStrategy} * @return The operator with the modified chaining strategy */ @PublicEvolving privat...
3.26
flink_SingleOutputStreamOperator_m0_rdh
/** * Sets the parallelism for this operator. * * @param parallelism * The parallelism for this operator. * @return The operator with set parallelism. */ public SingleOutputStreamOperator<T> m0(int parallelism) { OperatorValidationUtils.validateParallelism(parallelism, canBeParallel()); transformation.s...
3.26
flink_SingleOutputStreamOperator_setDescription_rdh
/** * Sets the description for this operation. * * <p>Description is used in json plan and web ui, but not in logging and metrics where only * name is available. Description is expected to provide detailed information about the sink, * while name is expected to be more simple, providing summary information only, s...
3.26
flink_SingleOutputStreamOperator_slotSharingGroup_rdh
/** * Sets the slot sharing group of this operation. Parallel instances of operations that are in * the same slot sharing group will be co-located in the same TaskManager slot, if possible. * * <p>Operations inherit the slot sharing group of input operations if all input operations are * in the same slot sharing g...
3.26
flink_SingleOutputStreamOperator_setUidHash_rdh
/** * Sets an user provided hash for this operator. This will be used AS IS the create the * JobVertexID. * * <p>The user provided hash is an alternative to the generated hashes, that is considered when * identifying an operator through the default hash mechanics fails (e.g. because of changes * between Flink ver...
3.26
flink_SingleOutputStreamOperator_disableChaining_rdh
/** * Turns off chaining for this operator so thread co-location will not be used as an * optimization. * * <p>Chaining can be turned off for the whole job by {@link StreamExecutionEnvironment#disableOperatorChaining()} however it is not advised for * performance considerations. * * @return The operator with cha...
3.26
flink_SingleOutputStreamOperator_forceNonParallel_rdh
/** * Sets the parallelism and maximum parallelism of this operator to one. And mark this operator * cannot set a non-1 degree of parallelism. * * @return The operator with only one parallelism. */@PublicEvolving public SingleOutputStreamOperator<T> forceNonParallel() { transformation.setParallelism(1); tr...
3.26
flink_SingleOutputStreamOperator_cache_rdh
/** * Cache the intermediate result of the transformation. Only support bounded streams and * currently only block mode is supported. The cache is generated lazily at the first time the * intermediate result is computed. The cache will be clear when {@link CachedDataStream#invalidate()} called or the {@link StreamEx...
3.26