name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
flink_ResultPartitionType_isHybridResultPartition_rdh
/** * {@link #isHybridResultPartition()} is used to judge whether it is the specified {@link #HYBRID_FULL} or {@link #HYBRID_SELECTIVE} resultPartitionType. * * <p>this method suitable for judgment conditions related to the specific implementation of * {@link ResultPartitionType}. * * <p>this method not related t...
3.26
flink_ResultPartitionType_isBounded_rdh
/** * Whether this partition uses a limited number of (network) buffers or not. * * @return <tt>true</tt> if the number of buffers should be bound to some limit */ public boolean isBounded() { return isBounded; }
3.26
flink_ResultPartitionType_isPipelinedOrPipelinedBoundedResultPartition_rdh
/** * {@link #isPipelinedOrPipelinedBoundedResultPartition()} is used to judge whether it is the * specified {@link #PIPELINED} or {@link #PIPELINED_BOUNDED} resultPartitionType. * * <p>This method suitable for judgment conditions related to the specific implementation of * {@link ResultPartitionType}. * * <p>Th...
3.26
flink_ResultPartitionType_isBlockingOrBlockingPersistentResultPartition_rdh
/** * {@link #isBlockingOrBlockingPersistentResultPartition()} is used to judge whether it is the * specified {@link #BLOCKING} or {@link #BLOCKING_PERSISTENT} resultPartitionType. * * <p>this method suitable for judgment conditions related to the specific implementation of * {@link ResultPartitionType}. * * <p>...
3.26
flink_ResultPartitionType_mustBePipelinedConsumed_rdh
/** * return if this partition's upstream and downstream must be scheduled in the same time. */ public boolean mustBePipelinedConsumed() { return f1 == ConsumingConstraint.MUST_BE_PIPELINED;}
3.26
flink_ExternalPythonKeyedCoProcessOperator_processTimer_rdh
/** * It is responsible to send timer data to python worker when a registered timer is fired. The * input data is a Row containing 4 fields: TimerFlag 0 for proc time, 1 for event time; * Timestamp of the fired timer; Current watermark and the key of the timer. * * @param timeDomain * The type of the timer. * ...
3.26
flink_ExternalPythonKeyedCoProcessOperator_setCurrentKey_rdh
/** * As the beam state gRPC service will access the KeyedStateBackend in parallel with this * operator, we must override this method to prevent changing the current key of the * KeyedStateBackend while the beam service is handling requests. */ @Override public void setCurrentKey(Object key) { if (inBatchExecut...
3.26
flink_GSFileSystemOptions_getWriterTemporaryBucketName_rdh
/** * The temporary bucket name to use for recoverable writes, if different from the final bucket * name. */public Optional<String> getWriterTemporaryBucketName() { return flinkConfig.getOptional(WRITER_TEMPORARY_BUCKET_NAME); }
3.26
flink_GSFileSystemOptions_m0_rdh
/** * The chunk size to use for writes on the underlying Google WriteChannel. */ public Optional<MemorySize> m0() { return flinkConfig.getOptional(WRITER_CHUNK_SIZE); }
3.26
flink_TableSource_explainSource_rdh
/** * Describes the table source. * * @return A String explaining the {@link TableSource}. */default String explainSource() { return TableConnectorUtils.generateRuntimeName(getClass(), getTableSchema().getFieldNames()); }
3.26
flink_TableSource_getProducedDataType_rdh
/** * Returns the {@link DataType} for the produced data of the {@link TableSource}. * * @return The data type of the returned {@code DataStream}. */ default DataType getProducedDataType() { final TypeInformation<T> legacyType = getReturnType(); if (legacyType == null) { throw new TableException(...
3.26
flink_RichSqlInsert_isUpsert_rdh
// ~ Tools ------------------------------------------------------------------ public static boolean isUpsert(List<SqlLiteral> keywords) { for (SqlNode keyword : keywords) { SqlInsertKeyword keyword2 = ((SqlLiteral) (keyword)).symbolValue(SqlInsertKeyword.class); if (keyword2 == SqlInsertKeyword.UP...
3.26
flink_RichSqlInsert_isOverwrite_rdh
/** * Returns whether the insert mode is overwrite (for whole table or for specific partitions). * * @return true if this is overwrite mode */ public boolean isOverwrite() { return getModifierNode(RichSqlInsertKeyword.OVERWRITE) != null; }
3.26
flink_RichSqlInsert_getTableHints_rdh
/** * Returns the table hints as list of {@code SqlNode} for current insert node. */ public SqlNodeList getTableHints() { return this.tableHints; }
3.26
flink_TupleComparator_hash_rdh
// -------------------------------------------------------------------------------------------- // Comparator Methods // -------------------------------------------------------------------------------------------- @SuppressWarnings("unchecked") @Override public int hash(T value) { int i = 0; try { int ...
3.26
flink_SourceProvider_of_rdh
/** * Helper method for creating a Source provider with a provided source parallelism. */ static SourceProvider of(Source<RowData, ?, ?> source, @Nullable Integer sourceParallelism) { return new SourceProvider() { @Override public Source<RowData, ?, ?> createSource() { return source; ...
3.26
flink_RowSerializer_getPositionByName_rdh
// -------------------------------------------------------------------------------------------- private int getPositionByName(String fieldName) { assert positionByName != null; final Integer targetPos = positionByName.get(fieldName); if (targetPos == null) { throw new RuntimeException(String.format("Unknown field name ...
3.26
flink_RowSerializer_snapshotConfiguration_rdh
// -------------------------------------------------------------------------------------------- // Serializer configuration snapshoting & compatibility // -------------------------------------------------------------------------------------------- @Override public TypeSerializerSnapshot<Row> snapshotConfiguration() { r...
3.26
flink_RowSerializer_fillMask_rdh
// -------------------------------------------------------------------------------------------- // Serialization utilities // -------------------------------------------------------------------------------------------- private static void fillMask(int fieldLength, Row row, boolean[] mask, boolean supportsRowKind, int...
3.26
flink_AbstractStreamTableEnvironmentImpl_execEnv_rdh
/** * This is a temporary workaround for Python API. Python API should not use * StreamExecutionEnvironment at all. */ @Internalpublic StreamExecutionEnvironment execEnv() { return executionEnvironment; }
3.26
flink_AbstractID_longToByteArray_rdh
/** * Converts a long to a byte array. * * @param l * the long variable to be converted * @param ba * the byte array to store the result the of the conversion * @param offset * offset indicating at what position inside the byte array the result of the * conversion shall be stored */ private static voi...
3.26
flink_AbstractID_getBytes_rdh
/** * Gets the bytes underlying this ID. * * @return The bytes underlying this ID. */public byte[] getBytes() { byte[] bytes = new byte[SIZE]; longToByteArray(lowerPart, bytes, 0); longToByteArray(upperPart, bytes, SIZE_OF_LONG); return bytes; }
3.26
flink_AbstractID_toHexString_rdh
/** * Returns pure String representation of the ID in hexadecimal. This method should be used to * construct things like paths etc., that require a stable representation and is therefore * final. */ public final String toHexString() { if (this.hexString == null) { final byte[] ba = new byte[SIZE]; ...
3.26
flink_AbstractID_equals_rdh
// -------------------------------------------------------------------------------------------- // Standard Utilities // -------------------------------------------------------------------------------------------- @Override public boolean equals(Object obj) { if (obj == this) { return true; } else ...
3.26
flink_AbstractID_getLowerPart_rdh
// -------------------------------------------------------------------------------------------- /** * Gets the lower 64 bits of the ID. * * @return The lower 64 bits of the ID. */ public long getLowerPart() { return lowerPart; }
3.26
flink_AbstractID_byteArrayToLong_rdh
// Conversion Utilities // -------------------------------------------------------------------------------------------- /** * Converts the given byte array to a long. * * @param ba * the byte array to be converted * @param offset * the offset indicating at which byte inside the array the conversion shall begi...
3.26
flink_AvroUtils_getAvroUtils_rdh
/** * Returns either the default {@link AvroUtils} which throw an exception in cases where Avro * would be needed or loads the specific utils for Avro from flink-avro. */ public static AvroUtils getAvroUtils() { // try and load the special AvroUtils from the flink-avro package try { Class<?> clazz ...
3.26
flink_PlannerCallProcedureOperation_toExternal_rdh
/** * Convert the value with internal representation to the value with external representation. */private Object toExternal(Object internalValue, DataType inputType, ClassLoader classLoader) { if (!DataTypeUtils.isInternal(inputType)) { // if the expected input type of the procedure is...
3.26
flink_PlannerCallProcedureOperation_procedureResultToTableResult_rdh
/** * Convert the result of procedure to table result . */ private TableResultInternal procedureResultToTableResult(Object procedureResult, TableConfig tableConfig, ClassLoader userClassLoader) { // get result converter ZoneId zoneId = tableConfig.getLocalTimeZone(); DataType v21 = outputType; // if is not compos...
3.26
flink_WordCount_main_rdh
// ************************************************************************* // PROGRAM // ************************************************************************* public static void main(String[] args) throws Exception { LOGGER.warn(DATASET_DEPRECATION_INFO); final MultipleParameterToo...
3.26
flink_CoProcessFunction_onTimer_rdh
/** * Called when a timer set using {@link TimerService} fires. * * @param timestamp * The timestamp of the firing timer. * @param ctx * An {@link OnTimerContext} that allows querying the timestamp of the firing timer, * querying the {@link TimeDomain} of the firing timer and getting a {@link TimerService}...
3.26
flink_ListAggWithRetractAggFunction_getArgumentDataTypes_rdh
// -------------------------------------------------------------------------------------------- // Planning // -------------------------------------------------------------------------------------------- @Override public List<DataType> getArgumentDataTypes() { return Collections.singletonList(DataTypes.STRING().bri...
3.26
flink_BatchShuffleReadBufferPool_requestBuffers_rdh
/** * Requests a collection of buffers (determined by {@link #numBuffersPerRequest}) from this * buffer pool. */ public List<MemorySegment> requestBuffers() throws Exception { List<MemorySegment> allocated = new ArrayList...
3.26
flink_BatchShuffleReadBufferPool_destroy_rdh
/** * Destroys this buffer pool and after which, no buffer can be allocated any more. */ public void destroy() { synchronized(buffers) { destroyed = true; buffers.clear(); buffers.notifyAll(); } }
3.26
flink_BatchShuffleReadBufferPool_initialize_rdh
/** * Initializes this buffer pool which allocates all the buffers. */ public void initialize() { synchronized(buffers) { checkState(!destroyed, "Buffer pool is already destroyed."); if (initialized) { return; } initialized = true; tr...
3.26
flink_BatchShuffleReadBufferPool_recycle_rdh
/** * Recycles a collection of buffers to this buffer pool. This method should never throw any * exception. */ public void recycle(Collection<MemorySegment> segments) { checkArgument(segments != null, "Buffer list must be not null."); if (segments.isEmpty()) { return; }synchronized(buffers) { ...
3.26
flink_JobGraph_getCheckpointingSettings_rdh
/** * Gets the settings for asynchronous snapshots. This method returns null, when checkpointing is * not enabled. * * @return The snapshot settings */ public JobCheckpointingSettings getCheckpointingSettings() { return snapshotSettings;}
3.26
flink_JobGraph_addVertex_rdh
/** * Adds a new task vertex to the job graph if it is not already included. * * @param vertex * the new task vertex to be added */ public void addVertex(JobVertex vertex) { final JobVertexID id = vertex.getID(); JobVertex previous = taskVertices.put(id, vertex); // if we had a prior association, restore and ...
3.26
flink_JobGraph_addUserJarBlobKey_rdh
/** * Adds the BLOB referenced by the key to the JobGraph's dependencies. * * @param key * path of the JAR file required to run the job on a task manager */ public void addUserJarBlobKey(PermanentBlobKey key) { if (key == null) { throw new IllegalArgumentException(); } if (!userJarBlobKeys.co...
3.26
flink_JobGraph_getJobConfiguration_rdh
/** * Returns the configuration object for this job. Job-wide parameters should be set into that * configuration object. * * @return The configuration object for this job. */ public Configuration getJobConfiguration() { return this.jobConfiguration; }
3.26
flink_JobGraph_setJobID_rdh
/** * Sets the ID of the job. */ public void setJobID(JobID jobID) { this.jobID = jobID; }
3.26
flink_JobGraph_setClasspaths_rdh
/** * Sets the classpaths required to run the job on a task manager. * * @param paths * paths of the directories/JAR files required to run the job on a task manager */ public void setClasspaths(List<URL> paths) { classpaths = paths; }
3.26
flink_JobGraph_getVertices_rdh
/** * Returns an Iterable to iterate all vertices registered with the job graph. * * @return an Iterable to iterate all vertices registered with the job graph */public Iterable<JobVertex> getVertices() { return this.taskVertices.values(); }
3.26
flink_JobGraph_getNumberOfVertices_rdh
/** * Returns the number of all vertices. * * @return The number of all vertices. */ public int getNumberOfVertices() {return this.taskVertices.size(); }
3.26
flink_JobGraph_setExecutionConfig_rdh
/** * Sets the execution config. This method eagerly serialized the ExecutionConfig for future RPC * transport. Further modification of the referenced ExecutionConfig object will not affect this * serialized copy. * * @param executionConfig * The ExecutionConfig to be serialized. * @throws IOException * Thr...
3.26
flink_JobGraph_getMaximumParallelism_rdh
/** * Gets the maximum parallelism of all operations in this job graph. * * @return The maximum parallelism of this job graph */public int getMaximumParallelism() { int maxParallelism = -1; for (JobVertex v6 : taskVertices.values()) { maxParallelism = Math.max(v6.getParallelism(), maxParallelism); } return ...
3.26
flink_JobGraph_addJars_rdh
/** * Adds the given jar files to the {@link JobGraph} via {@link JobGraph#addJar}. * * @param jarFilesToAttach * a list of the {@link URL URLs} of the jar files to attach to the * jobgraph. * @throws RuntimeException * if a jar URL is not valid. */ public void addJars(final List<URL> jarFilesToAttach) { ...
3.26
flink_JobGraph_setSavepointRestoreSettings_rdh
/** * Sets the savepoint restore settings. * * @param settings * The savepoint restore settings. */ public void setSavepointRestoreSettings(SavepointRestoreSettings settings) { this.savepointRestoreSettings = checkNotNull(settings, "Savepoint restore settings"); }
3.26
flink_JobGraph_getVerticesSortedTopologicallyFromSources_rdh
// -------------------------------------------------------------------------------------------- public List<JobVertex> getVerticesSortedTopologicallyFromSources() throws InvalidProgramException { // early out on empty lists if (this.taskVertices.isEmpty()) { return Collections.emptyList(); } List<JobVertex> sorted...
3.26
flink_JobGraph_addUserArtifact_rdh
/** * Adds the path of a custom file required to run the job on a task manager. * * @param name * a name under which this artifact will be accessible through {@link DistributedCache} * @param file * path of a custom file required to run the job on a task manager */ public void addUserArtifact(String name, D...
3.26
flink_JobGraph_getJobID_rdh
// -------------------------------------------------------------------------------------------- /** * Returns the ID of the job. * * @return the ID of the job */ public JobID getJobID() { return this.jobID; }
3.26
flink_JobGraph_getSerializedExecutionConfig_rdh
/** * Returns the {@link ExecutionConfig}. * * @return ExecutionConfig */ public SerializedValue<ExecutionConfig> getSerializedExecutionConfig() { return serializedExecutionConfig;}
3.26
flink_JobGraph_addJar_rdh
// -------------------------------------------------------------------------------------------- // Handling of attached JAR files // -------------------------------------------------------------------------------------------- /** * Adds the path of a JAR file required to run the job on a task manager. * * @param jar...
3.26
flink_JobGraph_m1_rdh
/** * Returns an array of all job vertices that are registered with the job graph. The order in * which the vertices appear in the list is not defined. * * @return an array of all job vertices that are registered with the job graph */ public JobVertex[] m1() { return this.taskVertices.values().toArray(new JobVerte...
3.26
flink_JobGraph_getUserArtifacts_rdh
/** * Gets the list of assigned user jar paths. * * @return The list of assigned user jar paths */ public Map<String, DistributedCache.DistributedCacheEntry> getUserArtifacts() { return userArtifacts; }
3.26
flink_JobGraph_getSavepointRestoreSettings_rdh
/** * Returns the configured savepoint restore setting. * * @return The configured savepoint restore settings. */ public SavepointRestoreSettings getSavepointRestoreSettings() { return savepointRestoreSettings; }
3.26
flink_JobGraph_getCoLocationGroups_rdh
/** * Returns all {@link CoLocationGroup} instances associated with this {@code JobGraph}. * * @return The associated {@code CoLocationGroup} instances. */ public Set<CoLocationGroup> getCoLocationGroups() { final Set<CoLocationGroup> coLocationGroups = IterableUtils.toStream(getVertices()).map(JobVertex::getCoLo...
3.26
flink_JobGraph_isCheckpointingEnabled_rdh
/** * Checks if the checkpointing was enabled for this job graph. * * @return true if checkpointing enabled */ public boolean isCheckpointingEnabled() { if (snapshotSettings == null) { return false; } return snapshotSettings.getCheckpointCoordinatorConfiguration().isCheckpointingEnabled(); }
3.26
flink_ApplicationDispatcherBootstrap_runApplicationEntryPoint_rdh
/** * Runs the user program entrypoint and completes the given {@code jobIdsFuture} with the {@link JobID JobIDs} of the submitted jobs. * * <p>This should be executed in a separate thread (or task). */ private void runApplicationEntryPoint(final CompletableFuture<List<JobID>> jobIdsFuture, final Set<JobID> tolerat...
3.26
flink_ApplicationDispatcherBootstrap_m0_rdh
/** * Runs the user program entrypoint by scheduling a task on the given {@code scheduledExecutor}. * The returned {@link CompletableFuture} completes when all jobs of the user application * succeeded. if any of them fails, or if job submission fails. */ private CompletableFuture<Void> m0(final DispatcherGateway di...
3.26
flink_ApplicationDispatcherBootstrap_finishBootstrapTasks_rdh
/** * Logs final application status and invokes error handler in case of unexpected failures. * Optionally shuts down the given dispatcherGateway when the application completes (either * successfully or in case of failure), depending on the corresponding config option. */ private Completable...
3.26
flink_ApplicationDispatcherBootstrap_unwrapJobResultException_rdh
/** * If the given {@link JobResult} indicates success, this passes through the {@link JobResult}. * Otherwise, this returns a future that is finished exceptionally (potentially with an * exception from the {@link JobResult}). */ private CompletableFuture<JobResult> unwrapJobResultException(final CompletableFuture<...
3.26
flink_HsSubpartitionConsumerMemoryDataManager_addBuffer_rdh
// this method only called from subpartitionMemoryDataManager with write lock. @GuardedBy("consumerLock") public boolean addBuffer(HsBufferContext bufferContext) { tryIncreaseBacklog(bufferContext.getBuffer()); unConsumedBuffers.add(bufferContext); trimHeadingReleasedBuffers(); return unCons...
3.26
flink_HsSubpartitionConsumerMemoryDataManager_getBacklog_rdh
// Un-synchronized get the backlog to provide memory data backlog, this will make the // result greater than or equal to the actual backlog, but obtaining an accurate backlog will // bring too much extra overhead. @SuppressWarnings("FieldAccessNotGuarded") @Override public int getBacklog() {return backlog; }
3.26
flink_HsSubpartitionConsumerMemoryDataManager_addInitialBuffers_rdh
// this method only called from subpartitionMemoryDataManager with write lock. @GuardedBy("consumerLock") public void addInitialBuffers(Deque<HsBufferContext> buffers) { for (HsBufferContext bufferContext : buffers) { tryIncreaseBacklog(bufferContext.getBuffer()); } unConsumedBuffers.addAll(buffers...
3.26
flink_ServiceType_classify_rdh
// Helper method public static ServiceExposedType classify(Service service) { KubernetesConfigOptions.ServiceExposedType type = KubernetesConfigOptions.ServiceExposedType.valueOf(service.getSpec().getType()); if (type == ServiceExposedType.ClusterIP) { if (HeadlessClusterIPService.HEADLESS_CLUSTER_...
3.26
flink_ServiceType_buildUpExternalRestService_rdh
/** * Build up the external rest service template, according to the jobManager parameters. * * @param kubernetesJobManagerParameters * the parameters of jobManager. * @return the external rest service */ public Service buildUpExternalRestService(KubernetesJobManagerParameters kubernetesJobManagerParameters) { ...
3.26
flink_ServiceType_getRestPortFromExternalService_rdh
/** * Get rest port from the external Service. */ public int getRestPortFromExternalService(Service externalService) { final List<ServicePort> servicePortCandidates = externalService.getSpec().getPorts().stream().filter(x -> x.getName().equals(Constants.REST_PORT_NAME)).collect(Collectors.toList()); if (servi...
3.26
flink_Tuple12_equals_rdh
/** * Deep equality for tuples by calling equals() on the tuple members. * * @param o * the object checked for equality * @return true if this is equal to o. */ @Override public boolean equals(Object o) { if (this == o) { return true; } if (!(o instanceof Tuple12)) { return false; ...
3.26
flink_Tuple12_m1_rdh
/** * Sets new values to all fields of the tuple. * * @param f0 * The value for field 0 * @param f1 * The value for field 1 * @param f2 * The value for field 2 * @param f3 * The value for field 3 * @param f4 * The value for field 4 * @param f5 * The value for field 5 * @param f6 * The valu...
3.26
flink_Tuple12_toString_rdh
// ------------------------------------------------------------------------------------------------- // standard utilities // ------------------------------------------------------------------------------------------------- /** * Creates a string representation of the tuple in the form (f0, f1, f2, f3, f4, f5, f6, f7,...
3.26
flink_Tuple12_of_rdh
/** * Creates a new tuple and assigns the given values to the tuple's fields. This is more * convenient than using the constructor, because the compiler can infer the generic type * arguments implicitly. For example: {@code Tuple3.of(n, x, s)} instead of {@code new * Tuple3<Integer, Double, String>(n, x, s)} */ pu...
3.26
flink_Tuple12_copy_rdh
/** * Shallow tuple copy. * * @return A new Tuple with the same fields as this. */ @Override @SuppressWarnings("unchecked") public Tuple12<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11> copy() {return new Tuple12<>(this.f0, this.f1, this.f2, this.f3, this.f4, this.f5, this.f6, this.f7, this.f8, this.f9, thi...
3.26
flink_ReduceDriver_prepare_rdh
// -------------------------------------------------------------------------------------------- @Override public void prepare() throws Exception { TaskConfig config = this.taskContext.getTaskConfig(); if (config.getDriverStrategy() != DriverStrategy.SORTED_REDUCE) { throw new Exception("Unrecognized dri...
3.26
flink_ReduceDriver_setup_rdh
// ------------------------------------------------------------------------ @Override public void setup(TaskContext<ReduceFunction<T>, T> context) { this.taskContext = context; this.running = true; }
3.26
flink_OggJsonDecodingFormat_listReadableMetadata_rdh
// -------------------------------------------------------------------------------------------- // Metadata handling // -------------------------------------------------------------------------------------------- @Override public Map<String, DataType> listReadableMetadata() { final Map<String, DataType> metadataMa...
3.26
flink_SharedStateRegistry_registerReference_rdh
/** * Shortcut for {@link #registerReference(SharedStateRegistryKey, StreamStateHandle, long, * boolean)} with preventDiscardingCreatedCheckpoint = false. */ default StreamStateHandle registerReference(SharedStateRegistryKey registrationKey, StreamStateHandle state, long checkpointID) { return registerReference(...
3.26
flink_KeyGroupStatePartitionStreamProvider_getKeyGroupId_rdh
/** * Returns the key group that corresponds to the data in the provided stream. */ public int getKeyGroupId() { return keyGroupId; }
3.26
flink_SegmentPartitionFileWriter_flush_rdh
/** * This method is only called by the flushing thread. */ private void flush(TieredStoragePartitionId partitionId, int subpartitionId, int segmentId, List<Tuple2<Buffer, Integer>> buffersToFlush) { try { writeBuffers(partitionId, subpartitionId, segmentId, buffersToFlush, getTotalBytes(buffersToFlush)...
3.26
flink_SegmentPartitionFileWriter_flushOrFinishSegment_rdh
// ------------------------------------------------------------------------ // Internal Methods // ------------------------------------------------------------------------ private void flushOrFinishSegment(TieredStoragePartitionId partitionId, int subpartitionId, SegmentBufferContext segmentBufferConte...
3.26
flink_SegmentPartitionFileWriter_writeSegmentFinishFile_rdh
/** * Writing a segment-finish file when the current segment is complete. The downstream can * determine if the current segment is complete by checking for the existence of the * segment-finish file. * * <p>Note that the method is only called by the flushing thread. */ private void writeSegmentFinishFile(Tiered...
3.26
flink_PythonEnvironmentManagerUtils_pipInstallRequirements_rdh
/** * Installs the 3rd party libraries listed in the user-provided requirements file. An optional * requirements cached directory can be provided to support offline installation. In order not * to populate the public environment, the libraries will be installed to the specified * directory, and added to the PYTHONP...
3.26
flink_FailoverStrategyFactoryLoader_loadFailoverStrategyFactory_rdh
/** * Loads a {@link FailoverStrategy.Factory} from the given configuration. * * @param config * which specifies the failover strategy factory to load * @return failover strategy factory loaded */public static Factory loadFailoverStrategyFactory(final Configuration config) { checkNotNull(config); final...
3.26
flink_BinaryType_ofEmptyLiteral_rdh
/** * The SQL standard defines that character string literals are allowed to be zero-length strings * (i.e., to contain no characters) even though it is not permitted to declare a type that is * zero. For consistent behavior, the same logic applies to binary strings. * * <p>This method enables this special kind of...
3.26
flink_DelegatingConfiguration_read_rdh
// -------------------------------------------------------------------------------------------- @Override public void read(DataInputView in) throws IOException { this.prefix = in.readUTF(); this.backingConfig.read(in); }
3.26
flink_DelegatingConfiguration_getString_rdh
// -------------------------------------------------------------------------------------------- @Override public String getString(String key, String defaultValue) { return this.backingConfig.getString(this.prefix + key, defaultValue); }
3.26
flink_DelegatingConfiguration_hashCode_rdh
// -------------------------------------------------------------------------------------------- @Override public int hashCode() {return this.prefix.hashCode() ^ this.backingConfig.hashCode(); }
3.26
flink_DelegatingConfiguration_prefixOption_rdh
// -------------------------------------------------------------------------------------------- private static <T> ConfigOption<T> prefixOption(ConfigOption<T> option, String prefix) { String key = prefix + option.key();List<FallbackKey> v11; ...
3.26
flink_SchedulerNG_updateJobResourceRequirements_rdh
/** * Update {@link JobResourceRequirements job resource requirements}. * * @param jobResourceRequirements * new resource requirements */ default void updateJobResourceRequirements(JobResourceRequirements jobResourceRequirements) { throw new UnsupportedOperationException(String.format("The %s does not su...
3.26
flink_SchedulerNG_requestJobResourceRequirements_rdh
/** * Read current {@link JobResourceRequirements job resource requirements}. * * @return Current resource requirements. */ default JobResourceRequirements requestJobResourceRequirements() { throw new UnsupportedOperationException(String.format("The %s does not support changing the parallelism without a job re...
3.26
flink_StateConfigUtil_createTtlConfig_rdh
/** * Creates a {@link StateTtlConfig} depends on retentionTime parameter. * * @param retentionTime * State ttl time which unit is MILLISECONDS. */ public static StateTtlConfig createTtlConfig(long retentionTime) { if (retentionTime > 0) { return StateTtlConfig.newBuilder(Time.milliseconds(retentionT...
3.26
flink_JoinDriver_setup_rdh
// ------------------------------------------------------------------------ @Override public void setup(TaskContext<FlatJoinFunction<IT1, IT2, OT>, OT> context) { this.taskContext = context; this.running = true; }
3.26
flink_FileRecordFormat_getCheckpointedPosition_rdh
/** * Optionally returns the current position of the reader. This can be implemented by readers * that want to speed up recovery from a checkpoint. * * <p>The current position of the reader is the position of the next record that will be * returned in a call to {@link #read()}. This can be implemented by readers t...
3.26
flink_MemoryBackendCheckpointStorageAccess_supportsHighlyAvailableStorage_rdh
// ------------------------------------------------------------------------ // Checkpoint Storage // ------------------------------------------------------------------------ @Override public boolean supportsHighlyAvailableStorage() { return checkpointsDirectory != null; }
3.26
flink_MemoryBackendCheckpointStorageAccess_toString_rdh
// ------------------------------------------------------------------------ // Utilities // ------------------------------------------------------------------------ @Override public String toString() { return (((((("MemoryBackendCheckpointStorage {" + "checkpointsDirectory=") + checkpointsDirectory) + ", fileSys...
3.26
flink_WebLogDataGenerator_genDocs_rdh
/** * Generates the files for the documents relation. The entries apply the following format: <br> * <code>URL | Content</code> * * @param noDocs * Number of entries for the documents relation * @param filterKeyWords * A list of keywords that should be contained * @param words * A list of words to fill t...
3.26
flink_WebLogDataGenerator_main_rdh
/** * Main method to generate data for the {@link WebLogAnalysis} example program. * * <p>The generator creates to files: * * <ul> * <li><code>{tmp.dir}/documents</code> for the web documents * <li><code>{tmp.dir}/ranks</code> for the ranks of the web documents * <li><code>{tmp.dir}/visits</code> for the ...
3.26
flink_WebLogDataGenerator_m0_rdh
/** * Generates the files for the visits relation. The visits entries apply the following format: * <br> * <code>IP Address | URL | Date (YYYY-MM-DD) | Misc. Data (e.g. User-Agent) |\n</code> * * @param noVisits * Number of entries for the visits relation * @param noDocs * Number of entries in the documents...
3.26
flink_WebLogDataGenerator_genRanks_rdh
/** * Generates the files for the ranks relation. The ranks entries apply the following format: * <br> * <code>Rank | URL | Average Duration |\n</code> * * @param noDocs * Number of entries in the documents relation * @param path * Output path for the ranks relation ...
3.26
flink_MemoryTierProducerAgent_releaseResources_rdh
// ------------------------------------------------------------------------ // Internal Methods // ------------------------------------------------------------------------ private void releaseResources() { Arrays.stream(subpartitionProducerAgents).forEach(MemoryTierSubpartitionProducerAgent::release); }
3.26