name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_ResultPartitionType_isHybridResultPartition_rdh | /**
* {@link #isHybridResultPartition()} is used to judge whether it is the specified {@link #HYBRID_FULL} or {@link #HYBRID_SELECTIVE} resultPartitionType.
*
* <p>this method suitable for judgment conditions related to the specific implementation of
* {@link ResultPartitionType}.
*
* <p>this method not related t... | 3.26 |
flink_ResultPartitionType_isBounded_rdh | /**
* Whether this partition uses a limited number of (network) buffers or not.
*
* @return <tt>true</tt> if the number of buffers should be bound to some limit
*/
public boolean isBounded() {
return isBounded;
} | 3.26 |
flink_ResultPartitionType_isPipelinedOrPipelinedBoundedResultPartition_rdh | /**
* {@link #isPipelinedOrPipelinedBoundedResultPartition()} is used to judge whether it is the
* specified {@link #PIPELINED} or {@link #PIPELINED_BOUNDED} resultPartitionType.
*
* <p>This method suitable for judgment conditions related to the specific implementation of
* {@link ResultPartitionType}.
*
* <p>Th... | 3.26 |
flink_ResultPartitionType_isBlockingOrBlockingPersistentResultPartition_rdh | /**
* {@link #isBlockingOrBlockingPersistentResultPartition()} is used to judge whether it is the
* specified {@link #BLOCKING} or {@link #BLOCKING_PERSISTENT} resultPartitionType.
*
* <p>this method suitable for judgment conditions related to the specific implementation of
* {@link ResultPartitionType}.
*
* <p>... | 3.26 |
flink_ResultPartitionType_mustBePipelinedConsumed_rdh | /**
* return if this partition's upstream and downstream must be scheduled in the same time.
*/
public boolean mustBePipelinedConsumed() {
return f1 == ConsumingConstraint.MUST_BE_PIPELINED;} | 3.26 |
flink_ExternalPythonKeyedCoProcessOperator_processTimer_rdh | /**
* It is responsible to send timer data to python worker when a registered timer is fired. The
* input data is a Row containing 4 fields: TimerFlag 0 for proc time, 1 for event time;
* Timestamp of the fired timer; Current watermark and the key of the timer.
*
* @param timeDomain
* The type of the timer.
* ... | 3.26 |
flink_ExternalPythonKeyedCoProcessOperator_setCurrentKey_rdh | /**
* As the beam state gRPC service will access the KeyedStateBackend in parallel with this
* operator, we must override this method to prevent changing the current key of the
* KeyedStateBackend while the beam service is handling requests.
*/
@Override
public void setCurrentKey(Object key) {
if (inBatchExecut... | 3.26 |
flink_GSFileSystemOptions_getWriterTemporaryBucketName_rdh | /**
* The temporary bucket name to use for recoverable writes, if different from the final bucket
* name.
*/public Optional<String> getWriterTemporaryBucketName() {
return flinkConfig.getOptional(WRITER_TEMPORARY_BUCKET_NAME);
} | 3.26 |
flink_GSFileSystemOptions_m0_rdh | /**
* The chunk size to use for writes on the underlying Google WriteChannel.
*/
public Optional<MemorySize> m0() {
return flinkConfig.getOptional(WRITER_CHUNK_SIZE);
} | 3.26 |
flink_TableSource_explainSource_rdh | /**
* Describes the table source.
*
* @return A String explaining the {@link TableSource}.
*/default String explainSource() { return TableConnectorUtils.generateRuntimeName(getClass(), getTableSchema().getFieldNames());
} | 3.26 |
flink_TableSource_getProducedDataType_rdh | /**
* Returns the {@link DataType} for the produced data of the {@link TableSource}.
*
* @return The data type of the returned {@code DataStream}.
*/
default DataType getProducedDataType() {
final TypeInformation<T> legacyType = getReturnType();
if (legacyType == null) {
throw new TableException(... | 3.26 |
flink_RichSqlInsert_isUpsert_rdh | // ~ Tools ------------------------------------------------------------------
public static boolean isUpsert(List<SqlLiteral> keywords) {
for (SqlNode keyword : keywords) {
SqlInsertKeyword keyword2 = ((SqlLiteral) (keyword)).symbolValue(SqlInsertKeyword.class);
if (keyword2 == SqlInsertKeyword.UP... | 3.26 |
flink_RichSqlInsert_isOverwrite_rdh | /**
* Returns whether the insert mode is overwrite (for whole table or for specific partitions).
*
* @return true if this is overwrite mode
*/
public boolean isOverwrite() {
return getModifierNode(RichSqlInsertKeyword.OVERWRITE) != null;
} | 3.26 |
flink_RichSqlInsert_getTableHints_rdh | /**
* Returns the table hints as list of {@code SqlNode} for current insert node.
*/
public SqlNodeList getTableHints() {
return this.tableHints;
} | 3.26 |
flink_TupleComparator_hash_rdh | // --------------------------------------------------------------------------------------------
// Comparator Methods
// --------------------------------------------------------------------------------------------
@SuppressWarnings("unchecked")
@Override public int hash(T value) {
int i = 0;
try {
int ... | 3.26 |
flink_SourceProvider_of_rdh | /**
* Helper method for creating a Source provider with a provided source parallelism.
*/
static SourceProvider of(Source<RowData, ?, ?> source, @Nullable
Integer sourceParallelism) {
return new SourceProvider() {
@Override
public Source<RowData, ?, ?> createSource() {
return source;
... | 3.26 |
flink_RowSerializer_getPositionByName_rdh | // --------------------------------------------------------------------------------------------
private int getPositionByName(String fieldName) {
assert positionByName != null;
final Integer targetPos = positionByName.get(fieldName);
if (targetPos == null) {
throw new RuntimeException(String.format("Unknown field name ... | 3.26 |
flink_RowSerializer_snapshotConfiguration_rdh | // --------------------------------------------------------------------------------------------
// Serializer configuration snapshoting & compatibility
// --------------------------------------------------------------------------------------------
@Override
public TypeSerializerSnapshot<Row> snapshotConfiguration() {
r... | 3.26 |
flink_RowSerializer_fillMask_rdh | // --------------------------------------------------------------------------------------------
// Serialization utilities
// --------------------------------------------------------------------------------------------
private static void fillMask(int fieldLength, Row row, boolean[] mask, boolean supportsRowKind, int... | 3.26 |
flink_AbstractStreamTableEnvironmentImpl_execEnv_rdh | /**
* This is a temporary workaround for Python API. Python API should not use
* StreamExecutionEnvironment at all.
*/
@Internalpublic StreamExecutionEnvironment execEnv() {
return executionEnvironment;
} | 3.26 |
flink_AbstractID_longToByteArray_rdh | /**
* Converts a long to a byte array.
*
* @param l
* the long variable to be converted
* @param ba
* the byte array to store the result the of the conversion
* @param offset
* offset indicating at what position inside the byte array the result of the
* conversion shall be stored
*/
private static voi... | 3.26 |
flink_AbstractID_getBytes_rdh | /**
* Gets the bytes underlying this ID.
*
* @return The bytes underlying this ID.
*/public byte[] getBytes() {
byte[] bytes = new byte[SIZE];
longToByteArray(lowerPart, bytes, 0);
longToByteArray(upperPart, bytes, SIZE_OF_LONG);
return bytes;
} | 3.26 |
flink_AbstractID_toHexString_rdh | /**
* Returns pure String representation of the ID in hexadecimal. This method should be used to
* construct things like paths etc., that require a stable representation and is therefore
* final.
*/
public final String toHexString() {
if (this.hexString == null) {
final byte[] ba = new byte[SIZE];
... | 3.26 |
flink_AbstractID_equals_rdh | // --------------------------------------------------------------------------------------------
// Standard Utilities
// --------------------------------------------------------------------------------------------
@Override
public boolean equals(Object obj) {
if (obj
== this) {
return true;
} else ... | 3.26 |
flink_AbstractID_getLowerPart_rdh | // --------------------------------------------------------------------------------------------
/**
* Gets the lower 64 bits of the ID.
*
* @return The lower 64 bits of the ID.
*/
public long getLowerPart() {
return lowerPart;
} | 3.26 |
flink_AbstractID_byteArrayToLong_rdh | // Conversion Utilities
// --------------------------------------------------------------------------------------------
/**
* Converts the given byte array to a long.
*
* @param ba
* the byte array to be converted
* @param offset
* the offset indicating at which byte inside the array the conversion shall begi... | 3.26 |
flink_AvroUtils_getAvroUtils_rdh | /**
* Returns either the default {@link AvroUtils} which throw an exception in cases where Avro
* would be needed or loads the specific utils for Avro from flink-avro.
*/
public static AvroUtils getAvroUtils() {
// try and load the special AvroUtils from the flink-avro package
try {
Class<?> clazz ... | 3.26 |
flink_PlannerCallProcedureOperation_toExternal_rdh | /**
* Convert the value with internal representation to the value with external representation.
*/private Object toExternal(Object internalValue, DataType inputType, ClassLoader classLoader) {
if (!DataTypeUtils.isInternal(inputType)) {
// if the expected input type of the procedure is... | 3.26 |
flink_PlannerCallProcedureOperation_procedureResultToTableResult_rdh | /**
* Convert the result of procedure to table result .
*/
private TableResultInternal procedureResultToTableResult(Object procedureResult, TableConfig tableConfig, ClassLoader userClassLoader) {
// get result converter
ZoneId zoneId = tableConfig.getLocalTimeZone();
DataType v21 = outputType;
// if is not compos... | 3.26 |
flink_WordCount_main_rdh | // *************************************************************************
// PROGRAM
// *************************************************************************
public static void main(String[] args) throws Exception {
LOGGER.warn(DATASET_DEPRECATION_INFO);
final MultipleParameterToo... | 3.26 |
flink_CoProcessFunction_onTimer_rdh | /**
* Called when a timer set using {@link TimerService} fires.
*
* @param timestamp
* The timestamp of the firing timer.
* @param ctx
* An {@link OnTimerContext} that allows querying the timestamp of the firing timer,
* querying the {@link TimeDomain} of the firing timer and getting a {@link TimerService}... | 3.26 |
flink_ListAggWithRetractAggFunction_getArgumentDataTypes_rdh | // --------------------------------------------------------------------------------------------
// Planning
// --------------------------------------------------------------------------------------------
@Override
public List<DataType>
getArgumentDataTypes() {
return Collections.singletonList(DataTypes.STRING().bri... | 3.26 |
flink_BatchShuffleReadBufferPool_requestBuffers_rdh | /**
* Requests a collection of buffers (determined by {@link #numBuffersPerRequest}) from this
* buffer pool.
*/
public List<MemorySegment> requestBuffers() throws Exception {
List<MemorySegment> allocated = new ArrayList... | 3.26 |
flink_BatchShuffleReadBufferPool_destroy_rdh | /**
* Destroys this buffer pool and after which, no buffer can be allocated any more.
*/
public void destroy() {
synchronized(buffers) {
destroyed
= true;
buffers.clear();
buffers.notifyAll();
}
} | 3.26 |
flink_BatchShuffleReadBufferPool_initialize_rdh | /**
* Initializes this buffer pool which allocates all the buffers.
*/
public void initialize() {
synchronized(buffers) {
checkState(!destroyed, "Buffer pool is already destroyed.");
if (initialized) {
return;
}
initialized
= true;
tr... | 3.26 |
flink_BatchShuffleReadBufferPool_recycle_rdh | /**
* Recycles a collection of buffers to this buffer pool. This method should never throw any
* exception.
*/
public void recycle(Collection<MemorySegment> segments) {
checkArgument(segments != null, "Buffer list must be not null.");
if (segments.isEmpty()) {
return;
}synchronized(buffers) {
... | 3.26 |
flink_JobGraph_getCheckpointingSettings_rdh | /**
* Gets the settings for asynchronous snapshots. This method returns null, when checkpointing is
* not enabled.
*
* @return The snapshot settings
*/
public JobCheckpointingSettings getCheckpointingSettings() {
return snapshotSettings;} | 3.26 |
flink_JobGraph_addVertex_rdh | /**
* Adds a new task vertex to the job graph if it is not already included.
*
* @param vertex
* the new task vertex to be added
*/
public void addVertex(JobVertex vertex) {
final JobVertexID id = vertex.getID();
JobVertex previous = taskVertices.put(id,
vertex);
// if we had a prior association, restore and ... | 3.26 |
flink_JobGraph_addUserJarBlobKey_rdh | /**
* Adds the BLOB referenced by the key to the JobGraph's dependencies.
*
* @param key
* path of the JAR file required to run the job on a task manager
*/
public void addUserJarBlobKey(PermanentBlobKey key) {
if (key == null) {
throw new IllegalArgumentException();
}
if (!userJarBlobKeys.co... | 3.26 |
flink_JobGraph_getJobConfiguration_rdh | /**
* Returns the configuration object for this job. Job-wide parameters should be set into that
* configuration object.
*
* @return The configuration object for this job.
*/
public Configuration getJobConfiguration()
{
return this.jobConfiguration;
} | 3.26 |
flink_JobGraph_setJobID_rdh | /**
* Sets the ID of the job.
*/
public void setJobID(JobID jobID) {
this.jobID = jobID;
} | 3.26 |
flink_JobGraph_setClasspaths_rdh | /**
* Sets the classpaths required to run the job on a task manager.
*
* @param paths
* paths of the directories/JAR files required to run the job on a task manager
*/
public void setClasspaths(List<URL> paths) {
classpaths = paths;
} | 3.26 |
flink_JobGraph_getVertices_rdh | /**
* Returns an Iterable to iterate all vertices registered with the job graph.
*
* @return an Iterable to iterate all vertices registered with the job graph
*/public Iterable<JobVertex> getVertices() {
return this.taskVertices.values();
} | 3.26 |
flink_JobGraph_getNumberOfVertices_rdh | /**
* Returns the number of all vertices.
*
* @return The number of all vertices.
*/
public int getNumberOfVertices() {return this.taskVertices.size();
} | 3.26 |
flink_JobGraph_setExecutionConfig_rdh | /**
* Sets the execution config. This method eagerly serialized the ExecutionConfig for future RPC
* transport. Further modification of the referenced ExecutionConfig object will not affect this
* serialized copy.
*
* @param executionConfig
* The ExecutionConfig to be serialized.
* @throws IOException
* Thr... | 3.26 |
flink_JobGraph_getMaximumParallelism_rdh | /**
* Gets the maximum parallelism of all operations in this job graph.
*
* @return The maximum parallelism of this job graph
*/public int getMaximumParallelism()
{
int maxParallelism = -1;
for (JobVertex v6 : taskVertices.values()) {
maxParallelism
= Math.max(v6.getParallelism(), maxParallelism); }
return ... | 3.26 |
flink_JobGraph_addJars_rdh | /**
* Adds the given jar files to the {@link JobGraph} via {@link JobGraph#addJar}.
*
* @param jarFilesToAttach
* a list of the {@link URL URLs} of the jar files to attach to the
* jobgraph.
* @throws RuntimeException
* if a jar URL is not valid.
*/
public void addJars(final List<URL> jarFilesToAttach) {
... | 3.26 |
flink_JobGraph_setSavepointRestoreSettings_rdh | /**
* Sets the savepoint restore settings.
*
* @param settings
* The savepoint restore settings.
*/
public void setSavepointRestoreSettings(SavepointRestoreSettings settings) {
this.savepointRestoreSettings = checkNotNull(settings, "Savepoint restore settings");
} | 3.26 |
flink_JobGraph_getVerticesSortedTopologicallyFromSources_rdh | // --------------------------------------------------------------------------------------------
public List<JobVertex> getVerticesSortedTopologicallyFromSources() throws InvalidProgramException {
// early out on empty lists
if (this.taskVertices.isEmpty()) {
return Collections.emptyList();
}
List<JobVertex> sorted... | 3.26 |
flink_JobGraph_addUserArtifact_rdh | /**
* Adds the path of a custom file required to run the job on a task manager.
*
* @param name
* a name under which this artifact will be accessible through {@link DistributedCache}
* @param file
* path of a custom file required to run the job on a task manager
*/
public void addUserArtifact(String name, D... | 3.26 |
flink_JobGraph_getJobID_rdh | // --------------------------------------------------------------------------------------------
/**
* Returns the ID of the job.
*
* @return the ID of the job
*/
public JobID getJobID() {
return this.jobID;
} | 3.26 |
flink_JobGraph_getSerializedExecutionConfig_rdh | /**
* Returns the {@link ExecutionConfig}.
*
* @return ExecutionConfig
*/
public SerializedValue<ExecutionConfig> getSerializedExecutionConfig() {
return serializedExecutionConfig;} | 3.26 |
flink_JobGraph_addJar_rdh | // --------------------------------------------------------------------------------------------
// Handling of attached JAR files
// --------------------------------------------------------------------------------------------
/**
* Adds the path of a JAR file required to run the job on a task manager.
*
* @param jar... | 3.26 |
flink_JobGraph_m1_rdh | /**
* Returns an array of all job vertices that are registered with the job graph. The order in
* which the vertices appear in the list is not defined.
*
* @return an array of all job vertices that are registered with the job graph
*/
public JobVertex[] m1() {
return this.taskVertices.values().toArray(new JobVerte... | 3.26 |
flink_JobGraph_getUserArtifacts_rdh | /**
* Gets the list of assigned user jar paths.
*
* @return The list of assigned user jar paths
*/
public Map<String, DistributedCache.DistributedCacheEntry> getUserArtifacts()
{
return userArtifacts;
} | 3.26 |
flink_JobGraph_getSavepointRestoreSettings_rdh | /**
* Returns the configured savepoint restore setting.
*
* @return The configured savepoint restore settings.
*/
public SavepointRestoreSettings getSavepointRestoreSettings() {
return savepointRestoreSettings;
} | 3.26 |
flink_JobGraph_getCoLocationGroups_rdh | /**
* Returns all {@link CoLocationGroup} instances associated with this {@code JobGraph}.
*
* @return The associated {@code CoLocationGroup} instances.
*/
public Set<CoLocationGroup> getCoLocationGroups() {
final Set<CoLocationGroup> coLocationGroups = IterableUtils.toStream(getVertices()).map(JobVertex::getCoLo... | 3.26 |
flink_JobGraph_isCheckpointingEnabled_rdh | /**
* Checks if the checkpointing was enabled for this job graph.
*
* @return true if checkpointing enabled
*/
public boolean isCheckpointingEnabled() {
if (snapshotSettings == null) {
return false;
}
return snapshotSettings.getCheckpointCoordinatorConfiguration().isCheckpointingEnabled();
} | 3.26 |
flink_ApplicationDispatcherBootstrap_runApplicationEntryPoint_rdh | /**
* Runs the user program entrypoint and completes the given {@code jobIdsFuture} with the {@link JobID JobIDs} of the submitted jobs.
*
* <p>This should be executed in a separate thread (or task).
*/
private void runApplicationEntryPoint(final CompletableFuture<List<JobID>> jobIdsFuture,
final Set<JobID> tolerat... | 3.26 |
flink_ApplicationDispatcherBootstrap_m0_rdh | /**
* Runs the user program entrypoint by scheduling a task on the given {@code scheduledExecutor}.
* The returned {@link CompletableFuture} completes when all jobs of the user application
* succeeded. if any of them fails, or if job submission fails.
*/
private CompletableFuture<Void> m0(final DispatcherGateway di... | 3.26 |
flink_ApplicationDispatcherBootstrap_finishBootstrapTasks_rdh | /**
* Logs final application status and invokes error handler in case of unexpected failures.
* Optionally shuts down the given dispatcherGateway when the application completes (either
* successfully or in case of failure), depending on the corresponding config option.
*/
private Completable... | 3.26 |
flink_ApplicationDispatcherBootstrap_unwrapJobResultException_rdh | /**
* If the given {@link JobResult} indicates success, this passes through the {@link JobResult}.
* Otherwise, this returns a future that is finished exceptionally (potentially with an
* exception from the {@link JobResult}).
*/
private CompletableFuture<JobResult> unwrapJobResultException(final CompletableFuture<... | 3.26 |
flink_HsSubpartitionConsumerMemoryDataManager_addBuffer_rdh | // this method only called from subpartitionMemoryDataManager with write lock.
@GuardedBy("consumerLock")
public boolean addBuffer(HsBufferContext bufferContext) {
tryIncreaseBacklog(bufferContext.getBuffer());
unConsumedBuffers.add(bufferContext);
trimHeadingReleasedBuffers();
return unCons... | 3.26 |
flink_HsSubpartitionConsumerMemoryDataManager_getBacklog_rdh | // Un-synchronized get the backlog to provide memory data backlog, this will make the
// result greater than or equal to the actual backlog, but obtaining an accurate backlog will
// bring too much extra overhead.
@SuppressWarnings("FieldAccessNotGuarded")
@Override
public int getBacklog() {return backlog;
} | 3.26 |
flink_HsSubpartitionConsumerMemoryDataManager_addInitialBuffers_rdh | // this method only called from subpartitionMemoryDataManager with write lock.
@GuardedBy("consumerLock")
public void addInitialBuffers(Deque<HsBufferContext> buffers) {
for (HsBufferContext bufferContext : buffers) {
tryIncreaseBacklog(bufferContext.getBuffer());
}
unConsumedBuffers.addAll(buffers... | 3.26 |
flink_ServiceType_classify_rdh | // Helper method
public static ServiceExposedType classify(Service service) {
KubernetesConfigOptions.ServiceExposedType type =
KubernetesConfigOptions.ServiceExposedType.valueOf(service.getSpec().getType());
if (type == ServiceExposedType.ClusterIP) {
if (HeadlessClusterIPService.HEADLESS_CLUSTER_... | 3.26 |
flink_ServiceType_buildUpExternalRestService_rdh | /**
* Build up the external rest service template, according to the jobManager parameters.
*
* @param kubernetesJobManagerParameters
* the parameters of jobManager.
* @return the external rest service
*/
public Service buildUpExternalRestService(KubernetesJobManagerParameters kubernetesJobManagerParameters) {
... | 3.26 |
flink_ServiceType_getRestPortFromExternalService_rdh | /**
* Get rest port from the external Service.
*/
public int getRestPortFromExternalService(Service externalService) {
final List<ServicePort> servicePortCandidates = externalService.getSpec().getPorts().stream().filter(x -> x.getName().equals(Constants.REST_PORT_NAME)).collect(Collectors.toList());
if (servi... | 3.26 |
flink_Tuple12_equals_rdh | /**
* Deep equality for tuples by calling equals() on the tuple members.
*
* @param o
* the object checked for equality
* @return true if this is equal to o.
*/
@Override
public boolean equals(Object o) {
if (this == o) {
return true;
}
if (!(o instanceof Tuple12)) {
return false;
... | 3.26 |
flink_Tuple12_m1_rdh | /**
* Sets new values to all fields of the tuple.
*
* @param f0
* The value for field 0
* @param f1
* The value for field 1
* @param f2
* The value for field 2
* @param f3
* The value for field 3
* @param f4
* The value for field 4
* @param f5
* The value for field 5
* @param f6
* The valu... | 3.26 |
flink_Tuple12_toString_rdh | // -------------------------------------------------------------------------------------------------
// standard utilities
// -------------------------------------------------------------------------------------------------
/**
* Creates a string representation of the tuple in the form (f0, f1, f2, f3, f4, f5, f6, f7,... | 3.26 |
flink_Tuple12_of_rdh | /**
* Creates a new tuple and assigns the given values to the tuple's fields. This is more
* convenient than using the constructor, because the compiler can infer the generic type
* arguments implicitly. For example: {@code Tuple3.of(n, x, s)} instead of {@code new
* Tuple3<Integer, Double, String>(n, x, s)}
*/
pu... | 3.26 |
flink_Tuple12_copy_rdh | /**
* Shallow tuple copy.
*
* @return A new Tuple with the same fields as this.
*/
@Override
@SuppressWarnings("unchecked")
public Tuple12<T0, T1, T2, T3, T4, T5, T6, T7,
T8, T9, T10, T11> copy() {return new
Tuple12<>(this.f0, this.f1, this.f2, this.f3, this.f4, this.f5, this.f6, this.f7, this.f8, this.f9, thi... | 3.26 |
flink_ReduceDriver_prepare_rdh | // --------------------------------------------------------------------------------------------
@Override
public void prepare() throws Exception {
TaskConfig config = this.taskContext.getTaskConfig();
if (config.getDriverStrategy() != DriverStrategy.SORTED_REDUCE) {
throw new Exception("Unrecognized dri... | 3.26 |
flink_ReduceDriver_setup_rdh | // ------------------------------------------------------------------------
@Override
public void setup(TaskContext<ReduceFunction<T>, T> context) {
this.taskContext = context;
this.running = true;
} | 3.26 |
flink_OggJsonDecodingFormat_listReadableMetadata_rdh | // --------------------------------------------------------------------------------------------
// Metadata handling
// --------------------------------------------------------------------------------------------
@Override
public Map<String, DataType> listReadableMetadata() {
final Map<String, DataType> metadataMa... | 3.26 |
flink_SharedStateRegistry_registerReference_rdh | /**
* Shortcut for {@link #registerReference(SharedStateRegistryKey, StreamStateHandle, long,
* boolean)} with preventDiscardingCreatedCheckpoint = false.
*/
default StreamStateHandle registerReference(SharedStateRegistryKey registrationKey, StreamStateHandle state, long checkpointID) {
return registerReference(... | 3.26 |
flink_KeyGroupStatePartitionStreamProvider_getKeyGroupId_rdh | /**
* Returns the key group that corresponds to the data in the provided stream.
*/
public int getKeyGroupId() {
return keyGroupId;
} | 3.26 |
flink_SegmentPartitionFileWriter_flush_rdh | /**
* This method is only called by the flushing thread.
*/
private void flush(TieredStoragePartitionId partitionId, int subpartitionId, int segmentId, List<Tuple2<Buffer, Integer>> buffersToFlush) {
try {
writeBuffers(partitionId, subpartitionId, segmentId, buffersToFlush, getTotalBytes(buffersToFlush)... | 3.26 |
flink_SegmentPartitionFileWriter_flushOrFinishSegment_rdh | // ------------------------------------------------------------------------
// Internal Methods
// ------------------------------------------------------------------------
private void flushOrFinishSegment(TieredStoragePartitionId partitionId, int subpartitionId, SegmentBufferContext segmentBufferConte... | 3.26 |
flink_SegmentPartitionFileWriter_writeSegmentFinishFile_rdh | /**
* Writing a segment-finish file when the current segment is complete. The downstream can
* determine if the current segment is complete by checking for the existence of the
* segment-finish file.
*
* <p>Note that the method is only called by the flushing thread.
*/
private void writeSegmentFinishFile(Tiered... | 3.26 |
flink_PythonEnvironmentManagerUtils_pipInstallRequirements_rdh | /**
* Installs the 3rd party libraries listed in the user-provided requirements file. An optional
* requirements cached directory can be provided to support offline installation. In order not
* to populate the public environment, the libraries will be installed to the specified
* directory, and added to the PYTHONP... | 3.26 |
flink_FailoverStrategyFactoryLoader_loadFailoverStrategyFactory_rdh | /**
* Loads a {@link FailoverStrategy.Factory} from the given configuration.
*
* @param config
* which specifies the failover strategy factory to load
* @return failover strategy factory loaded
*/public static Factory loadFailoverStrategyFactory(final Configuration config) {
checkNotNull(config);
final... | 3.26 |
flink_BinaryType_ofEmptyLiteral_rdh | /**
* The SQL standard defines that character string literals are allowed to be zero-length strings
* (i.e., to contain no characters) even though it is not permitted to declare a type that is
* zero. For consistent behavior, the same logic applies to binary strings.
*
* <p>This method enables this special kind of... | 3.26 |
flink_DelegatingConfiguration_read_rdh | // --------------------------------------------------------------------------------------------
@Override
public void read(DataInputView in) throws IOException {
this.prefix = in.readUTF();
this.backingConfig.read(in);
} | 3.26 |
flink_DelegatingConfiguration_getString_rdh | // --------------------------------------------------------------------------------------------
@Override
public String getString(String key, String defaultValue) {
return this.backingConfig.getString(this.prefix + key, defaultValue);
} | 3.26 |
flink_DelegatingConfiguration_hashCode_rdh | // --------------------------------------------------------------------------------------------
@Override
public int hashCode() {return this.prefix.hashCode() ^ this.backingConfig.hashCode();
} | 3.26 |
flink_DelegatingConfiguration_prefixOption_rdh | // --------------------------------------------------------------------------------------------
private static <T> ConfigOption<T> prefixOption(ConfigOption<T> option, String prefix) {
String key = prefix + option.key();List<FallbackKey> v11;
... | 3.26 |
flink_SchedulerNG_updateJobResourceRequirements_rdh | /**
* Update {@link JobResourceRequirements job resource requirements}.
*
* @param jobResourceRequirements
* new resource requirements
*/
default void updateJobResourceRequirements(JobResourceRequirements jobResourceRequirements) {
throw new
UnsupportedOperationException(String.format("The %s does not su... | 3.26 |
flink_SchedulerNG_requestJobResourceRequirements_rdh | /**
* Read current {@link JobResourceRequirements job resource requirements}.
*
* @return Current resource requirements.
*/
default JobResourceRequirements requestJobResourceRequirements() {
throw new UnsupportedOperationException(String.format("The %s does not support changing the parallelism without a job re... | 3.26 |
flink_StateConfigUtil_createTtlConfig_rdh | /**
* Creates a {@link StateTtlConfig} depends on retentionTime parameter.
*
* @param retentionTime
* State ttl time which unit is MILLISECONDS.
*/
public static StateTtlConfig createTtlConfig(long retentionTime) {
if (retentionTime > 0) {
return StateTtlConfig.newBuilder(Time.milliseconds(retentionT... | 3.26 |
flink_JoinDriver_setup_rdh | // ------------------------------------------------------------------------
@Override
public void setup(TaskContext<FlatJoinFunction<IT1, IT2, OT>, OT> context) {
this.taskContext = context;
this.running = true;
} | 3.26 |
flink_FileRecordFormat_getCheckpointedPosition_rdh | /**
* Optionally returns the current position of the reader. This can be implemented by readers
* that want to speed up recovery from a checkpoint.
*
* <p>The current position of the reader is the position of the next record that will be
* returned in a call to {@link #read()}. This can be implemented by readers t... | 3.26 |
flink_MemoryBackendCheckpointStorageAccess_supportsHighlyAvailableStorage_rdh | // ------------------------------------------------------------------------
// Checkpoint Storage
// ------------------------------------------------------------------------
@Override
public boolean supportsHighlyAvailableStorage() {
return checkpointsDirectory != null;
} | 3.26 |
flink_MemoryBackendCheckpointStorageAccess_toString_rdh | // ------------------------------------------------------------------------
// Utilities
// ------------------------------------------------------------------------
@Override
public String toString() {
return (((((("MemoryBackendCheckpointStorage {" + "checkpointsDirectory=") + checkpointsDirectory) + ", fileSys... | 3.26 |
flink_WebLogDataGenerator_genDocs_rdh | /**
* Generates the files for the documents relation. The entries apply the following format: <br>
* <code>URL | Content</code>
*
* @param noDocs
* Number of entries for the documents relation
* @param filterKeyWords
* A list of keywords that should be contained
* @param words
* A list of words to fill t... | 3.26 |
flink_WebLogDataGenerator_main_rdh | /**
* Main method to generate data for the {@link WebLogAnalysis} example program.
*
* <p>The generator creates to files:
*
* <ul>
* <li><code>{tmp.dir}/documents</code> for the web documents
* <li><code>{tmp.dir}/ranks</code> for the ranks of the web documents
* <li><code>{tmp.dir}/visits</code> for the ... | 3.26 |
flink_WebLogDataGenerator_m0_rdh | /**
* Generates the files for the visits relation. The visits entries apply the following format:
* <br>
* <code>IP Address | URL | Date (YYYY-MM-DD) | Misc. Data (e.g. User-Agent) |\n</code>
*
* @param noVisits
* Number of entries for the visits relation
* @param noDocs
* Number of entries in the documents... | 3.26 |
flink_WebLogDataGenerator_genRanks_rdh | /**
* Generates the files for the ranks relation. The ranks entries apply the following format:
* <br>
* <code>Rank | URL | Average Duration |\n</code>
*
* @param noDocs
* Number of entries in the documents relation
* @param path
* Output path for the ranks relation
... | 3.26 |
flink_MemoryTierProducerAgent_releaseResources_rdh | // ------------------------------------------------------------------------
// Internal Methods
// ------------------------------------------------------------------------
private void releaseResources() {
Arrays.stream(subpartitionProducerAgents).forEach(MemoryTierSubpartitionProducerAgent::release);
} | 3.26 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.