name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_ZooKeeperStateHandleStore_hasLock_rdh | // ---------------------------------------------------------------------------------------------------------
// Protected methods
// ---------------------------------------------------------------------------------------------------------
/**
* Checks whether a lock is created for this instance on the passed ZooKeeper... | 3.26 |
flink_ZooKeeperStateHandleStore_writeStoreHandleTransactionally_rdh | // this method is provided for the sole purpose of easier testing
@VisibleForTesting
void writeStoreHandleTransactionally(String path, byte[] serializedStoreHandle) throws Exception {
// Write state handle (not the actual state) to ZooKeeper. This is expected to be smaller
// than the state itself. This level o... | 3.26 |
flink_FsStateBackend_getBasePath_rdh | // ------------------------------------------------------------------------
// Properties
// ------------------------------------------------------------------------
/**
* Gets the base directory where all the checkpoints are stored. The job-specific checkpoint
* directory is created inside this directory.
*
* @ret... | 3.26 |
flink_FsStateBackend_toString_rdh | // ------------------------------------------------------------------------
// utilities
// ------------------------------------------------------------------------
@Override
public String toString() {
return (((((("File State Backend (" + "checkpoints: '") + getCheckpointPath()) + "', savepoints: '") + getSavepoi... | 3.26 |
flink_FsStateBackend_createCheckpointStorage_rdh | // ------------------------------------------------------------------------
// initialization and cleanup
// ------------------------------------------------------------------------
@Override
public CheckpointStorageAccess createCheckpointStorage(JobID jobId) throws IOException {
checkNotNull(jobId, "jobId");
... | 3.26 |
flink_FsStateBackend_createKeyedStateBackend_rdh | // ------------------------------------------------------------------------
// state holding structures
// ------------------------------------------------------------------------
@Override
public <K> AbstractKeyedStateBackend<K> createKeyedStateBackend(Environment env, JobID jobID, String operatorIdentifier, TypeSeria... | 3.26 |
flink_FsStateBackend_getWriteBufferSize_rdh | /**
* Gets the write buffer size for created checkpoint stream.
*
* <p>If not explicitly configured, this is the default value of {@link CheckpointingOptions#FS_WRITE_BUFFER_SIZE}.
*
* @return The write buffer size, in bytes.
*/
public int getWriteBufferSize() {
return writeBufferSize >= 0 ? writeBufferSize... | 3.26 |
flink_FsStateBackend_m0_rdh | /**
* Gets whether the key/value data structures are asynchronously snapshotted, which is always
* true for this state backend.
*/
public boolean m0() {
return true;
} | 3.26 |
flink_FsStateBackend_getCheckpointPath_rdh | /**
* Gets the base directory where all the checkpoints are stored. The job-specific checkpoint
* directory is created inside this directory.
*
* @return The base directory for checkpoints.
*/
@Nonnull
@Override
public Path getCheckpointPath() {
// we know that this can never be null by the way of constructor ... | 3.26 |
flink_FsStateBackend_getMinFileSizeThreshold_rdh | /**
* Gets the threshold below which state is stored as part of the metadata, rather than in files.
* This threshold ensures that the backend does not create a large amount of very small files,
* where potentially the file pointers are larger than the state itself.
*
* <p>If not explicitly configured, this is the ... | 3.26 |
flink_FsStateBackend_configure_rdh | // ------------------------------------------------------------------------
// Reconfiguration
// ------------------------------------------------------------------------
/**
* Creates a copy of this state backend that uses the values defined in the configuration for
* fields where that were not specified in this sta... | 3.26 |
flink_CompiledPlan_printJsonString_rdh | /**
* Like {@link #asJsonString()}, but prints the result to {@link System#out}.
*/
default CompiledPlan printJsonString() {
System.out.println(this.asJsonString());
return this;
} | 3.26 |
flink_CompiledPlan_writeToFile_rdh | /**
* Writes this plan to a file using the JSON representation. This operation will fail if the
* file already exists, even if the content is different from this plan.
*
* @param file
* the target file
* @throws TableException
* if the file cannot be written.
*/
default void writeToFile(File file) {
wr... | 3.26 |
flink_ThreadInfoSample_m0_rdh | /**
* Constructs a collection of {@link ThreadInfoSample}s from a collection of {@link ThreadInfo}
* samples.
*
* @param threadInfos
* the collection of {@link ThreadInfo}.
* @return the collection of the corresponding {@link ThreadInfoSample}s.
*/
public static Map<Long, ThreadInfoSample> m0(Collection<Thread... | 3.26 |
flink_RecordWriter_close_rdh | /**
* This is used to broadcast streaming Watermarks in-band with records.
*/public abstract void broadcastEmit(T record) throws IOException;
/**
* Closes the writer. This stops the flushing thread (if there is one).
*/
public void close() {
// make sure we terminate the thread in any case
if (outputFlush... | 3.26 |
flink_RecordWriter_notifyFlusherException_rdh | /**
* Notifies the writer that the output flusher thread encountered an exception.
*
* @param t
* The exception to report.
*/
private void notifyFlusherException(Throwable t) {
if (flusherException == null) {
LOG.error("An exception happened while flushing the outputs", t);
flusherException ... | 3.26 |
flink_RecordWriter_setMetricGroup_rdh | /**
* Sets the metric group for this RecordWriter.
*/
public void setMetricGroup(TaskIOMetricGroup metrics) {
targetPartition.setMetricGroup(metrics);
} | 3.26 |
flink_RecordWriter_m2_rdh | /**
* This is used to send LatencyMarks to a random target channel.
*/
public void
m2(T record) throws IOException {
checkErroneous();
int targetSubpartition = rng.nextInt(numberOfChannels);
emit(record, targetSubpartition);
} | 3.26 |
flink_ThrowingRunnable_unchecked_rdh | /**
* Converts a {@link ThrowingRunnable} into a {@link Runnable} which throws all checked
* exceptions as unchecked.
*
* @param throwingRunnable
* to convert into a {@link Runnable}
* @return {@link Runnable} which throws all checked exceptions as unchecked.
*/static Runnable unchecked(ThrowingRunnable<?> th... | 3.26 |
flink_RocksDBPriorityQueueConfig_getRocksDBPriorityQueueSetCacheSize_rdh | /**
* Gets the cache size of rocksDB priority queue set. It will fall back to the default value if
* it is not explicitly set.
*/
public int getRocksDBPriorityQueueSetCacheSize() {
return rocksDBPriorityQueueSetCacheSize == UNDEFINED_ROCKSDB_PRIORITY_QUEUE_SET_CACHE_SIZE ? ROCKSDB_TIMER_SERVICE_FACTORY_CACHE_SIZ... | 3.26 |
flink_RocksDBPriorityQueueConfig_getPriorityQueueStateType_rdh | /**
* Gets the type of the priority queue state. It will fall back to the default value if it is
* not explicitly set.
*/
public PriorityQueueStateType getPriorityQueueStateType() {
return priorityQueueStateType ==
null ? TIMER_SERVICE_FACTORY.defaultValue() : priorityQueueStateType;
} | 3.26 |
flink_ProjectedRowData_from_rdh | /**
* Create an empty {@link ProjectedRowData} starting from a {@link Projection}.
*
* <p>Throws {@link IllegalStateException} if the provided {@code projection} array contains
* nested projections, which are not supported by {@link ProjectedRowData}.
*
* @see Projection
* @see ProjectedRowData
*/
public static... | 3.26 |
flink_ProjectedRowData_getArity_rdh | // ---------------------------------------------------------------------------------------------
@Override
public int getArity() {
return indexMapping.length;
} | 3.26 |
flink_WebMonitorUtils_loadWebSubmissionExtension_rdh | /**
* Loads the {@link WebMonitorExtension} which enables web submission.
*
* @param leaderRetriever
* to retrieve the leader
* @param timeout
* for asynchronous requests
* @param responseHeaders
* for the web submission handlers
* @param localAddressFuture
* of the underlying REST server endpoint
* ... | 3.26 |
flink_WebMonitorUtils_find_rdh | /**
* Finds the Flink log directory using log.file Java property that is set during startup.
*/
public static LogFileLocation find(Configuration config) {
final String logEnv = "log.file";
String logFilePath = System.getProperty(logEnv); if (logFilePath == nul... | 3.26 |
flink_WebMonitorUtils_resolveFileLocation_rdh | /**
* Verify log file location.
*
* @param logFilePath
* Path to log file
* @return File or null if not a valid log file
*/
private static File resolveFileLocation(String logFilePath) {
File logFile = new File(logFilePath);
return logFile.exists() && logFile.canRead() ? logFile : null;
} | 3.26 |
flink_WebMonitorUtils_isFlinkRuntimeWebInClassPath_rdh | /**
* Returns {@code true} if the optional dependency {@code flink-runtime-web} is in the
* classpath.
*/
private static boolean isFlinkRuntimeWebInClassPath() {
try {
Class.forName(WEB_FRONTEND_BOOTSTRAP_CLASS_FQN); return true;
} catch (ClassNotFoundException e) {
// class not found means that there is no ... | 3.26 |
flink_MemoryMappedBoundedData_finishWrite_rdh | /**
* Finishes the current region and prevents further writes. After calling this method, further
* calls to {@link #writeBuffer(Buffer)} will fail.
*/
@Override
public void finishWrite() throws IOException {
assert currentBuffer != null;
currentBuffer.flip();
fullBuffers.add(currentBuffer);
current... | 3.26 |
flink_MemoryMappedBoundedData_createWithRegionSize_rdh | /**
* Creates new MemoryMappedBoundedData, creating a memory mapped file at the given path. Each
* mapped region (= ByteBuffer) will be of the given size.
*/
public static MemoryMappedBoundedData createWithRegionSize(Path memMappedFilePath, int regionSize) throws IOException {
final FileChannel fileChannel = FileCh... | 3.26 |
flink_MemoryMappedBoundedData_create_rdh | // ------------------------------------------------------------------------
// Factories
// ------------------------------------------------------------------------
/**
* Creates new MemoryMappedBoundedData, creating a memory mapped file at the given path.
*/
public static MemoryMappedBoundedData create(Path
memMapp... | 3.26 |
flink_MemoryMappedBoundedData_alignSize_rdh | /**
* Rounds the size down to the next multiple of the {@link #PAGE_SIZE}. We need to round down
* here to not exceed the original maximum size value. Otherwise, values like INT_MAX would
* round up to overflow the valid maximum size of a memory mapping region in Java.
*/
private static int alignSize(int maxRegion... | 3.26 |
flink_MemoryMappedBoundedData_getSize_rdh | /**
* Gets the number of bytes of all written data (including the metadata in the buffer headers).
*/
@Override
public long getSize() {
long size = 0L;
for (ByteBuffer bb : fullBuffers) {
size += bb.remaining();
}
if (currentBuffer != null) {
size += currentBuffer.position();
}
... | 3.26 |
flink_MemoryMappedBoundedData_close_rdh | /**
* Unmaps the file from memory and deletes the file. After calling this method, access to any
* ByteBuffer obtained from this instance will cause a segmentation fault.
*/
public void close() throws IOException {
IOUtils.closeQuietly(file);// in case we dispose before finishing writes
for (ByteBuffer bb... | 3.26 |
flink_WindowedOperatorTransformation_reduce_rdh | /**
* Applies the given window function to each window. The window function is called for each
* evaluation of the window for each key individually. The output of the window function is
* interpreted as a regular non-windowed stream.
*
* <p>Arriving data is incrementally aggregated using the given reducer.
*
* @... | 3.26 |
flink_WindowedOperatorTransformation_process_rdh | /**
* Applies the given window function to each window. The window function is called for each
* evaluation of the window for each key individually. The output of the window function is
* interpreted as a regular non-windowed stream.
*
* <p>Note that this function requires that all data in the windows is buffered ... | 3.26 |
flink_WindowedOperatorTransformation_apply_rdh | /**
* Applies the given window function to each window. The window function is called for each
* evaluation of the window for each key individually. The output of the window function is
* interpreted as a regular non-windowed stream.
*
* <p>Note that this function requires that all data in the windows is buffered ... | 3.26 |
flink_WindowedOperatorTransformation_m0_rdh | /**
* Applies the given window function to each window. The window function is called for each
* evaluation of the window for each key individually. The output of the window function is
* interpreted as a regular non-windowed stream.
*
* <p>Arriving data is incrementally aggregated using the given aggregate functi... | 3.26 |
flink_WindowedOperatorTransformation_evictor_rdh | /**
* Sets the {@code Evictor} that should be used to evict elements from a window before emission.
*
* <p>Note: When using an evictor window performance will degrade significantly, since
* incremental aggregation of window results cannot be used.
*/
@PublicEvolving
public WindowedOperatorTransformation<T, K, W> ... | 3.26 |
flink_WindowedOperatorTransformation_aggregate_rdh | /**
* Applies the given window function to each window. The window function is called for each
* evaluation of the window for each key individually. The output of the window function is
* interpreted as a regular non-windowed stream.
*
* <p>Arriving data is incrementally aggregated using the given aggregate functi... | 3.26 |
flink_WindowedOperatorTransformation_trigger_rdh | /**
* Sets the {@code Trigger} that should be used to trigger window emission.
*/
@PublicEvolving
public WindowedOperatorTransformation<T, K, W> trigger(Trigger<? super T, ? super W> trigger) {
builder.trigger(trigger);
return this;
} | 3.26 |
flink_PartitioningProperty_isComputablyPartitioned_rdh | /**
* Checks, if this property represents a partitioning that is computable. A computable
* partitioning can be recreated through an algorithm. If two sets of data are to be
* co-partitioned, it is crucial, that the partitioning schemes are computable.
*
* <p>Examples for computable partitioning schemes are hash- ... | 3.26 |
flink_PartitioningProperty_isPartitionedOnKey_rdh | /**
* Checks if this property presents a partitioning that is not random, but on a partitioning
* key.
*
* @return True, if the data is partitioned on a key.
*/
public boolean isPartitionedOnKey() {
return isPartitioned() && (this != RANDOM_PARTITIONED);
} | 3.26 |
flink_TableColumn_physical_rdh | /**
* Creates a regular table column that represents physical data.
*/
public static PhysicalColumn physical(String name, DataType type) {
Preconditions.checkNotNull(name, "Column name can not be null.");
Preconditions.checkNotNull(type, "Column type can not be null.");
return new PhysicalColumn(name, ty... | 3.26 |
flink_TableColumn_getName_rdh | /**
* Returns the name of this column.
*/
public String getName() {
return name;
} | 3.26 |
flink_TableColumn_getType_rdh | /**
* Returns the data type of this column.
*/
public DataType getType() {
return this.type;
} | 3.26 |
flink_TableColumn_metadata_rdh | /**
* Creates a metadata column from metadata of the given column name or from metadata of the
* given alias (if not null).
*
* <p>Allows to specify whether the column is virtual or not.
*/public static MetadataColumn
metadata(String name, DataType type, @Nullable
String metadataAlias, boolean isVirtual) {
Pre... | 3.26 |
flink_TableColumn_asSummaryString_rdh | /**
* Returns a string that summarizes this column for printing to a console.
*/
public String asSummaryString() {
final StringBuilder sb = new StringBuilder();
sb.append(name);
sb.append(": ");
sb.append(type);
explainExtras().ifPresent(e -> {
sb.append(" ");
sb.append(e);
});... | 3.26 |
flink_TableColumn_computed_rdh | /**
* Creates a computed column that is computed from the given SQL expression.
*/
public static ComputedColumn computed(String name, DataType type, String expression) {
Preconditions.checkNotNull(name, "Column name can not be null.");
Preconditions.checkNotNull(type, "Column type can not be null.");
Prec... | 3.26 |
flink_TableColumn_of_rdh | /**
*
* @deprecated Use {@link #computed(String, DataType, String)} instead.
*/
@Deprecated
public static TableColumn of(String name, DataType type, String expression) {
return computed(name, type, expression);
} | 3.26 |
flink_TableColumn_m0_rdh | /**
* Creates a metadata column from metadata of the given column name.
*
* <p>Allows to specify whether the column is virtual or not.
*/
public static MetadataColumn m0(String name, DataType type, boolean isVirtual) {
return metadata(name, type, null, isVirtual);
} | 3.26 |
flink_DownTimeGauge_getValue_rdh | // ------------------------------------------------------------------------
@Override
public Long getValue() {
final JobStatus status = jobStatusProvider.getState();
// not running any more -> finished or not on leader
if (status.isTerminalState()) {
return NO_LONGER_RUNNING;
}
final long r... | 3.26 |
flink_Plan_getDefaultParallelism_rdh | /**
* Gets the default parallelism for this job. That degree is always used when an operator is not
* explicitly given a parallelism.
*
* @return The default parallelism for the plan.
*/
public int getDefaultParallelism() {
return this.defaultParallelism;
} | 3.26 |
flink_Plan_getJobId_rdh | /**
* Gets the ID of the job that the dataflow plan belongs to. If this ID is not set, then the
* dataflow represents its own independent job.
*
* @return The ID of the job that the dataflow plan belongs to.
*/
public JobID getJobId() {
return jobId;
} | 3.26 |
flink_Plan_getPostPassClassName_rdh | /**
* Gets the optimizer post-pass class for this job. The post-pass typically creates utility
* classes for data types and is specific to a particular data model (record, tuple, Scala, ...)
*
* @return The name of the class implementing the optimizer post-pass.
*/
public String getPostPassClassName() {
return "or... | 3.26 |
flink_Plan_registerCachedFile_rdh | /**
* Register cache files at program level.
*
* @param entry
* contains all relevant information
* @param name
* user defined name of that file
* @throws java.io.IOException
*/
public void registerCachedFile(String name, DistributedCacheEntry entry) throws IOException {
if (!this.cacheFile.containsKey(... | 3.26 |
flink_Plan_setJobId_rdh | /**
* Sets the ID of the job that the dataflow plan belongs to. If this ID is set to {@code null},
* then the dataflow represents its own independent job.
*
* @param jobId
* The ID of the job that the dataflow plan belongs to.
*/
public void setJobId(JobID jobId) {
this.jobId = jobId;
} | 3.26 |
flink_Plan_setJobName_rdh | /**
* Sets the jobName for this Plan.
*
* @param jobName
* The jobName to set.
*/
public void setJobName(String jobName) {
checkNotNull(jobName, "The job name must not be null.");this.jobName = jobName;
} | 3.26 |
flink_Plan_getDataSinks_rdh | /**
* Gets all the data sinks of this job.
*
* @return All sinks of the program.
*/ public Collection<? extends GenericDataSinkBase<?>> getDataSinks() {
return this.sinks;
} | 3.26 |
flink_Plan_accept_rdh | // ------------------------------------------------------------------------
/**
* Traverses the job depth first from all data sinks on towards the sources.
*
* @see Visitable#accept(Visitor)
*/
@Override
public void accept(Visitor<Operator<?>> visitor) {
for (GenericDataSinkBase<?> sink : this.sinks) {
... | 3.26 |
flink_Plan_getJobName_rdh | /**
* Gets the name of this job.
*
* @return The name of the job.
*/
public String getJobName() {
return this.jobName;
} | 3.26 |
flink_Plan_getCachedFiles_rdh | /**
* Return the registered cached files.
*
* @return Set of (name, filePath) pairs
*/
public Set<Entry<String, DistributedCacheEntry>> getCachedFiles() {
return this.cacheFile.entrySet();
} | 3.26 |
flink_Plan_setExecutionConfig_rdh | /**
* Sets the runtime config object defining execution parameters.
*
* @param executionConfig
* The execution config to use.
*/
public void setExecutionConfig(ExecutionConfig executionConfig) {
this.executionConfig = executionConfig;
} | 3.26 |
flink_Plan_setDefaultParallelism_rdh | /**
* Sets the default parallelism for this plan. That degree is always used when an operator is
* not explicitly given a parallelism.
*
* @param defaultParallelism
* The default parallelism for the plan.
*/
public void setDefaultParallelism(int defaultParallelism) {
checkArgument((defaultParallelism >= 1) || (... | 3.26 |
flink_Plan_addDataSink_rdh | // ------------------------------------------------------------------------
/**
* Adds a data sink to the set of sinks in this program.
*
* @param sink
* The data sink to add.
*/public void addDataSink(GenericDataSinkBase<?>
sink) {
checkNotNull(sink, "The data sink must not be null.");
if (!this.sinks.contains(... | 3.26 |
flink_CachingLookupFunction_lookupByDelegate_rdh | // -------------------------------- Helper functions ------------------------------
private Collection<RowData> lookupByDelegate(RowData keyRow) throws IOException {
try {
Preconditions.checkState(delegate != null, "User's lookup function can't be null, if there are possible cache misses.");
long l... | 3.26 |
flink_CachingLookupFunction_open_rdh | /**
* Open the {@link CachingLookupFunction}.
*
* <p>In order to reduce the memory usage of the cache, {@link LookupCacheManager} is used to
* provide a shared cache instance across subtasks of this function. Here we use {@link #functionIdentifier()} as the id of the cache, which is generated by MD5 of serialized b... | 3.26 |
flink_StreamRecord_equals_rdh | // Utilities
// ------------------------------------------------------------------------
@Override
public boolean equals(Object
o) {
if (this == o) {
return true;
} else if ((o != null) && (getClass() == o.getClass())) {
StreamRecord<?> that = ((StreamRecord<?>) (o));
return ((t... | 3.26 |
flink_StreamRecord_getTimestamp_rdh | /**
* Returns the timestamp associated with this stream value in milliseconds.
*/
public long getTimestamp() {
if (hasTimestamp) {return timestamp;
} else {
return Long.MIN_VALUE;
// throw new IllegalStateException(
// "Record has no timestamp. Is the time characterist... | 3.26 |
flink_StreamRecord_replace_rdh | /**
* Replace the currently stored value by the given new value and the currently stored timestamp
* with the new timestamp. This returns a StreamElement with the generic type parameter that
* matches the new value.
*
* @param value
* The new value to wrap in this StreamRecord
* @param timestamp
* The new t... | 3.26 |
flink_StreamRecord_copy_rdh | // ------------------------------------------------------------------------
// Copying
// ------------------------------------------------------------------------
/**
* Creates a copy of this stream record. Uses the copied value as the value for the new record,
* i.e., only copies timestamp fields.
*/
public StreamR... | 3.26 |
flink_StreamRecord_getValue_rdh | // ------------------------------------------------------------------------
// Accessors
// ------------------------------------------------------------------------
/**
* Returns the value wrapped in this stream value.
*/
public T getValue() {
return value;
} | 3.26 |
flink_ErrorInfo_getException_rdh | /**
* Returns the serialized form of the original exception.
*/
public SerializedThrowable getException() {
return exception;
} | 3.26 |
flink_ErrorInfo_handleMissingThrowable_rdh | /**
* Utility method to cover FLINK-21376.
*
* @param throwable
* The actual exception.
* @return a {@link FlinkException} if no exception was passed.
*/
public static Throwable handleMissingThrowable(@Nullable
Throwable throwable) {
return throwable != null ? throwable : new FlinkException("Unknown caus... | 3.26 |
flink_WindowSavepointReader_process_rdh | /**
* Reads window state generated without any preaggregation such as {@code WindowedStream#apply}
* and {@code WindowedStream#process}.
*
* @param uid
* The uid of the operator.
* @param readerFunction
* The window reader function.
* @param keyType
* The key type of the window.
* @param stateType
* ... | 3.26 |
flink_WindowSavepointReader_evictor_rdh | /**
* Reads from a window that uses an evictor.
*/
public EvictingWindowSavepointReader<W> evictor() {
return new EvictingWindowSavepointReader<>(env, metadata, stateBackend, windowSerializer);
} | 3.26 |
flink_WindowSavepointReader_aggregate_rdh | /**
* Reads window state generated using an {@link AggregateFunction}.
*
* @param uid
* The uid of the operator.
* @param aggregateFunction
* The aggregate function used to create the window.
* @param readerFunction
* The window reader function.
* @param keyType
* The key type of the window.
* @param... | 3.26 |
flink_WindowSavepointReader_reduce_rdh | /**
* Reads window state generated using a {@link ReduceFunction}.
*
* @param uid
* The uid of the operator.
* @param function
* The reduce function used to create the window.
* @param readerFunction
* The window reader function.
* @param keyType
* The key type of the window.
* @param reduceType
* ... | 3.26 |
flink_TestingSinkSettings_getCheckpointingMode_rdh | /**
* Checkpointing mode required for the sink.
*/
public CheckpointingMode getCheckpointingMode() {
return checkpointingMode;
} | 3.26 |
flink_CountWindow_getId_rdh | /**
* Gets the id (0-based) of the window.
*/
public long getId() {
return id;
} | 3.26 |
flink_CountWindow_snapshotConfiguration_rdh | // ------------------------------------------------------------------------
@Override
public TypeSerializerSnapshot<CountWindow> snapshotConfiguration() {
return new CountWindow.Serializer.CountWindowSerializerSnapshot();
} | 3.26 |
flink_CoGroupDriver_setup_rdh | // ------------------------------------------------------------------------
@Override
public void setup(TaskContext<CoGroupFunction<IT1, IT2, OT>, OT> context) {this.taskContext = context;
this.running = true;
} | 3.26 |
flink_EmbeddedRocksDBStateBackend_getRocksDBOptions_rdh | /**
* Gets {@link org.rocksdb.Options} for the RocksDB instances.
*
* <p>The options created by the factory here are applied on top of the pre-defined options
* profile selected via {@link #setPredefinedOptions(PredefinedOptions)}. If the pre-defined
* options profile is the default ({@link PredefinedOptions#DEFAU... | 3.26 |
flink_EmbeddedRocksDBStateBackend_setDbStoragePaths_rdh | /**
* Sets the directories in which the local RocksDB database puts its files (like SST and
* metadata files). These directories do not need to be persistent, they can be ephemeral,
* meaning that they are lost on a machine failure, because state in RocksDB is persisted in
* checkpoints.
*
* <p>If nothing is conf... | 3.26 |
flink_EmbeddedRocksDBStateBackend_setPriorityQueueStateType_rdh | /**
* Sets the type of the priority queue state. It will fallback to the default value, if it is
* not explicitly set.
*/
public void setPriorityQueueStateType(PriorityQueueStateType priorityQueueStateType) {
this.priorityQueueConfig.setPriorityQueueStateType(priorityQueueStateType);
} | 3.26 |
flink_EmbeddedRocksDBStateBackend_setRocksDBMemoryFactory_rdh | /**
* Set RocksDBMemoryFactory.
*/
public void setRocksDBMemoryFactory(RocksDBMemoryFactory rocksDBMemoryFactory) {
this.rocksDBMemoryFactory = checkNotNull(rocksDBMemoryFactory);
} | 3.26 |
flink_EmbeddedRocksDBStateBackend_mergeConfigurableOptions_rdh | // ------------------------------------------------------------------------
// utilities
// ------------------------------------------------------------------------
private ReadableConfig mergeConfigurableOptions(ReadableConfig base, ReadableConfig onTop) {
if (base == null) {
base = new Configuration();
}
Configuratio... | 3.26 |
flink_EmbeddedRocksDBStateBackend_setWriteBatchSize_rdh | /**
* Sets the max batch size will be used in {@link RocksDBWriteBatchWrapper}, no positive value
* will disable memory size controller, just use item count controller.
*
* @param writeBatchSize
* The size will used to be used in {@link RocksDBWriteBatchWrapper}.
*/
public void setWriteBatchSize(long writeBatch... | 3.26 |
flink_EmbeddedRocksDBStateBackend_setDbStoragePath_rdh | /**
* Sets the path where the RocksDB local database files should be stored on the local file
* system. Setting this path overrides the default behavior, where the files are stored across
* the configured temp directories.
*
* <p>Passing {@code null} to this function restores the default behavior, where the config... | 3.26 |
flink_EmbeddedRocksDBStateBackend_setRocksDBOptions_rdh | /**
* Sets {@link org.rocksdb.Options} for the RocksDB instances. Because the options are not
* serializable and hold native code references, they must be specified through a factory.
*
* <p>The options created by the factory here are applied on top of the pre-defined options
* profile selected via {@link #setPred... | 3.26 |
flink_EmbeddedRocksDBStateBackend_getNumberOfTransferThreads_rdh | /**
* Gets the number of threads used to transfer files while snapshotting/restoring.
*/
public int getNumberOfTransferThreads() {
return numberOfTransferThreads == UNDEFINED_NUMBER_OF_TRANSFER_THREADS ? CHECKPOINT_TRANSFER_THREAD_NUM.defaultValue() : numberOfTransferThreads;
} | 3.26 |
flink_EmbeddedRocksDBStateBackend_m0_rdh | /**
* Gets the configured local DB storage paths, or null, if none were configured.
*
* <p>Under these directories on the TaskManager, RocksDB stores its SST files and metadata
* files. These directories do not need to be persistent, they can be ephermeral, meaning that
* they are lost on a machine failure, becaus... | 3.26 |
flink_EmbeddedRocksDBStateBackend_configure_rdh | // ------------------------------------------------------------------------
// Reconfiguration
// ------------------------------------------------------------------------
/**
* Creates a copy of this state backend that uses the values defined in the configuration for
* fields where that were not yet specified in this... | 3.26 |
flink_EmbeddedRocksDBStateBackend_createKeyedStateBackend_rdh | // ------------------------------------------------------------------------
// State holding data structures
// ------------------------------------------------------------------------
@Overridepublic <K> AbstractKeyedStateBackend<K> createKeyedStateBackend(Environment env, JobID jobID, String operatorIdentifier, TypeS... | 3.26 |
flink_EmbeddedRocksDBStateBackend_getMemoryConfiguration_rdh | // ------------------------------------------------------------------------
// Parameters
// ------------------------------------------------------------------------
/**
* Gets the memory configuration object, which offers settings to control RocksDB's memory
* usage.
*/public RocksDBMemoryConfiguration getMemoryCon... | 3.26 |
flink_EmbeddedRocksDBStateBackend_setPredefinedOptions_rdh | // ------------------------------------------------------------------------
// Parametrize with RocksDB Options
// ------------------------------------------------------------------------
/**
* Sets the predefined options for RocksDB.
*
* <p>If user-configured options within {@link RocksDBConfigurableOptions} is se... | 3.26 |
flink_EmbeddedRocksDBStateBackend_ensureRocksDBIsLoaded_rdh | // ------------------------------------------------------------------------
// static library loading utilities
// ------------------------------------------------------------------------
@VisibleForTesting
static void ensureRocksDBIsLoaded(String tempDirectory) throws IOException {
ensureRocksDBIsLoaded(tempDirectory,... | 3.26 |
flink_EmbeddedRocksDBStateBackend_supportsNoClaimRestoreMode_rdh | // ------------------------------------------------------------------------
// State backend methods
// ------------------------------------------------------------------------
@Override
public boolean supportsNoClaimRestoreMode() {
// We are able to create CheckpointType#FULL_CHECKPOINT. (we might potentially reup... | 3.26 |
flink_EmbeddedRocksDBStateBackend_isIncrementalCheckpointsEnabled_rdh | /**
* Gets whether incremental checkpoints are enabled for this state backend.
*/
public boolean isIncrementalCheckpointsEnabled() {
return enableIncrementalCheckpointing.getOrDefault(CheckpointingOptions.INCREMENTAL_CHECKPOINTS.defaultValue());
} | 3.26 |
flink_EmbeddedRocksDBStateBackend_setNumberOfTransferThreads_rdh | /**
* Sets the number of threads used to transfer files while snapshotting/restoring.
*
* @param numberOfTransferThreads
* The number of threads used to transfer files while
* snapshotting/restoring.
*/
public void setNumberOfTransferThreads(int numberOfTransferThreads) {
Preconditions.checkArgument(numberOfT... | 3.26 |
flink_EmbeddedRocksDBStateBackend_getPredefinedOptions_rdh | /**
* Gets the currently set predefined options for RocksDB. The default options (if nothing was
* set via {@link #setPredefinedOptions(PredefinedOptions)}) are {@link PredefinedOptions#DEFAULT}.
*
* <p>If user-configured options within {@link RocksDBConfigurableOptions} is set (through
* flink-conf.yaml) of a use... | 3.26 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.