name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_TableChange_modifyPhysicalColumnType_rdh | /**
* A table change that modify the physical column data type.
*
* <p>It is equal to the following statement:
*
* <pre>
* ALTER TABLE <table_name> MODIFY <column_name> <new_column_type>
* </pre>
*
* @param oldColumn
* the definition of the old column.
* @param newType
* the type of... | 3.26 |
flink_TableChange_modifyColumnComment_rdh | /**
* A table change to modify the column comment.
*
* <p>It is equal to the following statement:
*
* <pre>
* ALTER TABLE <table_name> MODIFY <column_name> <original_column_type> COMMENT '<new_column_comment>'
* </pre>
*
* @param oldColumn
* the definition of the old column.
* @p... | 3.26 |
flink_TableChange_add_rdh | /**
* A table change to add a watermark.
*
* <p>It is equal to the following statement:
*
* <pre>
* ALTER TABLE <table_name> ADD WATERMARK FOR <row_time> AS <row_time_expression>
* </pre>
*
* @param watermarkSpec
* the added watermark definition.
* @return a TableChange represents the ... | 3.26 |
flink_TableChange_modifyColumnName_rdh | /**
* A table change to modify the column name.
*
* <p>It is equal to the following statement:
*
* <pre>
* ALTER TABLE <table_name> RENAME <old_column_name> TO <new_column_name>
* </pre>
*
* @param oldColumn
* the definition of the old column.
* @param newName
* the name of the new ... | 3.26 |
flink_TableChange_m0_rdh | /**
* A table change to add a unique constraint.
*
* <p>It is equal to the following statement:
*
* <pre>
* ALTER TABLE <table_name> ADD PRIMARY KEY (<column_name>...) NOT ENFORCED
* </pre>
*
* @param constraint
* the added constraint definition.
* @return a TableChange represents the modifi... | 3.26 |
flink_TableChange_getConstraint_rdh | /**
* Returns the unique constraint to add.
*/
public UniqueConstraint getConstraint() {
return constraint;
} | 3.26 |
flink_TableChange_getValue_rdh | /**
* Returns the Option value to set.
*/
public String getValue() {
return value;
} | 3.26 |
flink_TableChange_getConstraintName_rdh | /**
* Returns the constraint name.
*/
public String getConstraintName() {
return constraintName;
} | 3.26 |
flink_TableChange_dropColumn_rdh | /**
* A table change to drop column.
*
* <p>It is equal to the following statement:
*
* <pre>
* ALTER TABLE <table_name> DROP COLUMN <column_name>
* </pre>
*
* @param columnName
* the column to drop.
... | 3.26 |
flink_TableChange_getNewPosition_rdh | /**
* Returns the position of the modified {@link Column} instance. When the return value is
* null, it means modify the column at the original position. When the return value is
* FIRST, it means move the modified column to the first. When the return value is AFTER, it
* means move the column after the referred co... | 3.26 |
flink_TableChange_getNewConstraint_rdh | /**
* Returns the modified unique constraint.
*/
public UniqueConstraint getNewConstraint() {
return newConstraint;
} | 3.26 |
flink_TableChange_after_rdh | /**
* Get the position to place the column after the specified column.
*/
static ColumnPosition after(String column) {
return new After(column);
} | 3.26 |
flink_TableChange_getKey_rdh | /**
* Returns the Option key to reset.
*/ public String getKey() {return key;
} | 3.26 |
flink_TableChange_getNewColumnName_rdh | /**
* Returns the new column name after renaming the column name.
*/
public String getNewColumnName() {return newColumn.getName();
} | 3.26 |
flink_TableChange_modify_rdh | /**
* A table change to modify a watermark.
*
* <p>It is equal to the following statement:
*
* <pre>
* ALTER TABLE <table_name> MODIFY WATERMARK FOR <row_time> AS <row_time_expression>
* </pre>
*
* @param newWatermarkSpec
* the modified watermark definition.
* @return a TableChange rep... | 3.26 |
flink_TableChange_dropConstraint_rdh | /**
* A table change to drop constraint.
*
* <p>It is equal to the following statement:
*
* <pre>
* ALTER TABLE <table_name> DROP CONSTRAINT <constraint_name>
* </pre>
*
* @param constraintName
* the constraint to drop.
* @return a TableChange represents the modification.
*/
static DropCons... | 3.26 |
flink_TableChange_getOldColumnName_rdh | /**
* Returns the origin column name.
*/
public String getOldColumnName() {
return oldColumn.getName();
} | 3.26 |
flink_TableChange_getNewWatermark_rdh | /**
* Returns the modified watermark.
*/
public WatermarkSpec getNewWatermark() {
return newWatermark;
} | 3.26 |
flink_TableChange_getNewColumn_rdh | /**
* Returns the modified {@link Column} instance.
*/
public Column getNewColumn() {
return newColumn;
} | 3.26 |
flink_TableChange_getColumnName_rdh | /**
* Returns the column name.
*/
public String getColumnName() {
return columnName;
} | 3.26 |
flink_TableChange_getOldColumn_rdh | /**
* Returns the original {@link Column} instance.
*/
public Column getOldColumn() {
return oldColumn;
} | 3.26 |
flink_TableChange_set_rdh | /**
* A table change to set the table option.
*
* <p>It is equal to the following statement:
*
* <pre>
* ALTER TABLE <table_name> SET '<key>' = '<value>';
* </pre>
*
* @param key
* the option name to set.
* @param value
* the option value to set.
* @return a TableChange represents ... | 3.26 |
flink_TableChange_modifyColumnPosition_rdh | /**
* A table change to modify the column position.
*
* <p>It is equal to the following statement:
*
* <pre>
* ALTER TABLE <table_name> MODIFY <column_name> <original_column_type> <column_position>
* </pre>
*
* @param oldColumn
* the definition of the old column.
* @param columnP... | 3.26 |
flink_TableChange_getWatermark_rdh | /**
* Returns the watermark to add.
*/
public WatermarkSpec getWatermark() {
return watermarkSpec;
} | 3.26 |
flink_TableChange_first_rdh | /**
* Get the position to place the column at the first.
*/
static ColumnPosition first() {
return First.INSTANCE;
} | 3.26 |
flink_TableChange_reset_rdh | /**
* A table change to reset the table option.
*
* <p>It is equal to the following statement:
*
* <pre>
* ALTER TABLE <table_name> RESET '<key>'
* </pre>
*
* @param key
* the option name to reset.
* @return a TableChange represents the modification.
*/
static ResetOption reset(String key) ... | 3.26 |
flink_TaskManagerRuntimeInfo_getTaskManagerBindAddress_rdh | /**
* Gets the bind address of the Taskmanager.
*
* @return The bind address of the TaskManager.
*/
default String getTaskManagerBindAddress() {
return m0().getString(TaskManagerOptions.BIND_HOST);
} | 3.26 |
flink_HiveFunctionWrapper_deserializeUDF_rdh | /**
* Deserialize UDF used the udfSerializedString held on.
*
* @return the UDF deserialized
*/
@SuppressWarnings("unchecked")
private UDFType deserializeUDF() {
return ((UDFType) (deserializeObjectFromKryo(udfSerializedBytes, ((Class<Serializable>) (getUDFClass())))));
} | 3.26 |
flink_HiveFunctionWrapper_createFunction_rdh | /**
* Instantiate a Hive function instance.
*
* @return a Hive function instance
*/
public UDFType createFunction() {
if (udfSerializedBytes != null) {
// deserialize the string to udf instance
return deserializeUDF();
} else if (instance != null) { return
instance;
} else {
... | 3.26 |
flink_EitherTypeInfo_getLeftType_rdh | // --------------------------------------------------------------------------------------------
public TypeInformation<L> getLeftType() {
return leftType;
} | 3.26 |
flink_ConfigOptions_durationType_rdh | /**
* Defines that the value of the option should be of {@link Duration} type.
*/
public TypedConfigOptionBuilder<Duration> durationType() {
return new TypedConfigOptionBuilder<>(key, Duration.class);
} | 3.26 |
flink_ConfigOptions_defaultValue_rdh | /**
* Creates a ConfigOption with the given default value.
*
* @param value
* The default value for the config option
* @return The config option with the default value.
*/
public ConfigOption<T> defaultValue(T value) {
return new ConfigOption<>(key, clazz,
ConfigOption.EMPTY_DESCRIPTION, value, false);... | 3.26 |
flink_ConfigOptions_mapType_rdh | /**
* Defines that the value of the option should be a set of properties, which can be
* represented as {@code Map<String, String>}.
*/
public TypedConfigOptionBuilder<Map<String, String>> mapType() {
return new TypedConfigOptionBuilder<>(key, PROPERTIES_MAP_CLASS);
}
/**
* Creates a ConfigOption with the give... | 3.26 |
flink_ConfigOptions_asList_rdh | /**
* Defines that the option's type should be a list of previously defined atomic type.
*/
public ListConfigOptionBuilder<T> asList() {
return new ListConfigOptionBuilder<>(key, clazz);
} | 3.26 |
flink_ConfigOptions_booleanType_rdh | /**
* Defines that the value of the option should be of {@link Boolean} type.
*/
public TypedConfigOptionBuilder<Boolean> booleanType() {
return new TypedConfigOptionBuilder<>(key, Boolean.class);} | 3.26 |
flink_ConfigOptions_enumType_rdh | /**
* Defines that the value of the option should be of {@link Enum} type.
*
* @param enumClass
* Concrete type of the expected enum.
*/
public <T extends Enum<T>> TypedConfigOptionBuilder<T> enumType(Class<T> enumClass) {
return new TypedConfigOptionBuilder<>(key, enumClass);
} | 3.26 |
flink_ConfigOptions_intType_rdh | /**
* Defines that the value of the option should be of {@link Integer} type.
*/
public TypedConfigOptionBuilder<Integer> intType() {
return new TypedConfigOptionBuilder<>(key, Integer.class);
} | 3.26 |
flink_ConfigOptions_defaultValues_rdh | /**
* Creates a ConfigOption with the given default value.
*
* @param values
* The list of default values for the config option
* @return The config option with the default value.
*/
@SafeVarargs
public final ConfigOption<List<E>> defaultValues(E... values) {
return new ConfigOption<>(key, clazz, ConfigOpti... | 3.26 |
flink_ConfigOptions_doubleType_rdh | /**
* Defines that the value of the option should be of {@link Double} type.
*/
public TypedConfigOptionBuilder<Double> doubleType() {
return new TypedConfigOptionBuilder<>(key, Double.class);
} | 3.26 |
flink_ConfigOptions_floatType_rdh | /**
* Defines that the value of the option should be of {@link Float} type.
*/
public TypedConfigOptionBuilder<Float> floatType() {
return new TypedConfigOptionBuilder<>(key, Float.class);
} | 3.26 |
flink_ConfigOptions_memoryType_rdh | /**
* Defines that the value of the option should be of {@link MemorySize} type.
*/
public TypedConfigOptionBuilder<MemorySize> memoryType() {
return new TypedConfigOptionBuilder<>(key, MemorySize.class);
} | 3.26 |
flink_ConfigOptions_key_rdh | /**
* Starts building a new {@link ConfigOption}.
*
* @param key
* The key for the config option.
* @return The builder for the config option with the given key.
*/
public static OptionBuilder key(String key)
{
checkNotNull(key);
return new OptionBuilder(key);
} | 3.26 |
flink_ConfigOptions_longType_rdh | /**
* Defines that the value of the option should be of {@link Long} type.
*/
public TypedConfigOptionBuilder<Long> longType() {
return new TypedConfigOptionBuilder<>(key, Long.class);
} | 3.26 |
flink_ConfigOptions_stringType_rdh | /**
* Defines that the value of the option should be of {@link String} type.
*/
public TypedConfigOptionBuilder<String> stringType() {
return new TypedConfigOptionBuilder<>(key, String.class);
} | 3.26 |
flink_OverWindowPartitionedOrdered_as_rdh | /**
* Assigns an alias for this window that the following {@code select()} clause can refer to.
*
* @param alias
* alias for this over window
* @return the fully defined over window
*/
public OverWindow as(Expression alias) {
return new OverWindow(alias, partitionBy, orderBy, valueLiteral(OverWindowRange.UN... | 3.26 |
flink_OverWindowPartitionedOrdered_preceding_rdh | /**
* Set the preceding offset (based on time or row-count intervals) for over window.
*
* @param preceding
* preceding offset relative to the current row.
* @return an over window with defined preceding
*/
public OverWindowPartitionedOrderedPreceding preceding(Expression preceding) {
return new OverWindowP... | 3.26 |
flink_JarManifestParser_findOnlyEntryClass_rdh | /**
* Returns a JAR file with its entry class as specified in the manifest.
*
* @param jarFiles
* JAR files to parse
* @throws NoSuchElementException
* if no JAR file contains an entry class attribute
* @throws IllegalArgumentException
* if multiple JAR files contain an entry class manifest
* attribute... | 3.26 |
flink_JarManifestParser_findEntryClass_rdh | /**
* Returns the entry class as specified in the manifest of the provided JAR file.
*
* <p>The following manifest attributes are checked in order to find the entry class:
*
* <ol>
* <li>{@link PackagedProgram#MANIFEST_ATTRIBUTE_ASSEMBLER_CLASS}
* <li>{@link PackagedProgram#MANIFEST_ATTRIBUTE_MAIN_CLASS}
* ... | 3.26 |
flink_AvroRowDeserializationSchema_convertAvroRecordToRow_rdh | // --------------------------------------------------------------------------------------------
private Row convertAvroRecordToRow(Schema schema, RowTypeInfo typeInfo, IndexedRecord record) {
final List<Schema.Field> fields = schema.getFields();
final TypeInformation<?>[] fieldInfo = typeInfo.getFieldTypes();
final in... | 3.26 |
flink_SqlShowPartitions_getPartitionKVs_rdh | /**
* Get partition spec as key-value strings.
*/public LinkedHashMap<String, String> getPartitionKVs() {
return SqlPartitionUtils.getPartitionKVs(getPartitionSpec());
} | 3.26 |
flink_SpillingBuffer_moveAll_rdh | /**
* Utility method that moves elements. It avoids copying the data into a dedicated array first,
* as the {@link ArrayList#addAll(java.util.Collection)} method does.
*
* @param <E>
* @param source
* @param target
*/
private static final <E> void moveAll(ArrayList<E> source, ArrayList<E> target) {
target.e... | 3.26 |
flink_SpillingBuffer_close_rdh | /**
*
* @return A list with all memory segments that have been taken from the memory segment source.
*/
public List<MemorySegment> close() throws IOException
{
final ArrayList<MemorySegment> segments = new ArrayList<MemorySegment>(this.fullSegments.size() + this.numMemorySegmentsInWriter);
// if the buffer i... | 3.26 |
flink_ProcessorArchitecture_getProcessorArchitecture_rdh | // ------------------------------------------------------------------------
/**
* Gets the ProcessorArchitecture of the system running this process.
*/
public static ProcessorArchitecture getProcessorArchitecture() {return
CURRENT;
} | 3.26 |
flink_ProcessorArchitecture_getAlternativeNames_rdh | /**
* Gets the alternative names for the processor architecture. Alternative names are for example
* "i586" for "x86", or "x86_64" for "amd64".
*/
public List<String> getAlternativeNames() {
return alternativeNames;
} | 3.26 |
flink_ProcessorArchitecture_getAddressSize_rdh | /**
* Gets the address size of the memory (32 bit, 64 bit).
*/
public MemoryAddressSize getAddressSize() {
return addressSize;
} | 3.26 |
flink_SchedulerFactory_create_rdh | /**
* Create a {@link ScheduledThreadPoolExecutor} using the provided corePoolSize. The following
* behaviour is configured:
*
* <ul>
* <li>rejected executions are logged if the executor is {@link java.util.concurrent.ThreadPoolExecutor#isShutdown shutdown}
* <li>otherwise, {@link ... | 3.26 |
flink_TaskSlot_add_rdh | // ----------------------------------------------------------------------------------
// State changing methods
// ----------------------------------------------------------------------------------
/**
* Add the given task to the task slot. This is only possible if there is not ... | 3.26 |
flink_TaskSlot_markInactive_rdh | /**
* Mark the slot as inactive/allocated. A slot can only be marked as inactive/allocated if it's
* in state allocated or active.
*
* @return True if the new state of the slot is allocated; otherwise false
*/
public boolean markInactive() {
if ((TaskSlotState.ACTIVE == state) || (TaskSlotState.ALLOCATED == s... | 3.26 |
flink_TaskSlot_generateSlotOffer_rdh | /**
* Generate the slot offer from this TaskSlot.
*
* @return The sot offer which this task slot can provide
*/
public SlotOffer generateSlotOffer() {Preconditions.checkState((TaskSlotState.ACTIVE == state) || (TaskSlotState.ALLOCATED == state), "The task slot is not in state active or allocated.");
Preconditio... | 3.26 |
flink_TaskSlot_getTasks_rdh | /**
* Get all tasks running in this task slot.
*
* @return Iterator to all currently contained tasks in this task slot.
*/
public Iterator<T> getTasks() {return tasks.values().iterator();
} | 3.26 |
flink_TaskSlot_m1_rdh | /**
* Remove the task identified by the given execution attempt id.
*
* @param executionAttemptId
* identifying the task to be removed
* @return The removed task if there was any; otherwise null.
*/
public T m1(ExecutionAttemptID executionAttemptId) {
return tasks.remove(executionAttemptId);
} | 3.26 |
flink_TaskSlot_markActive_rdh | /**
* Mark this slot as active. A slot can only be marked active if it's in state allocated.
*
* <p>The method returns true if the slot was set to active. Otherwise it returns false.
*
* @return True if the new state of the slot is active; otherwise false
*/
public boolean markActi... | 3.26 |
flink_TaskSlot_clear_rdh | /**
* Removes all tasks from this task slot.
*/
public void clear() {
tasks.clear();} | 3.26 |
flink_TaskSlot_getIndex_rdh | // ----------------------------------------------------------------------------------
// State accessors
// ----------------------------------------------------------------------------------
public int getIndex() {
return index;
} | 3.26 |
flink_TaskSlot_closeAsync_rdh | /**
* Close the task slot asynchronously.
*
* <p>Slot is moved to {@link TaskSlotState#RELEASING} state and only once. If there are active
* tasks running in the slot then they are failed. The future of all tasks terminated and slot
* cleaned up is initiated only once and always returned in case of multiple attemp... | 3.26 |
flink_Executors_newDirectExecutorServiceWithNoOpShutdown_rdh | /**
* Creates a new {@link ExecutorService} that runs the passed tasks in the calling thread but
* doesn't implement proper shutdown behavior. Tasks can be still submitted even after {@link ExecutorService#shutdown()} is called.
*
* @see #newDirectExecutorService()
*/
public static ExecutorService newDirectExecuto... | 3.26 |
flink_Executors_directExecutor_rdh | /**
* Return a direct executor. The direct executor directly executes the runnable in the calling
* thread.
*
* @return Direct executor
*/
public static Executor directExecutor() {
return
DirectExecutor.INSTANCE;
} | 3.26 |
flink_Executors_newDirectExecutorService_rdh | /**
* Creates a more {@link ExecutorService} that runs the passed task in the calling thread.
*/public
static ExecutorService newDirectExecutorService() {
return new
DirectExecutorService(true);
} | 3.26 |
flink_CompositeBuffer_getFullBufferData_rdh | /**
* Returns the full buffer data in one piece of {@link MemorySegment}. If there is multiple
* partial buffers, the partial data will be copied to the given target {@link MemorySegment}.
*/
public Buffer getFullBufferData(MemorySegment
segment) {
checkState(!partialBuffers.isEmpty());
checkState(currentLen... | 3.26 |
flink_LongParser_parseField_rdh | /**
* Static utility to parse a field of type long from a byte sequence that represents text
* characters (such as when read from a file stream).
*
* @param bytes
* The bytes containing the text data that should be parsed.
* @param startPos
* The offset to start the parsing.
* @param length
* The length ... | 3.26 |
flink_NonReusingBuildSecondHashJoinIterator_open_rdh | // --------------------------------------------------------------------------------------------
@Override
public void open() throws IOException, MemoryAllocationException, InterruptedException {
this.hashJoin.open(this.secondInput, this.firstInput,
buildSideOuterJoin);
} | 3.26 |
flink_JobGraphGenerator_connectJobVertices_rdh | // ------------------------------------------------------------------------
// Connecting Vertices
// ------------------------------------------------------------------------
/**
* NOTE: The channel for global and local strategies are different if we connect a union. The
* global strategy channel is then the channel ... | 3.26 |
flink_JobGraphGenerator_compileJobGraph_rdh | /**
* Translates a {@link org.apache.flink.optimizer.plan.OptimizedPlan} into a {@link org.apache.flink.runtime.jobgraph.JobGraph}.
*
* @param program
* Optimized plan that is translated into a JobGraph.
* @return JobGraph generated from the plan.
*/
public JobGraph compileJobGraph(OptimizedPlan program) {
... | 3.26 |
flink_JobGraphGenerator_postVisit_rdh | /**
* This method implements the post-visit during the depth-first traversal. When the post visit
* happens, all of the descendants have been processed, so this method connects all of the
* current node's predecessors to the current node.
*
* @param node
* The node currently processed during the post-visit.
* ... | 3.26 |
flink_JobGraphGenerator_preVisit_rdh | /**
* This methods implements the pre-visiting during a depth-first traversal. It create the job
* vertex and sets local strategy.
*
* @param node
* The node that is currently processed.
* @return True, if the visitor should descend to the node's children, false if not.
* @see org.apache.flink.util.Visitor#pre... | 3.26 |
flink_JobGraphGenerator_createSingleInputVertex_rdh | // ------------------------------------------------------------------------
// Methods for creating individual vertices
// ------------------------------------------------------------------------
private JobVertex createSingleInputVertex(SingleInputPlanNode node) throws CompilerException {
final String taskName = node.... | 3.26 |
flink_AsyncWaitOperator_outputCompletedElement_rdh | /**
* Outputs one completed element. Watermarks are always completed if it's their turn to be
* processed.
*
* <p>This method will be called from {@link #processWatermark(Watermark)} and from a mail
* processing the result of an async function call.
*/
private void outputCompletedElement() {
if (queue.hasComp... | 3.26 |
flink_AsyncWaitOperator_tryOnce_rdh | /**
* Increments number of attempts and fire the attempt.
*/
private void tryOnce(RetryableResultHandlerDelegator resultHandlerDelegator) throws Exception {
// increment current attempt number
resultHandlerDelegator.currentAttempts++;
// fire a new attempt
userFunction.asyncInvoke(resultHandlerDelega... | 3.26 |
flink_AsyncWaitOperator_timerTriggered_rdh | /**
* Rewrite the timeout process to deal with retry state.
*/
private void timerTriggered() throws Exception {
if (!resultHandler.completed.get()) {
// cancel delayed retry timer first
cancelRetryTimer();
// force reset retryAwaiting to prevent the handler to trigger retry unnecessaril... | 3.26 |
flink_AsyncWaitOperator_registerTimer_rdh | /**
* Utility method to register timeout timer.
*/
private ScheduledFuture<?> registerTimer(ProcessingTimeService processingTimeService, long
timeout, ThrowingConsumer<Void,
Exception> callback) {
final long timeoutTimestamp = timeout + processingTimeService.getCurrentProcessingTime();... | 3.26 |
flink_FromJarEntryClassInformationProvider_getJarFile_rdh | /**
* Returns the specified {@code jarFile}.
*
* @return The specified {@code jarFile}.
* @see #getJobClassName()
*/
@Override
public Optional<File> getJarFile() {
return Optional.of(jarFile);
}
/**
* Returns the specified job class name that is either available in the corresponding {@code jarFile}. It can ... | 3.26 |
flink_FromJarEntryClassInformationProvider_createFromCustomJar_rdh | /**
* Creates a {@code FromJarEntryClassInformationProvider} for a custom Jar archive. At least the
* {@code jarFile} or the {@code jobClassName} has to be set.
*
* @param jarFile
* The Jar archive.
* @param jobClassName
* The name of the job class.
* @return The {@code FromJarEntryClassInformationProvider}... | 3.26 |
flink_FromJarEntryClassInformationProvider_createFromPythonJar_rdh | /**
* Creates a {@code FromJarEntryClassInformationProvider} for a job implemented in Python.
*
* @return A {@code FromJarEntryClassInformationProvider} for a job implemented in Python
*/
public static FromJarEntryClassInformationProvider createFromPythonJar() {
return new FromJarEntryClassInformationProvider(... | 3.26 |
flink_ListView_setList_rdh | /**
* Replaces the entire view's content with the content of the given {@link List}.
*/
public void setList(List<T> list) {
this.list = list;
} | 3.26 |
flink_ListView_getList_rdh | /**
* Returns the entire view's content as an instance of {@link List}.
*/
public List<T> getList() {
return list;
} | 3.26 |
flink_ListView_newListViewDataType_rdh | // --------------------------------------------------------------------------------------------
// Utilities
// --------------------------------------------------------------------------------------------
/**
* Utility method for creating a {@link DataType} of {@link ListView} explicitly.
*/
public static DataType ne... | 3.26 |
flink_ListView_add_rdh | /**
* Adds the given value to the list.
*
* @throws Exception
* Thrown if the system cannot add data.
* @param value
* The element to be appended to this list view.
*/
public void add(T value) throws Exception {
list.add(value);
} | 3.26 |
flink_ListView_m0_rdh | /**
* Returns an iterable of the list view.
*
* @throws Exception
* Thrown if the system cannot get data.
* @return The iterable of the list.
*/
public Iterable<T> m0() throws Exception {
return list;
} | 3.26 |
flink_ListView_clear_rdh | /**
* Removes all of the elements from this list view.
*/
@Override
public void clear() {
list.clear();
} | 3.26 |
flink_ExecutionEnvironment_m0_rdh | // ----------------------------------- Generic Input Format
// ---------------------------------------
/**
* Generic method to create an input {@link DataSet} with in {@link InputFormat}. The DataSet
* will not be immediately created - instead, this method returns a DataSet that will be lazily
* created from the inp... | 3.26 |
flink_ExecutionEnvironment_generateSequence_rdh | /**
* Creates a new data set that contains a sequence of numbers. The data set will be created in
* parallel, so there is no guarantee about the order of the elements.
*
* @param from
* The number to start at (inclusive).
* @param to
* The number to stop at (inclusive).
* @return A DataSet, containing all n... | 3.26 |
flink_ExecutionEnvironment_fromElements_rdh | /**
* Creates a new data set that contains the given elements. The framework will determine the
* type according to the based type user supplied. The elements should be the same or be the
* subclass to the based type. The sequence of elements must not be empty. Note that this
* operation will result in a non-parall... | 3.26 |
flink_ExecutionEnvironment_registerJobListener_rdh | /**
* Register a {@link JobListener} in this environment. The {@link JobListener} will be notified
* on specific job status changed.
*/
@PublicEvolving
public void registerJobListener(JobListener jobListener) {
checkNotNull(jobListener, "JobListener cannot be null");
jobListen... | 3.26 |
flink_ExecutionEnvironment_getConfig_rdh | // --------------------------------------------------------------------------------------------
// Properties
// --------------------------------------------------------------------------------------------
/**
* Gets the config object that defines execution parameters.
*
* @return The environment's execution configu... | 3.26 |
flink_ExecutionEnvironment_addDefaultKryoSerializer_rdh | /**
* Adds a new Kryo default serializer to the Runtime.
*
* @param type
* The class of the types serialized with the given serializer.
* @param serializerClass
* The class of the serializer to use.
*/
public void addDefaultKryoSerializer(Class<?> type, Class<? extends Serializer<?>> serializerClass) {confi... | 3.26 |
flink_ExecutionEnvironment_getExecutionPlan_rdh | /**
* Creates the plan with which the system will execute the program, and returns it as a String
* using a JSON representation of the execution data flow graph.
*
* @return The execution plan of the program, as a JSON String.
* @throws Exception
* Thrown, if the compiler could not be instantiated.
*/
public S... | 3.26 |
flink_ExecutionEnvironment_registerDataSink_rdh | /**
* Adds the given sink to this environment. Only sinks that have been added will be executed
* once the {@link #execute()} or {@link #execute(String)} method is called.
*
* @param sink
* The sink to add for execution.
*/
@Internal
void registerDataSink(DataSink<?> sink) {this.sinks.add(sink);
} | 3.26 |
flink_ExecutionEnvironment_readTextFile_rdh | /**
* Creates a {@link DataSet} that represents the Strings produced by reading the given file line
* wise. The {@link java.nio.charset.Charset} with the given name will be used to read the
* files.
*
* @param filePath
* The path of the file, as a URI (e.g., "file:///some/local/file" or
* "hdfs://host:port/f... | 3.26 |
flink_ExecutionEnvironment_m1_rdh | /**
* Creates a new data set that contains elements in the iterator. The iterator is splittable,
* allowing the framework to create a parallel data source that returns the elements in the
* iterator.
*
* <p>Because the iterator will remain unmodified until the actual execution happens, the type
* of data returned... | 3.26 |
flink_ExecutionEnvironment_createInput_rdh | /**
* Generic method to create an input DataSet with in {@link InputFormat}. The {@link DataSet}
* will not be immediately created - instead, this method returns a {@link DataSet} that will be
* lazily created from the input format once the program is executed.
*
* <p>The {@link DataSet} is typed to the given Type... | 3.26 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.