name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_CrossOperator_projectTuple12_rdh | /**
* Projects a pair of crossed elements to a {@link Tuple} with the previously selected
* fields.
*
* @return The projected data set.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11> ProjectCross<I1, I2, Tuple12<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9,
T10, T11>> proje... | 3.26 |
flink_CrossOperator_projectTuple23_rdh | /**
* Projects a pair of crossed elements to a {@link Tuple} with the previously selected
* fields.
*
* @return The projected data set.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12,
T13, T14, T15, T16, T17, T18, T19, T20, T21, T22> ProjectCross<I1, I2, Tuple23<T0, ... | 3.26 |
flink_CrossOperator_projectTuple4_rdh | /**
* Projects a pair of crossed elements to a {@link Tuple} with the previously selected
* fields.
*
* @return The projected data set.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3> ProjectCross<I1, I2, Tuple4<T0, T1, T2, T3>> projectTuple4() {
TypeInformation<?>[] fTypes = m3(fieldIndexes);
T... | 3.26 |
flink_CrossOperator_projectTuple16_rdh | /**
* Projects a pair of crossed elements to a {@link Tuple} with the previously selected
* fields.
*
* @return The projected data set.
* @see Tuple
* @see DataSet
... | 3.26 |
flink_CrossOperator_projectTuple8_rdh | /**
* Projects a pair of crossed elements to a {@link Tuple} with the previously selected
* fields.
*
* @return The projected data set.
* @see Tuple
* @see DataSet
*/
public <T0, T1,
T2, T3, T4, T5, T6, T7> ProjectCross<I1, I2, Tuple8<T0, T1, T2, T3, T4, T5, T6, T7>> projectTuple8() {
TypeInformation<?>[] fT... | 3.26 |
flink_CrossOperator_with_rdh | /**
* Finalizes a Cross transformation by applying a {@link CrossFunction} to each pair of
* crossed elements.
*
* <p>Each CrossFunction call returns exactly one element.
*
* @param function
* The CrossFunction that is called for each pair of crossed elements.
* @return An CrossOperator that represents the cr... | 3.26 |
flink_CrossOperator_m3_rdh | // END_OF_TUPLE_DEPENDENT_CODE
// -----------------------------------------------------------------------------------------
private TypeInformation<?>[]
m3(int[] fields) {
TypeInformation<?>[] fieldTypes = new TypeInformation[fields.length];
... | 3.26 |
flink_MapTypeInfo_isBasicType_rdh | // ------------------------------------------------------------------------
// TypeInformation implementation
// ------------------------------------------------------------------------
@Override
public boolean isBasicType() {
return false;
} | 3.26 |
flink_MapTypeInfo_getKeyTypeInfo_rdh | // ------------------------------------------------------------------------
// MapTypeInfo specific properties
// ------------------------------------------------------------------------
/**
* Gets the type information for the keys in the map
*/
public TypeInformation<K> getKeyTypeInfo() { return keyTypeInfo;
} | 3.26 |
flink_ChainedFlatMapDriver_getStub_rdh | // --------------------------------------------------------------------------------------------
public Function getStub() {
return this.mapper;
} | 3.26 |
flink_ChainedFlatMapDriver_collect_rdh | // --------------------------------------------------------------------------------------------
@Override
public void collect(IT record) {
try {
this.numRecordsIn.inc();
this.mapper.flatMap(record, this.outputCollector);
} catch (Exception ex) {throw new ExceptionInChainedStubException(this.task... | 3.26 |
flink_ChainedFlatMapDriver_setup_rdh | // --------------------------------------------------------------------------------------------
@Override
public void setup(AbstractInvokable parent) {
@SuppressWarnings("unchecked")
final FlatMapFunction<IT, OT> mapper = BatchTask.instantiateUserCode(this.config, userCodeClassLoader, FlatMapFunction.class);
... | 3.26 |
flink_TableDescriptor_toCatalogTable_rdh | // ---------------------------------------------------------------------------------------------
/**
* Converts this descriptor into a {@link CatalogTable}.
*/
public CatalogTable toCatalogTable() {
final Schema schema =
getSchema().orElseThrow(() -> new ValidationException((("Missing schema in TableDescrip... | 3.26 |
flink_TableDescriptor_getSchema_rdh | // ---------------------------------------------------------------------------------------------
public Optional<Schema> getSchema() {
return Optional.ofNullable(schema);
} | 3.26 |
flink_TableDescriptor_option_rdh | /**
* Sets the given option on the table.
*
* <p>Option keys must be fully specified. When defining options for a {@link Format
* format}, use {@link #format(FormatDescriptor)} instead.
*
* <p>Example:
*
* <pre>{@code TableDescriptor.forCon... | 3.26 |
flink_TableDescriptor_build_rdh | /**
* Returns an immutable instance of {@link TableDescriptor}.
*/
public TableDescriptor build() {
return new TableDescriptor(schema, options, partitionKeys, comment);
} | 3.26 |
flink_TableDescriptor_toString_rdh | // ---------------------------------------------------------------------------------------------
@Override
public String toString() {
final String escapedPartitionKeys = partitionKeys.stream().map(EncodingUtils::escapeIdentifier).collect(Collectors.joining(", "));
final String partitionedBy = (!partitionKeys.i... | 3.26 |
flink_TableDescriptor_format_rdh | /**
* Defines the format to be used for this table.
*
* <p>Note that not every connector requires a format to be specified, while others may use
* multiple formats.
*
* <p>Options of the provided {@param formatDescriptor} are automatically prefixed. For
* example,
*
* <pre>{@code descriptorBuilder.format(Kafka... | 3.26 |
flink_TableDescriptor_comment_rdh | /**
* Define the comment for this table.
*/
public Builder comment(@Nullable
String comment) {
this.comment = comment;
return this;
} | 3.26 |
flink_TableDescriptor_schema_rdh | /**
* Define the schema of the {@link TableDescriptor}.
*
* <p>The schema is typically required. It is optional only in cases where the schema can be
* inferred, e.g. {@link Table#insertInto(TableDescriptor)}.
*/
public Builder schema(@Nullable
Schema schema) {
this.schema = schema;
return this;
} | 3.26 |
flink_TableDescriptor_m0_rdh | /**
* Creates a new {@link Builder} for a managed table.
*
* @deprecated This method will be removed soon. Please see FLIP-346 for more details.
*/
@Deprecatedpublic static Builder m0() {
return new Builder();
} | 3.26 |
flink_TableDescriptor_toBuilder_rdh | /**
* Converts this immutable instance into a mutable {@link Builder}.
*/public Builder toBuilder() {
return new Builder(this);} | 3.26 |
flink_TableDescriptor_forConnector_rdh | /**
* Creates a new {@link Builder} for a table using the given connector.
*
* @param connector
* The factory identifier for the connector.
*/public static Builder forConnector(String connector) {
Preconditions.checkNotNull(connector, "Table descriptors require a connector identifier.");
final Builder de... | 3.26 |
flink_TableDescriptor_partitionedBy_rdh | /**
* Define which columns this table is partitioned by.
*/
public Builder partitionedBy(String... partitionKeys) {
this.partitionKeys.addAll(Arrays.asList(partitionKeys));
return this;
} | 3.26 |
flink_FloatParser_parseField_rdh | /**
* Static utility to parse a field of type float from a byte sequence that represents text
* characters (such as when read from a file stream).
*
* @param bytes
* The bytes containing the text data that should be parsed.
* @param startPos
* The offset to start the parsing.
* @param length
* The length... | 3.26 |
flink_UploadThrottle_hasCapacity_rdh | /**
* Test whether some capacity is available.
*/
public boolean hasCapacity() {
return inFlightBytesCounter < maxBytesInFlight;
} | 3.26 |
flink_UploadThrottle_seizeCapacity_rdh | /**
* Seize <b>bytes</b> capacity. It is the caller responsibility to ensure at least some capacity
* {@link #hasCapacity() is available}. <strong>After</strong> this call, the caller is allowed
* to actually use the seized capacity. When the capacity is not needed anymore, the caller is
* required to {@link #relea... | 3.26 |
flink_UploadThrottle_releaseCapacity_rdh | /**
* Release previously {@link #seizeCapacity(long) seized} capacity. Called by {@link BatchingStateChangeUploadScheduler} (IO thread).
*/
public void releaseCapacity(long bytes) {
inFlightBytesCounter -= bytes;
} | 3.26 |
flink_TypeInferenceExtractor_forAsyncTableFunction_rdh | /**
* Extracts a type inference from a {@link AsyncTableFunction}.
*/
public static TypeInference forAsyncTableFunction(DataTypeFactory typeFactory, Class<? extends AsyncTableFunction<?>> function) {
final FunctionMappingExtractor mappingExtractor = new FunctionMappingExtractor(typeFactory, function, UserD... | 3.26 |
flink_TypeInferenceExtractor_forTableFunction_rdh | /**
* Extracts a type inference from a {@link TableFunction}.
*/
public static TypeInference forTableFunction(DataTypeFactory typeFactory, Class<? extends TableFunction<?>> function) {
final FunctionMappingExtractor mappingExtractor = new FunctionMappingExtractor(typeFactory, function, UserDef... | 3.26 |
flink_TypeInferenceExtractor_forProcedure_rdh | /**
* Extracts a type in inference from a {@link Procedure}.
*/
public static TypeInference forProcedure(DataTypeFactory typeFactory, Class<? extends Procedure> procedure) {
final ProcedureMappingExtractor mappingExtractor = new ProcedureMappingExtractor(typeFactory, procedure, ProcedureDefinition.PROCEDURE_CALL... | 3.26 |
flink_TypeInferenceExtractor_forTableAggregateFunction_rdh | /**
* Extracts a type inference from a {@link TableAggregateFunction}.
*/ public static TypeInference forTableAggregateFunction(DataTypeFactory typeFactory, Class<? extends TableAggregateFunction<?, ?>> function) {
final FunctionMappingExtractor mappingExtractor =
new FunctionMappingExtractor(typeFactory, fu... | 3.26 |
flink_TypeInferenceExtractor_forAggregateFunction_rdh | /**
* Extracts a type inference from a {@link AggregateFunction}.
*/
public static TypeInference forAggregateFunction(DataTypeFactory typeFactory, Class<? extends AggregateFunction<?, ?>> function) {
final FunctionMappingExtractor mappingExtractor = new FunctionMappingExtractor(typeFactory, function, UserDefined... | 3.26 |
flink_TypeInferenceExtractor_m0_rdh | /**
* Extracts a type inference from a {@link ScalarFunction}.
*/
public static TypeInference m0(DataTypeFactory typeFactory, Class<? extends ScalarFunction> function) {
final FunctionMappingExtractor mappingExtractor = new
FunctionMappingExtractor(typeFactory, function, UserDefinedFunctionHelper.SCALAR_EVAL,... | 3.26 |
flink_CheckpointStatsCounts_incrementCompletedCheckpoints_rdh | /**
* Increments the number of successfully completed checkpoints.
*
* <p>It is expected that this follows a previous call to {@link #incrementInProgressCheckpoints()}.
*/
void incrementCompletedCheckpoints() {
if (canDecrementOfInProgressCheckpointsNumber()) {
numInProgressCheckpoints--;
}
numC... | 3.26 |
flink_CheckpointStatsCounts_getNumberOfRestoredCheckpoints_rdh | /**
* Returns the number of restored checkpoints.
*
* @return Number of restored checkpoints.
*/public long getNumberOfRestoredCheckpoints() {
return numRestoredCheckpoints;
} | 3.26 |
flink_CheckpointStatsCounts_getNumberOfFailedCheckpoints_rdh | /**
* Returns the number of failed checkpoints.
*
* @return Number of failed checkpoints.
*/
public long getNumberOfFailedCheckpoints() {
return numFailedCheckpoints;
} | 3.26 |
flink_CheckpointStatsCounts_incrementFailedCheckpoints_rdh | /**
* Increments the number of failed checkpoints.
*
* <p>It is expected that this follows a previous call to {@link #incrementInProgressCheckpoints()}.
*/
void incrementFailedCheckpoints() {
if (canDecrementOfInProgressCheckpointsNumber()) {
numInProgressCheckpoints--;
}
numFailedCheckpoints++;... | 3.26 |
flink_CheckpointStatsCounts_incrementRestoredCheckpoints_rdh | /**
* Increments the number of restored checkpoints.
*/
void incrementRestoredCheckpoints() {
numRestoredCheckpoints++;
} | 3.26 |
flink_CheckpointStatsCounts_createSnapshot_rdh | /**
* Creates a snapshot of the current state.
*
* @return Snapshot of the current state.
*/CheckpointStatsCounts createSnapshot() {
return new CheckpointStatsCounts(numRestoredCheckpoints, numTotalCheckpoints, numInProgressCheckpoints, numCompletedCheckpoints, numFailedCheckpoints);
} | 3.26 |
flink_CheckpointStatsCounts_getTotalNumberOfCheckpoints_rdh | /**
* Returns the total number of checkpoints (in progress, completed, failed).
*
* @return Total number of checkpoints.
*/
public long getTotalNumberOfCheckpoints() {
return numTotalCheckpoints;
} | 3.26 |
flink_CheckpointStatsCounts_incrementInProgressCheckpoints_rdh | /**
* Increments the number of total and in progress checkpoints.
*/
void incrementInProgressCheckpoints() {
numInProgressCheckpoints++;
numTotalCheckpoints++;
} | 3.26 |
flink_StreamingSink_sink_rdh | /**
* Create a sink from file writer. Decide whether to add the node to commit partitions according
* to options.
*/public static DataStreamSink<?> sink(ProviderContext providerContext, DataStream<PartitionCommitInfo> writer, Path locationPath, ObjectIdentifier identifier, List<String> partitionKeys, TableMetaStoreF... | 3.26 |
flink_StreamingSink_compactionWriter_rdh | /**
* Create a file writer with compaction operators by input stream. In addition, it can emit
* {@link PartitionCommitInfo} to down stream.
*/
public static <T> DataStream<PartitionCommitInfo> compactionWriter(ProviderContext providerContext, DataStream<T> inputStream, long bucketCheckInterval, BucketsBuilder<T, ... | 3.26 |
flink_StreamingSink_m0_rdh | /**
* Create a file writer by input stream. This is similar to {@link StreamingFileSink}, in
* addition, it can emit {@link PartitionCommitInfo} to down stream.
*/
public static <T> DataStream<PartitionCommitInfo> m0(ProviderContext providerContext, DataStream<T> inputStream, long bucketCheckInterval, BucketsBuilder... | 3.26 |
flink_FlatMapIterator_flatMap_rdh | // --------------------------------------------------------------------------------------------
/**
* Delegates calls to the {@link #flatMap(Object)} method.
*/
@Override
public final void flatMap(IN value, Collector<OUT> out) throws Exception {for (Iterator<OUT> iter = flatMap(value); iter.hasNext();) {out.collect... | 3.26 |
flink_KubernetesSessionCli_repStep_rdh | /**
* Check whether need to continue or kill the cluster.
*
* @param in
* input buffer reader
* @return f0, whether need to continue read from input. f1, whether need to kill the cluster.
*/
private Tuple2<Boolean, Boolean> repStep(BufferedReader in) throws IOException, InterruptedException {
final long sta... | 3.26 |
flink_BufferManager_releaseAll_rdh | /**
* The floating buffer is recycled to local buffer pool directly, and the exclusive buffer
* will be gathered to return to global buffer pool later.
*
* @param exclusiveSegments
* The list that we will add exclusive segments into.
*/
void releaseAll(List<MemorySegment> exclusiveSegments) {
Buffer buffer;
whi... | 3.26 |
flink_BufferManager_unsynchronizedGetNumberOfRequiredBuffers_rdh | // ------------------------------------------------------------------------
// Getter properties
// ------------------------------------------------------------------------
@VisibleForTesting
int unsynchronizedGetNumberOfRequiredBuffers() {return numRequiredBuffers;
} | 3.26 |
flink_BufferManager_requestFloatingBuffers_rdh | /**
* Requests floating buffers from the buffer pool based on the given required amount, and
* returns the actual requested amount. If the required amount is not fully satisfied, it will
* register as a listener.
*/
int requestFloatingBuffers(int numRequired) {
int numRequestedBuffers = 0;
synchronized(buff... | 3.26 |
flink_BufferManager_addExclusiveBuffer_rdh | /**
* Adds an exclusive buffer (back) into the queue and releases one floating buffer if the
* number of available buffers in queue is more than the required amount. If floating buffer
* is released, the total amount of available buffers after adding this exclusive buffer has
* not changed, and no new buffers are a... | 3.26 |
flink_BufferManager_releaseAllBuffers_rdh | /**
* Recycles all the exclusive and floating buffers from the given buffer queue.
*/
void releaseAllBuffers(ArrayDeque<Buffer> buffers) throws IOException {
// Gather all exclusive buffers and recycle them to global pool in batch, because
// we do not want to trigger redistribution of buffers after each recycle.
fi... | 3.26 |
flink_BufferManager_requestBuffer_rdh | // ------------------------------------------------------------------------
// Buffer request
// ------------------------------------------------------------------------
@Nullable
Buffer requestBuffer() {
synchronized(bufferQueue) {
// decrease the number of buffers require to avoid the possibility of
... | 3.26 |
flink_BufferManager_requestExclusiveBuffers_rdh | /**
* Requests exclusive buffers from the provider.
*/
void requestExclusiveBuffers(int numExclusiveBuffers) throws IOException {
checkArgument(numExclusiveBuffers >= 0, "Num exclusive buffers must be non-negative.");
if (numExclusiveBuffers == 0) {
return;
}
Collection<MemorySegment> segmen... | 3.26 |
flink_BufferManager_recycle_rdh | // ------------------------------------------------------------------------
// Buffer recycle
// ------------------------------------------------------------------------
/**
* Exclusive buffer is recycled to this channel manager directly and it may trigger return extra
* floating buffer based on <tt>numRequiredBuffer... | 3.26 |
flink_BufferManager_notifyBufferAvailable_rdh | // ------------------------------------------------------------------------
// Buffer listener notification
// ------------------------------------------------------------------------
/**
* The buffer pool notifies this listener of an available floating buffer. If the listener is
* released or currently does not need... | 3.26 |
flink_TaskManagerExceptionUtils_tryEnrichTaskManagerError_rdh | /**
* Tries to enrich the passed exception or its causes with additional information.
*
* <p>This method improves error messages for direct and metaspace {@link OutOfMemoryError}. It
* adds descriptions about possible causes and ways of resolution.
*
* @param root
* The Throwable of which the cause tree shall ... | 3.26 |
flink_MultipleIdsMessageAcknowledgingSourceBase_snapshotState_rdh | // ------------------------------------------------------------------------
// Checkpointing the data
// ------------------------------------------------------------------------
@Override
public void snapshotState(FunctionSnapshotContext context) throws Exception {
f0.add(new Tuple2<>(context.getCheckpointId(), ses... | 3.26 |
flink_MultipleIdsMessageAcknowledgingSourceBase_acknowledgeIDs_rdh | // ------------------------------------------------------------------------
// ID Checkpointing
// ------------------------------------------------------------------------
/**
* Acknowledges the session ids.
*
* @param checkpointId
* The id of the current checkout to acknowledge ids for.
* @param uniqueIds
* ... | 3.26 |
flink_TextOutputFormat_toString_rdh | // --------------------------------------------------------------------------------------------
@Override
public String toString() {
return (("TextOutputFormat (" + getOutputFilePath()) + ") - ") + this.charsetName;
} | 3.26 |
flink_TextOutputFormat_open_rdh | // --------------------------------------------------------------------------------------------
@Override
public void open(int taskNumber, int numTasks) throws IOException {
super.open(taskNumber, numTasks);
try {
this.charset = Charset.forName(charsetName);
} catch (IllegalCharsetNameException e) {... | 3.26 |
flink_CustomizedConverter_checkArgumentNumber_rdh | // ---------------------------------------------------------------------------------------------
protected static void checkArgumentNumber(CallExpression call, int... validArgumentCounts) {boolean hasValidArgumentCount = false;
for (int argumentCount : validArgumentCounts) {
if (call.getChildren().size()... | 3.26 |
flink_HiveTableMetaStoreFactory_listDataFileRecursively_rdh | /**
* List data files recursively.
*/
private List<FileStatus> listDataFileRecursively(FileSystem fileSystem, Path f) throws IOException {
List<FileStatus> fileStatusList = new ArrayList<>();
for (FileStatus fileStatus : fileSystem.listStatus(f)) {
if (fileStatus.isDir() && (!isStagingDir(fileStatus.g... | 3.26 |
flink_CollectionInputFormat_toString_rdh | // --------------------------------------------------------------------------------------------
@Override
public String toString() {
StringBuilder sb = new StringBuilder();
sb.append('[');
int num = 0;
for (T e : dataSet) {
sb.append(e);
if (num != (dataSet.size() - 1)) {
... | 3.26 |
flink_CollectionInputFormat_writeObject_rdh | // --------------------------------------------------------------------------------------------
private void writeObject(ObjectOutputStream out) throws IOException {
out.defaultWriteObject();
final int size = dataSet.size();
out.writeInt(size);
if (size > 0) {
DataOutputViewStreamWrapper wrapper... | 3.26 |
flink_CollectionInputFormat_checkCollection_rdh | // --------------------------------------------------------------------------------------------
public static <X> void checkCollection(Collection<X> elements, Class<X> viewedAs) {
if ((elements == null) || (viewedAs == null)) {
throw new NullPointerException();
}
for (X elem : elements) {
i... | 3.26 |
flink_CallBindingCallContext_m0_rdh | // --------------------------------------------------------------------------------------------
@Nullable
private static DataType m0(SqlCallBinding binding, @Nullable
RelDataType returnType) {
if (((returnType == null) || returnType.equals(binding.getValidator().getUnknownType())) || (returnType.getSqlTypeName() ==... | 3.26 |
flink_TimestampUtil_createVectorFromConstant_rdh | // creates a Hive ColumnVector of constant timestamp value
public static ColumnVector createVectorFromConstant(int batchSize, Object value) {
if (hiveTSColVectorClz != null) {
return OrcTimestampColumnVector.createFromConstant(batchSize, value);
} else {
return OrcLegacyTimestampColumnVector.cr... | 3.26 |
flink_TimestampUtil_isHiveTimestampColumnVector_rdh | // whether a ColumnVector is the new TimestampColumnVector
public static boolean isHiveTimestampColumnVector(ColumnVector vector) {return (hiveTSColVectorClz != null) && hiveTSColVectorClz.isAssignableFrom(vector.getClass());
} | 3.26 |
flink_EmptyMutableObjectIterator_get_rdh | /**
* Gets a singleton instance of the empty iterator.
*
* @param <E>
* The type of the objects (not) returned by the iterator.
* @return An instance of the iterator.
*/
public static <E> MutableObjectIterator<E> get() {
@SuppressWarnings("unchecked")
MutableObjectIterator<E> iter = ((MutableObjectItera... | 3.26 |
flink_Tuple0_getArity_rdh | // ------------------------------------------------------------------------
@Overridepublic int getArity() {
return 0;
} | 3.26 |
flink_Tuple0_toString_rdh | // -------------------------------------------------------------------------------------------------
// standard utilities
// -------------------------------------------------------------------------------------------------
/**
* Creates a string representation of the tuple in the form "()".
*
* @return The string r... | 3.26 |
flink_Tuple0_equals_rdh | /**
* Deep equality for tuples by calling equals() on the tuple members.
*
* @param o
* the object checked for equality
* @return true if this is equal to o.
*/
@Override
public boolean equals(Object o) {
return (this == o) || (o instanceof Tuple0);
} | 3.26 |
flink_Tuple0_readResolve_rdh | // singleton deserialization
private Object readResolve() throws ObjectStreamException {
return INSTANCE; } | 3.26 |
flink_JobSchedulingPlan_empty_rdh | /**
* Create an empty {@link JobSchedulingPlan} with no information about vertices or allocations.
*/public static JobSchedulingPlan empty() {
return new JobSchedulingPlan(VertexParallelism.empty(), Collections.emptyList());
} | 3.26 |
flink_AbstractHeapVector_getDictionaryIds_rdh | /**
* Returns the underlying integer column for ids of dictionary.
*/
@Override
public HeapIntVector getDictionaryIds() {
return dictionaryIds;
} | 3.26 |
flink_AbstractHeapVector_reset_rdh | /**
* Resets the column to default state. - fills the isNull array with false. - sets noNulls to
* true.
*/
@Override
public void reset()
{
if (!noNulls) {
Arrays.fill(isNull, false);
}
noNulls = true;
} | 3.26 |
flink_SqlColumnPosSpec_symbol_rdh | /**
* Creates a parse-tree node representing an occurrence of this keyword at a particular position
* in the parsed text.
*/
public SqlLiteral symbol(SqlParserPos
pos) {
return SqlLiteral.createSymbol(this, pos);
} | 3.26 |
flink_MasterState_name_rdh | // ------------------------------------------------------------------------
public String name() {
return name;
} | 3.26 |
flink_MasterState_m0_rdh | // ------------------------------------------------------------------------
@Override
public String m0() {
return (((("name: " + name) + " ; version: ") + version) + " ; bytes: ") + Arrays.toString(bytes);
} | 3.26 |
flink_LegacySourceTransformationTranslator_translateForBatchInternal_rdh | /**
* A {@link TransformationTranslator} for the {@link LegacySourceTransformation}.
*
* @param <OUT>
* The type of the elements that the {@link LegacySourceTransformation} we are
* translating is producing.
*/
@Internalpublic class LegacySourceTransformationTranslator<OUT> extends SimpleTransformationTransl... | 3.26 |
flink_FlinkPreparingTableBase_getStatistic_rdh | // ~ Methods ----------------------------------------------------------------
/**
* Returns the statistic of this table.
*/
public FlinkStatistic getStatistic() {
return this.statistic;
} | 3.26 |
flink_FlinkPreparingTableBase_getNames_rdh | /**
* Returns the table path in the {@link RelOptSchema}. Different with {@link #getQualifiedName()}, the latter is mainly used for table digest.
*/
public List<String> getNames() {
return names;
} | 3.26 |
flink_FlinkPreparingTableBase_getRowType_rdh | /**
* Returns the type of rows returned by this table.
*/
public RelDataType getRowType() {
return f0;
} | 3.26 |
flink_FlinkPreparingTableBase_getDistribution_rdh | /**
* Returns a description of the physical distribution of the rows in this table.
*
* @see org.apache.calcite.rel.metadata.RelMetadataQuery#distribution
*/
public RelDistribution getDistribution() {
return null;
}
/**
* Returns whether the given columns are a key or a superset of a unique key of this tabl... | 3.26 |
flink_FlinkPreparingTableBase_m0_rdh | /**
* Obtains whether the ordinal column has a default value, which is not supported now.
*
* @param rowType
* Row type of field
* @param ordinal
* Index of the given column
* @param initializerContext
* Context for {@link org.apache.calcite.sql2rel.InitializerExpressionFactory}
* @return true if the col... | 3.26 |
flink_FlinkPreparingTableBase_getExpression_rdh | /**
* Generates code for this table, which is not supported now.
*
* @param clazz
* The desired collection class, for example {@link org.apache.calcite.linq4j.Queryable}
*/
public Expression getExpression(Class clazz) {
throw new UnsupportedOperationException();
} | 3.26 |
flink_FlinkPreparingTableBase_getRowCount_rdh | /**
* Returns an estimate of the number of rows in the table.
*/
public double getRowCount() {
Double rowCnt
= getStatistic().getRowCount();
return rowCnt
== null ? DEFAULT_ROWCOUNT
: rowCnt;
} | 3.26 |
flink_FlinkPreparingTableBase_getMonotonicity_rdh | /**
* Obtains whether a given column is monotonic.
*
* @param columnName
* Column name
* @return True if the given column is monotonic
*/
public SqlMonotonicity getMonotonicity(String columnName) {
return SqlMonotonicity.NOT_MONOTONIC;
} | 3.26 |
flink_FlinkPreparingTableBase_explainSourceAsString_rdh | // ~ Tools ------------------------------------------------------------------
/**
* Returns the digest of the {@link TableSource} instance.
*/
protected List<String> explainSourceAsString(TableSource<?> ts) {
String tsDigest = ts.explainSource();
if (!Strings.isNullOrEmpty(tsDigest)) {
return Immutabl... | 3.26 |
flink_FlinkPreparingTableBase_getCollationList_rdh | /**
* Returns a description of the physical ordering (or orderings) of the rows returned from this
* table.
*
* @see org.apache.calcite.rel.metadata.RelMetadataQuery#collations(RelNode)
*/
public List<RelCollation> getCollationList() {
return ImmutableList.of();} | 3.26 |
flink_FlinkPreparingTableBase_isTemporal_rdh | /**
* We recognize all tables in FLink are temporal as they are changeable.
*/
public boolean isTemporal() {
return true;
} | 3.26 |
flink_FlinkPreparingTableBase_m1_rdh | /**
* Returns unique keySets of current table.
*/
public Optional<Set<ImmutableBitSet>> m1() {Set<? extends Set<String>> uniqueKeys = statistic.getUniqueKeys();
if (uniqueKeys == null) {
return Optional.empty();
} else if (uniqueKeys.size() == 0) {
return Optional.of(Immuta... | 3.26 |
flink_FlinkPreparingTableBase_getAllowedAccess_rdh | /**
* Obtains the access type of the table.
*
* @return all access types including SELECT/UPDATE/INSERT/DELETE
*/
public SqlAccessType getAllowedAccess() {
return SqlAccessType.ALL;
} | 3.26 |
flink_LatencyMarker_equals_rdh | // ------------------------------------------------------------------------
@Override
public boolean equals(Object o) {
if (this == o) {return true;
}
if ((o == null) || (getClass()
!= o.getClass())) {
return false;
}
LatencyMarker that = ((LatencyMarker) (o));
if (markedTime != t... | 3.26 |
flink_LatencyMarker_getMarkedTime_rdh | /**
* Returns the timestamp marked by the LatencyMarker.
*/
public long getMarkedTime() {
return markedTime;
} | 3.26 |
flink_BiFunctionWithException_unchecked_rdh | /**
* Convert at {@link BiFunctionWithException} into a {@link BiFunction}.
*
* @param biFunctionWithException
* function with exception to convert into a function
* @param <A>
* input type
* @param <B>
* output type
* @return {@link BiFunction} which throws all checked exception as an unchecked exceptio... | 3.26 |
flink_OperationExecutor_runClusterAction_rdh | /**
* Retrieves the {@link ClusterClient} from the session and runs the given {@link ClusterAction}
* against it.
*
* @param configuration
* the combined configuration of {@code sessionConf} and {@code executionConfig}.
* @param handle
* the specified operation handle
* @param clusterAction
* the cluster... | 3.26 |
flink_FlinkCalciteCatalogReader_toPreparingTable_rdh | /**
* Translate this {@link CatalogSchemaTable} into Flink source table.
*/
private static FlinkPreparingTableBase toPreparingTable(RelOptSchema relOptSchema, List<String> names, RelDataType rowType, CatalogSchemaTable schemaTable) {
final ResolvedCatalogB... | 3.26 |
flink_FlinkCalciteCatalogReader_isLegacySourceOptions_rdh | /**
* Checks whether the {@link CatalogTable} uses legacy connector source options.
*/
private static boolean isLegacySourceOptions(CatalogSchemaTable schemaTable) {
// normalize option keys
DescriptorProperties properties = new
DescriptorProperties(true);
properties.putProperties(schemaTable.getConte... | 3.26 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.