name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
flink_CrossOperator_projectTuple12_rdh
/** * Projects a pair of crossed elements to a {@link Tuple} with the previously selected * fields. * * @return The projected data set. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11> ProjectCross<I1, I2, Tuple12<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11>> proje...
3.26
flink_CrossOperator_projectTuple23_rdh
/** * Projects a pair of crossed elements to a {@link Tuple} with the previously selected * fields. * * @return The projected data set. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21, T22> ProjectCross<I1, I2, Tuple23<T0, ...
3.26
flink_CrossOperator_projectTuple4_rdh
/** * Projects a pair of crossed elements to a {@link Tuple} with the previously selected * fields. * * @return The projected data set. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3> ProjectCross<I1, I2, Tuple4<T0, T1, T2, T3>> projectTuple4() { TypeInformation<?>[] fTypes = m3(fieldIndexes); T...
3.26
flink_CrossOperator_projectTuple16_rdh
/** * Projects a pair of crossed elements to a {@link Tuple} with the previously selected * fields. * * @return The projected data set. * @see Tuple * @see DataSet ...
3.26
flink_CrossOperator_projectTuple8_rdh
/** * Projects a pair of crossed elements to a {@link Tuple} with the previously selected * fields. * * @return The projected data set. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5, T6, T7> ProjectCross<I1, I2, Tuple8<T0, T1, T2, T3, T4, T5, T6, T7>> projectTuple8() { TypeInformation<?>[] fT...
3.26
flink_CrossOperator_with_rdh
/** * Finalizes a Cross transformation by applying a {@link CrossFunction} to each pair of * crossed elements. * * <p>Each CrossFunction call returns exactly one element. * * @param function * The CrossFunction that is called for each pair of crossed elements. * @return An CrossOperator that represents the cr...
3.26
flink_CrossOperator_m3_rdh
// END_OF_TUPLE_DEPENDENT_CODE // ----------------------------------------------------------------------------------------- private TypeInformation<?>[] m3(int[] fields) { TypeInformation<?>[] fieldTypes = new TypeInformation[fields.length]; ...
3.26
flink_MapTypeInfo_isBasicType_rdh
// ------------------------------------------------------------------------ // TypeInformation implementation // ------------------------------------------------------------------------ @Override public boolean isBasicType() { return false; }
3.26
flink_MapTypeInfo_getKeyTypeInfo_rdh
// ------------------------------------------------------------------------ // MapTypeInfo specific properties // ------------------------------------------------------------------------ /** * Gets the type information for the keys in the map */ public TypeInformation<K> getKeyTypeInfo() { return keyTypeInfo; }
3.26
flink_ChainedFlatMapDriver_getStub_rdh
// -------------------------------------------------------------------------------------------- public Function getStub() { return this.mapper; }
3.26
flink_ChainedFlatMapDriver_collect_rdh
// -------------------------------------------------------------------------------------------- @Override public void collect(IT record) { try { this.numRecordsIn.inc(); this.mapper.flatMap(record, this.outputCollector); } catch (Exception ex) {throw new ExceptionInChainedStubException(this.task...
3.26
flink_ChainedFlatMapDriver_setup_rdh
// -------------------------------------------------------------------------------------------- @Override public void setup(AbstractInvokable parent) { @SuppressWarnings("unchecked") final FlatMapFunction<IT, OT> mapper = BatchTask.instantiateUserCode(this.config, userCodeClassLoader, FlatMapFunction.class); ...
3.26
flink_TableDescriptor_toCatalogTable_rdh
// --------------------------------------------------------------------------------------------- /** * Converts this descriptor into a {@link CatalogTable}. */ public CatalogTable toCatalogTable() { final Schema schema = getSchema().orElseThrow(() -> new ValidationException((("Missing schema in TableDescrip...
3.26
flink_TableDescriptor_getSchema_rdh
// --------------------------------------------------------------------------------------------- public Optional<Schema> getSchema() { return Optional.ofNullable(schema); }
3.26
flink_TableDescriptor_option_rdh
/** * Sets the given option on the table. * * <p>Option keys must be fully specified. When defining options for a {@link Format * format}, use {@link #format(FormatDescriptor)} instead. * * <p>Example: * * <pre>{@code TableDescriptor.forCon...
3.26
flink_TableDescriptor_build_rdh
/** * Returns an immutable instance of {@link TableDescriptor}. */ public TableDescriptor build() { return new TableDescriptor(schema, options, partitionKeys, comment); }
3.26
flink_TableDescriptor_toString_rdh
// --------------------------------------------------------------------------------------------- @Override public String toString() { final String escapedPartitionKeys = partitionKeys.stream().map(EncodingUtils::escapeIdentifier).collect(Collectors.joining(", ")); final String partitionedBy = (!partitionKeys.i...
3.26
flink_TableDescriptor_format_rdh
/** * Defines the format to be used for this table. * * <p>Note that not every connector requires a format to be specified, while others may use * multiple formats. * * <p>Options of the provided {@param formatDescriptor} are automatically prefixed. For * example, * * <pre>{@code descriptorBuilder.format(Kafka...
3.26
flink_TableDescriptor_comment_rdh
/** * Define the comment for this table. */ public Builder comment(@Nullable String comment) { this.comment = comment; return this; }
3.26
flink_TableDescriptor_schema_rdh
/** * Define the schema of the {@link TableDescriptor}. * * <p>The schema is typically required. It is optional only in cases where the schema can be * inferred, e.g. {@link Table#insertInto(TableDescriptor)}. */ public Builder schema(@Nullable Schema schema) { this.schema = schema; return this; }
3.26
flink_TableDescriptor_m0_rdh
/** * Creates a new {@link Builder} for a managed table. * * @deprecated This method will be removed soon. Please see FLIP-346 for more details. */ @Deprecatedpublic static Builder m0() { return new Builder(); }
3.26
flink_TableDescriptor_toBuilder_rdh
/** * Converts this immutable instance into a mutable {@link Builder}. */public Builder toBuilder() { return new Builder(this);}
3.26
flink_TableDescriptor_forConnector_rdh
/** * Creates a new {@link Builder} for a table using the given connector. * * @param connector * The factory identifier for the connector. */public static Builder forConnector(String connector) { Preconditions.checkNotNull(connector, "Table descriptors require a connector identifier."); final Builder de...
3.26
flink_TableDescriptor_partitionedBy_rdh
/** * Define which columns this table is partitioned by. */ public Builder partitionedBy(String... partitionKeys) { this.partitionKeys.addAll(Arrays.asList(partitionKeys)); return this; }
3.26
flink_FloatParser_parseField_rdh
/** * Static utility to parse a field of type float from a byte sequence that represents text * characters (such as when read from a file stream). * * @param bytes * The bytes containing the text data that should be parsed. * @param startPos * The offset to start the parsing. * @param length * The length...
3.26
flink_UploadThrottle_hasCapacity_rdh
/** * Test whether some capacity is available. */ public boolean hasCapacity() { return inFlightBytesCounter < maxBytesInFlight; }
3.26
flink_UploadThrottle_seizeCapacity_rdh
/** * Seize <b>bytes</b> capacity. It is the caller responsibility to ensure at least some capacity * {@link #hasCapacity() is available}. <strong>After</strong> this call, the caller is allowed * to actually use the seized capacity. When the capacity is not needed anymore, the caller is * required to {@link #relea...
3.26
flink_UploadThrottle_releaseCapacity_rdh
/** * Release previously {@link #seizeCapacity(long) seized} capacity. Called by {@link BatchingStateChangeUploadScheduler} (IO thread). */ public void releaseCapacity(long bytes) { inFlightBytesCounter -= bytes; }
3.26
flink_TypeInferenceExtractor_forAsyncTableFunction_rdh
/** * Extracts a type inference from a {@link AsyncTableFunction}. */ public static TypeInference forAsyncTableFunction(DataTypeFactory typeFactory, Class<? extends AsyncTableFunction<?>> function) { final FunctionMappingExtractor mappingExtractor = new FunctionMappingExtractor(typeFactory, function, UserD...
3.26
flink_TypeInferenceExtractor_forTableFunction_rdh
/** * Extracts a type inference from a {@link TableFunction}. */ public static TypeInference forTableFunction(DataTypeFactory typeFactory, Class<? extends TableFunction<?>> function) { final FunctionMappingExtractor mappingExtractor = new FunctionMappingExtractor(typeFactory, function, UserDef...
3.26
flink_TypeInferenceExtractor_forProcedure_rdh
/** * Extracts a type in inference from a {@link Procedure}. */ public static TypeInference forProcedure(DataTypeFactory typeFactory, Class<? extends Procedure> procedure) { final ProcedureMappingExtractor mappingExtractor = new ProcedureMappingExtractor(typeFactory, procedure, ProcedureDefinition.PROCEDURE_CALL...
3.26
flink_TypeInferenceExtractor_forTableAggregateFunction_rdh
/** * Extracts a type inference from a {@link TableAggregateFunction}. */ public static TypeInference forTableAggregateFunction(DataTypeFactory typeFactory, Class<? extends TableAggregateFunction<?, ?>> function) { final FunctionMappingExtractor mappingExtractor = new FunctionMappingExtractor(typeFactory, fu...
3.26
flink_TypeInferenceExtractor_forAggregateFunction_rdh
/** * Extracts a type inference from a {@link AggregateFunction}. */ public static TypeInference forAggregateFunction(DataTypeFactory typeFactory, Class<? extends AggregateFunction<?, ?>> function) { final FunctionMappingExtractor mappingExtractor = new FunctionMappingExtractor(typeFactory, function, UserDefined...
3.26
flink_TypeInferenceExtractor_m0_rdh
/** * Extracts a type inference from a {@link ScalarFunction}. */ public static TypeInference m0(DataTypeFactory typeFactory, Class<? extends ScalarFunction> function) { final FunctionMappingExtractor mappingExtractor = new FunctionMappingExtractor(typeFactory, function, UserDefinedFunctionHelper.SCALAR_EVAL,...
3.26
flink_CheckpointStatsCounts_incrementCompletedCheckpoints_rdh
/** * Increments the number of successfully completed checkpoints. * * <p>It is expected that this follows a previous call to {@link #incrementInProgressCheckpoints()}. */ void incrementCompletedCheckpoints() { if (canDecrementOfInProgressCheckpointsNumber()) { numInProgressCheckpoints--; } numC...
3.26
flink_CheckpointStatsCounts_getNumberOfRestoredCheckpoints_rdh
/** * Returns the number of restored checkpoints. * * @return Number of restored checkpoints. */public long getNumberOfRestoredCheckpoints() { return numRestoredCheckpoints; }
3.26
flink_CheckpointStatsCounts_getNumberOfFailedCheckpoints_rdh
/** * Returns the number of failed checkpoints. * * @return Number of failed checkpoints. */ public long getNumberOfFailedCheckpoints() { return numFailedCheckpoints; }
3.26
flink_CheckpointStatsCounts_incrementFailedCheckpoints_rdh
/** * Increments the number of failed checkpoints. * * <p>It is expected that this follows a previous call to {@link #incrementInProgressCheckpoints()}. */ void incrementFailedCheckpoints() { if (canDecrementOfInProgressCheckpointsNumber()) { numInProgressCheckpoints--; } numFailedCheckpoints++;...
3.26
flink_CheckpointStatsCounts_incrementRestoredCheckpoints_rdh
/** * Increments the number of restored checkpoints. */ void incrementRestoredCheckpoints() { numRestoredCheckpoints++; }
3.26
flink_CheckpointStatsCounts_createSnapshot_rdh
/** * Creates a snapshot of the current state. * * @return Snapshot of the current state. */CheckpointStatsCounts createSnapshot() { return new CheckpointStatsCounts(numRestoredCheckpoints, numTotalCheckpoints, numInProgressCheckpoints, numCompletedCheckpoints, numFailedCheckpoints); }
3.26
flink_CheckpointStatsCounts_getTotalNumberOfCheckpoints_rdh
/** * Returns the total number of checkpoints (in progress, completed, failed). * * @return Total number of checkpoints. */ public long getTotalNumberOfCheckpoints() { return numTotalCheckpoints; }
3.26
flink_CheckpointStatsCounts_incrementInProgressCheckpoints_rdh
/** * Increments the number of total and in progress checkpoints. */ void incrementInProgressCheckpoints() { numInProgressCheckpoints++; numTotalCheckpoints++; }
3.26
flink_StreamingSink_sink_rdh
/** * Create a sink from file writer. Decide whether to add the node to commit partitions according * to options. */public static DataStreamSink<?> sink(ProviderContext providerContext, DataStream<PartitionCommitInfo> writer, Path locationPath, ObjectIdentifier identifier, List<String> partitionKeys, TableMetaStoreF...
3.26
flink_StreamingSink_compactionWriter_rdh
/** * Create a file writer with compaction operators by input stream. In addition, it can emit * {@link PartitionCommitInfo} to down stream. */ public static <T> DataStream<PartitionCommitInfo> compactionWriter(ProviderContext providerContext, DataStream<T> inputStream, long bucketCheckInterval, BucketsBuilder<T, ...
3.26
flink_StreamingSink_m0_rdh
/** * Create a file writer by input stream. This is similar to {@link StreamingFileSink}, in * addition, it can emit {@link PartitionCommitInfo} to down stream. */ public static <T> DataStream<PartitionCommitInfo> m0(ProviderContext providerContext, DataStream<T> inputStream, long bucketCheckInterval, BucketsBuilder...
3.26
flink_FlatMapIterator_flatMap_rdh
// -------------------------------------------------------------------------------------------- /** * Delegates calls to the {@link #flatMap(Object)} method. */ @Override public final void flatMap(IN value, Collector<OUT> out) throws Exception {for (Iterator<OUT> iter = flatMap(value); iter.hasNext();) {out.collect...
3.26
flink_KubernetesSessionCli_repStep_rdh
/** * Check whether need to continue or kill the cluster. * * @param in * input buffer reader * @return f0, whether need to continue read from input. f1, whether need to kill the cluster. */ private Tuple2<Boolean, Boolean> repStep(BufferedReader in) throws IOException, InterruptedException { final long sta...
3.26
flink_BufferManager_releaseAll_rdh
/** * The floating buffer is recycled to local buffer pool directly, and the exclusive buffer * will be gathered to return to global buffer pool later. * * @param exclusiveSegments * The list that we will add exclusive segments into. */ void releaseAll(List<MemorySegment> exclusiveSegments) { Buffer buffer; whi...
3.26
flink_BufferManager_unsynchronizedGetNumberOfRequiredBuffers_rdh
// ------------------------------------------------------------------------ // Getter properties // ------------------------------------------------------------------------ @VisibleForTesting int unsynchronizedGetNumberOfRequiredBuffers() {return numRequiredBuffers; }
3.26
flink_BufferManager_requestFloatingBuffers_rdh
/** * Requests floating buffers from the buffer pool based on the given required amount, and * returns the actual requested amount. If the required amount is not fully satisfied, it will * register as a listener. */ int requestFloatingBuffers(int numRequired) { int numRequestedBuffers = 0; synchronized(buff...
3.26
flink_BufferManager_addExclusiveBuffer_rdh
/** * Adds an exclusive buffer (back) into the queue and releases one floating buffer if the * number of available buffers in queue is more than the required amount. If floating buffer * is released, the total amount of available buffers after adding this exclusive buffer has * not changed, and no new buffers are a...
3.26
flink_BufferManager_releaseAllBuffers_rdh
/** * Recycles all the exclusive and floating buffers from the given buffer queue. */ void releaseAllBuffers(ArrayDeque<Buffer> buffers) throws IOException { // Gather all exclusive buffers and recycle them to global pool in batch, because // we do not want to trigger redistribution of buffers after each recycle. fi...
3.26
flink_BufferManager_requestBuffer_rdh
// ------------------------------------------------------------------------ // Buffer request // ------------------------------------------------------------------------ @Nullable Buffer requestBuffer() { synchronized(bufferQueue) { // decrease the number of buffers require to avoid the possibility of ...
3.26
flink_BufferManager_requestExclusiveBuffers_rdh
/** * Requests exclusive buffers from the provider. */ void requestExclusiveBuffers(int numExclusiveBuffers) throws IOException { checkArgument(numExclusiveBuffers >= 0, "Num exclusive buffers must be non-negative."); if (numExclusiveBuffers == 0) { return; } Collection<MemorySegment> segmen...
3.26
flink_BufferManager_recycle_rdh
// ------------------------------------------------------------------------ // Buffer recycle // ------------------------------------------------------------------------ /** * Exclusive buffer is recycled to this channel manager directly and it may trigger return extra * floating buffer based on <tt>numRequiredBuffer...
3.26
flink_BufferManager_notifyBufferAvailable_rdh
// ------------------------------------------------------------------------ // Buffer listener notification // ------------------------------------------------------------------------ /** * The buffer pool notifies this listener of an available floating buffer. If the listener is * released or currently does not need...
3.26
flink_TaskManagerExceptionUtils_tryEnrichTaskManagerError_rdh
/** * Tries to enrich the passed exception or its causes with additional information. * * <p>This method improves error messages for direct and metaspace {@link OutOfMemoryError}. It * adds descriptions about possible causes and ways of resolution. * * @param root * The Throwable of which the cause tree shall ...
3.26
flink_MultipleIdsMessageAcknowledgingSourceBase_snapshotState_rdh
// ------------------------------------------------------------------------ // Checkpointing the data // ------------------------------------------------------------------------ @Override public void snapshotState(FunctionSnapshotContext context) throws Exception { f0.add(new Tuple2<>(context.getCheckpointId(), ses...
3.26
flink_MultipleIdsMessageAcknowledgingSourceBase_acknowledgeIDs_rdh
// ------------------------------------------------------------------------ // ID Checkpointing // ------------------------------------------------------------------------ /** * Acknowledges the session ids. * * @param checkpointId * The id of the current checkout to acknowledge ids for. * @param uniqueIds * ...
3.26
flink_TextOutputFormat_toString_rdh
// -------------------------------------------------------------------------------------------- @Override public String toString() { return (("TextOutputFormat (" + getOutputFilePath()) + ") - ") + this.charsetName; }
3.26
flink_TextOutputFormat_open_rdh
// -------------------------------------------------------------------------------------------- @Override public void open(int taskNumber, int numTasks) throws IOException { super.open(taskNumber, numTasks); try { this.charset = Charset.forName(charsetName); } catch (IllegalCharsetNameException e) {...
3.26
flink_CustomizedConverter_checkArgumentNumber_rdh
// --------------------------------------------------------------------------------------------- protected static void checkArgumentNumber(CallExpression call, int... validArgumentCounts) {boolean hasValidArgumentCount = false; for (int argumentCount : validArgumentCounts) { if (call.getChildren().size()...
3.26
flink_HiveTableMetaStoreFactory_listDataFileRecursively_rdh
/** * List data files recursively. */ private List<FileStatus> listDataFileRecursively(FileSystem fileSystem, Path f) throws IOException { List<FileStatus> fileStatusList = new ArrayList<>(); for (FileStatus fileStatus : fileSystem.listStatus(f)) { if (fileStatus.isDir() && (!isStagingDir(fileStatus.g...
3.26
flink_CollectionInputFormat_toString_rdh
// -------------------------------------------------------------------------------------------- @Override public String toString() { StringBuilder sb = new StringBuilder(); sb.append('['); int num = 0; for (T e : dataSet) { sb.append(e); if (num != (dataSet.size() - 1)) { ...
3.26
flink_CollectionInputFormat_writeObject_rdh
// -------------------------------------------------------------------------------------------- private void writeObject(ObjectOutputStream out) throws IOException { out.defaultWriteObject(); final int size = dataSet.size(); out.writeInt(size); if (size > 0) { DataOutputViewStreamWrapper wrapper...
3.26
flink_CollectionInputFormat_checkCollection_rdh
// -------------------------------------------------------------------------------------------- public static <X> void checkCollection(Collection<X> elements, Class<X> viewedAs) { if ((elements == null) || (viewedAs == null)) { throw new NullPointerException(); } for (X elem : elements) { i...
3.26
flink_CallBindingCallContext_m0_rdh
// -------------------------------------------------------------------------------------------- @Nullable private static DataType m0(SqlCallBinding binding, @Nullable RelDataType returnType) { if (((returnType == null) || returnType.equals(binding.getValidator().getUnknownType())) || (returnType.getSqlTypeName() ==...
3.26
flink_TimestampUtil_createVectorFromConstant_rdh
// creates a Hive ColumnVector of constant timestamp value public static ColumnVector createVectorFromConstant(int batchSize, Object value) { if (hiveTSColVectorClz != null) { return OrcTimestampColumnVector.createFromConstant(batchSize, value); } else { return OrcLegacyTimestampColumnVector.cr...
3.26
flink_TimestampUtil_isHiveTimestampColumnVector_rdh
// whether a ColumnVector is the new TimestampColumnVector public static boolean isHiveTimestampColumnVector(ColumnVector vector) {return (hiveTSColVectorClz != null) && hiveTSColVectorClz.isAssignableFrom(vector.getClass()); }
3.26
flink_EmptyMutableObjectIterator_get_rdh
/** * Gets a singleton instance of the empty iterator. * * @param <E> * The type of the objects (not) returned by the iterator. * @return An instance of the iterator. */ public static <E> MutableObjectIterator<E> get() { @SuppressWarnings("unchecked") MutableObjectIterator<E> iter = ((MutableObjectItera...
3.26
flink_Tuple0_getArity_rdh
// ------------------------------------------------------------------------ @Overridepublic int getArity() { return 0; }
3.26
flink_Tuple0_toString_rdh
// ------------------------------------------------------------------------------------------------- // standard utilities // ------------------------------------------------------------------------------------------------- /** * Creates a string representation of the tuple in the form "()". * * @return The string r...
3.26
flink_Tuple0_equals_rdh
/** * Deep equality for tuples by calling equals() on the tuple members. * * @param o * the object checked for equality * @return true if this is equal to o. */ @Override public boolean equals(Object o) { return (this == o) || (o instanceof Tuple0); }
3.26
flink_Tuple0_readResolve_rdh
// singleton deserialization private Object readResolve() throws ObjectStreamException { return INSTANCE; }
3.26
flink_JobSchedulingPlan_empty_rdh
/** * Create an empty {@link JobSchedulingPlan} with no information about vertices or allocations. */public static JobSchedulingPlan empty() { return new JobSchedulingPlan(VertexParallelism.empty(), Collections.emptyList()); }
3.26
flink_AbstractHeapVector_getDictionaryIds_rdh
/** * Returns the underlying integer column for ids of dictionary. */ @Override public HeapIntVector getDictionaryIds() { return dictionaryIds; }
3.26
flink_AbstractHeapVector_reset_rdh
/** * Resets the column to default state. - fills the isNull array with false. - sets noNulls to * true. */ @Override public void reset() { if (!noNulls) { Arrays.fill(isNull, false); } noNulls = true; }
3.26
flink_SqlColumnPosSpec_symbol_rdh
/** * Creates a parse-tree node representing an occurrence of this keyword at a particular position * in the parsed text. */ public SqlLiteral symbol(SqlParserPos pos) { return SqlLiteral.createSymbol(this, pos); }
3.26
flink_MasterState_name_rdh
// ------------------------------------------------------------------------ public String name() { return name; }
3.26
flink_MasterState_m0_rdh
// ------------------------------------------------------------------------ @Override public String m0() { return (((("name: " + name) + " ; version: ") + version) + " ; bytes: ") + Arrays.toString(bytes); }
3.26
flink_LegacySourceTransformationTranslator_translateForBatchInternal_rdh
/** * A {@link TransformationTranslator} for the {@link LegacySourceTransformation}. * * @param <OUT> * The type of the elements that the {@link LegacySourceTransformation} we are * translating is producing. */ @Internalpublic class LegacySourceTransformationTranslator<OUT> extends SimpleTransformationTransl...
3.26
flink_FlinkPreparingTableBase_getStatistic_rdh
// ~ Methods ---------------------------------------------------------------- /** * Returns the statistic of this table. */ public FlinkStatistic getStatistic() { return this.statistic; }
3.26
flink_FlinkPreparingTableBase_getNames_rdh
/** * Returns the table path in the {@link RelOptSchema}. Different with {@link #getQualifiedName()}, the latter is mainly used for table digest. */ public List<String> getNames() { return names; }
3.26
flink_FlinkPreparingTableBase_getRowType_rdh
/** * Returns the type of rows returned by this table. */ public RelDataType getRowType() { return f0; }
3.26
flink_FlinkPreparingTableBase_getDistribution_rdh
/** * Returns a description of the physical distribution of the rows in this table. * * @see org.apache.calcite.rel.metadata.RelMetadataQuery#distribution */ public RelDistribution getDistribution() { return null; } /** * Returns whether the given columns are a key or a superset of a unique key of this tabl...
3.26
flink_FlinkPreparingTableBase_m0_rdh
/** * Obtains whether the ordinal column has a default value, which is not supported now. * * @param rowType * Row type of field * @param ordinal * Index of the given column * @param initializerContext * Context for {@link org.apache.calcite.sql2rel.InitializerExpressionFactory} * @return true if the col...
3.26
flink_FlinkPreparingTableBase_getExpression_rdh
/** * Generates code for this table, which is not supported now. * * @param clazz * The desired collection class, for example {@link org.apache.calcite.linq4j.Queryable} */ public Expression getExpression(Class clazz) { throw new UnsupportedOperationException(); }
3.26
flink_FlinkPreparingTableBase_getRowCount_rdh
/** * Returns an estimate of the number of rows in the table. */ public double getRowCount() { Double rowCnt = getStatistic().getRowCount(); return rowCnt == null ? DEFAULT_ROWCOUNT : rowCnt; }
3.26
flink_FlinkPreparingTableBase_getMonotonicity_rdh
/** * Obtains whether a given column is monotonic. * * @param columnName * Column name * @return True if the given column is monotonic */ public SqlMonotonicity getMonotonicity(String columnName) { return SqlMonotonicity.NOT_MONOTONIC; }
3.26
flink_FlinkPreparingTableBase_explainSourceAsString_rdh
// ~ Tools ------------------------------------------------------------------ /** * Returns the digest of the {@link TableSource} instance. */ protected List<String> explainSourceAsString(TableSource<?> ts) { String tsDigest = ts.explainSource(); if (!Strings.isNullOrEmpty(tsDigest)) { return Immutabl...
3.26
flink_FlinkPreparingTableBase_getCollationList_rdh
/** * Returns a description of the physical ordering (or orderings) of the rows returned from this * table. * * @see org.apache.calcite.rel.metadata.RelMetadataQuery#collations(RelNode) */ public List<RelCollation> getCollationList() { return ImmutableList.of();}
3.26
flink_FlinkPreparingTableBase_isTemporal_rdh
/** * We recognize all tables in FLink are temporal as they are changeable. */ public boolean isTemporal() { return true; }
3.26
flink_FlinkPreparingTableBase_m1_rdh
/** * Returns unique keySets of current table. */ public Optional<Set<ImmutableBitSet>> m1() {Set<? extends Set<String>> uniqueKeys = statistic.getUniqueKeys(); if (uniqueKeys == null) { return Optional.empty(); } else if (uniqueKeys.size() == 0) { return Optional.of(Immuta...
3.26
flink_FlinkPreparingTableBase_getAllowedAccess_rdh
/** * Obtains the access type of the table. * * @return all access types including SELECT/UPDATE/INSERT/DELETE */ public SqlAccessType getAllowedAccess() { return SqlAccessType.ALL; }
3.26
flink_LatencyMarker_equals_rdh
// ------------------------------------------------------------------------ @Override public boolean equals(Object o) { if (this == o) {return true; } if ((o == null) || (getClass() != o.getClass())) { return false; } LatencyMarker that = ((LatencyMarker) (o)); if (markedTime != t...
3.26
flink_LatencyMarker_getMarkedTime_rdh
/** * Returns the timestamp marked by the LatencyMarker. */ public long getMarkedTime() { return markedTime; }
3.26
flink_BiFunctionWithException_unchecked_rdh
/** * Convert at {@link BiFunctionWithException} into a {@link BiFunction}. * * @param biFunctionWithException * function with exception to convert into a function * @param <A> * input type * @param <B> * output type * @return {@link BiFunction} which throws all checked exception as an unchecked exceptio...
3.26
flink_OperationExecutor_runClusterAction_rdh
/** * Retrieves the {@link ClusterClient} from the session and runs the given {@link ClusterAction} * against it. * * @param configuration * the combined configuration of {@code sessionConf} and {@code executionConfig}. * @param handle * the specified operation handle * @param clusterAction * the cluster...
3.26
flink_FlinkCalciteCatalogReader_toPreparingTable_rdh
/** * Translate this {@link CatalogSchemaTable} into Flink source table. */ private static FlinkPreparingTableBase toPreparingTable(RelOptSchema relOptSchema, List<String> names, RelDataType rowType, CatalogSchemaTable schemaTable) { final ResolvedCatalogB...
3.26
flink_FlinkCalciteCatalogReader_isLegacySourceOptions_rdh
/** * Checks whether the {@link CatalogTable} uses legacy connector source options. */ private static boolean isLegacySourceOptions(CatalogSchemaTable schemaTable) { // normalize option keys DescriptorProperties properties = new DescriptorProperties(true); properties.putProperties(schemaTable.getConte...
3.26