name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_ResolvedSchema_getColumnNames_rdh | /**
* Returns all column names. It does not distinguish between different kinds of columns.
*/
public List<String> getColumnNames() {
return columns.stream().map(Column::getName).collect(Collectors.toList());
} | 3.26 |
flink_ResolvedSchema_toRowDataType_rdh | // --------------------------------------------------------------------------------------------
private DataType toRowDataType(Predicate<Column> columnPredicate) {
return // the row should never be null
columns.stream().filter(columnPredicate).map(ResolvedSchema::columnToField).collect(Collectors.collectingAnd... | 3.26 |
flink_ResolvedSchema_getPrimaryKey_rdh | /**
* Returns the primary key if it has been defined.
*/
public Optional<UniqueConstraint> getPrimaryKey() {
return
Optional.ofNullable(primaryKey);
} | 3.26 |
flink_ResolvedSchema_m0_rdh | /**
* Shortcut for a resolved schema of only physical columns.
*/
public static ResolvedSchema m0(List<String> columnNames, List<DataType> columnDataTypes)
{
Preconditions.checkArgument(columnNames.size() == columnDataTypes.size(), "Mismatch between number of columns names and data types.");
final List<Column... | 3.26 |
flink_ResolvedSchema_getColumn_rdh | /**
* Returns the {@link Column} instance for the given column name.
*
* @param columnName
* the name of the column
*/
public Optional<Column> getColumn(String columnName) {
return this.columns.stream().filter(column -> column.getName().equals(columnName)).findFirst();
} | 3.26 |
flink_ResolvedSchema_of_rdh | /**
* Shortcut for a resolved schema of only columns.
*/
public static ResolvedSchema of(Column... columns) {
return ResolvedSchema.of(Arrays.asList(columns));
} | 3.26 |
flink_ResolvedSchema_toSourceRowDataType_rdh | /**
* Converts all columns of this schema into a (possibly nested) row data type.
*
* <p>This method returns the <b>source-to-query schema</b>.
*
* <p>Note: The returned row data type contains physical, computed, and metadata columns. Be
* careful when using this method in a table source or table sink. In many ca... | 3.26 |
flink_ResolvedSchema_getColumns_rdh | /**
* Returns all {@link Column}s of this schema.
*/
public List<Column> getColumns() {return columns;
} | 3.26 |
flink_ResolvedSchema_m1_rdh | /**
* Returns a list of watermark specifications each consisting of a rowtime attribute and
* watermark strategy expression.
*
* <p>Note: Currently, there is at most one {@link WatermarkSpec} in the list, because we don't
* support multiple watermark definitions yet.
*/
public List<WatermarkSpec> m1() {
retu... | 3.26 |
flink_FileChannelOutputView_closeAndDelete_rdh | /**
* Closes this output, writing pending data and releasing the memory.
*
* @throws IOException
* Thrown, if the pending data could not be written.
*/
public void closeAndDelete() throws IOException {
close(true);
} | 3.26 |
flink_FileChannelOutputView_getBlockCount_rdh | // --------------------------------------------------------------------------------------------
/**
* Gets the number of blocks written by this output view.
*
* @return The number of blocks written by this output view.
*/
public int getBlockCount() {
return numBlocksWritten;
} | 3.26 |
flink_FileChannelOutputView_close_rdh | // --------------------------------------------------------------------------------------------
/**
* Closes this output, writing pending data and releasing the memory.
*
* @throws IOException
* Thrown, if the pending data could not be written.
*/
public void close() throws IOException {
close(false);
} | 3.26 |
flink_AbstractBytesMultiMap_writePointer_rdh | /**
* Write value into the output view, and return offset of the value.
*/
private long writePointer(SimpleCollectingOutputView outputView, int value) throws IOException {
int oldPosition = ((int) (outputView.getCurrentOffset()));int skip = checkSkipWriteForPointer(outputView);
outputView.getCurrentSegment().putInt... | 3.26 |
flink_AbstractBytesMultiMap_checkSkipWriteForPointer_rdh | /**
* For pointer needing update, skip unaligned part (4 bytes) for convenient updating.
*/
private int checkSkipWriteForPointer(AbstractPagedOutputView outView) throws IOException {
// skip if there is no enough size.
int available = outView.getSegmentSize() - outView.getCurrentPositionInSegment();
if (available... | 3.26 |
flink_AbstractBytesMultiMap_append_rdh | // ----------------------- Public Interface -----------------------
/**
* Append an value into the hash map's record area.
*/
public void append(LookupInfo<K, Iterator<RowData>> lookupInfo, BinaryRowData value) throws IOException {
try {
if (lookupInfo.found) {
// append value only if there ex... | 3.26 |
flink_AbstractBytesMultiMap_appendRecord_rdh | /**
* The key is not exist before. Add key and first value to key area.
*/
@Override
public int appendRecord(LookupInfo<K, Iterator<RowData>> lookupInfo, BinaryRowData value)
throws IOException {
int lastPosition = ((int) (keyOutView.getCurrentOffset()));
// write key to keyOutView
int skip = keySerialize... | 3.26 |
flink_AbstractBytesMultiMap_free_rdh | /**
*
* @param reservedFixedMemory
* reserved fixed memory or not.
*/
@Override
public void free(boolean reservedFixedMemory) {
recordArea.release();
numKeys = 0;
super.free(reservedFixedMemory);
} | 3.26 |
flink_AbstractBytesMultiMap_getNumKeys_rdh | // ----------------------- Abstract Interface -----------------------
@Override
public long getNumKeys() {
return numKeys;
} | 3.26 |
flink_AbstractBytesMultiMap_reset_rdh | /**
* reset the map's record and bucket area's memory segments for reusing.
*/
@Override
public void reset() {
super.reset();
// reset the record segments.
recordArea.reset();
numKeys = 0;
} | 3.26 |
flink_AbstractBytesMultiMap_updateValuePointer_rdh | /**
* Update the content from specific offset.
*/
private void
updateValuePointer(RandomAccessInputView view, int newPointer, int ptrOffset) throws IOException {
view.setReadPosition(ptrOffset);
int v8 = view.getCurrentPositionInSegment();
view.getCurrentSegment().putInt(v8, newPointer);
} | 3.26 |
flink_MetadataOutputStreamWrapper_closeForCommit_rdh | /**
* The function will check output stream valid. If it has been closed before, it will throw
* {@link IOException}. If not, it will invoke {@code closeForCommitAction()} and mark it
* closed.
*/final void closeForCommit() throws IOException
{
if (closed) {
throw new IOException("The output stream has ... | 3.26 |
flink_ResourceSpec_setExtendedResource_rdh | /**
* Add the given extended resource. The old value with the same resource name will be
* replaced if present.
*/
public Builder setExtendedResource(ExternalResource extendedResource) {
this.extendedResources.put(extendedResource.getName(), extendedResource);
return this;} | 3.26 |
flink_ResourceSpec_newBuilder_rdh | // ------------------------------------------------------------------------
// builder
// ------------------------------------------------------------------------
public static Builder newBuilder(double cpuCores, int taskHeapMemoryMB)
{
return new Builder(new
CPUResource(cpuCores), MemorySize.ofMebiBytes(taskHe... | 3.26 |
flink_ResourceSpec_subtract_rdh | /**
* Subtracts another resource spec from this one.
*
* @param other
* The other resource spec to subtract.
* @return The subtracted resource spec.
*/public ResourceSpec subtract(final ResourceSpec other) {
checkNotNull(other, "Cannot subtract null resources");
if (this.equals(UNKNOWN) || other.equals(... | 3.26 |
flink_ResourceSpec_setExtendedResources_rdh | /**
* Add the given extended resources. This will discard all the previous added extended
* resources.
*/
public Builder setExtendedResources(Collection<ExternalResource> extendedResources) {
this.extendedResources = extendedResources.stream().collect(Collectors.toMap(ExternalResource::getName, Function.identity... | 3.26 |
flink_ResourceSpec_merge_rdh | /**
* Used by system internally to merge the other resources of chained operators when generating
* the job graph.
*
* @param other
* Reference to resource to merge in.
* @return The new resource with merged values.
*/
public ResourceSpec merge(final ResourceSpec other) {
checkNotNull(other, "Cannot mer... | 3.26 |
flink_ResourceSpec_readResolve_rdh | // ------------------------------------------------------------------------
// serialization
// ------------------------------------------------------------------------
private Object readResolve() {
// try to preserve the singleton property for UNKNOWN
return this.equals(UNKNOWN) ? UNKNOWN : this;
} | 3.26 |
flink_CommittableCollector_isFinished_rdh | /**
* Returns whether all {@link CheckpointCommittableManager} currently hold by the collector are
* either committed or failed.
*
* @return state of the {@link CheckpointCommittableManager}
*/
public boolean isFinished() {
return checkpointCommittables.values().stream().allMatch(CheckpointCommittableManagerIm... | 3.26 |
flink_CommittableCollector_copy_rdh | /**
* Returns a new committable collector that deep copies all internals.
*
* @return {@link CommittableCollector}
*/
public CommittableCollector<CommT> copy() {
return new CommittableCollector<>(checkpointCommittables.entrySet().stream().map(e -> Tuple2.of(e.getKey(), e.getValue().copy())).collect(Collectors... | 3.26 |
flink_CommittableCollector_addMessage_rdh | /**
* Adds a {@link CommittableMessage} to the collector to hold it until emission.
*
* @param message
* either {@link CommittableSummary} or {@link CommittableWithLineage}
*/
public void addMessage(CommittableMessage<CommT> message) {
if (message instanceof CommittableSummary) {
addSummary(((Committ... | 3.26 |
flink_CommittableCollector_of_rdh | /**
* Creates a {@link CommittableCollector} based on the current runtime information. This method
* should be used for to instantiate a collector for all Sink V2.
*
* @param context
* holding runtime of information
* @param metricGroup
* storing the committable metrics
* @param <CommT>
* type of the com... | 3.26 |
flink_CommittableCollector_merge_rdh | /**
* Merges all information from an external collector into this collector.
*
* <p>This method is important during recovery from existing state.
*
* @param cc
* other {@link CommittableCollector}
*/
public void merge(CommittableCollector<CommT> cc) {
for (Entry<Long, CheckpointCommittableManagerImpl<CommT... | 3.26 |
flink_CommittableCollector_getSubtaskId_rdh | /**
* Returns subtask id.
*
* @return subtask id.
*/
public int getSubtaskId() {
return subtaskId;
} | 3.26 |
flink_CommittableCollector_m0_rdh | /**
* Returns all {@link CheckpointCommittableManager} until the requested checkpoint id.
*
* @param checkpointId
* counter
* @return collection of {@link CheckpointCommittableManager}
*/
public Collection<? extends CheckpointCommittableManager<CommT>> m0(long checkpointId) {// clean up fully committed previous... | 3.26 |
flink_CommittableCollector_ofLegacy_rdh | /**
* Creates a {@link CommittableCollector} for a list of committables. This method is mainly used
* to create a collector from the state of Sink V1.
*
* @param committables
* list of committables
* @param metricGroup
* storing the committable metrics
* @param <CommT>
* type of committables
* @return {... | 3.26 |
flink_TaskStateStats_getSummaryStats_rdh | /**
*
* @return Summary of the subtask stats.
*/
public TaskStateStatsSummary getSummaryStats() {
return summaryStats;
} | 3.26 |
flink_TaskStateStats_getJobVertexId_rdh | /**
*
* @return ID of the operator the statistics belong to.
*/
public JobVertexID getJobVertexId() {
return f0;
} | 3.26 |
flink_TaskStateStats_getStateSize_rdh | /**
*
* @return Total checkpoint state size over all subtasks.
*/
public long getStateSize() {
return summaryStats.m0().getSum();
} | 3.26 |
flink_TaskStateStats_getCheckpointedSize_rdh | /**
*
* @return Total persisted size over all subtasks of this checkpoint.
*/
public long getCheckpointedSize() {return summaryStats.getCheckpointedSize().getSum();
} | 3.26 |
flink_TaskStateStats_getSubtaskStats_rdh | /**
* Returns the stats for all subtasks.
*
* <p>Elements of the returned array are <code>null</code> if no stats are available yet for the
* respective subtask.
*
* <p>Note: The returned array must not be modified.
*... | 3.26 |
flink_StringColumnSummary_getMaxLength_rdh | /**
* Longest String length.
*/public Integer getMaxLength() {
return maxLength;
} | 3.26 |
flink_StringColumnSummary_m0_rdh | /**
* Number of empty strings e.g. java.lang.String.isEmpty().
*/
public long m0() {
return emptyCount;
} | 3.26 |
flink_StringColumnSummary_getMinLength_rdh | /**
* Shortest String length.
*/
public Integer getMinLength() {
return minLength;} | 3.26 |
flink_KeyedStream_getKeySelector_rdh | // ------------------------------------------------------------------------
// properties
// ------------------------------------------------------------------------
/**
* Gets the key selector that can get the key by which the stream if partitioned from the
* elements.
*
* @return The key selector for the key.
*/... | 3.26 |
flink_KeyedStream_between_rdh | /**
* Specifies the time boundaries over which the join operation works, so that
*
* <pre>
* leftElement.timestamp + lowerBound <= rightElement.timestamp <= leftElement.timestamp + upperBound
* </pre>
*
* <p>By default both the lower and the upper bound are inclusive. This can be configured
* with {@link Interv... | 3.26 |
flink_KeyedStream_inEventTime_rdh | /**
* Sets the time characteristic to event time.
*/
public IntervalJoin<T1, T2, KEY> inEventTime() {
timeBehaviour =
TimeBehaviour.EventTime;
return this;
} | 3.26 |
flink_KeyedStream_window_rdh | /**
* Windows this data stream to a {@code WindowedStream}, which evaluates windows over a key
* grouped stream. Elements are put into windows by a {@link WindowAssigner}. The grouping of
* elements is done both by key and by window.
*
* <p>A {@link org.apache.flink.streaming.api.windowing.triggers.Trigger} can be... | 3.26 |
flink_KeyedStream_maxBy_rdh | /**
* Applies an aggregation that gives the current element with the maximum value at the given
* position by the given key. An independent aggregate is kept per key. If more elements have
* the maximum value at the given position, the operator returns either the first or last one,
* depending on the parameter set.... | 3.26 |
flink_KeyedStream_sum_rdh | /**
* Applies an aggregation that gives the current sum of the data stream at the given field by
* the given key. An independent aggregate is kept per key.
*
* @param field
* In case of a POJO, Scala case class, or Tuple type, the name of the (public)
* field on which to perform the aggregation. Additionally,... | 3.26 |
flink_KeyedStream_upperBoundExclusive_rdh | /**
* Set the upper bound to be exclusive.
*/
@PublicEvolvingpublic IntervalJoined<IN1, IN2, KEY> upperBoundExclusive()
{
this.upperBoundInclusive = false;
return this;
} | 3.26 |
flink_KeyedStream_doTransform_rdh | // ------------------------------------------------------------------------
// basic transformations
// ------------------------------------------------------------------------
@Override
protected <R> SingleOutputStreamOperator<R> doTransform(final String operatorName, final TypeInformation<R> outTypeInfo, final Stream... | 3.26 |
flink_KeyedStream_validateKeyType_rdh | /**
* Validates that a given type of element (as encoded by the provided {@link TypeInformation})
* can be used as a key in the {@code DataStream.keyBy()} operation. This is done by searching
* depth-first the key type and checking if each of the composite types satisfies the required
* conditions (see {@link #vali... | 3.26 |
flink_KeyedStream_min_rdh | /**
* Applies an aggregation that gives the current minimum of the data stream at the given field
* expression by the given key. An independent aggregate is kept per key. A field expression is
* either the name of a public field or a getter method with parentheses of the {@link DataStream}'s underlying type. A dot c... | 3.26 |
flink_KeyedStream_max_rdh | /**
* Applies an aggregation that gives the current maximum of the data stream at the given field
* expression by the given key. An independent aggregate is kept per key. A field expression is
* either the name of a public field or a getter method with parentheses of the {@link DataStream}'s underlying type. A dot c... | 3.26 |
flink_KeyedStream_countWindow_rdh | /**
* Windows this {@code KeyedStream} into sliding count windows.
*
* @param size
* The size of the windows in number of elements.
* @param slide
* The slide interval in number of elements.
*/
public WindowedStream<T, KEY, GlobalWindow> countWindow(long size, long slide) {
return window(GlobalWindows.crea... | 3.26 |
flink_KeyedStream_inProcessingTime_rdh | /**
* Sets the time characteristic to processing time.
*/
public IntervalJoin<T1, T2, KEY> inProcessingTime() {
timeBehaviour = TimeBehaviour.ProcessingTime;
return this;
} | 3.26 |
flink_KeyedStream_m0_rdh | /**
* Applies the given {@link ProcessFunction} on the input stream, thereby creating a transformed
* output stream.
*
* <p>The function will be called for every element in the input streams and can produce zero or
* more output elements. Contrary to the {@link DataStream#flatMap(FlatMapFunction)} function,
* thi... | 3.26 |
flink_KeyedStream_lowerBoundExclusive_rdh | /**
* Set the lower bound to be exclusive.
*/
@PublicEvolving
public IntervalJoined<IN1, IN2, KEY> lowerBoundExclusive() {
this.lowerBoundInclusive = false;
return this;
} | 3.26 |
flink_KeyedStream_process_rdh | /**
* Completes the join operation with the given user function that is executed for each
* joined pair of elements. This methods allows for passing explicit type information for
* the output type.
*
* @param processJoinFunction
* The user-defined process join function.
* @param outputType
* The type inform... | 3.26 |
flink_KeyedStream_sideOutputLeftLateData_rdh | /**
* Send late arriving left-side data to the side output identified by the given {@link OutputTag}. Data is considered late after the watermark
*/@PublicEvolving
public IntervalJoined<IN1, IN2, KEY> sideOutputLeftLateData(OutputTag<IN1> outputTag) {
outputTag = left.getExecutionEnvironment().clean(outputTag);
this... | 3.26 |
flink_KeyedStream_intervalJoin_rdh | // ------------------------------------------------------------------------
// Joining
// ------------------------------------------------------------------------
/**
* Join elements of this {@link KeyedStream} with elements of another {@link KeyedStream} over a
* time interval that can be specified with {@link Inter... | 3.26 |
flink_KeyedStream_minBy_rdh | /**
* Applies an aggregation that gives the current element with the minimum value at the given
* position by the given key. An independent aggregate is kept per key. If more elements have
* the minimum value at the given position, the operator returns either the first or last one,
* depending on the parameter set.... | 3.26 |
flink_KeyedStream_reduce_rdh | // ------------------------------------------------------------------------
// Non-Windowed aggregation operations
// ------------------------------------------------------------------------
/**
* Applies a reduce transformation on the grouped data stream grouped on by the given key
* position. The {@link ReduceFunct... | 3.26 |
flink_KeyedStream_m1_rdh | /**
* Send late arriving right-side data to the side output identified by the given {@link OutputTag}. Data is considered late after the watermark
*/
@PublicEvolving
public IntervalJoined<IN1, IN2, KEY> m1(OutputTag<IN2> outputTag) {
outputTag = right.getExecutionEnvironment().clean(outputTag);
this.rightLateDataO... | 3.26 |
flink_DefaultExecutionGraphFactory_tryRestoreExecutionGraphFromSavepoint_rdh | /**
* Tries to restore the given {@link ExecutionGraph} from the provided {@link SavepointRestoreSettings}, iff checkpointing is enabled.
*
* @param executionGraphToRestore
* {@link ExecutionGraph} which is supposed to be restored
* @param savepointRestoreSettings
* {@link SavepointRestoreSettings} containing... | 3.26 |
flink_SchedulingPipelinedRegionComputeUtil_mergeRegionsOnCycles_rdh | /**
* Merge the regions base on <a
* href="https://en.wikipedia.org/wiki/Tarjan%27s_strongly_connected_components_algorithm">
* Tarjan's strongly connected components algorithm</a>. For more details please see <a
* href="https://issues.apache.org/jira/browse/FLINK-17330">FLINK-17330</a>.
*/... | 3.26 |
flink_RocksDBFullRestoreOperation_restore_rdh | /**
* Restores all key-groups data that is referenced by the passed state handles.
*/@Override
public RocksDBRestoreResult restore() throws IOException, StateMigrationException, RocksDBException {rocksHandle.openDB();
try (ThrowingIterator<SavepointRestoreResult> restore = savepointRestoreOperation.restore()) {
... | 3.26 |
flink_RocksDBFullRestoreOperation_restoreKVStateData_rdh | /**
* Restore the KV-state / ColumnFamily data for all key-groups referenced by the current state
* handle.
*/
private void restoreKVStateData(ThrowingIterator<KeyGroup> keyGroups, Map<Integer, ColumnFamilyHandle> columnFamilies) throws IOException, RocksDBException, StateMigrationException {
... | 3.26 |
flink_IOUtils_closeAll_rdh | /**
* Closes all {@link AutoCloseable} objects in the parameter, suppressing exceptions. Exception
* will be emitted after calling close() on every object.
*
* @param closeables
* iterable with closeables to close.
* @param suppressedException
* ... | 3.26 |
flink_IOUtils_deleteFilesRecursively_rdh | /**
* Delete the given directory or file recursively.
*/
public static void deleteFilesRecursively(Path path) throws Exception {
File[] files
= path.toFile().listFiles();
if ((files == null) || (files.length == 0)) {
return;
}
for (File file : files) {
if (!file.isDirectory()) {
... | 3.26 |
flink_IOUtils_closeSocket_rdh | /**
* Closes the socket ignoring {@link IOException}.
*
* @param sock
* the socket to close
*/
public static void closeSocket(final Socket sock) {
// avoids try { close() } dance
if (sock != null) {
try {
sock.close();
} catch (IOException ignored) {
}
}
} | 3.26 |
flink_IOUtils_skipFully_rdh | /**
* Similar to readFully(). Skips bytes in a loop.
*
* @param in
* The InputStream to skip bytes from
* @param len
* number of bytes to skip
* @throws IOException
* if it could not skip requested number of bytes for any reason (including
* EOF)
*/
public static void skipFully(final InputStream in, l... | 3.26 |
flink_IOUtils_cleanup_rdh | // ------------------------------------------------------------------------
// Silent I/O cleanup / closing
// ------------------------------------------------------------------------
/**
* Close the AutoCloseable objects and <b>ignore</b> any {@link Exception} or null pointers.
* Must only be used for cleanup in exc... | 3.26 |
flink_IOUtils_tryReadFully_rdh | /**
* Similar to {@link #readFully(InputStream, byte[], int, int)}. Returns the total number of
* bytes read into the buffer.
*
* @param in
* The InputStream to read from
* @param buf
* The buffer to fill
* @return The total number of bytes read into the buffer
* @throws IOException
* If the first byte ... | 3.26 |
flink_IOUtils_closeAllQuietly_rdh | /**
* Closes all elements in the iterable with closeQuietly().
*/
public static void closeAllQuietly(Iterable<? extends AutoCloseable> closeables) {
if (null
!= closeables) {
for (AutoCloseable closeable : closeables) {
closeQuietly(closeable);
}
}
} | 3.26 |
flink_IOUtils_closeStream_rdh | /**
* Closes the stream ignoring {@link IOException}. Must only be called in cleaning up from
* exception handlers.
*
* @param stream
* the stream to close
*/
public static void
closeStream(final Closeable stream) {
cleanup(null, stream);
} | 3.26 |
flink_IOUtils_copyBytes_rdh | /**
* Copies from one stream to another.
*
* @param in
* InputStream to read from
* @param out
* OutputStream to write to
* @param close
* whether or not close the InputStream and OutputStream at the end. The streams
* are closed in the finally clause.
* @throws IOException
* thrown if an I/O error... | 3.26 |
flink_IOUtils_readFully_rdh | // ------------------------------------------------------------------------
// Stream input skipping
// ------------------------------------------------------------------------
/**
* Reads len bytes in a loop.
*
* @param in
* The InputStream to read from
* @param buf
* The buffer to fill
* @param off
* of... | 3.26 |
flink_HiveParserQB_setTabAlias_rdh | /**
* Maintain table alias -> (originTableName, qualifiedName).
*
* @param alias
* table alias
* @param originTableName
* table name that be actually specified, may be "table", "db.table",
* "catalog.db.table"
* @param qualifiedName
* table name with full path, always is "catalog.db.table"
*/
public v... | 3.26 |
flink_HiveParserQB_containsQueryWithoutSourceTable_rdh | /**
* returns true, if the query block contains any query, or subquery without a source table. Like
* select current_user(), select current_database()
*
* @return true, if the query block contains any query without a source table
*/
public boolean containsQueryWithoutSourceTable() {
for (HiveParserQBExpr qbexp... | 3.26 |
flink_HiveParserQB_getSkewedColumnNames_rdh | /**
* Retrieve skewed column name for a table.
*/
public List<String> getSkewedColumnNames(String alias) {
// currently, skew column means nothing for flink, so we just return an empty list.
return Collections.emptyList();
} | 3.26 |
flink_HiveParserQB_getAppendedAliasFromId_rdh | // For sub-queries, the id. and alias should be appended since same aliases can be re-used
// within different sub-queries.
// For a query like:
// select ...
// (select * from T1 a where ...) subq1
// join
// (select * from T2 a where ...) subq2
// ..
// the alias is modified to subq1:a and subq2:a from a, to identify... | 3.26 |
flink_HiveParserQB_isSimpleSelectQuery_rdh | // to find target for fetch task conversion optimizer (not allows subqueries)
public boolean isSimpleSelectQuery() {
if (((!qbp.isSimpleSelectQuery()) || isCTAS()) || qbp.isAnalyzeCommand()) {
return false;
}
for (HiveParserQBExpr qbexpr : aliasToSubq.values()) {
if (!qbexpr.isSimpleSelect... | 3.26 |
flink_TaskMailboxImpl_put_rdh | // ------------------------------------------------------------------------------------------------------------------
@Override
public void put(@Nonnull
Mail
mail) {final ReentrantLock lock = this.lock;
lock.lock();
try {
checkPutStateConditions();
queue.addLast(mail);
hasNe... | 3.26 |
flink_TaskMailboxImpl_createBatch_rdh | // ------------------------------------------------------------------------------------------------------------------
@Override
public boolean createBatch() {
checkIsMailboxThread();
if (!hasNewMail) {
// batch is usually depleted by previous MailboxProces... | 3.26 |
flink_FsJobArchivist_getArchivedJsons_rdh | /**
* Reads the given archive file and returns a {@link Collection} of contained {@link ArchivedJson}.
*
* @param file
* archive to extract
* @return collection of archived jsons
* @throws IOException
* if the file can't be opened, read or doesn't contain valid json
*/
public static Collection<ArchivedJson>... | 3.26 |
flink_FsJobArchivist_archiveJob_rdh | /**
* Writes the given {@link AccessExecutionGraph} to the {@link FileSystem} pointed to by {@link JobManagerOptions#ARCHIVE_DIR}.
*
* @param rootPath
* directory to which the archive should be written to
* @param jobId
* job id
* @param jsonToArchive
* collection of json-path pairs to that should be arch... | 3.26 |
flink_OptimizerNode_areBranchCompatible_rdh | /**
* Checks whether to candidate plans for the sub-plan of this node are comparable. The two
* alternative plans are comparable, if
*
* <p>a) There is no branch in the sub-plan of this node b) Both candidates have the same
* candidate as the child at the last open branch.
*
* @param plan1
* The root node of ... | 3.26 |
flink_OptimizerNode_initId_rdh | /**
* Sets the ID of this node.
*
* @param id
* The id for this node.
*/
public void initId(int id) {
if (id <= 0) {
throw new IllegalArgumentException();
}
if (this.id == (-1)) {
this.id = id;
} else {
throw new IllegalStateException("Id has already been initialized.");
... | 3.26 |
flink_OptimizerNode_readStubAnnotations_rdh | // ------------------------------------------------------------------------
// Reading of stub annotations
// ------------------------------------------------------------------------
/**
* Reads all stub annotations, i.e. which fields remain constant, what cardinality bounds the
* functions have, which fields remain ... | 3.26 |
flink_OptimizerNode_hasUnclosedBranches_rdh | // --------------------------------------------------------------------------------------------
// Handling of branches
// --------------------------------------------------------------------------------------------
public boolean hasUnclosedBranches() {
return (this.openBranches != null) && (!this.openBranches.isE... | 3.26 |
flink_OptimizerNode_setParallelism_rdh | /**
* Sets the parallelism for this optimizer node. The parallelism denotes how many parallel
* instances of the operator will be spawned during the execution.
*
* @param parallelism
* The parallelism to set. If this value is {@link Executio... | 3.26 |
flink_OptimizerNode_getUniqueFields_rdh | // ------------------------------------------------------------------------
// Access of stub annotations
// ------------------------------------------------------------------------
/**
* Gets the FieldSets which are unique in the output of the node.
*/
public Set<FieldSet> getUniqueFields() {
return this.unique... | 3.26 |
flink_OptimizerNode_getBroadcastConnectionNames_rdh | /**
* Return the list of names associated with broadcast inputs for this node.
*/
public List<String> getBroadcastConnectionNames() {
return this.broadcastConnectionNames;
} | 3.26 |
flink_OptimizerNode_setBroadcastInputs_rdh | /**
* This function connects the operators that produce the broadcast inputs to this operator.
*
* @param operatorToNode
* The map from program operators to optimizer nodes.
* @param defaultExchangeMode
* The data exchange mode to use, if the operator does not specify
* one.
* @throws CompilerException
*... | 3.26 |
flink_OptimizerNode_getOperator_rdh | /**
* Gets the operator represented by this optimizer node.
*
* @return This node's operator.
*/
public Operator<?> getOperator() {
return this.operator;
} | 3.26 |
flink_OptimizerNode_prunePlanAlternatives_rdh | // --------------------------------------------------------------------------------------------
// Pruning
// --------------------------------------------------------------------------------------------
protected void prunePlanAlternatives(List<PlanNode> plans) {
if (plans.isEmpty()) {
throw new CompilerExc... | 3.26 |
flink_OptimizerNode_computeOutputEstimates_rdh | /**
* Causes this node to compute its output estimates (such as number of rows, size in bytes)
* based on the inputs and the compiler hints. The compiler hints are instantiated with
* conservative default values which are used if no other values are provided. Nodes may access
* the statistics to determine relevant ... | 3.26 |
flink_OptimizerNode_getInterestingProperties_rdh | /**
* Gets the properties that are interesting for this node to produce.
*
* @return The interesting properties for this node, or null, if not yet computed.
*/
public InterestingProperties getInterestingProperties() {
return this.intProps;
} | 3.26 |
flink_OptimizerNode_computeUnionOfInterestingPropertiesFromSuccessors_rdh | /**
* Computes all the interesting properties that are relevant to this node. The interesting
* properties are a union of the interesting properties on each outgoing connection. However, if
* two interesting properties on the outgoing connections overlap, the interesting properties
* will occur only once in this se... | 3.26 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.