name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
flink_ResolvedSchema_getColumnNames_rdh
/** * Returns all column names. It does not distinguish between different kinds of columns. */ public List<String> getColumnNames() { return columns.stream().map(Column::getName).collect(Collectors.toList()); }
3.26
flink_ResolvedSchema_toRowDataType_rdh
// -------------------------------------------------------------------------------------------- private DataType toRowDataType(Predicate<Column> columnPredicate) { return // the row should never be null columns.stream().filter(columnPredicate).map(ResolvedSchema::columnToField).collect(Collectors.collectingAnd...
3.26
flink_ResolvedSchema_getPrimaryKey_rdh
/** * Returns the primary key if it has been defined. */ public Optional<UniqueConstraint> getPrimaryKey() { return Optional.ofNullable(primaryKey); }
3.26
flink_ResolvedSchema_m0_rdh
/** * Shortcut for a resolved schema of only physical columns. */ public static ResolvedSchema m0(List<String> columnNames, List<DataType> columnDataTypes) { Preconditions.checkArgument(columnNames.size() == columnDataTypes.size(), "Mismatch between number of columns names and data types."); final List<Column...
3.26
flink_ResolvedSchema_getColumn_rdh
/** * Returns the {@link Column} instance for the given column name. * * @param columnName * the name of the column */ public Optional<Column> getColumn(String columnName) { return this.columns.stream().filter(column -> column.getName().equals(columnName)).findFirst(); }
3.26
flink_ResolvedSchema_of_rdh
/** * Shortcut for a resolved schema of only columns. */ public static ResolvedSchema of(Column... columns) { return ResolvedSchema.of(Arrays.asList(columns)); }
3.26
flink_ResolvedSchema_toSourceRowDataType_rdh
/** * Converts all columns of this schema into a (possibly nested) row data type. * * <p>This method returns the <b>source-to-query schema</b>. * * <p>Note: The returned row data type contains physical, computed, and metadata columns. Be * careful when using this method in a table source or table sink. In many ca...
3.26
flink_ResolvedSchema_getColumns_rdh
/** * Returns all {@link Column}s of this schema. */ public List<Column> getColumns() {return columns; }
3.26
flink_ResolvedSchema_m1_rdh
/** * Returns a list of watermark specifications each consisting of a rowtime attribute and * watermark strategy expression. * * <p>Note: Currently, there is at most one {@link WatermarkSpec} in the list, because we don't * support multiple watermark definitions yet. */ public List<WatermarkSpec> m1() { retu...
3.26
flink_FileChannelOutputView_closeAndDelete_rdh
/** * Closes this output, writing pending data and releasing the memory. * * @throws IOException * Thrown, if the pending data could not be written. */ public void closeAndDelete() throws IOException { close(true); }
3.26
flink_FileChannelOutputView_getBlockCount_rdh
// -------------------------------------------------------------------------------------------- /** * Gets the number of blocks written by this output view. * * @return The number of blocks written by this output view. */ public int getBlockCount() { return numBlocksWritten; }
3.26
flink_FileChannelOutputView_close_rdh
// -------------------------------------------------------------------------------------------- /** * Closes this output, writing pending data and releasing the memory. * * @throws IOException * Thrown, if the pending data could not be written. */ public void close() throws IOException { close(false); }
3.26
flink_AbstractBytesMultiMap_writePointer_rdh
/** * Write value into the output view, and return offset of the value. */ private long writePointer(SimpleCollectingOutputView outputView, int value) throws IOException { int oldPosition = ((int) (outputView.getCurrentOffset()));int skip = checkSkipWriteForPointer(outputView); outputView.getCurrentSegment().putInt...
3.26
flink_AbstractBytesMultiMap_checkSkipWriteForPointer_rdh
/** * For pointer needing update, skip unaligned part (4 bytes) for convenient updating. */ private int checkSkipWriteForPointer(AbstractPagedOutputView outView) throws IOException { // skip if there is no enough size. int available = outView.getSegmentSize() - outView.getCurrentPositionInSegment(); if (available...
3.26
flink_AbstractBytesMultiMap_append_rdh
// ----------------------- Public Interface ----------------------- /** * Append an value into the hash map's record area. */ public void append(LookupInfo<K, Iterator<RowData>> lookupInfo, BinaryRowData value) throws IOException { try { if (lookupInfo.found) { // append value only if there ex...
3.26
flink_AbstractBytesMultiMap_appendRecord_rdh
/** * The key is not exist before. Add key and first value to key area. */ @Override public int appendRecord(LookupInfo<K, Iterator<RowData>> lookupInfo, BinaryRowData value) throws IOException { int lastPosition = ((int) (keyOutView.getCurrentOffset())); // write key to keyOutView int skip = keySerialize...
3.26
flink_AbstractBytesMultiMap_free_rdh
/** * * @param reservedFixedMemory * reserved fixed memory or not. */ @Override public void free(boolean reservedFixedMemory) { recordArea.release(); numKeys = 0; super.free(reservedFixedMemory); }
3.26
flink_AbstractBytesMultiMap_getNumKeys_rdh
// ----------------------- Abstract Interface ----------------------- @Override public long getNumKeys() { return numKeys; }
3.26
flink_AbstractBytesMultiMap_reset_rdh
/** * reset the map's record and bucket area's memory segments for reusing. */ @Override public void reset() { super.reset(); // reset the record segments. recordArea.reset(); numKeys = 0; }
3.26
flink_AbstractBytesMultiMap_updateValuePointer_rdh
/** * Update the content from specific offset. */ private void updateValuePointer(RandomAccessInputView view, int newPointer, int ptrOffset) throws IOException { view.setReadPosition(ptrOffset); int v8 = view.getCurrentPositionInSegment(); view.getCurrentSegment().putInt(v8, newPointer); }
3.26
flink_MetadataOutputStreamWrapper_closeForCommit_rdh
/** * The function will check output stream valid. If it has been closed before, it will throw * {@link IOException}. If not, it will invoke {@code closeForCommitAction()} and mark it * closed. */final void closeForCommit() throws IOException { if (closed) { throw new IOException("The output stream has ...
3.26
flink_ResourceSpec_setExtendedResource_rdh
/** * Add the given extended resource. The old value with the same resource name will be * replaced if present. */ public Builder setExtendedResource(ExternalResource extendedResource) { this.extendedResources.put(extendedResource.getName(), extendedResource); return this;}
3.26
flink_ResourceSpec_newBuilder_rdh
// ------------------------------------------------------------------------ // builder // ------------------------------------------------------------------------ public static Builder newBuilder(double cpuCores, int taskHeapMemoryMB) { return new Builder(new CPUResource(cpuCores), MemorySize.ofMebiBytes(taskHe...
3.26
flink_ResourceSpec_subtract_rdh
/** * Subtracts another resource spec from this one. * * @param other * The other resource spec to subtract. * @return The subtracted resource spec. */public ResourceSpec subtract(final ResourceSpec other) { checkNotNull(other, "Cannot subtract null resources"); if (this.equals(UNKNOWN) || other.equals(...
3.26
flink_ResourceSpec_setExtendedResources_rdh
/** * Add the given extended resources. This will discard all the previous added extended * resources. */ public Builder setExtendedResources(Collection<ExternalResource> extendedResources) { this.extendedResources = extendedResources.stream().collect(Collectors.toMap(ExternalResource::getName, Function.identity...
3.26
flink_ResourceSpec_merge_rdh
/** * Used by system internally to merge the other resources of chained operators when generating * the job graph. * * @param other * Reference to resource to merge in. * @return The new resource with merged values. */ public ResourceSpec merge(final ResourceSpec other) { checkNotNull(other, "Cannot mer...
3.26
flink_ResourceSpec_readResolve_rdh
// ------------------------------------------------------------------------ // serialization // ------------------------------------------------------------------------ private Object readResolve() { // try to preserve the singleton property for UNKNOWN return this.equals(UNKNOWN) ? UNKNOWN : this; }
3.26
flink_CommittableCollector_isFinished_rdh
/** * Returns whether all {@link CheckpointCommittableManager} currently hold by the collector are * either committed or failed. * * @return state of the {@link CheckpointCommittableManager} */ public boolean isFinished() { return checkpointCommittables.values().stream().allMatch(CheckpointCommittableManagerIm...
3.26
flink_CommittableCollector_copy_rdh
/** * Returns a new committable collector that deep copies all internals. * * @return {@link CommittableCollector} */ public CommittableCollector<CommT> copy() { return new CommittableCollector<>(checkpointCommittables.entrySet().stream().map(e -> Tuple2.of(e.getKey(), e.getValue().copy())).collect(Collectors...
3.26
flink_CommittableCollector_addMessage_rdh
/** * Adds a {@link CommittableMessage} to the collector to hold it until emission. * * @param message * either {@link CommittableSummary} or {@link CommittableWithLineage} */ public void addMessage(CommittableMessage<CommT> message) { if (message instanceof CommittableSummary) { addSummary(((Committ...
3.26
flink_CommittableCollector_of_rdh
/** * Creates a {@link CommittableCollector} based on the current runtime information. This method * should be used for to instantiate a collector for all Sink V2. * * @param context * holding runtime of information * @param metricGroup * storing the committable metrics * @param <CommT> * type of the com...
3.26
flink_CommittableCollector_merge_rdh
/** * Merges all information from an external collector into this collector. * * <p>This method is important during recovery from existing state. * * @param cc * other {@link CommittableCollector} */ public void merge(CommittableCollector<CommT> cc) { for (Entry<Long, CheckpointCommittableManagerImpl<CommT...
3.26
flink_CommittableCollector_getSubtaskId_rdh
/** * Returns subtask id. * * @return subtask id. */ public int getSubtaskId() { return subtaskId; }
3.26
flink_CommittableCollector_m0_rdh
/** * Returns all {@link CheckpointCommittableManager} until the requested checkpoint id. * * @param checkpointId * counter * @return collection of {@link CheckpointCommittableManager} */ public Collection<? extends CheckpointCommittableManager<CommT>> m0(long checkpointId) {// clean up fully committed previous...
3.26
flink_CommittableCollector_ofLegacy_rdh
/** * Creates a {@link CommittableCollector} for a list of committables. This method is mainly used * to create a collector from the state of Sink V1. * * @param committables * list of committables * @param metricGroup * storing the committable metrics * @param <CommT> * type of committables * @return {...
3.26
flink_TaskStateStats_getSummaryStats_rdh
/** * * @return Summary of the subtask stats. */ public TaskStateStatsSummary getSummaryStats() { return summaryStats; }
3.26
flink_TaskStateStats_getJobVertexId_rdh
/** * * @return ID of the operator the statistics belong to. */ public JobVertexID getJobVertexId() { return f0; }
3.26
flink_TaskStateStats_getStateSize_rdh
/** * * @return Total checkpoint state size over all subtasks. */ public long getStateSize() { return summaryStats.m0().getSum(); }
3.26
flink_TaskStateStats_getCheckpointedSize_rdh
/** * * @return Total persisted size over all subtasks of this checkpoint. */ public long getCheckpointedSize() {return summaryStats.getCheckpointedSize().getSum(); }
3.26
flink_TaskStateStats_getSubtaskStats_rdh
/** * Returns the stats for all subtasks. * * <p>Elements of the returned array are <code>null</code> if no stats are available yet for the * respective subtask. * * <p>Note: The returned array must not be modified. *...
3.26
flink_StringColumnSummary_getMaxLength_rdh
/** * Longest String length. */public Integer getMaxLength() { return maxLength; }
3.26
flink_StringColumnSummary_m0_rdh
/** * Number of empty strings e.g. java.lang.String.isEmpty(). */ public long m0() { return emptyCount; }
3.26
flink_StringColumnSummary_getMinLength_rdh
/** * Shortest String length. */ public Integer getMinLength() { return minLength;}
3.26
flink_KeyedStream_getKeySelector_rdh
// ------------------------------------------------------------------------ // properties // ------------------------------------------------------------------------ /** * Gets the key selector that can get the key by which the stream if partitioned from the * elements. * * @return The key selector for the key. */...
3.26
flink_KeyedStream_between_rdh
/** * Specifies the time boundaries over which the join operation works, so that * * <pre> * leftElement.timestamp + lowerBound <= rightElement.timestamp <= leftElement.timestamp + upperBound * </pre> * * <p>By default both the lower and the upper bound are inclusive. This can be configured * with {@link Interv...
3.26
flink_KeyedStream_inEventTime_rdh
/** * Sets the time characteristic to event time. */ public IntervalJoin<T1, T2, KEY> inEventTime() { timeBehaviour = TimeBehaviour.EventTime; return this; }
3.26
flink_KeyedStream_window_rdh
/** * Windows this data stream to a {@code WindowedStream}, which evaluates windows over a key * grouped stream. Elements are put into windows by a {@link WindowAssigner}. The grouping of * elements is done both by key and by window. * * <p>A {@link org.apache.flink.streaming.api.windowing.triggers.Trigger} can be...
3.26
flink_KeyedStream_maxBy_rdh
/** * Applies an aggregation that gives the current element with the maximum value at the given * position by the given key. An independent aggregate is kept per key. If more elements have * the maximum value at the given position, the operator returns either the first or last one, * depending on the parameter set....
3.26
flink_KeyedStream_sum_rdh
/** * Applies an aggregation that gives the current sum of the data stream at the given field by * the given key. An independent aggregate is kept per key. * * @param field * In case of a POJO, Scala case class, or Tuple type, the name of the (public) * field on which to perform the aggregation. Additionally,...
3.26
flink_KeyedStream_upperBoundExclusive_rdh
/** * Set the upper bound to be exclusive. */ @PublicEvolvingpublic IntervalJoined<IN1, IN2, KEY> upperBoundExclusive() { this.upperBoundInclusive = false; return this; }
3.26
flink_KeyedStream_doTransform_rdh
// ------------------------------------------------------------------------ // basic transformations // ------------------------------------------------------------------------ @Override protected <R> SingleOutputStreamOperator<R> doTransform(final String operatorName, final TypeInformation<R> outTypeInfo, final Stream...
3.26
flink_KeyedStream_validateKeyType_rdh
/** * Validates that a given type of element (as encoded by the provided {@link TypeInformation}) * can be used as a key in the {@code DataStream.keyBy()} operation. This is done by searching * depth-first the key type and checking if each of the composite types satisfies the required * conditions (see {@link #vali...
3.26
flink_KeyedStream_min_rdh
/** * Applies an aggregation that gives the current minimum of the data stream at the given field * expression by the given key. An independent aggregate is kept per key. A field expression is * either the name of a public field or a getter method with parentheses of the {@link DataStream}'s underlying type. A dot c...
3.26
flink_KeyedStream_max_rdh
/** * Applies an aggregation that gives the current maximum of the data stream at the given field * expression by the given key. An independent aggregate is kept per key. A field expression is * either the name of a public field or a getter method with parentheses of the {@link DataStream}'s underlying type. A dot c...
3.26
flink_KeyedStream_countWindow_rdh
/** * Windows this {@code KeyedStream} into sliding count windows. * * @param size * The size of the windows in number of elements. * @param slide * The slide interval in number of elements. */ public WindowedStream<T, KEY, GlobalWindow> countWindow(long size, long slide) { return window(GlobalWindows.crea...
3.26
flink_KeyedStream_inProcessingTime_rdh
/** * Sets the time characteristic to processing time. */ public IntervalJoin<T1, T2, KEY> inProcessingTime() { timeBehaviour = TimeBehaviour.ProcessingTime; return this; }
3.26
flink_KeyedStream_m0_rdh
/** * Applies the given {@link ProcessFunction} on the input stream, thereby creating a transformed * output stream. * * <p>The function will be called for every element in the input streams and can produce zero or * more output elements. Contrary to the {@link DataStream#flatMap(FlatMapFunction)} function, * thi...
3.26
flink_KeyedStream_lowerBoundExclusive_rdh
/** * Set the lower bound to be exclusive. */ @PublicEvolving public IntervalJoined<IN1, IN2, KEY> lowerBoundExclusive() { this.lowerBoundInclusive = false; return this; }
3.26
flink_KeyedStream_process_rdh
/** * Completes the join operation with the given user function that is executed for each * joined pair of elements. This methods allows for passing explicit type information for * the output type. * * @param processJoinFunction * The user-defined process join function. * @param outputType * The type inform...
3.26
flink_KeyedStream_sideOutputLeftLateData_rdh
/** * Send late arriving left-side data to the side output identified by the given {@link OutputTag}. Data is considered late after the watermark */@PublicEvolving public IntervalJoined<IN1, IN2, KEY> sideOutputLeftLateData(OutputTag<IN1> outputTag) { outputTag = left.getExecutionEnvironment().clean(outputTag); this...
3.26
flink_KeyedStream_intervalJoin_rdh
// ------------------------------------------------------------------------ // Joining // ------------------------------------------------------------------------ /** * Join elements of this {@link KeyedStream} with elements of another {@link KeyedStream} over a * time interval that can be specified with {@link Inter...
3.26
flink_KeyedStream_minBy_rdh
/** * Applies an aggregation that gives the current element with the minimum value at the given * position by the given key. An independent aggregate is kept per key. If more elements have * the minimum value at the given position, the operator returns either the first or last one, * depending on the parameter set....
3.26
flink_KeyedStream_reduce_rdh
// ------------------------------------------------------------------------ // Non-Windowed aggregation operations // ------------------------------------------------------------------------ /** * Applies a reduce transformation on the grouped data stream grouped on by the given key * position. The {@link ReduceFunct...
3.26
flink_KeyedStream_m1_rdh
/** * Send late arriving right-side data to the side output identified by the given {@link OutputTag}. Data is considered late after the watermark */ @PublicEvolving public IntervalJoined<IN1, IN2, KEY> m1(OutputTag<IN2> outputTag) { outputTag = right.getExecutionEnvironment().clean(outputTag); this.rightLateDataO...
3.26
flink_DefaultExecutionGraphFactory_tryRestoreExecutionGraphFromSavepoint_rdh
/** * Tries to restore the given {@link ExecutionGraph} from the provided {@link SavepointRestoreSettings}, iff checkpointing is enabled. * * @param executionGraphToRestore * {@link ExecutionGraph} which is supposed to be restored * @param savepointRestoreSettings * {@link SavepointRestoreSettings} containing...
3.26
flink_SchedulingPipelinedRegionComputeUtil_mergeRegionsOnCycles_rdh
/** * Merge the regions base on <a * href="https://en.wikipedia.org/wiki/Tarjan%27s_strongly_connected_components_algorithm"> * Tarjan's strongly connected components algorithm</a>. For more details please see <a * href="https://issues.apache.org/jira/browse/FLINK-17330">FLINK-17330</a>. */...
3.26
flink_RocksDBFullRestoreOperation_restore_rdh
/** * Restores all key-groups data that is referenced by the passed state handles. */@Override public RocksDBRestoreResult restore() throws IOException, StateMigrationException, RocksDBException {rocksHandle.openDB(); try (ThrowingIterator<SavepointRestoreResult> restore = savepointRestoreOperation.restore()) { ...
3.26
flink_RocksDBFullRestoreOperation_restoreKVStateData_rdh
/** * Restore the KV-state / ColumnFamily data for all key-groups referenced by the current state * handle. */ private void restoreKVStateData(ThrowingIterator<KeyGroup> keyGroups, Map<Integer, ColumnFamilyHandle> columnFamilies) throws IOException, RocksDBException, StateMigrationException { ...
3.26
flink_IOUtils_closeAll_rdh
/** * Closes all {@link AutoCloseable} objects in the parameter, suppressing exceptions. Exception * will be emitted after calling close() on every object. * * @param closeables * iterable with closeables to close. * @param suppressedException * ...
3.26
flink_IOUtils_deleteFilesRecursively_rdh
/** * Delete the given directory or file recursively. */ public static void deleteFilesRecursively(Path path) throws Exception { File[] files = path.toFile().listFiles(); if ((files == null) || (files.length == 0)) { return; } for (File file : files) { if (!file.isDirectory()) { ...
3.26
flink_IOUtils_closeSocket_rdh
/** * Closes the socket ignoring {@link IOException}. * * @param sock * the socket to close */ public static void closeSocket(final Socket sock) { // avoids try { close() } dance if (sock != null) { try { sock.close(); } catch (IOException ignored) { } } }
3.26
flink_IOUtils_skipFully_rdh
/** * Similar to readFully(). Skips bytes in a loop. * * @param in * The InputStream to skip bytes from * @param len * number of bytes to skip * @throws IOException * if it could not skip requested number of bytes for any reason (including * EOF) */ public static void skipFully(final InputStream in, l...
3.26
flink_IOUtils_cleanup_rdh
// ------------------------------------------------------------------------ // Silent I/O cleanup / closing // ------------------------------------------------------------------------ /** * Close the AutoCloseable objects and <b>ignore</b> any {@link Exception} or null pointers. * Must only be used for cleanup in exc...
3.26
flink_IOUtils_tryReadFully_rdh
/** * Similar to {@link #readFully(InputStream, byte[], int, int)}. Returns the total number of * bytes read into the buffer. * * @param in * The InputStream to read from * @param buf * The buffer to fill * @return The total number of bytes read into the buffer * @throws IOException * If the first byte ...
3.26
flink_IOUtils_closeAllQuietly_rdh
/** * Closes all elements in the iterable with closeQuietly(). */ public static void closeAllQuietly(Iterable<? extends AutoCloseable> closeables) { if (null != closeables) { for (AutoCloseable closeable : closeables) { closeQuietly(closeable); } } }
3.26
flink_IOUtils_closeStream_rdh
/** * Closes the stream ignoring {@link IOException}. Must only be called in cleaning up from * exception handlers. * * @param stream * the stream to close */ public static void closeStream(final Closeable stream) { cleanup(null, stream); }
3.26
flink_IOUtils_copyBytes_rdh
/** * Copies from one stream to another. * * @param in * InputStream to read from * @param out * OutputStream to write to * @param close * whether or not close the InputStream and OutputStream at the end. The streams * are closed in the finally clause. * @throws IOException * thrown if an I/O error...
3.26
flink_IOUtils_readFully_rdh
// ------------------------------------------------------------------------ // Stream input skipping // ------------------------------------------------------------------------ /** * Reads len bytes in a loop. * * @param in * The InputStream to read from * @param buf * The buffer to fill * @param off * of...
3.26
flink_HiveParserQB_setTabAlias_rdh
/** * Maintain table alias -> (originTableName, qualifiedName). * * @param alias * table alias * @param originTableName * table name that be actually specified, may be "table", "db.table", * "catalog.db.table" * @param qualifiedName * table name with full path, always is "catalog.db.table" */ public v...
3.26
flink_HiveParserQB_containsQueryWithoutSourceTable_rdh
/** * returns true, if the query block contains any query, or subquery without a source table. Like * select current_user(), select current_database() * * @return true, if the query block contains any query without a source table */ public boolean containsQueryWithoutSourceTable() { for (HiveParserQBExpr qbexp...
3.26
flink_HiveParserQB_getSkewedColumnNames_rdh
/** * Retrieve skewed column name for a table. */ public List<String> getSkewedColumnNames(String alias) { // currently, skew column means nothing for flink, so we just return an empty list. return Collections.emptyList(); }
3.26
flink_HiveParserQB_getAppendedAliasFromId_rdh
// For sub-queries, the id. and alias should be appended since same aliases can be re-used // within different sub-queries. // For a query like: // select ... // (select * from T1 a where ...) subq1 // join // (select * from T2 a where ...) subq2 // .. // the alias is modified to subq1:a and subq2:a from a, to identify...
3.26
flink_HiveParserQB_isSimpleSelectQuery_rdh
// to find target for fetch task conversion optimizer (not allows subqueries) public boolean isSimpleSelectQuery() { if (((!qbp.isSimpleSelectQuery()) || isCTAS()) || qbp.isAnalyzeCommand()) { return false; } for (HiveParserQBExpr qbexpr : aliasToSubq.values()) { if (!qbexpr.isSimpleSelect...
3.26
flink_TaskMailboxImpl_put_rdh
// ------------------------------------------------------------------------------------------------------------------ @Override public void put(@Nonnull Mail mail) {final ReentrantLock lock = this.lock; lock.lock(); try { checkPutStateConditions(); queue.addLast(mail); hasNe...
3.26
flink_TaskMailboxImpl_createBatch_rdh
// ------------------------------------------------------------------------------------------------------------------ @Override public boolean createBatch() { checkIsMailboxThread(); if (!hasNewMail) { // batch is usually depleted by previous MailboxProces...
3.26
flink_FsJobArchivist_getArchivedJsons_rdh
/** * Reads the given archive file and returns a {@link Collection} of contained {@link ArchivedJson}. * * @param file * archive to extract * @return collection of archived jsons * @throws IOException * if the file can't be opened, read or doesn't contain valid json */ public static Collection<ArchivedJson>...
3.26
flink_FsJobArchivist_archiveJob_rdh
/** * Writes the given {@link AccessExecutionGraph} to the {@link FileSystem} pointed to by {@link JobManagerOptions#ARCHIVE_DIR}. * * @param rootPath * directory to which the archive should be written to * @param jobId * job id * @param jsonToArchive * collection of json-path pairs to that should be arch...
3.26
flink_OptimizerNode_areBranchCompatible_rdh
/** * Checks whether to candidate plans for the sub-plan of this node are comparable. The two * alternative plans are comparable, if * * <p>a) There is no branch in the sub-plan of this node b) Both candidates have the same * candidate as the child at the last open branch. * * @param plan1 * The root node of ...
3.26
flink_OptimizerNode_initId_rdh
/** * Sets the ID of this node. * * @param id * The id for this node. */ public void initId(int id) { if (id <= 0) { throw new IllegalArgumentException(); } if (this.id == (-1)) { this.id = id; } else { throw new IllegalStateException("Id has already been initialized."); ...
3.26
flink_OptimizerNode_readStubAnnotations_rdh
// ------------------------------------------------------------------------ // Reading of stub annotations // ------------------------------------------------------------------------ /** * Reads all stub annotations, i.e. which fields remain constant, what cardinality bounds the * functions have, which fields remain ...
3.26
flink_OptimizerNode_hasUnclosedBranches_rdh
// -------------------------------------------------------------------------------------------- // Handling of branches // -------------------------------------------------------------------------------------------- public boolean hasUnclosedBranches() { return (this.openBranches != null) && (!this.openBranches.isE...
3.26
flink_OptimizerNode_setParallelism_rdh
/** * Sets the parallelism for this optimizer node. The parallelism denotes how many parallel * instances of the operator will be spawned during the execution. * * @param parallelism * The parallelism to set. If this value is {@link Executio...
3.26
flink_OptimizerNode_getUniqueFields_rdh
// ------------------------------------------------------------------------ // Access of stub annotations // ------------------------------------------------------------------------ /** * Gets the FieldSets which are unique in the output of the node. */ public Set<FieldSet> getUniqueFields() { return this.unique...
3.26
flink_OptimizerNode_getBroadcastConnectionNames_rdh
/** * Return the list of names associated with broadcast inputs for this node. */ public List<String> getBroadcastConnectionNames() { return this.broadcastConnectionNames; }
3.26
flink_OptimizerNode_setBroadcastInputs_rdh
/** * This function connects the operators that produce the broadcast inputs to this operator. * * @param operatorToNode * The map from program operators to optimizer nodes. * @param defaultExchangeMode * The data exchange mode to use, if the operator does not specify * one. * @throws CompilerException *...
3.26
flink_OptimizerNode_getOperator_rdh
/** * Gets the operator represented by this optimizer node. * * @return This node's operator. */ public Operator<?> getOperator() { return this.operator; }
3.26
flink_OptimizerNode_prunePlanAlternatives_rdh
// -------------------------------------------------------------------------------------------- // Pruning // -------------------------------------------------------------------------------------------- protected void prunePlanAlternatives(List<PlanNode> plans) { if (plans.isEmpty()) { throw new CompilerExc...
3.26
flink_OptimizerNode_computeOutputEstimates_rdh
/** * Causes this node to compute its output estimates (such as number of rows, size in bytes) * based on the inputs and the compiler hints. The compiler hints are instantiated with * conservative default values which are used if no other values are provided. Nodes may access * the statistics to determine relevant ...
3.26
flink_OptimizerNode_getInterestingProperties_rdh
/** * Gets the properties that are interesting for this node to produce. * * @return The interesting properties for this node, or null, if not yet computed. */ public InterestingProperties getInterestingProperties() { return this.intProps; }
3.26
flink_OptimizerNode_computeUnionOfInterestingPropertiesFromSuccessors_rdh
/** * Computes all the interesting properties that are relevant to this node. The interesting * properties are a union of the interesting properties on each outgoing connection. However, if * two interesting properties on the outgoing connections overlap, the interesting properties * will occur only once in this se...
3.26