name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_ExecNodeContext_withId_rdh | /**
* Set the unique ID of the node, so that the {@link ExecNodeContext}, together with the type
* related {@link #name} and {@link #version}, stores all the necessary info to uniquely
* reconstruct the {@link ExecNode}, and avoid storing the {@link #id} independently as a field
* in {@link ExecNodeBase}.
*/
publi... | 3.26 |
flink_ExecNodeContext_getId_rdh | /**
* The unique identifier for each ExecNode in the JSON plan.
*/
int getId() {
return checkNotNull(id); } | 3.26 |
flink_ExecNodeContext_getName_rdh | /**
* The type identifying an ExecNode in the JSON plan. See {@link ExecNodeMetadata#name()}.
*/
public String getName() {
return name;
}
/**
* The version of the ExecNode in the JSON plan. See {@link ExecNodeMetadata#version()} | 3.26 |
flink_ExecNodeContext_resetIdCounter_rdh | /**
* Reset the id counter to 0.
*/
@VisibleForTesting
public static void resetIdCounter() {
idCounter.set(0);} | 3.26 |
flink_LocalTimeComparator_compareSerializedLocalTime_rdh | // --------------------------------------------------------------------------------------------
// Static Helpers for Date Comparison
// --------------------------------------------------------------------------------------------
public static int compareSerializedLocalTime(DataInputView firstSource, DataInputView seco... | 3.26 |
flink_FlinkSemiAntiJoinProjectTransposeRule_onMatch_rdh | // ~ Methods ----------------------------------------------------------------
public void onMatch(RelOptRuleCall call) {
LogicalJoin join =
call.rel(0);
LogicalProject
project = call.rel(1);
// convert the semi/anti join condition to reflect the LHS with the project
// pulled up
RexNode newC... | 3.26 |
flink_FlinkSemiAntiJoinProjectTransposeRule_adjustCondition_rdh | /**
* Pulls the project above the semi/anti join and returns the resulting semi/anti join
* condition. As a result, the semi/anti join condition should be modified such that references
* to the LHS of a semi/anti join should now reference the children of the project that's on the
* LHS.
*
* @param project
* Lo... | 3.26 |
flink_LongSumAggregator_aggregate_rdh | /**
* Adds the given value to the current aggregate.
*
* @param value
* The value to add to the aggregate.
*/
public void aggregate(long value) {
sum += value;
} | 3.26 |
flink_GenericInMemoryCatalog_dropTable_rdh | // ------ tables and views ------
@Override
public void dropTable(ObjectPath tablePath, boolean ignoreIfNotExists) throws
TableNotExistException {
checkNotNull(tablePath);
if (tableExists(tablePath)) {tables.remove(tablePath);
tableStats.remove(tablePath);
tableColumnStats.remove(tablePath);
... | 3.26 |
flink_GenericInMemoryCatalog_createDatabase_rdh | // ------ databases ------
@Override
public void createDatabase(String databaseName, CatalogDatabase db, boolean ignoreIfExists) throws DatabaseAlreadyExistException {
checkArgument(!StringUtils.isNullOrWhitespaceOnly(databaseName));
checkNotNull(db);
if (databaseExists(databaseName)) {
if (!ignore... | 3.26 |
flink_GenericInMemoryCatalog_isPartitionedTable_rdh | /**
* Check if the given table is a partitioned table. Note that "false" is returned if the table
* doesn't exists.
*/
private boolean isPartitionedTable(ObjectPath tablePath) {
CatalogBaseTable table = null;
try {
... | 3.26 |
flink_GenericInMemoryCatalog_createFunction_rdh | // ------ functions ------
@Override
public void
createFunction(ObjectPath path, CatalogFunction function, boolean ignoreIfExists) throws FunctionAlreadyExistException, DatabaseNotExistException {
checkNotNull(path);
checkNotNull(function);
ObjectPath functionPath = normalize(path);
if (!databaseExists(... | 3.26 |
flink_GenericInMemoryCatalog_isFullPartitionSpec_rdh | /**
* Check if the given partitionSpec is full partition spec for the given table.
*/
private boolean isFullPartitionSpec(ObjectPath tablePath, CatalogPartitionSpec partitionSpec) throws TableNotExistException {
CatalogBaseTable baseTable = getTable(tablePath);
if (!(baseTable instanceof CatalogTable)) {
... | 3.26 |
flink_GenericInMemoryCatalog_getTableStatistics_rdh | // ------ statistics ------
@Override
public CatalogTableStatistics getTableStatistics(ObjectPath tablePath) throws TableNotExistException {
checkNotNull(tablePath);
if (!tableExists(tablePath)) {
throw new TableNotExistException(getName(), tablePath);
}
i... | 3.26 |
flink_GenericInMemoryCatalog_createTable_rdh | // ------ tables ------
@Override
public void createTable(ObjectPath tablePath, CatalogBaseTable table, boolean ignoreIfExists) throws TableAlreadyExistException, DatabaseNotExistException {
checkNotNull(tablePath);
checkNotNull(table);
if (!databaseExists(tablePath.getDatabaseName()))
{
throw... | 3.26 |
flink_GenericInMemoryCatalog_createPartition_rdh | // ------ partitions ------
@Override
public void createPartition(ObjectPath tablePath, CatalogPartitionSpec partitionSpec, CatalogPartition partition, boolean ignoreIfExists) throws TableNotExistException, TableNotPartitionedException, PartitionSpecInvalidException, PartitionAlreadyExistsException, CatalogException {
... | 3.26 |
flink_ReduceCombineDriver_setup_rdh | // ------------------------------------------------------------------------
@Override
public void setup(TaskContext<ReduceFunction<T>, T> context) {
taskContext = context;
running = true;
} | 3.26 |
flink_HiveTableSource_toHiveTablePartition_rdh | /**
* Convert partition to HiveTablePartition.
*/
public HiveTablePartition toHiveTablePartition(Partition partition)
{
return HivePartitionUtils.toHiveTablePartition(partitionKeys, tableProps, partition);
} | 3.26 |
flink_CheckpointStatistics_generateCheckpointStatistics_rdh | // -------------------------------------------------------------------------
// Static factory methods
// -------------------------------------------------------------------------
public static CheckpointStatistics generateCheckpointStatistics(AbstractCheckpointStats checkpointStats, boolean includeTaskCheckpointStati... | 3.26 |
flink_OrcShim_defaultShim_rdh | /**
* Default with orc dependent, we should use v2.3.0.
*/
static OrcShim<VectorizedRowBatch> defaultShim() {
return new OrcShimV230();
} | 3.26 |
flink_OrcShim_createShim_rdh | /**
* Create shim from hive version.
*/
static OrcShim<VectorizedRowBatch> createShim(String hiveVersion) {
if
(hiveVersion.startsWith("2.0")) {
return new OrcShimV200();
} else if (hiveVersion.startsWith("2.1")) {
return new OrcShimV210();
} else if ((hiveVersion.startsWith("2.2") |... | 3.26 |
flink_BeamStateRequestHandler_of_rdh | /**
* Create a {@link BeamStateRequestHandler}.
*
* @param keyedStateBackend
* if null, {@link BeamStateRequestHandler} would throw an error when
* receive keyed-state requests.
* @param operatorStateBackend
* if null, {@link BeamStateRequestHandler} would throw an error
* when receive operator-state re... | 3.26 |
flink_JobIdsWithStatusOverview_hashCode_rdh | // ------------------------------------------------------------------------
@Override
public int hashCode() {
return jobsWithStatus.hashCode();
} | 3.26 |
flink_JobIdsWithStatusOverview_combine_rdh | // ------------------------------------------------------------------------
private static Collection<JobIdWithStatus> combine(Collection<JobIdWithStatus> first, Collection<JobIdWithStatus> second) {
checkNotNull(first);
checkNotNull(second);
ArrayList<JobIdWithStatus> result = new ArrayList<>(first.size() ... | 3.26 |
flink_HeaderlessChannelWriterOutputView_close_rdh | /**
* Closes this OutputView, closing the underlying writer. And return number bytes in last memory
* segment.
*/
@Override
public int close() throws IOException {
if (!writer.isClosed()) {
int currentPositionInSegment = getCurrentPositionInSegment();
// write last segment
writer.writeBlock(getCurrentSeg... | 3.26 |
flink_ResultRetryStrategy_fixedDelayRetry_rdh | /**
* Create a fixed-delay retry strategy by given params.
*/
public static ResultRetryStrategy fixedDelayRetry(int maxAttempts, long backoffTimeMillis,
Predicate<Collection<RowData>> resultPredicate) {
return new ResultRetryStrategy(new AsyncRetryStrategies.FixedDelayRetryStrategyBuilder(maxAttempts, backoffTime... | 3.26 |
flink_SSLUtils_createRestSSLContext_rdh | /**
* Creates an SSL context for clients against the external REST endpoint.
*/
@Nullable
@VisibleForTesting
public static SSLContext createRestSSLContext(Configuration config, boolean clientMode) throws Exception {
ClientAuth clientAuth = (SecurityOptions.isRe... | 3.26 |
flink_SSLUtils_getAndCheckOption_rdh | // ------------------------------------------------------------------------
// Utilities
// ------------------------------------------------------------------------
private static String getAndCheckOption(Configuration config, ConfigOption<String> primaryOption, ConfigOption<String> fallbackOption)
{
String value =... | 3.26 |
flink_SSLUtils_createSSLClientSocketFactory_rdh | /**
* Creates a factory for SSL Client Sockets from the given configuration. SSL Client Sockets are
* always part of internal communication.
*/
public static SocketFactory createSSLClientSocketFactory(Configuration config) throws Exception {
SSLContext sslContext = createInternalSSLContext(config, true);
if ... | 3.26 |
flink_SSLUtils_m0_rdh | /**
* Creates a {@link SSLHandlerFactory} to be used by the REST Servers.
*
* @param config
* The application configuration.
*/
public static SSLHandlerFactory m0(final Configuration config) throws Exception {
ClientAuth clientAuth = (SecurityOptions.isRestSSLAuthenticationEnabled(config)) ? ClientAuth.REQUI... | 3.26 |
flink_SSLUtils_createInternalClientSSLEngineFactory_rdh | /**
* Creates a SSLEngineFactory to be used by internal communication client endpoints.
*/
public static SSLHandlerFactory createInternalClientSSLEngineFactory(final Configuration config) throws Exception {
SslContext sslContext = m1(config, true);
if (sslContext == null) {
throw new IllegalConfigurat... | 3.26 |
flink_SSLUtils_createSSLServerSocketFactory_rdh | /**
* Creates a factory for SSL Server Sockets from the given configuration. SSL Server Sockets are
* always part of internal communication.
*/
public static ServerSocketFactory createSSLServerSocketFactory(Configuration config) throws Exception {
SSLContext
sslContext = createInternalSSLContext(config, fal... | 3.26 |
flink_SSLUtils_createRestNettySSLContext_rdh | /**
* Creates an SSL context for the external REST SSL. If mutual authentication is configured the
* client and the server side configuration are identical.
*/
@Nullable
public static SslContext createRestNettySSLContext(Configuration config, boolean clientMode, ClientAuth
clientAuth, SslProvider provider) throws Ex... | 3.26 |
flink_SSLUtils_m2_rdh | /**
* Creates the SSL Context for internal SSL, if internal SSL is configured. For internal SSL,
* the client and server side configuration are identical, because of mutual authentication.
*/
@Nullable
private static SslContext m2(Configuration config, boolean clientMode, SslProvider provider) throws Exception {
... | 3.26 |
flink_SSLUtils_createRestClientSSLEngineFactory_rdh | /**
* Creates a {@link SSLHandlerFactory} to be used by the REST Clients.
*
* @param config
* The application configuration.
*/ public static SSLHandlerFactory createRestClientSSLEngineFactory(final Configuration config) throws Exception {
ClientAuth clientAuth
= (SecurityOptions.isRestSSLAuthenticationE... | 3.26 |
flink_SSLUtils_createInternalServerSSLEngineFactory_rdh | /**
* Creates a SSLEngineFactory to be used by internal communication server endpoints.
*/
public static SSLHandlerFactory createInternalServerSSLEngineFactory(final Configuration config) throws Exception {
SslContext sslContext = m1(config, false);
if (sslContext == null) {
throw new IllegalConfigura... | 3.26 |
flink_SSLUtils_createInternalSSLContext_rdh | /**
* Creates the SSL Context for internal SSL, if internal SSL is configured. For internal SSL,
* the client and server side configuration are identical, because of mutual authentication.
*/
@Nullable
private static SSLContext createInternalSSLContext(Configuration config, boolean clientMode) throws Exception {
... | 3.26 |
flink_HiveJdbcParameterUtils_m0_rdh | /**
* Use the {@param parameters} to set {@param hiveConf} or {@param hiveVariables} according to
* what kinds of the parameter belongs.
*/
public static void m0(HiveConf
hiveConf, Map<String, String> sessionConfigs, Map<String, String> parameters) {
for (Map.Entry<String, String> en... | 3.26 |
flink_YarnTaskExecutorRunner_main_rdh | // ------------------------------------------------------------------------
// Program entry point
// ------------------------------------------------------------------------
/**
* The entry point for the YARN task executor runner.
*
* @param args
* The command line arguments.
*/
public static void main(String[]... | 3.26 |
flink_YarnTaskExecutorRunner_runTaskManagerSecurely_rdh | /**
* The instance entry point for the YARN task executor. Obtains user group information and calls
* the main work method {@link TaskManagerRunner#runTaskManager(Configuration, PluginManager)}
* as a privileged action.
*
* @param args
* The command line arguments.
*/
private static void runTaskManagerSecurely... | 3.26 |
flink_FlinkS3FileSystem_getEntropyInjectionKey_rdh | // ------------------------------------------------------------------------
@Nullable
@Override
public String getEntropyInjectionKey() {return entropyInjectionKey;
} | 3.26 |
flink_TableConfigUtils_isOperatorDisabled_rdh | /**
* Returns whether the given operator type is disabled.
*
* @param tableConfig
* TableConfig object
* @param operatorType
* operator type to check
* @return true if the given operator is disabled.
*/
public static boolean isOperatorDisabled(TableConfig tableConfig,... | 3.26 |
flink_TableConfigUtils_getCalciteConfig_rdh | /**
* Returns {@link CalciteConfig} wraps in the given TableConfig.
*
* @param tableConfig
* TableConfig object
* @return wrapped CalciteConfig.
*/
public static CalciteConfig getCalciteConfig(TableConfig tableConfig) {
return tableConfig.getPlannerConfig().unwrap(CalciteConfig.class).orElse(CalciteConfig$.... | 3.26 |
flink_TableConfigUtils_getMaxIdleStateRetentionTime_rdh | /**
* Similar to {@link TableConfig#getMaxIdleStateRetentionTime()}.
*
* @see TableConfig#getMaxIdleStateRetentionTime()
*/
@Deprecated
public static long getMaxIdleStateRetentionTime(ReadableConfig tableConfig) {
return (tableConfig.get(ExecutionConfigOptions.IDLE_STATE_RETENTION).toMillis() * 3) / 2;
} | 3.26 |
flink_TableConfigUtils_m0_rdh | /**
* Returns the aggregate phase strategy configuration.
*
* @param tableConfig
* TableConfig object
* @return the aggregate phase strategy
*/
public static AggregatePhaseStrategy m0(ReadableConfig tableConfig) {
String aggPhaseConf = tableConfig.get(TABLE_OPTIMIZER_AGG_PHASE_STRATEGY).trim();
if (agg... | 3.26 |
flink_TableConfigUtils_getLocalTimeZone_rdh | /**
* Similar to {@link TableConfig#getLocalTimeZone()} but extracting it from a generic {@link ReadableConfig}.
*
* @see TableConfig#getLocalTimeZone()
*/
public static ZoneId getLocalTimeZone(ReadableConfig tableConfig) {
final String zone = tableConfig.get(TableConfigOptions.LOCAL_TIME_ZONE);
if (TableCo... | 3.26 |
flink_OperatorCoordinatorHolder_start_rdh | // ------------------------------------------------------------------------
// OperatorCoordinator Interface
// ------------------------------------------------------------------------
public void start() throws Exception {
mainThreadExecutor.assertRunningInMainThread();
checkState(context.isInitialized(), "Coo... | 3.26 |
flink_OperatorCoordinatorHolder_coordinator_rdh | // ------------------------------------------------------------------------
// Properties
// ------------------------------------------------------------------------
public OperatorCoordinator coordinator() {
return coordinator;
} | 3.26 |
flink_OperatorCoordinatorHolder_abortCurrentTriggering_rdh | // ------------------------------------------------------------------------
// Checkpointing Callbacks
// ------------------------------------------------------------------------
@Override
public void abortCurrentTriggering() {
// unfortunately, this method does not run in the scheduler executor, but in the
// ... | 3.26 |
flink_OperatorCoordinatorHolder_setupAllSubtaskGateways_rdh | // ------------------------------------------------------------------------
// miscellaneous helpers
// ------------------------------------------------------------------------
private void setupAllSubtaskGateways() {
for (int i = 0; i < operatorParallelism; i++) {
setupSubtaskGateway(i);
}
} | 3.26 |
flink_TimeUtils_toDuration_rdh | /**
* Translates {@link Time} to {@link Duration}.
*
* @param time
* time to transform into duration
* @return duration equal to the given time
*/
public static Duration
toDuration(Time time) {
return Duration.of(time.getSize(), toChronoUnit(time.getUnit()));
} | 3.26 |
flink_TimeUtils_formatWithHighestUnit_rdh | /**
* Pretty prints the duration as a lowest granularity unit that does not lose precision.
*
* <p>Examples:
*
* <pre>{@code Duration.ofMilliseconds(60000) will be printed as 1 min
* Duration.ofHours(1).plusSeconds(1) will be printed as 3601 s}</pre>
*
* <b>NOTE:</b> It supports only durations that fit into lon... | 3.26 |
flink_TimeUtils_m0_rdh | /**
*
* @param duration
* to convert to string
* @return duration string in millis
*/
public static String m0(final Duration duration) {
return duration.toMillis() + TimeUnit.MILLISECONDS.labels.get(0); } | 3.26 |
flink_TimeUtils_parseDuration_rdh | /**
* Parse the given string to a java {@link Duration}. The string is in format "{length
* value}{time unit label}", e.g. "123ms", "321 s". If no time unit label is specified, it will
* be considered as milliseconds.
*
* <p>Supported time unit labels are:
*
* <ul>
* <li>DA... | 3.26 |
flink_TimeUtils_singular_rdh | /**
*
* @param label
* the original label
* @return the singular format of the original label
*/
private static String[] singular(String label) {
return new String[]{ label };
} | 3.26 |
flink_Order_isOrdered_rdh | // --------------------------------------------------------------------------------------------
/**
* Checks, if this enum constant represents in fact an order. That is, whether this property is
* not equal to <tt>Order.NONE</tt>.
*
* @return True, if this enum constant is unequal to <tt>Order.NONE</tt>, false othe... | 3.26 |
flink_TypeInference_m0_rdh | /**
* Sets the list of argument types for specifying a fixed, not overloaded, not vararg input
* signature explicitly.
*
* <p>This information is useful for optional arguments with default value. In particular,
* the number of arguments that need to be filled with a default value and their types is
* important.
... | 3.26 |
flink_TypeInference_outputTypeStrategy_rdh | /**
* Sets the strategy for inferring the final output data type of a function call.
*
* <p>Required.
*/
public Builder outputTypeStrategy(TypeStrategy outputTypeStrategy) {
this.outputTypeStrategy = Preconditions.checkNotNull(outputTypeStrategy, "Output type strategy must not be null.");
return this;
} | 3.26 |
flink_TypeInference_inputTypeStrategy_rdh | /**
* Sets the strategy for inferring and validating input arguments in a function call.
*
* <p>A {@link InputTypeStrategies#WILDCARD} strategy function is assumed by default.
*/
public Builder inputTypeStrategy(InputTypeStrategy inputTypeStrategy) {
this.inputTypeStrategy = Preconditions.checkNotNull(inputType... | 3.26 |
flink_TypeInference_typedArguments_rdh | /**
*
* @see #typedArguments(List)
*/
public Builder typedArguments(DataType... argumentTypes) {
return m0(Arrays.asList(argumentTypes));} | 3.26 |
flink_TypeInference_namedArguments_rdh | /**
*
* @see #namedArguments(List)
*/
public Builder namedArguments(String... argumentNames) {
return namedArguments(Arrays.asList(argumentNames));
} | 3.26 |
flink_TypeInference_accumulatorTypeStrategy_rdh | /**
* Sets the strategy for inferring the intermediate accumulator data type of a function
* call.
*/
public Builder accumulatorTypeStrategy(TypeStrategy accumulatorTypeStrategy) {
this.f0 = Preconditions.checkNotNull(accumulatorTypeStrategy, "Accumulator type strategy must not be null.");
return this;
} | 3.26 |
flink_TypeInference_newBuilder_rdh | /**
* Builder for configuring and creating instances of {@link TypeInference}.
*/
public static TypeInference.Builder newBuilder() {
return new TypeInference.Builder();
} | 3.26 |
flink_AfterMatchSkipStrategy_noSkip_rdh | /**
* Every possible match will be emitted.
*
* @return the created AfterMatchSkipStrategy
*/
public static NoSkipStrategy noSkip() {
return NoSkipStrategy.INSTANCE;
} | 3.26 |
flink_AfterMatchSkipStrategy_skipToLast_rdh | /**
* Discards every partial match that started before the last event of emitted match mapped to
* *PatternName*.
*
* @param patternName
* the pattern name to skip to
* @return the created AfterMatchSkipStrategy
*/
public static SkipToLastStrategy skipToLast(String patternName) {
return new SkipToLastStrat... | 3.26 |
flink_AfterMatchSkipStrategy_skipPastLastEvent_rdh | /**
* Discards every partial match that started before emitted match ended.
*
* @return the created AfterMatchSkipStrategy
*/
public static SkipPastLastStrategy skipPastLastEvent() {
return SkipPastLastStrategy.INSTANCE;
} | 3.26 |
flink_AfterMatchSkipStrategy_skipToFirst_rdh | /**
* Discards every partial match that started before the first event of emitted match mapped to
* *PatternName*.
*
* @param patternName
* the pattern name to skip to
* @return the created AfterMatchSkipStrategy
*/
public static SkipToFirstStrategy skipToFirst(String patternName) {
return new SkipToFirstS... | 3.26 |
flink_AfterMatchSkipStrategy_getPatternName_rdh | /**
* Name of pattern that processing will be skipped to.
*/
public Optional<String> getPatternName() {
return Optional.empty();
} | 3.26 |
flink_Savepoint_load_rdh | /**
* Loads an existing savepoint. Useful if you want to query, modify, or extend the state of an
* existing application.
*
* @param env
* The execution environment used to transform the savepoint.
* @param path
* The path to an existing savepoint on disk.
* @param stateBackend
* The state backend of the... | 3.26 |
flink_Savepoint_create_rdh | /**
* Creates a new savepoint.
*
* @param stateBackend
* The state backend of the savepoint used for keyed state.
* @param maxParallelism
* The max parallelism of the savepoint.
* @return A new savepoint.
* @see #create(int)
*/
public static NewSavepoint create(StateBackend stateBackend, int maxParallelism... | 3.26 |
flink_AbstractRuntimeUDFContext_getAccumulator_rdh | // --------------------------------------------------------------------------------------------
@SuppressWarnings("unchecked")
private <V, A extends Serializable> Accumulator<V, A> getAccumulator(String name, Class<? extends Accumulator<V, A>> accumulatorClass)
{
Accumulator<?, ?> accumulator = accumulators.get(n... | 3.26 |
flink_SelectByMaxFunction_reduce_rdh | /**
* Reduce implementation, returns bigger tuple or value1 if both tuples are equal. Comparison
* highly depends on the order and amount of fields chosen as indices. All given fields (at
* construction time) are checked in the same order as defined (at construction time). If both
* tuples are equal in one index, t... | 3.26 |
flink_GlobalWindow_snapshotConfiguration_rdh | // ------------------------------------------------------------------------
@Overridepublic TypeSerializerSnapshot<GlobalWindow> snapshotConfiguration() {
return new GlobalWindowSerializerSnapshot();
} | 3.26 |
flink_WorkerResourceSpec_setExtendedResource_rdh | /**
* Add the given extended resource. The old value with the same resource name will be
* replaced if present.
*/
public Builder setExtendedResource(ExternalResource extendedResource) {
this.extendedResources.put(extendedResource.getName(), extendedResource);
return this;
} | 3.26 |
flink_WorkerResourceSpec_setExtendedResources_rdh | /**
* Add the given extended resources. This will discard all the previous added extended
* resources.
*/
public Builder setExtendedResources(Collection<ExternalResource> extendedResources) {
this.extendedResources = extendedResources.stream().collect(Collectors.toMap(ExternalResource::getName, Function.identi... | 3.26 |
flink_SessionContext_isStatementSetState_rdh | // --------------------------------------------------------------------------------------------
// Begin statement set
// --------------------------------------------------------------------------------------------
public boolean isStatementSetState() {
return isStatementSetState;} | 3.26 |
flink_SessionContext_create_rdh | // --------------------------------------------------------------------------------------------
// Utilities
// --------------------------------------------------------------------------------------------
public static SessionContext create(DefaultContext defaultContext, SessionHandle sessionId, SessionEnvironment envi... | 3.26 |
flink_SessionContext_close_rdh | // --------------------------------------------------------------------------------------------
/**
* Close resources, e.g. catalogs.
*/
public void close() {
operationManager.close();
for (String name : sessionState.catalogManager.listCatalogs()) {
try {
sessionState.catalogManager.getCat... | 3.26 |
flink_SessionContext_getSessionId_rdh | // --------------------------------------------------------------------------------------------
// Getter/Setter
// --------------------------------------------------------------------------------------------
public SessionHandle getSessionId() {
return this.sessionId;
} | 3.26 |
flink_SessionContext_initializeConfiguration_rdh | // ------------------------------------------------------------------------------------------------------------------
// Helpers
// ------------------------------------------------------------------------------------------------------------------
protected static Configuration initializeConfiguration(DefaultContext de... | 3.26 |
flink_SessionContext_createOperationExecutor_rdh | // --------------------------------------------------------------------------------------------
// Method to execute commands
// --------------------------------------------------------------------------------------------
public OperationExecutor createOperationExecutor(Configuration executionConfig) {
return new ... | 3.26 |
flink_KerberosLoginProvider_doLoginAndReturnUGI_rdh | /**
* Does kerberos login and doesn't set current user, just returns a new UGI instance. Must be
* called when isLoginPossible returns true.
*/
public UserGroupInformation doLoginAndReturnUGI() throws IOException {
UserGroupInformation currentUser = UserGroupInformation.getCurrentUser();
if (principal != nul... | 3.26 |
flink_KerberosLoginProvider_doLogin_rdh | /**
* Does kerberos login and sets current user. Must be called when isLoginPossible returns true.
*/public void doLogin(boolean supportProxyUser) throws IOException {
if (principal != null) {
LOG.info("Attempting to login to KDC using principal: {} keytab: {}", principal, keytab);
UserGroupInform... | 3.26 |
flink_DatabaseMetaDataUtils_createSchemasResultSet_rdh | /**
* Create result set for schemas. The schema columns are:
*
* <ul>
* <li>TABLE_SCHEM String => schema name
* <li>TABLE_CATALOG String => catalog name (may be null)
* </ul>
*
* <p>The results are ordered by TABLE_CATALOG and TABLE_SCHEM.
*
* @param statement
* The statement for database meta data
* ... | 3.26 |
streampipes_SpServiceDefinitionBuilder_m0_rdh | /**
* Include migrations in the service definition.
* <br>
* Please refrain from providing {@link IModelMigrator}s with overlapping version definitions for one application id.
*
* @param migrations
* List of migrations to be registered
* @return {@link SpServiceDefinitionBuilder}
*/
public SpServiceDefinition... | 3.26 |
streampipes_SpOpcUaConfigExtractor_extractAdapterConfig_rdh | /**
* Creates {@link OpcUaAdapterConfig} instance in accordance with the given
* {@link org.apache.streampipes.sdk.extractor.StaticPropertyExtractor}.
*
* @param extractor
* extractor for user inputs
* @return {@link OpcUaAdapterConfig} instance based on information from {@code extractor}
*/
public static Op... | 3.26 |
streampipes_AbstractProcessingElementBuilder_naryMappingPropertyWithoutRequirement_rdh | /**
* Adds a new {@link org.apache.streampipes.model.staticproperty.MappingPropertyNary}
* to the pipeline element definition which is not linked to a specific input property.
* Use this method if you want to present users a selection (in form of a Checkbox Group)
* of all available input event properties.
*
* @p... | 3.26 |
streampipes_AbstractProcessingElementBuilder_requiredStream_rdh | /**
* Set a new stream requirement by adding restrictions on this stream. Use
* {@link StreamRequirementsBuilder} to create requirements for a single stream.
*
* @param streamRequirements
* A bundle of collected {@link CollectedStreamRequirements}
* @return this
*/
public K... | 3.26 |
streampipes_AbstractProcessingElementBuilder_supportedProtocols_rdh | /**
* Assigns supported communication/transport protocols to the pipeline elements that can be handled at runtime (e.g.,
* Kafka or JMS).
*
* @param protocol
* An arbitrary number of supported
* {@link org.apache.streampipes.model.grounding.TransportProtocol}s.
* Use {@link org.apache.streampipes.sdk.helpe... | 3.26 |
streampipes_AbstractProcessingElementBuilder_setStream2_rdh | /**
*
* @deprecated Use {@link #requiredStream(CollectedStreamRequirements)} instead
*/
@Deprecated(since = "0.90.0", forRemoval = true)
public K setStream2() {
f1 = true;
return me();
} | 3.26 |
streampipes_AbstractProcessingElementBuilder_setStream1_rdh | /**
*
* @deprecated Use {@link #requiredStream(CollectedStreamRequirements)} instead
*/
@Deprecated(since = "0.90.0", forRemoval = true)public K setStream1() {
stream1 = true;
return me();} | 3.26 |
streampipes_AbstractProcessingElementBuilder_supportedFormats_rdh | /**
* Assigns supported transport formats to the pipeline elements that can be handled at runtime (e.g.,
* JSON or XMl).
*
* @param formats
* A list of supported {@link org.apache.streampipes.model.grounding.TransportFormat}s. Use
* {@link org.apache.streampipes.sdk.helpers.SupportedFormats} to assign formats... | 3.26 |
streampipes_PipelineApi_delete_rdh | /**
* Deletes the pipeline with a given id
*
* @param pipelineId
* The id of the pipeline
*/
@Overridepublic void delete(String pipelineId) {delete(getBaseResourcePath().addToPath(pipelineId), Message.class);
} | 3.26 |
streampipes_PipelineApi_all_rdh | /**
* Receives all pipelines owned by the current user
*
* @return (list) {@link org.apache.streampipes.model.pipeline.Pipeline} a list of all pipelines
*/
@Override
public List<Pipeline> all() {
return getAll(getBaseResourcePath());
} | 3.26 |
streampipes_PipelineApi_start_rdh | /**
* Starts a pipeline by given id
*
* @param pipeline
* The pipeline
* @return {@link org.apache.streampipes.model.pipeline.PipelineOperationStatus} the status message after invocation
*/
@Override
public PipelineOperationStatus start(Pipeline pipeline) {
return start(pipeline.getPipelineId());
} | 3.26 |
streampipes_PipelineApi_stop_rdh | /**
* Stops a pipeline by given id
*
* @param pipeline
* The pipeline
* @return {@link org.apache.streampipes.model.pipeline.PipelineOperationStatus} the status message after detach
*/
@Override
public PipelineOperationStatus stop(Pipeline pipeline) {
return stop(pipeline.getPipelineId());
}
/**
* Stops a... | 3.26 |
streampipes_Options_from_rdh | /**
* Creates a new list of options by using the provided string values.
*
* @param optionLabel
* An arbitrary number of option labels.
* @return */
public static List<Option> from(String... optionLabel) {
return Arrays.stream(optionLabel).map(Option::new).collect(Collectors.toList());
} | 3.26 |
streampipes_TokenizerProcessor_declareModel_rdh | // TODO: Maybe change outputStrategy to an array instead of tons of different strings
@Override
public DataProcessorDescription declareModel() {
return ProcessingElementBuilder.create("org.apache.streampipes.processors.textmining.jvm.tokenizer").category(DataProcessorType.ENRICH_TEXT).withAssets(Assets.DOCUMENTATIO... | 3.26 |
streampipes_OutputStrategies_append_rdh | /**
* Creates a {@link org.apache.streampipes.model.output.AppendOutputStrategy}. Append output strategies add additional
* properties to an input event stream.
*
* @param appendProperties
* An arbitrary number of event properties that are appended to any input stream.
* @return AppendOutputStrategy
*/
public ... | 3.26 |
streampipes_OutputStrategies_custom_rdh | /**
* Creates a {@link org.apache.streampipes.model.output.CustomOutputStrategy}.
*
* @param outputBoth
* If two input streams are expected by a pipeline element, you can use outputBoth to indicate
* whether the properties of both input streams should be available to the pipeline developer for
* selection.
... | 3.26 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.