name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
flink_ExecNodeContext_withId_rdh
/** * Set the unique ID of the node, so that the {@link ExecNodeContext}, together with the type * related {@link #name} and {@link #version}, stores all the necessary info to uniquely * reconstruct the {@link ExecNode}, and avoid storing the {@link #id} independently as a field * in {@link ExecNodeBase}. */ publi...
3.26
flink_ExecNodeContext_getId_rdh
/** * The unique identifier for each ExecNode in the JSON plan. */ int getId() { return checkNotNull(id); }
3.26
flink_ExecNodeContext_getName_rdh
/** * The type identifying an ExecNode in the JSON plan. See {@link ExecNodeMetadata#name()}. */ public String getName() { return name; } /** * The version of the ExecNode in the JSON plan. See {@link ExecNodeMetadata#version()}
3.26
flink_ExecNodeContext_resetIdCounter_rdh
/** * Reset the id counter to 0. */ @VisibleForTesting public static void resetIdCounter() { idCounter.set(0);}
3.26
flink_LocalTimeComparator_compareSerializedLocalTime_rdh
// -------------------------------------------------------------------------------------------- // Static Helpers for Date Comparison // -------------------------------------------------------------------------------------------- public static int compareSerializedLocalTime(DataInputView firstSource, DataInputView seco...
3.26
flink_FlinkSemiAntiJoinProjectTransposeRule_onMatch_rdh
// ~ Methods ---------------------------------------------------------------- public void onMatch(RelOptRuleCall call) { LogicalJoin join = call.rel(0); LogicalProject project = call.rel(1); // convert the semi/anti join condition to reflect the LHS with the project // pulled up RexNode newC...
3.26
flink_FlinkSemiAntiJoinProjectTransposeRule_adjustCondition_rdh
/** * Pulls the project above the semi/anti join and returns the resulting semi/anti join * condition. As a result, the semi/anti join condition should be modified such that references * to the LHS of a semi/anti join should now reference the children of the project that's on the * LHS. * * @param project * Lo...
3.26
flink_LongSumAggregator_aggregate_rdh
/** * Adds the given value to the current aggregate. * * @param value * The value to add to the aggregate. */ public void aggregate(long value) { sum += value; }
3.26
flink_GenericInMemoryCatalog_dropTable_rdh
// ------ tables and views ------ @Override public void dropTable(ObjectPath tablePath, boolean ignoreIfNotExists) throws TableNotExistException { checkNotNull(tablePath); if (tableExists(tablePath)) {tables.remove(tablePath); tableStats.remove(tablePath); tableColumnStats.remove(tablePath); ...
3.26
flink_GenericInMemoryCatalog_createDatabase_rdh
// ------ databases ------ @Override public void createDatabase(String databaseName, CatalogDatabase db, boolean ignoreIfExists) throws DatabaseAlreadyExistException { checkArgument(!StringUtils.isNullOrWhitespaceOnly(databaseName)); checkNotNull(db); if (databaseExists(databaseName)) { if (!ignore...
3.26
flink_GenericInMemoryCatalog_isPartitionedTable_rdh
/** * Check if the given table is a partitioned table. Note that "false" is returned if the table * doesn't exists. */ private boolean isPartitionedTable(ObjectPath tablePath) { CatalogBaseTable table = null; try { ...
3.26
flink_GenericInMemoryCatalog_createFunction_rdh
// ------ functions ------ @Override public void createFunction(ObjectPath path, CatalogFunction function, boolean ignoreIfExists) throws FunctionAlreadyExistException, DatabaseNotExistException { checkNotNull(path); checkNotNull(function); ObjectPath functionPath = normalize(path); if (!databaseExists(...
3.26
flink_GenericInMemoryCatalog_isFullPartitionSpec_rdh
/** * Check if the given partitionSpec is full partition spec for the given table. */ private boolean isFullPartitionSpec(ObjectPath tablePath, CatalogPartitionSpec partitionSpec) throws TableNotExistException { CatalogBaseTable baseTable = getTable(tablePath); if (!(baseTable instanceof CatalogTable)) { ...
3.26
flink_GenericInMemoryCatalog_getTableStatistics_rdh
// ------ statistics ------ @Override public CatalogTableStatistics getTableStatistics(ObjectPath tablePath) throws TableNotExistException { checkNotNull(tablePath); if (!tableExists(tablePath)) { throw new TableNotExistException(getName(), tablePath); } i...
3.26
flink_GenericInMemoryCatalog_createTable_rdh
// ------ tables ------ @Override public void createTable(ObjectPath tablePath, CatalogBaseTable table, boolean ignoreIfExists) throws TableAlreadyExistException, DatabaseNotExistException { checkNotNull(tablePath); checkNotNull(table); if (!databaseExists(tablePath.getDatabaseName())) { throw...
3.26
flink_GenericInMemoryCatalog_createPartition_rdh
// ------ partitions ------ @Override public void createPartition(ObjectPath tablePath, CatalogPartitionSpec partitionSpec, CatalogPartition partition, boolean ignoreIfExists) throws TableNotExistException, TableNotPartitionedException, PartitionSpecInvalidException, PartitionAlreadyExistsException, CatalogException { ...
3.26
flink_ReduceCombineDriver_setup_rdh
// ------------------------------------------------------------------------ @Override public void setup(TaskContext<ReduceFunction<T>, T> context) { taskContext = context; running = true; }
3.26
flink_HiveTableSource_toHiveTablePartition_rdh
/** * Convert partition to HiveTablePartition. */ public HiveTablePartition toHiveTablePartition(Partition partition) { return HivePartitionUtils.toHiveTablePartition(partitionKeys, tableProps, partition); }
3.26
flink_CheckpointStatistics_generateCheckpointStatistics_rdh
// ------------------------------------------------------------------------- // Static factory methods // ------------------------------------------------------------------------- public static CheckpointStatistics generateCheckpointStatistics(AbstractCheckpointStats checkpointStats, boolean includeTaskCheckpointStati...
3.26
flink_OrcShim_defaultShim_rdh
/** * Default with orc dependent, we should use v2.3.0. */ static OrcShim<VectorizedRowBatch> defaultShim() { return new OrcShimV230(); }
3.26
flink_OrcShim_createShim_rdh
/** * Create shim from hive version. */ static OrcShim<VectorizedRowBatch> createShim(String hiveVersion) { if (hiveVersion.startsWith("2.0")) { return new OrcShimV200(); } else if (hiveVersion.startsWith("2.1")) { return new OrcShimV210(); } else if ((hiveVersion.startsWith("2.2") |...
3.26
flink_BeamStateRequestHandler_of_rdh
/** * Create a {@link BeamStateRequestHandler}. * * @param keyedStateBackend * if null, {@link BeamStateRequestHandler} would throw an error when * receive keyed-state requests. * @param operatorStateBackend * if null, {@link BeamStateRequestHandler} would throw an error * when receive operator-state re...
3.26
flink_JobIdsWithStatusOverview_hashCode_rdh
// ------------------------------------------------------------------------ @Override public int hashCode() { return jobsWithStatus.hashCode(); }
3.26
flink_JobIdsWithStatusOverview_combine_rdh
// ------------------------------------------------------------------------ private static Collection<JobIdWithStatus> combine(Collection<JobIdWithStatus> first, Collection<JobIdWithStatus> second) { checkNotNull(first); checkNotNull(second); ArrayList<JobIdWithStatus> result = new ArrayList<>(first.size() ...
3.26
flink_HeaderlessChannelWriterOutputView_close_rdh
/** * Closes this OutputView, closing the underlying writer. And return number bytes in last memory * segment. */ @Override public int close() throws IOException { if (!writer.isClosed()) { int currentPositionInSegment = getCurrentPositionInSegment(); // write last segment writer.writeBlock(getCurrentSeg...
3.26
flink_ResultRetryStrategy_fixedDelayRetry_rdh
/** * Create a fixed-delay retry strategy by given params. */ public static ResultRetryStrategy fixedDelayRetry(int maxAttempts, long backoffTimeMillis, Predicate<Collection<RowData>> resultPredicate) { return new ResultRetryStrategy(new AsyncRetryStrategies.FixedDelayRetryStrategyBuilder(maxAttempts, backoffTime...
3.26
flink_SSLUtils_createRestSSLContext_rdh
/** * Creates an SSL context for clients against the external REST endpoint. */ @Nullable @VisibleForTesting public static SSLContext createRestSSLContext(Configuration config, boolean clientMode) throws Exception { ClientAuth clientAuth = (SecurityOptions.isRe...
3.26
flink_SSLUtils_getAndCheckOption_rdh
// ------------------------------------------------------------------------ // Utilities // ------------------------------------------------------------------------ private static String getAndCheckOption(Configuration config, ConfigOption<String> primaryOption, ConfigOption<String> fallbackOption) { String value =...
3.26
flink_SSLUtils_createSSLClientSocketFactory_rdh
/** * Creates a factory for SSL Client Sockets from the given configuration. SSL Client Sockets are * always part of internal communication. */ public static SocketFactory createSSLClientSocketFactory(Configuration config) throws Exception { SSLContext sslContext = createInternalSSLContext(config, true); if ...
3.26
flink_SSLUtils_m0_rdh
/** * Creates a {@link SSLHandlerFactory} to be used by the REST Servers. * * @param config * The application configuration. */ public static SSLHandlerFactory m0(final Configuration config) throws Exception { ClientAuth clientAuth = (SecurityOptions.isRestSSLAuthenticationEnabled(config)) ? ClientAuth.REQUI...
3.26
flink_SSLUtils_createInternalClientSSLEngineFactory_rdh
/** * Creates a SSLEngineFactory to be used by internal communication client endpoints. */ public static SSLHandlerFactory createInternalClientSSLEngineFactory(final Configuration config) throws Exception { SslContext sslContext = m1(config, true); if (sslContext == null) { throw new IllegalConfigurat...
3.26
flink_SSLUtils_createSSLServerSocketFactory_rdh
/** * Creates a factory for SSL Server Sockets from the given configuration. SSL Server Sockets are * always part of internal communication. */ public static ServerSocketFactory createSSLServerSocketFactory(Configuration config) throws Exception { SSLContext sslContext = createInternalSSLContext(config, fal...
3.26
flink_SSLUtils_createRestNettySSLContext_rdh
/** * Creates an SSL context for the external REST SSL. If mutual authentication is configured the * client and the server side configuration are identical. */ @Nullable public static SslContext createRestNettySSLContext(Configuration config, boolean clientMode, ClientAuth clientAuth, SslProvider provider) throws Ex...
3.26
flink_SSLUtils_m2_rdh
/** * Creates the SSL Context for internal SSL, if internal SSL is configured. For internal SSL, * the client and server side configuration are identical, because of mutual authentication. */ @Nullable private static SslContext m2(Configuration config, boolean clientMode, SslProvider provider) throws Exception { ...
3.26
flink_SSLUtils_createRestClientSSLEngineFactory_rdh
/** * Creates a {@link SSLHandlerFactory} to be used by the REST Clients. * * @param config * The application configuration. */ public static SSLHandlerFactory createRestClientSSLEngineFactory(final Configuration config) throws Exception { ClientAuth clientAuth = (SecurityOptions.isRestSSLAuthenticationE...
3.26
flink_SSLUtils_createInternalServerSSLEngineFactory_rdh
/** * Creates a SSLEngineFactory to be used by internal communication server endpoints. */ public static SSLHandlerFactory createInternalServerSSLEngineFactory(final Configuration config) throws Exception { SslContext sslContext = m1(config, false); if (sslContext == null) { throw new IllegalConfigura...
3.26
flink_SSLUtils_createInternalSSLContext_rdh
/** * Creates the SSL Context for internal SSL, if internal SSL is configured. For internal SSL, * the client and server side configuration are identical, because of mutual authentication. */ @Nullable private static SSLContext createInternalSSLContext(Configuration config, boolean clientMode) throws Exception { ...
3.26
flink_HiveJdbcParameterUtils_m0_rdh
/** * Use the {@param parameters} to set {@param hiveConf} or {@param hiveVariables} according to * what kinds of the parameter belongs. */ public static void m0(HiveConf hiveConf, Map<String, String> sessionConfigs, Map<String, String> parameters) { for (Map.Entry<String, String> en...
3.26
flink_YarnTaskExecutorRunner_main_rdh
// ------------------------------------------------------------------------ // Program entry point // ------------------------------------------------------------------------ /** * The entry point for the YARN task executor runner. * * @param args * The command line arguments. */ public static void main(String[]...
3.26
flink_YarnTaskExecutorRunner_runTaskManagerSecurely_rdh
/** * The instance entry point for the YARN task executor. Obtains user group information and calls * the main work method {@link TaskManagerRunner#runTaskManager(Configuration, PluginManager)} * as a privileged action. * * @param args * The command line arguments. */ private static void runTaskManagerSecurely...
3.26
flink_FlinkS3FileSystem_getEntropyInjectionKey_rdh
// ------------------------------------------------------------------------ @Nullable @Override public String getEntropyInjectionKey() {return entropyInjectionKey; }
3.26
flink_TableConfigUtils_isOperatorDisabled_rdh
/** * Returns whether the given operator type is disabled. * * @param tableConfig * TableConfig object * @param operatorType * operator type to check * @return true if the given operator is disabled. */ public static boolean isOperatorDisabled(TableConfig tableConfig,...
3.26
flink_TableConfigUtils_getCalciteConfig_rdh
/** * Returns {@link CalciteConfig} wraps in the given TableConfig. * * @param tableConfig * TableConfig object * @return wrapped CalciteConfig. */ public static CalciteConfig getCalciteConfig(TableConfig tableConfig) { return tableConfig.getPlannerConfig().unwrap(CalciteConfig.class).orElse(CalciteConfig$....
3.26
flink_TableConfigUtils_getMaxIdleStateRetentionTime_rdh
/** * Similar to {@link TableConfig#getMaxIdleStateRetentionTime()}. * * @see TableConfig#getMaxIdleStateRetentionTime() */ @Deprecated public static long getMaxIdleStateRetentionTime(ReadableConfig tableConfig) { return (tableConfig.get(ExecutionConfigOptions.IDLE_STATE_RETENTION).toMillis() * 3) / 2; }
3.26
flink_TableConfigUtils_m0_rdh
/** * Returns the aggregate phase strategy configuration. * * @param tableConfig * TableConfig object * @return the aggregate phase strategy */ public static AggregatePhaseStrategy m0(ReadableConfig tableConfig) { String aggPhaseConf = tableConfig.get(TABLE_OPTIMIZER_AGG_PHASE_STRATEGY).trim(); if (agg...
3.26
flink_TableConfigUtils_getLocalTimeZone_rdh
/** * Similar to {@link TableConfig#getLocalTimeZone()} but extracting it from a generic {@link ReadableConfig}. * * @see TableConfig#getLocalTimeZone() */ public static ZoneId getLocalTimeZone(ReadableConfig tableConfig) { final String zone = tableConfig.get(TableConfigOptions.LOCAL_TIME_ZONE); if (TableCo...
3.26
flink_OperatorCoordinatorHolder_start_rdh
// ------------------------------------------------------------------------ // OperatorCoordinator Interface // ------------------------------------------------------------------------ public void start() throws Exception { mainThreadExecutor.assertRunningInMainThread(); checkState(context.isInitialized(), "Coo...
3.26
flink_OperatorCoordinatorHolder_coordinator_rdh
// ------------------------------------------------------------------------ // Properties // ------------------------------------------------------------------------ public OperatorCoordinator coordinator() { return coordinator; }
3.26
flink_OperatorCoordinatorHolder_abortCurrentTriggering_rdh
// ------------------------------------------------------------------------ // Checkpointing Callbacks // ------------------------------------------------------------------------ @Override public void abortCurrentTriggering() { // unfortunately, this method does not run in the scheduler executor, but in the // ...
3.26
flink_OperatorCoordinatorHolder_setupAllSubtaskGateways_rdh
// ------------------------------------------------------------------------ // miscellaneous helpers // ------------------------------------------------------------------------ private void setupAllSubtaskGateways() { for (int i = 0; i < operatorParallelism; i++) { setupSubtaskGateway(i); } }
3.26
flink_TimeUtils_toDuration_rdh
/** * Translates {@link Time} to {@link Duration}. * * @param time * time to transform into duration * @return duration equal to the given time */ public static Duration toDuration(Time time) { return Duration.of(time.getSize(), toChronoUnit(time.getUnit())); }
3.26
flink_TimeUtils_formatWithHighestUnit_rdh
/** * Pretty prints the duration as a lowest granularity unit that does not lose precision. * * <p>Examples: * * <pre>{@code Duration.ofMilliseconds(60000) will be printed as 1 min * Duration.ofHours(1).plusSeconds(1) will be printed as 3601 s}</pre> * * <b>NOTE:</b> It supports only durations that fit into lon...
3.26
flink_TimeUtils_m0_rdh
/** * * @param duration * to convert to string * @return duration string in millis */ public static String m0(final Duration duration) { return duration.toMillis() + TimeUnit.MILLISECONDS.labels.get(0); }
3.26
flink_TimeUtils_parseDuration_rdh
/** * Parse the given string to a java {@link Duration}. The string is in format "{length * value}{time unit label}", e.g. "123ms", "321 s". If no time unit label is specified, it will * be considered as milliseconds. * * <p>Supported time unit labels are: * * <ul> * <li>DA...
3.26
flink_TimeUtils_singular_rdh
/** * * @param label * the original label * @return the singular format of the original label */ private static String[] singular(String label) { return new String[]{ label }; }
3.26
flink_Order_isOrdered_rdh
// -------------------------------------------------------------------------------------------- /** * Checks, if this enum constant represents in fact an order. That is, whether this property is * not equal to <tt>Order.NONE</tt>. * * @return True, if this enum constant is unequal to <tt>Order.NONE</tt>, false othe...
3.26
flink_TypeInference_m0_rdh
/** * Sets the list of argument types for specifying a fixed, not overloaded, not vararg input * signature explicitly. * * <p>This information is useful for optional arguments with default value. In particular, * the number of arguments that need to be filled with a default value and their types is * important. ...
3.26
flink_TypeInference_outputTypeStrategy_rdh
/** * Sets the strategy for inferring the final output data type of a function call. * * <p>Required. */ public Builder outputTypeStrategy(TypeStrategy outputTypeStrategy) { this.outputTypeStrategy = Preconditions.checkNotNull(outputTypeStrategy, "Output type strategy must not be null."); return this; }
3.26
flink_TypeInference_inputTypeStrategy_rdh
/** * Sets the strategy for inferring and validating input arguments in a function call. * * <p>A {@link InputTypeStrategies#WILDCARD} strategy function is assumed by default. */ public Builder inputTypeStrategy(InputTypeStrategy inputTypeStrategy) { this.inputTypeStrategy = Preconditions.checkNotNull(inputType...
3.26
flink_TypeInference_typedArguments_rdh
/** * * @see #typedArguments(List) */ public Builder typedArguments(DataType... argumentTypes) { return m0(Arrays.asList(argumentTypes));}
3.26
flink_TypeInference_namedArguments_rdh
/** * * @see #namedArguments(List) */ public Builder namedArguments(String... argumentNames) { return namedArguments(Arrays.asList(argumentNames)); }
3.26
flink_TypeInference_accumulatorTypeStrategy_rdh
/** * Sets the strategy for inferring the intermediate accumulator data type of a function * call. */ public Builder accumulatorTypeStrategy(TypeStrategy accumulatorTypeStrategy) { this.f0 = Preconditions.checkNotNull(accumulatorTypeStrategy, "Accumulator type strategy must not be null."); return this; }
3.26
flink_TypeInference_newBuilder_rdh
/** * Builder for configuring and creating instances of {@link TypeInference}. */ public static TypeInference.Builder newBuilder() { return new TypeInference.Builder(); }
3.26
flink_AfterMatchSkipStrategy_noSkip_rdh
/** * Every possible match will be emitted. * * @return the created AfterMatchSkipStrategy */ public static NoSkipStrategy noSkip() { return NoSkipStrategy.INSTANCE; }
3.26
flink_AfterMatchSkipStrategy_skipToLast_rdh
/** * Discards every partial match that started before the last event of emitted match mapped to * *PatternName*. * * @param patternName * the pattern name to skip to * @return the created AfterMatchSkipStrategy */ public static SkipToLastStrategy skipToLast(String patternName) { return new SkipToLastStrat...
3.26
flink_AfterMatchSkipStrategy_skipPastLastEvent_rdh
/** * Discards every partial match that started before emitted match ended. * * @return the created AfterMatchSkipStrategy */ public static SkipPastLastStrategy skipPastLastEvent() { return SkipPastLastStrategy.INSTANCE; }
3.26
flink_AfterMatchSkipStrategy_skipToFirst_rdh
/** * Discards every partial match that started before the first event of emitted match mapped to * *PatternName*. * * @param patternName * the pattern name to skip to * @return the created AfterMatchSkipStrategy */ public static SkipToFirstStrategy skipToFirst(String patternName) { return new SkipToFirstS...
3.26
flink_AfterMatchSkipStrategy_getPatternName_rdh
/** * Name of pattern that processing will be skipped to. */ public Optional<String> getPatternName() { return Optional.empty(); }
3.26
flink_Savepoint_load_rdh
/** * Loads an existing savepoint. Useful if you want to query, modify, or extend the state of an * existing application. * * @param env * The execution environment used to transform the savepoint. * @param path * The path to an existing savepoint on disk. * @param stateBackend * The state backend of the...
3.26
flink_Savepoint_create_rdh
/** * Creates a new savepoint. * * @param stateBackend * The state backend of the savepoint used for keyed state. * @param maxParallelism * The max parallelism of the savepoint. * @return A new savepoint. * @see #create(int) */ public static NewSavepoint create(StateBackend stateBackend, int maxParallelism...
3.26
flink_AbstractRuntimeUDFContext_getAccumulator_rdh
// -------------------------------------------------------------------------------------------- @SuppressWarnings("unchecked") private <V, A extends Serializable> Accumulator<V, A> getAccumulator(String name, Class<? extends Accumulator<V, A>> accumulatorClass) { Accumulator<?, ?> accumulator = accumulators.get(n...
3.26
flink_SelectByMaxFunction_reduce_rdh
/** * Reduce implementation, returns bigger tuple or value1 if both tuples are equal. Comparison * highly depends on the order and amount of fields chosen as indices. All given fields (at * construction time) are checked in the same order as defined (at construction time). If both * tuples are equal in one index, t...
3.26
flink_GlobalWindow_snapshotConfiguration_rdh
// ------------------------------------------------------------------------ @Overridepublic TypeSerializerSnapshot<GlobalWindow> snapshotConfiguration() { return new GlobalWindowSerializerSnapshot(); }
3.26
flink_WorkerResourceSpec_setExtendedResource_rdh
/** * Add the given extended resource. The old value with the same resource name will be * replaced if present. */ public Builder setExtendedResource(ExternalResource extendedResource) { this.extendedResources.put(extendedResource.getName(), extendedResource); return this; }
3.26
flink_WorkerResourceSpec_setExtendedResources_rdh
/** * Add the given extended resources. This will discard all the previous added extended * resources. */ public Builder setExtendedResources(Collection<ExternalResource> extendedResources) { this.extendedResources = extendedResources.stream().collect(Collectors.toMap(ExternalResource::getName, Function.identi...
3.26
flink_SessionContext_isStatementSetState_rdh
// -------------------------------------------------------------------------------------------- // Begin statement set // -------------------------------------------------------------------------------------------- public boolean isStatementSetState() { return isStatementSetState;}
3.26
flink_SessionContext_create_rdh
// -------------------------------------------------------------------------------------------- // Utilities // -------------------------------------------------------------------------------------------- public static SessionContext create(DefaultContext defaultContext, SessionHandle sessionId, SessionEnvironment envi...
3.26
flink_SessionContext_close_rdh
// -------------------------------------------------------------------------------------------- /** * Close resources, e.g. catalogs. */ public void close() { operationManager.close(); for (String name : sessionState.catalogManager.listCatalogs()) { try { sessionState.catalogManager.getCat...
3.26
flink_SessionContext_getSessionId_rdh
// -------------------------------------------------------------------------------------------- // Getter/Setter // -------------------------------------------------------------------------------------------- public SessionHandle getSessionId() { return this.sessionId; }
3.26
flink_SessionContext_initializeConfiguration_rdh
// ------------------------------------------------------------------------------------------------------------------ // Helpers // ------------------------------------------------------------------------------------------------------------------ protected static Configuration initializeConfiguration(DefaultContext de...
3.26
flink_SessionContext_createOperationExecutor_rdh
// -------------------------------------------------------------------------------------------- // Method to execute commands // -------------------------------------------------------------------------------------------- public OperationExecutor createOperationExecutor(Configuration executionConfig) { return new ...
3.26
flink_KerberosLoginProvider_doLoginAndReturnUGI_rdh
/** * Does kerberos login and doesn't set current user, just returns a new UGI instance. Must be * called when isLoginPossible returns true. */ public UserGroupInformation doLoginAndReturnUGI() throws IOException { UserGroupInformation currentUser = UserGroupInformation.getCurrentUser(); if (principal != nul...
3.26
flink_KerberosLoginProvider_doLogin_rdh
/** * Does kerberos login and sets current user. Must be called when isLoginPossible returns true. */public void doLogin(boolean supportProxyUser) throws IOException { if (principal != null) { LOG.info("Attempting to login to KDC using principal: {} keytab: {}", principal, keytab); UserGroupInform...
3.26
flink_DatabaseMetaDataUtils_createSchemasResultSet_rdh
/** * Create result set for schemas. The schema columns are: * * <ul> * <li>TABLE_SCHEM String => schema name * <li>TABLE_CATALOG String => catalog name (may be null) * </ul> * * <p>The results are ordered by TABLE_CATALOG and TABLE_SCHEM. * * @param statement * The statement for database meta data * ...
3.26
streampipes_SpServiceDefinitionBuilder_m0_rdh
/** * Include migrations in the service definition. * <br> * Please refrain from providing {@link IModelMigrator}s with overlapping version definitions for one application id. * * @param migrations * List of migrations to be registered * @return {@link SpServiceDefinitionBuilder} */ public SpServiceDefinition...
3.26
streampipes_SpOpcUaConfigExtractor_extractAdapterConfig_rdh
/** * Creates {@link OpcUaAdapterConfig} instance in accordance with the given * {@link org.apache.streampipes.sdk.extractor.StaticPropertyExtractor}. * * @param extractor * extractor for user inputs * @return {@link OpcUaAdapterConfig} instance based on information from {@code extractor} */ public static Op...
3.26
streampipes_AbstractProcessingElementBuilder_naryMappingPropertyWithoutRequirement_rdh
/** * Adds a new {@link org.apache.streampipes.model.staticproperty.MappingPropertyNary} * to the pipeline element definition which is not linked to a specific input property. * Use this method if you want to present users a selection (in form of a Checkbox Group) * of all available input event properties. * * @p...
3.26
streampipes_AbstractProcessingElementBuilder_requiredStream_rdh
/** * Set a new stream requirement by adding restrictions on this stream. Use * {@link StreamRequirementsBuilder} to create requirements for a single stream. * * @param streamRequirements * A bundle of collected {@link CollectedStreamRequirements} * @return this */ public K...
3.26
streampipes_AbstractProcessingElementBuilder_supportedProtocols_rdh
/** * Assigns supported communication/transport protocols to the pipeline elements that can be handled at runtime (e.g., * Kafka or JMS). * * @param protocol * An arbitrary number of supported * {@link org.apache.streampipes.model.grounding.TransportProtocol}s. * Use {@link org.apache.streampipes.sdk.helpe...
3.26
streampipes_AbstractProcessingElementBuilder_setStream2_rdh
/** * * @deprecated Use {@link #requiredStream(CollectedStreamRequirements)} instead */ @Deprecated(since = "0.90.0", forRemoval = true) public K setStream2() { f1 = true; return me(); }
3.26
streampipes_AbstractProcessingElementBuilder_setStream1_rdh
/** * * @deprecated Use {@link #requiredStream(CollectedStreamRequirements)} instead */ @Deprecated(since = "0.90.0", forRemoval = true)public K setStream1() { stream1 = true; return me();}
3.26
streampipes_AbstractProcessingElementBuilder_supportedFormats_rdh
/** * Assigns supported transport formats to the pipeline elements that can be handled at runtime (e.g., * JSON or XMl). * * @param formats * A list of supported {@link org.apache.streampipes.model.grounding.TransportFormat}s. Use * {@link org.apache.streampipes.sdk.helpers.SupportedFormats} to assign formats...
3.26
streampipes_PipelineApi_delete_rdh
/** * Deletes the pipeline with a given id * * @param pipelineId * The id of the pipeline */ @Overridepublic void delete(String pipelineId) {delete(getBaseResourcePath().addToPath(pipelineId), Message.class); }
3.26
streampipes_PipelineApi_all_rdh
/** * Receives all pipelines owned by the current user * * @return (list) {@link org.apache.streampipes.model.pipeline.Pipeline} a list of all pipelines */ @Override public List<Pipeline> all() { return getAll(getBaseResourcePath()); }
3.26
streampipes_PipelineApi_start_rdh
/** * Starts a pipeline by given id * * @param pipeline * The pipeline * @return {@link org.apache.streampipes.model.pipeline.PipelineOperationStatus} the status message after invocation */ @Override public PipelineOperationStatus start(Pipeline pipeline) { return start(pipeline.getPipelineId()); }
3.26
streampipes_PipelineApi_stop_rdh
/** * Stops a pipeline by given id * * @param pipeline * The pipeline * @return {@link org.apache.streampipes.model.pipeline.PipelineOperationStatus} the status message after detach */ @Override public PipelineOperationStatus stop(Pipeline pipeline) { return stop(pipeline.getPipelineId()); } /** * Stops a...
3.26
streampipes_Options_from_rdh
/** * Creates a new list of options by using the provided string values. * * @param optionLabel * An arbitrary number of option labels. * @return */ public static List<Option> from(String... optionLabel) { return Arrays.stream(optionLabel).map(Option::new).collect(Collectors.toList()); }
3.26
streampipes_TokenizerProcessor_declareModel_rdh
// TODO: Maybe change outputStrategy to an array instead of tons of different strings @Override public DataProcessorDescription declareModel() { return ProcessingElementBuilder.create("org.apache.streampipes.processors.textmining.jvm.tokenizer").category(DataProcessorType.ENRICH_TEXT).withAssets(Assets.DOCUMENTATIO...
3.26
streampipes_OutputStrategies_append_rdh
/** * Creates a {@link org.apache.streampipes.model.output.AppendOutputStrategy}. Append output strategies add additional * properties to an input event stream. * * @param appendProperties * An arbitrary number of event properties that are appended to any input stream. * @return AppendOutputStrategy */ public ...
3.26
streampipes_OutputStrategies_custom_rdh
/** * Creates a {@link org.apache.streampipes.model.output.CustomOutputStrategy}. * * @param outputBoth * If two input streams are expected by a pipeline element, you can use outputBoth to indicate * whether the properties of both input streams should be available to the pipeline developer for * selection. ...
3.26