name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
open-banking-gateway_PsuAuthService_tryAuthenticateUser | /**
* Try to authenticate PSU give login and password
* @param login PSU login
* @param password PSU password
* @return PSU entity if user was successfully authenticated
* @throws PsuWrongCredentials Exception indicating user has provided wrong name or password.
*/
@Transactional
public Psu tryAuthenticateUser(St... | 3.68 |
hbase_AsyncRpcRetryingCallerFactory_start | /**
* Short cut for {@code build().start(HBaseRpcController, ScanResponse)}.
*/
public CompletableFuture<Boolean> start(HBaseRpcController controller,
ScanResponse respWhenOpen) {
return build().start(controller, respWhenOpen);
} | 3.68 |
dubbo_CompatibleTypeUtils_compatibleTypeConvert | /**
* Compatible type convert. Null value is allowed to pass in. If no conversion is needed, then the original value
* will be returned.
* <p>
* Supported compatible type conversions include (primary types and corresponding wrappers are not listed):
* <ul>
* <li> String -> char, enum, Date
* <li> byte, short, in... | 3.68 |
framework_LegacyWindow_removeComponent | /**
* This implementation removes the component from the content container (
* {@link #getContent()}) instead of from the actual UI.
*
* This method should only be called when the content is a
* {@link ComponentContainer} (default {@link VerticalLayout} or explicitly
* set).
*/
public void removeComponent(Compon... | 3.68 |
framework_VAbstractOrderedLayout_getCaptionPositionFromElement | /**
* Deducts the caption position by examining the wrapping element.
* <p>
* For internal use only. May be removed or replaced in the future.
*
* @param captionWrap
* The wrapping element
*
* @return The caption position
* @since 7.2
*/
public CaptionPosition getCaptionPositionFromElement(Element ... | 3.68 |
querydsl_Fetchable_stream | /**
* Get the projection as a typed closeable Stream.
*
* @return closeable stream
*/
default Stream<T> stream() {
final CloseableIterator<T> iterator = iterate();
final Spliterator<T> spliterator = Spliterators.spliteratorUnknownSize(iterator, Spliterator.ORDERED);
return StreamSupport.stream(spliterat... | 3.68 |
morf_AbstractSqlDialectTest_expectedSqlForMathOperations3 | /**
* @return expected SQL for math operation 3
*/
protected String expectedSqlForMathOperations3() {
return "a / (b + c)";
} | 3.68 |
morf_SchemaAdapter_getTable | /**
* @see org.alfasoftware.morf.metadata.Schema#getTable(java.lang.String)
*/
@Override
public Table getTable(String name) {
return delegate.getTable(name);
} | 3.68 |
flink_OneInputOperatorTransformation_keyBy | /**
* Partitions the operator state of a {@link OperatorTransformation} using field expressions. A
* field expression is either the name of a public field or a getter method with parentheses of
* the {@code OperatorTransformation}'s underlying type. A dot can be used to drill down into
* objects, as in {@code "fiel... | 3.68 |
flink_TableDescriptor_toBuilder | /** Converts this immutable instance into a mutable {@link Builder}. */
public Builder toBuilder() {
return new Builder(this);
} | 3.68 |
morf_AbstractSelectStatementBuilder_alias | /**
* Sets the alias for this select statement. This is useful if you are
* including multiple select statements in a single select (not to be confused
* with a join) and wish to reference the select statement itself.
*
* @param alias the alias to set.
* @return this, for method chaining.
*/
public T alias(Strin... | 3.68 |
flink_IntermediateResult_getConsumersParallelism | /**
* Currently, this method is only used to compute the maximum number of consumers. For dynamic
* graph, it should be called before adaptively deciding the downstream consumer parallelism.
*/
int getConsumersParallelism() {
List<JobEdge> consumers = intermediateDataSet.getConsumers();
checkState(!consumers... | 3.68 |
framework_DragSourceExtension_setDragImage | /**
* Set a custom drag image for the current drag source.
*
* @param imageResource
* Resource of the image to be displayed as drag image.
*/
public void setDragImage(Resource imageResource) {
setResource(DragSourceState.RESOURCE_DRAG_IMAGE, imageResource);
} | 3.68 |
hbase_ZkSplitLogWorkerCoordination_nodeChildrenChanged | /**
* Override handler from {@link ZKListener}
*/
@Override
public void nodeChildrenChanged(String path) {
if (path.equals(watcher.getZNodePaths().splitLogZNode)) {
if (LOG.isTraceEnabled()) {
LOG.trace("tasks arrived or departed on " + path);
}
synchronized (taskReadySeq) {
this.taskReadySe... | 3.68 |
framework_VAbstractCalendarPanel_renderCalendar | /**
* For internal use only. May be removed or replaced in the future.
*
* Updates the calendar and text field with the selected dates.
*
* @param updateDate
* The value false prevents setting the selected date of the
* calendar based on focusedDate. That can be used when only the
* ... | 3.68 |
graphhopper_DistanceCalcEarth_calcCircumference | /**
* Circumference of the earth at different latitudes (breitengrad)
*/
public double calcCircumference(double lat) {
return 2 * PI * R * cos(toRadians(lat));
} | 3.68 |
flink_Tuple6_setFields | /**
* Sets new values to all fields of the tuple.
*
* @param f0 The value for field 0
* @param f1 The value for field 1
* @param f2 The value for field 2
* @param f3 The value for field 3
* @param f4 The value for field 4
* @param f5 The value for field 5
*/
public void setFields(T0 f0, T1 f1, T2 f2, T3 f3, T4... | 3.68 |
flink_DataStream_global | /**
* Sets the partitioning of the {@link DataStream} so that the output values all go to the first
* instance of the next processing operator. Use this setting with care since it might cause a
* serious performance bottleneck in the application.
*
* @return The DataStream with shuffle partitioning set.
*/
@Publi... | 3.68 |
flink_HiveDDLUtils_disableConstraint | // returns a constraint trait that doesn't require ENABLE
public static byte disableConstraint(byte trait) {
return (byte) (trait & (~HIVE_CONSTRAINT_ENABLE));
} | 3.68 |
hadoop_AbstractS3ACommitter_getTaskAttemptPath | /**
* Compute the path where the output of a task attempt is stored until
* that task is committed. This may be the normal Task attempt path
* or it may be a subdirectory.
* The default implementation returns the value of
* {@link #getBaseTaskAttemptPath(TaskAttemptContext)};
* subclasses may return different val... | 3.68 |
hbase_CreateTableProcedure_addRegionsToMeta | /**
* Add the specified set of regions to the hbase:meta table.
*/
private static void addRegionsToMeta(final MasterProcedureEnv env,
final TableDescriptor tableDescriptor, final List<RegionInfo> regionInfos) throws IOException {
MetaTableAccessor.addRegionsToMeta(env.getMasterServices().getConnection(), regionIn... | 3.68 |
pulsar_FunctionApiResource_clientAuthData | /**
* @deprecated use {@link #authParams()} instead.
*/
@Deprecated
public AuthenticationDataSource clientAuthData() {
return (AuthenticationDataSource) httpRequest.getAttribute(AuthenticationFilter.AuthenticatedDataAttributeName);
} | 3.68 |
framework_AbstractComponent_getCustomAttributes | /**
* Returns a collection of attributes that should not be handled by the
* basic implementation of the {@link #readDesign(Element, DesignContext)}
* and {@link #writeDesign(Element, DesignContext)} methods. Typically these
* are handled in a custom way in the overridden versions of the above
* methods
*
* @sin... | 3.68 |
morf_H2Dialect_changePrimaryKeyColumns | /**
* @see org.alfasoftware.morf.jdbc.SqlDialect#changePrimaryKeyColumns(org.alfasoftware.morf.metadata.Table, java.util.List, java.util.List)
*/
@Override
public Collection<String> changePrimaryKeyColumns(Table table, List<String> oldPrimaryKeyColumns, List<String> newPrimaryKeyColumns) {
List<String> result = new... | 3.68 |
querydsl_SQLExpressions_percentileDisc | /**
* PERCENTILE_DISC is an inverse distribution function that assumes a discrete distribution model.
* It takes a percentile value and a sort specification and returns an element from the set.
* Nulls are ignored in the calculation.
*
* <p>This function takes as an argument any numeric datatype or any nonnumeric ... | 3.68 |
flink_SingleInputGate_retriggerPartitionRequest | /** Retriggers a partition request. */
public void retriggerPartitionRequest(
IntermediateResultPartitionID partitionId, int subpartitionIndex) throws IOException {
synchronized (requestLock) {
if (!closeFuture.isDone()) {
final InputChannel ch =
inputChannels.get(new... | 3.68 |
flink_ValueLiteralExpression_stringifyValue | /** Supports (nested) arrays and makes string values more explicit. */
private static String stringifyValue(Object value) {
if (value instanceof String[]) {
final String[] array = (String[]) value;
return Stream.of(array)
.map(ValueLiteralExpression::stringifyValue)
.... | 3.68 |
flink_CheckpointProperties_isUnclaimed | /** Returns whether the checkpoint should be restored in a {@link RestoreMode#NO_CLAIM} mode. */
public boolean isUnclaimed() {
return unclaimed;
} | 3.68 |
Activiti_TreeMethodExpression_isDeferred | /**
* Answer <code>true</code> if this is a deferred expression (starting with <code>#{</code>)
*/
public boolean isDeferred() {
return deferred;
} | 3.68 |
dubbo_SslContexts_findSslProvider | /**
* Returns OpenSSL if available, otherwise returns the JDK provider.
*/
private static SslProvider findSslProvider() {
if (OpenSsl.isAvailable()) {
logger.debug("Using OPENSSL provider.");
return SslProvider.OPENSSL;
}
if (checkJdkProvider()) {
logger.debug("Using JDK provider."... | 3.68 |
flink_SharedResources_getOrAllocateSharedResource | /**
* Gets the shared memory resource for the given owner and registers a lease. If the resource
* does not yet exist, it will be created via the given initializer function.
*
* <p>The resource must be released when no longer used. That releases the lease. When all
* leases are released, the resource is disposed.
... | 3.68 |
hbase_WALProcedureStore_initOldLog | /**
* Loads given log file and it's tracker.
*/
private ProcedureWALFile initOldLog(final FileStatus logFile, final Path walArchiveDir)
throws IOException {
final ProcedureWALFile log = new ProcedureWALFile(fs, logFile);
if (logFile.getLen() == 0) {
LOG.warn("Remove uninitialized log: {}", logFile);
log... | 3.68 |
hadoop_IOStatisticsStoreImpl_getMeanStatistic | /**
* Get a mean statistic.
* @param key statistic name
* @return the reference
* @throws NullPointerException if there is no entry of that name
*/
@Override
public MeanStatistic getMeanStatistic(String key) {
return lookup(meanStatisticMap, key);
} | 3.68 |
hbase_MetaTableAccessor_addRegionStateToPut | /**
* Set the column value corresponding to this {@code replicaId}'s {@link RegionState} to the
* provided {@code state}. Mutates the provided {@link Put}.
*/
public static Put addRegionStateToPut(Put put, int replicaId, RegionState.State state)
throws IOException {
put.add(CellBuilderFactory.create(CellBuilderT... | 3.68 |
zxing_ECIStringBuilder_appendECI | /**
* Appends ECI value to output.
*
* @param value ECI value to append, as an int
* @throws FormatException on invalid ECI value
*/
public void appendECI(int value) throws FormatException {
encodeCurrentBytesIfAny();
CharacterSetECI characterSetECI = CharacterSetECI.getCharacterSetECIByValue(value);
if (cha... | 3.68 |
flink_RocksDBIncrementalCheckpointUtils_deleteRange | /**
* Delete the record falls into [beginKeyBytes, endKeyBytes) of the db.
*
* @param db the target need to be clipped.
* @param columnFamilyHandles the column family need to be clipped.
* @param beginKeyBytes the begin key bytes
* @param endKeyBytes the end key bytes
*/
private static void deleteRange(
... | 3.68 |
AreaShop_AreaShop_getWorldGuard | /**
* Function to get the WorldGuard plugin.
* @return WorldGuardPlugin
*/
@Override
public WorldGuardPlugin getWorldGuard() {
return worldGuard;
} | 3.68 |
incubator-hugegraph-toolchain_DataTypeUtil_parseMultiValues | /**
* collection format: "obj1,obj2,...,objn" or "[obj1,obj2,...,objn]" ..etc
* TODO: After parsing to json, the order of the collection changed
* in some cases (such as list<date>)
**/
private static Object parseMultiValues(String key, Object values,
DataType dataType,
... | 3.68 |
pulsar_ClientCnxIdleState_tryMarkReleasing | /**
* Changes the idle-state of the connection to #{@link State#RELEASING}, This method only changes this
* connection from the #{@link State#IDLE} state to the #{@link State#RELEASING} state.
* @return Whether change idle-stat to #{@link State#RELEASING} success.
*/
public boolean tryMarkReleasing() {
return c... | 3.68 |
querydsl_ExpressionUtils_isNull | /**
* Create a {@code left is null} expression
*
* @param left operation argument
* @return left is null
*/
public static Predicate isNull(Expression<?> left) {
return predicate(Ops.IS_NULL, left);
} | 3.68 |
flink_NFACompiler_createMiddleStates | /**
* Creates all the states between Start and Final state.
*
* @param sinkState the state that last state should point to (always the Final state)
* @return the next state after Start in the resulting graph
*/
private State<T> createMiddleStates(final State<T> sinkState) {
State<T> lastSink = sinkState;
w... | 3.68 |
dubbo_RegistryBuilder_transport | /**
* @param transport
* @see #transporter(String)
* @deprecated
*/
@Deprecated
public RegistryBuilder transport(String transport) {
this.transporter = transport;
return getThis();
} | 3.68 |
hadoop_QuotaUsage_getTypeConsumed | /**
* Return storage type consumed.
*
* @param type storage type.
* @return type consumed.
*/
public long getTypeConsumed(StorageType type) {
return (typeConsumed != null) ? typeConsumed[type.ordinal()] : 0L;
} | 3.68 |
flink_TypeExtractor_createSubTypesInfo | /**
* Creates the TypeInformation for all elements of a type that expects a certain number of
* subtypes (e.g. TupleXX).
*
* @param originalType most concrete subclass
* @param definingType type that defines the number of subtypes (e.g. Tuple2 -> 2 subtypes)
* @param typeHierarchy necessary for type inference
* ... | 3.68 |
hbase_StateMachineProcedure_addChildProcedure | /**
* Add a child procedure to execute
* @param subProcedure the child procedure
*/
protected <T extends Procedure<TEnvironment>> void
addChildProcedure(@SuppressWarnings("unchecked") T... subProcedure) {
if (subProcedure == null) {
return;
}
final int len = subProcedure.length;
if (len == 0) {
ret... | 3.68 |
pulsar_ManagedLedgerImpl_isBkErrorNotRecoverable | /**
* return BK error codes that are considered not likely to be recoverable.
*/
private static boolean isBkErrorNotRecoverable(int rc) {
switch (rc) {
case Code.NoSuchLedgerExistsException:
case Code.NoSuchLedgerExistsOnMetadataServerException:
case Code.NoSuchEntryException:
return true;
... | 3.68 |
hbase_ClientTokenUtil_toToken | /**
* Converts a protobuf Token message back into a Token instance.
* @param proto the protobuf Token message
* @return the Token instance
*/
@InterfaceAudience.Private
static Token<AuthenticationTokenIdentifier> toToken(AuthenticationProtos.Token proto) {
return new Token<>(proto.hasIdentifier() ? proto.getIdent... | 3.68 |
flink_Configuration_get | /**
* Please check the java doc of {@link #getRawValueFromOption(ConfigOption)}. If no keys are
* found in {@link Configuration}, default value of the given option will return. Please make
* sure there will be at least one value available. Otherwise, a NPE will be thrown by Flink
* when the value is used.
*
* <p>... | 3.68 |
framework_Tree_getItemIdInto | /**
* If the event is on a node that can not have children (see
* {@link Tree#areChildrenAllowed(Object)}), this method returns the
* parent item id of the target item (see {@link #getItemIdOver()} ).
* The identifier of the parent node is also returned if the cursor is
* on the top part of node. Else this method ... | 3.68 |
pulsar_Topics_deletePartitionedTopicAsync | /**
* @see Topics#deletePartitionedTopic(String, boolean, boolean)
*/
default CompletableFuture<Void> deletePartitionedTopicAsync(String topic, boolean force) {
return deletePartitionedTopicAsync(topic, force, true);
} | 3.68 |
hbase_RegionInfo_encodeRegionName | /** Returns the encodedName */
@InterfaceAudience.Private
static String encodeRegionName(final byte[] regionName) {
String encodedName;
if (hasEncodedName(regionName)) {
// region is in new format:
// <tableName>,<startKey>,<regionIdTimeStamp>/encodedName/
encodedName =
Bytes.toString(regionName, ... | 3.68 |
hudi_BaseHoodieWriteClient_restoreToInstant | /**
* NOTE : This action requires all writers (ingest and compact) to a table to be stopped before proceeding. Revert
* the (inflight/committed) record changes for all commits after the provided instant time.
*
* @param savepointToRestoreTimestamp savepoint instant time to which restoration is requested
*/
public ... | 3.68 |
graphhopper_OSMReader_addEdge | /**
* This method is called for each segment an OSM way is split into during the second pass of {@link WaySegmentParser}.
*
* @param fromIndex a unique integer id for the first node of this segment
* @param toIndex a unique integer id for the last node of this segment
* @param pointList coordinates of this segme... | 3.68 |
flink_ForwardHashExchangeProcessor_updateOriginalEdgeInMultipleInput | /**
* Add new exchange node between the input node and the target node for the given edge, and
* reconnect the edges. So that the transformations can be connected correctly.
*/
private void updateOriginalEdgeInMultipleInput(
BatchExecMultipleInput multipleInput, int edgeIdx, BatchExecExchange newExchange) {
... | 3.68 |
flink_CopyOnWriteSkipListStateMap_getNodeStateHelper | /** Return the state of the node. null will be returned if the node is removed. */
private S getNodeStateHelper(long node) {
Node nodeStorage = getNodeSegmentAndOffset(node);
MemorySegment segment = nodeStorage.nodeSegment;
int offsetInSegment = nodeStorage.nodeOffset;
long valuePointer = SkipListUtils.... | 3.68 |
hadoop_ManifestSuccessData_getDescription | /**
* @return any description text.
*/
public String getDescription() {
return description;
} | 3.68 |
framework_AbstractDateField_setResolution | /**
* Sets the resolution of the DateField.
*
* The default resolution is {@link DateResolution#DAY} since Vaadin 7.0.
*
* @param resolution
* the resolution to set, not {@code null}
*/
public void setResolution(R resolution) {
if (!resolution.equals(this.resolution)) {
this.resolution = r... | 3.68 |
hbase_Client_head | /**
* Send a HEAD request
* @param cluster the cluster definition
* @param path the path or URI
* @param headers the HTTP headers to include in the request
* @return a Response object with response detail
*/
public Response head(Cluster cluster, String path, Header[] headers) throws IOException {
HttpHead me... | 3.68 |
hudi_OptionsResolver_isSchemaEvolutionEnabled | /**
* Returns whether comprehensive schema evolution enabled.
*/
public static boolean isSchemaEvolutionEnabled(Configuration conf) {
return conf.getBoolean(HoodieCommonConfig.SCHEMA_EVOLUTION_ENABLE.key(), HoodieCommonConfig.SCHEMA_EVOLUTION_ENABLE.defaultValue());
} | 3.68 |
flink_AllWindowedStream_sideOutputLateData | /**
* Send late arriving data to the side output identified by the given {@link OutputTag}. Data is
* considered late after the watermark has passed the end of the window plus the allowed
* lateness set using {@link #allowedLateness(Time)}.
*
* <p>You can get the stream of late data using {@link
* SingleOutputStr... | 3.68 |
hbase_HBaseFsckRepair_fixMetaHoleOnlineAndAddReplicas | /**
* Puts the specified RegionInfo into META with replica related columns
*/
public static void fixMetaHoleOnlineAndAddReplicas(Configuration conf, RegionInfo hri,
Collection<ServerName> servers, int numReplicas) throws IOException {
Connection conn = ConnectionFactory.createConnection(conf);
Table meta = conn... | 3.68 |
flink_FieldSet_clone | /**
* Since instances of FieldSet are strictly immutable, this method does not actually clone, but
* it only returns the original instance.
*
* @return This objects reference, unmodified.
*/
public FieldSet clone() {
return this;
} | 3.68 |
hbase_HttpServer_addServletWithAuth | /**
* Internal method to add a servlet to the HTTP server. Developers should not call this method
* directly, but invoke it via {@link #addUnprivilegedServlet(String, ServletHolder)} or
* {@link #addPrivilegedServlet(String, ServletHolder)}.
*/
void addServletWithAuth(String pathSpec, ServletHolder holder, boolean ... | 3.68 |
pulsar_JsonRecordBuilderImpl_clear | /**
* Clears the value of the given field.
*
* @param field the field to clear.
* @return a reference to the RecordBuilder.
*/
@Override
public GenericRecordBuilder clear(Field field) {
clear(field.getName());
return this;
} | 3.68 |
cron-utils_Preconditions_checkState | /**
* Ensures the truth of an expression involving the state of the calling instance, but not.
* involving any parameters to the calling method.
*
* @param expression a boolean expression
* @param errorMessageTemplate a template for the exception message should the check fail. The
* ... | 3.68 |
framework_TreeGrid_addCollapseListener | /**
* Adds a CollapseListener to this TreeGrid.
*
* @see CollapseEvent
*
* @param listener
* the listener to add
* @return a registration for the listener
*/
public Registration addCollapseListener(CollapseListener<T> listener) {
return addListener(CollapseEvent.class, listener,
Colla... | 3.68 |
morf_SqlDialect_legacyFetchSizeForBulkSelectsAllowingConnectionUseDuringStreaming | /**
* @deprecated this method returns the legacy value and is primarily for backwards compatibility.
* Please use {@link SqlDialect#fetchSizeForBulkSelectsAllowingConnectionUseDuringStreaming()} for the new recommended default value.
* @see SqlDialect#fetchSizeForBulkSelectsAllowingConnectionUseDuringStreaming()
*
... | 3.68 |
framework_Table_getVisibleColumns | /**
* Gets the array of visible column id:s, including generated columns.
*
* <p>
* The columns are show in the order of their appearance in this array.
* </p>
*
* @return an array of currently visible propertyIds and generated column
* ids.
*/
public Object[] getVisibleColumns() {
if (visibleColum... | 3.68 |
rocketmq-connect_WrapperStatusListener_onFailure | /**
* Invoked if the task raises an error. No shutdown event will follow.
*
* @param id The id of the task
* @param cause The error raised by the task.
*/
@Override
public void onFailure(ConnectorTaskId id, Throwable cause) {
managementService.putSafe(new TaskStatus(id, TaskStatus.State.FAILED, workerId, ge... | 3.68 |
framework_Overlay_getVisualViewportHeight | /**
* Gets the visual viewport height, which is useful for e.g iOS where the
* view can be zoomed in while keeping the layout viewport intact.
*
* Falls back to layout viewport; for those browsers/devices the difference
* is that the scrollbar with is included (if there is a scrollbar).
*
* @since 7.0.7
* @retu... | 3.68 |
druid_MySqlStatementParser_parseDeclareHandler | /**
* 定义异常处理程序
*
* @author zhujun [455910092@qq.com]
* 2016-04-16
*/
public MySqlDeclareHandlerStatement parseDeclareHandler() {
//DECLARE handler_type HANDLER FOR condition_value[,...] sp_statement
//handler_type 取值为 CONTINUE | EXIT | UNDO
//condition_value 取值为 SQLWARNING | NOT FOUND | SQLEXCEPTION | ... | 3.68 |
pulsar_BrokersBase_persistDynamicConfigurationAsync | /**
* if {@link ServiceConfiguration}-field is allowed to be modified dynamically, update configuration-map into zk, so
* all other brokers get the watch and can see the change and take appropriate action on the change.
*
* @param configName
* : configuration key
* @param configValue
* : co... | 3.68 |
hadoop_IOStatisticsBinding_copyMap | /**
* Copy into the dest map all the source entries.
* The destination is cleared first.
* @param <E> entry type
* @param dest destination of the copy
* @param source source
* @param copyFn function to copy entries
* @return the destination.
*/
private static <E> Map<String, E> copyMap(
Map<String, E> dest,... | 3.68 |
dubbo_LoggerFactory_getErrorTypeAwareLogger | /**
* Get error type aware logger by a String key.
*
* @param key the returned logger will be named after key
* @return error type aware logger
*/
public static ErrorTypeAwareLogger getErrorTypeAwareLogger(String key) {
return ConcurrentHashMapUtils.computeIfAbsent(
ERROR_TYPE_AWARE_LOGGERS, key, k... | 3.68 |
flink_RuntimeConverter_create | /**
* Creates a new instance of {@link Context}.
*
* @param classLoader runtime classloader for loading user-defined classes.
*/
static Context create(ClassLoader classLoader) {
return new Context() {
@Override
public ClassLoader getClassLoader() {
return classLoader;
}
}... | 3.68 |
hudi_SingleSparkJobExecutionStrategy_runClusteringForGroup | /**
* Submit job to execute clustering for the group.
*/
private Stream<WriteStatus> runClusteringForGroup(ClusteringGroupInfo clusteringOps, Map<String, String> strategyParams,
boolean preserveHoodieMetadata, SerializableSchema schema,
... | 3.68 |
dubbo_PortUnificationExchanger_getServers | // for test
public static ConcurrentMap<String, RemotingServer> getServers() {
return servers;
} | 3.68 |
flink_CreditBasedPartitionRequestClientHandler_checkError | /** Checks for an error and rethrows it if one was reported. */
@VisibleForTesting
void checkError() throws IOException {
final Throwable t = channelError.get();
if (t != null) {
if (t instanceof IOException) {
throw (IOException) t;
} else {
throw new IOException("There... | 3.68 |
open-banking-gateway_Xs2aFlowNameSelector_getNameForValidation | /**
* Sub-process name for current context (PSU/FinTech input) validation.
*/
public String getNameForValidation(Xs2aContext ctx) {
return actionName(ctx);
} | 3.68 |
flink_FromElementsFunction_getNumElementsEmitted | /**
* Gets the number of elements emitted so far.
*
* @return The number of elements emitted so far.
*/
public int getNumElementsEmitted() {
return numElementsEmitted;
} | 3.68 |
hadoop_RMContainerTokenSecretManager_rollMasterKey | /**
* Creates a new master-key and sets it as the primary.
*/
@Private
public void rollMasterKey() {
super.writeLock.lock();
try {
LOG.info("Rolling master-key for container-tokens");
if (this.currentMasterKey == null) { // Setting up for the first time.
this.currentMasterKey = createNewMasterKey();... | 3.68 |
hbase_TableMapReduceUtil_initCredentialsForCluster | /**
* Obtain an authentication token, for the specified cluster, on behalf of the current user and
* add it to the credentials for the given map reduce job.
* @param job The job that requires the permission.
* @param conf The configuration to use in connecting to the peer cluster
* @throws IOException When the au... | 3.68 |
flink_StreamTask_advanceToEndOfEventTime | /**
* Emits the {@link org.apache.flink.streaming.api.watermark.Watermark#MAX_WATERMARK
* MAX_WATERMARK} so that all registered timers are fired.
*
* <p>This is used by the source task when the job is {@code TERMINATED}. In the case, we want
* all the timers registered throughout the pipeline to fire and the relat... | 3.68 |
morf_UpdateStatement_drive | /**
* @see org.alfasoftware.morf.util.ObjectTreeTraverser.Driver#drive(ObjectTreeTraverser)
*/
@Override
public void drive(ObjectTreeTraverser traverser) {
traverser
.dispatch(getTable())
.dispatch(getWhereCriterion())
.dispatch(getFields())
.dispatch(getHints());
} | 3.68 |
pulsar_ResourceGroupService_unRegisterNameSpace | /**
* UnRegisters a namespace from a resource group.
*
* @param resourceGroupName
* @param fqNamespaceName i.e., in "tenant/Namespace" format)
* @throws if the RG does not exist, or if the NS does not references the RG yet.
*/
public void unRegisterNameSpace(String resourceGroupName, NamespaceName fqNamespaceName... | 3.68 |
zxing_EAN13Reader_determineFirstDigit | /**
* Based on pattern of odd-even ('L' and 'G') patterns used to encoded the explicitly-encoded
* digits in a barcode, determines the implicitly encoded first digit and adds it to the
* result string.
*
* @param resultString string to insert decoded first digit into
* @param lgPatternFound int whose bits indicat... | 3.68 |
hadoop_HeapSort_sort | /**
* Sort the given range of items using heap sort.
* {@inheritDoc}
*/
@Override
public void sort(IndexedSortable s, int p, int r) {
sort(s, p, r, null);
} | 3.68 |
hbase_Result_getRow | /**
* Method for retrieving the row key that corresponds to the row from which this Result was
* created.
*/
public byte[] getRow() {
if (this.row == null) {
this.row =
(this.cells == null || this.cells.length == 0) ? null : CellUtil.cloneRow(this.cells[0]);
}
return this.row;
} | 3.68 |
framework_Slot_getExpandRatio | /**
* Get the expand ratio for the slot. The expand ratio describes how the
* slot should be resized compared to other slots in the layout
*
* @return the expand ratio of the slot
*
* @see #setExpandRatio(double)
*/
public double getExpandRatio() {
return expandRatio;
} | 3.68 |
hbase_MultiByteBuff_getItemIndex | /*
* Returns in which sub ByteBuffer, the given element index will be available.
*/
private int getItemIndex(int elemIndex) {
if (elemIndex < 0) {
throw new IndexOutOfBoundsException();
}
int index = 1;
while (elemIndex >= this.itemBeginPos[index]) {
index++;
if (index == this.itemBeginPos.length)... | 3.68 |
hbase_HRegion_logFatLineOnFlush | /**
* Utility method broken out of internalPrepareFlushCache so that method is smaller.
*/
private void logFatLineOnFlush(Collection<HStore> storesToFlush, long sequenceId) {
if (!LOG.isInfoEnabled()) {
return;
}
// Log a fat line detailing what is being flushed.
StringBuilder perCfExtras = null;
if (!i... | 3.68 |
framework_VAbstractDropHandler_dragLeave | /**
* Default implementation does nothing. Implementors should clean possible
* emphasis or drag icons here.
*/
@Override
public void dragLeave(VDragEvent drag) {
} | 3.68 |
hudi_HoodieMetaSyncOperations_updateTableComments | /**
* Update the field comments for table in metastore, by using the ones from storage.
*
* @return
*/
default boolean updateTableComments(String tableName, List<FieldSchema> fromMetastore, List<FieldSchema> fromStorage) {
return false;
} | 3.68 |
aws-saas-boost_UpdateWorkflow_updateCloudFormationStack | // TODO all CloudFormation activities (reading params, updating stacks)
// should be extracted to a class for easier mocking/testing
private void updateCloudFormationStack(String stackName, Map<String, String> paramsMap, String yamlFile) {
List<Parameter> templateParameters = paramsMap.entrySet().stream()
... | 3.68 |
querydsl_SimpleExpression_isNotNull | /**
* Create a {@code this is not null} expression
*
* @return this is not null
*/
public BooleanExpression isNotNull() {
if (isnotnull == null) {
isnotnull = Expressions.booleanOperation(Ops.IS_NOT_NULL, mixin);
}
return isnotnull;
} | 3.68 |
hbase_Result_value | /**
* Returns the value of the first column in the Result.
* @return value of the first column
*/
public byte[] value() {
if (isEmpty()) {
return null;
}
return CellUtil.cloneValue(cells[0]);
} | 3.68 |
framework_GridRowDragger_setTargetDataProviderUpdater | /**
* Sets the target data provider updater, which handles adding the dropped
* items to the target grid.
* <p>
* By default, items are added to the index where they were dropped on for
* any {@link ListDataProvider}. If another type of data provider is used,
* this updater should be set to handle updating instea... | 3.68 |
querydsl_ExpressionUtils_count | /**
* Create a {@code count(source)} expression
*
* @param source source
* @return count(source)
*/
public static Expression<Long> count(Expression<?> source) {
return operation(Long.class, Ops.AggOps.COUNT_AGG, source);
} | 3.68 |
hadoop_AbstractConfigurableFederationPolicy_setPolicyInfo | /**
* Setter method for the configuration weights.
*
* @param policyInfo the {@link WeightedPolicyInfo} representing the policy
* configuration.
*/
public void setPolicyInfo(WeightedPolicyInfo policyInfo) {
this.policyInfo = policyInfo;
} | 3.68 |
framework_VAbsoluteLayout_setWidgetWrapperStyleNames | /**
* Sets style names for the wrapper wrapping the widget in the layout. The
* style names will be prefixed with v-absolutelayout-wrapper.
*
* @param widget
* The widget which wrapper we want to add the stylenames to
* @param stylenames
* The style names that should be added to the wrapper... | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.