name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_StringUtils_writeString | /**
* Writes a String to the given output. The written string can be read with {@link
* #readString(DataInputView)}.
*
* @param str The string to write
* @param out The output to write to
* @throws IOException Thrown, if the writing or the serialization fails.
*/
public static void writeString(@Nonnull String st... | 3.68 |
hbase_HBaseConfiguration_merge | /**
* Merge two configurations.
* @param destConf the configuration that will be overwritten with items from the srcConf
* @param srcConf the source configuration
**/
public static void merge(Configuration destConf, Configuration srcConf) {
for (Map.Entry<String, String> e : srcConf) {
destConf.set(e.getKey(... | 3.68 |
framework_VFilterSelect_updatePopupPositionOnScroll | /**
* Make the popup follow the position of the ComboBox when the page is
* scrolled.
*/
private void updatePopupPositionOnScroll() {
if (!scrollPending) {
AnimationScheduler.get()
.requestAnimationFrame(new AnimationCallback() {
public void execute(double timestamp) {... | 3.68 |
flink_StreamExecutionEnvironment_registerJobListener | /**
* Register a {@link JobListener} in this environment. The {@link JobListener} will be notified
* on specific job status changed.
*/
@PublicEvolving
public void registerJobListener(JobListener jobListener) {
checkNotNull(jobListener, "JobListener cannot be null");
jobListeners.add(jobListener);
} | 3.68 |
flink_FutureUtils_composeAfterwards | /**
* Run the given asynchronous action after the completion of the given future. The given future
* can be completed normally or exceptionally. In case of an exceptional completion, the
* asynchronous action's exception will be added to the initial exception.
*
* @param future to wait for its completion
* @param... | 3.68 |
framework_HierarchicalDataProvider_fetch | /**
* Fetches data from this HierarchicalDataProvider using given
* {@code query}. Only the immediate children of
* {@link HierarchicalQuery#getParent()} will be returned.
*
* @param query
* given query to request data with
* @return a stream of data objects resulting from the query
*
* @throws Ille... | 3.68 |
hbase_RegionSplitter_split2 | /**
* Divide 2 numbers in half (for split algorithm)
* @param a number #1
* @param b number #2
* @return the midpoint of the 2 numbers
*/
public BigInteger split2(BigInteger a, BigInteger b) {
return a.add(b).divide(BigInteger.valueOf(2)).abs();
} | 3.68 |
hudi_ClusteringCommand_runClustering | /**
* Run clustering table service.
* <p>
* Example:
* > connect --path {path to hudi table}
* > clustering scheduleAndExecute --sparkMaster local --sparkMemory 2g
*/
@ShellMethod(key = "clustering scheduleAndExecute", value = "Run Clustering. Make a cluster plan first and execute that plan immediately")
public S... | 3.68 |
querydsl_NumberExpression_divide | /**
* Create a {@code this / right} expression
*
* <p>Get the result of the operation this / right</p>
*
* @param right
* @return this / right
*/
public <N extends Number & Comparable<?>> NumberExpression<T> divide(N right) {
@SuppressWarnings("unchecked")
Class<T> type = (Class< T>) getDivisionType(getT... | 3.68 |
hbase_QuotaUtil_disableTableIfNotDisabled | /**
* Method to disable a table, if not already disabled. This method suppresses
* {@link TableNotEnabledException}, if thrown while disabling the table.
* @param conn connection to re-use
* @param tableName table name which has moved into space quota violation
*/
public static void disableTableIfNotDisabled(... | 3.68 |
hibernate-validator_ValidationXmlParser_run | /**
* Runs the given privileged action, using a privileged block if required.
* <p>
* <b>NOTE:</b> This must never be changed into a publicly available method to avoid execution of arbitrary
* privileged actions within HV's protection domain.
*/
@IgnoreForbiddenApisErrors(reason = "SecurityManager is deprecated in... | 3.68 |
hbase_RegionStateNode_offline | /**
* Put region into OFFLINE mode (set state and clear location).
* @return Last recorded server deploy
*/
public ServerName offline() {
setState(State.OFFLINE);
return setRegionLocation(null);
} | 3.68 |
flink_TableConfig_setPlannerConfig | /**
* Sets the configuration of Planner for Table API and SQL queries. Changing the configuration
* has no effect after the first query has been defined.
*/
public void setPlannerConfig(PlannerConfig plannerConfig) {
this.plannerConfig = Preconditions.checkNotNull(plannerConfig);
} | 3.68 |
framework_StringToBooleanConverter_getTrueString | /**
* Gets the locale-depended string representation for true. Default is
* locale-independent value provided by {@link #getTrueString()}
*
* @since 7.5.4
* @param locale
* to be used
* @return the string representation for true
*/
protected String getTrueString(Locale locale) {
return getTrueStr... | 3.68 |
hbase_ProcedureExecutor_submitProcedure | // ==========================================================================
// Submit/Abort Procedure
// ==========================================================================
/**
* Add a new root-procedure to the executor.
* @param proc the new procedure to execute.
* @return the procedure id, that can be use... | 3.68 |
flink_StreamOperatorWrapper_close | /** Close the operator. */
public void close() throws Exception {
closed = true;
wrapped.close();
} | 3.68 |
querydsl_StringExpression_trim | /**
* Create a {@code this.trim()} expression
*
* <p>Create a copy of the string, with leading and trailing whitespace
* omitted.</p>
*
* @return this.trim()
* @see java.lang.String#trim()
*/
public StringExpression trim() {
if (trim == null) {
trim = Expressions.stringOperation(Ops.TRIM, mixin);
... | 3.68 |
hadoop_FederationMembershipStateStoreInputValidator_checkTimestamp | /**
* Validate if the timestamp is positive or not.
*
* @param timestamp the timestamp to be verified
* @throws FederationStateStoreInvalidInputException if the timestamp is
* invalid
*/
private static void checkTimestamp(long timestamp)
throws FederationStateStoreInvalidInputException {
if (times... | 3.68 |
rocketmq-connect_WorkerSinkTask_pauseAll | //pause all consumer topic queue
private void pauseAll() {
consumer.pause(messageQueues);
} | 3.68 |
hadoop_NamenodeStatusReport_statsValid | /**
* If the statistics are valid.
*
* @return If the statistics are valid.
*/
public boolean statsValid() {
return this.statsValid;
} | 3.68 |
framework_AbsoluteLayoutRelativeSizeContent_createFullOnFixed | /**
* Creates an {@link AbsoluteLayout} of fixed size that contains a
* fixed-sized {@link AbsoluteLayout}.
*
* @return the created layout
*/
private Component createFullOnFixed() {
AbsoluteLayout absoluteLayout = new AbsoluteLayout();
absoluteLayout.setWidth(200, Unit.PIXELS);
absoluteLayout.setHeight... | 3.68 |
hadoop_BaseService_getServiceDependencies | /**
* Returns the service dependencies of this service. The service will be
* instantiated only if all the service dependencies are already initialized.
* <p>
* This method returns an empty array (size 0)
*
* @return an empty array (size 0).
*/
@Override
public Class[] getServiceDependencies() {
return new Cla... | 3.68 |
framework_CellReference_set | /**
* Sets the identifying information for this cell.
* <p>
* The difference between {@link #columnIndexDOM} and {@link #columnIndex}
* comes from hidden columns.
*
* @param columnIndexDOM
* the index of the column in the DOM
* @param columnIndex
* the index of the column
* @param column... | 3.68 |
flink_CsvBulkWriter_forPojo | /**
* Builds a writer based on a POJO class definition.
*
* @param pojoClass The class of the POJO.
* @param stream The output stream.
* @param <T> The type of the elements accepted by this writer.
*/
static <T> CsvBulkWriter<T, T, Void> forPojo(Class<T> pojoClass, FSDataOutputStream stream) {
final Converter... | 3.68 |
zxing_PDF417Writer_bitMatrixFromEncoder | /**
* Takes encoder, accounts for width/height, and retrieves bit matrix
*/
private static BitMatrix bitMatrixFromEncoder(PDF417 encoder,
String contents,
int errorCorrectionLevel,
... | 3.68 |
flink_WindowSavepointReader_process | /**
* Reads window state generated without any preaggregation such as {@code WindowedStream#apply}
* and {@code WindowedStream#process}.
*
* @param uid The uid of the operator.
* @param readerFunction The window reader function.
* @param keyType The key type of the window.
* @param stateType The type of records ... | 3.68 |
framework_VTree_getNextSibling | /**
* Gets the next sibling in the tree
*
* @param node
* The node to get the sibling for
* @return The sibling node or null if the node is the last sibling
*/
private TreeNode getNextSibling(TreeNode node) {
TreeNode parent = node.getParentNode();
List<TreeNode> children;
if (parent == nul... | 3.68 |
flink_RemoteInputChannel_notifyBufferAvailable | /**
* The unannounced credit is increased by the given amount and might notify increased credit to
* the producer.
*/
@Override
public void notifyBufferAvailable(int numAvailableBuffers) throws IOException {
if (numAvailableBuffers > 0 && unannouncedCredit.getAndAdd(numAvailableBuffers) == 0) {
notifyCre... | 3.68 |
flink_OrcLegacyTimestampColumnVector_createFromConstant | // creates a Hive ColumnVector of constant timestamp value
public static ColumnVector createFromConstant(int batchSize, Object value) {
LongColumnVector res = new LongColumnVector(batchSize);
if (value == null) {
res.noNulls = false;
res.isNull[0] = true;
res.isRepeating = true;
} el... | 3.68 |
framework_AbstractColorPicker_getSwatchesVisibility | /**
* Gets the visibility of the Swatches Tab.
*
* @since 7.5.0
* @return visibility of the swatches tab
*/
public boolean getSwatchesVisibility() {
return swatchesVisible;
} | 3.68 |
flink_AbstractKeyedStateBackend_getOrCreateKeyedState | /** @see KeyedStateBackend */
@Override
@SuppressWarnings("unchecked")
public <N, S extends State, V> S getOrCreateKeyedState(
final TypeSerializer<N> namespaceSerializer, StateDescriptor<S, V> stateDescriptor)
throws Exception {
checkNotNull(namespaceSerializer, "Namespace serializer");
checkNo... | 3.68 |
graphhopper_MiniPerfTest_getMin | /**
* @return minimum time of every call, in ms
*/
public double getMin() {
return min / NS_PER_MS;
} | 3.68 |
hadoop_BufferData_setDone | /**
* Indicates that this block is no longer of use and can be reclaimed.
*/
public synchronized void setDone() {
if (this.checksum != 0) {
if (getChecksum(this.buffer) != this.checksum) {
throw new IllegalStateException("checksum changed after setReady()");
}
}
this.state = State.DONE;
this.act... | 3.68 |
framework_VTabsheet_setFocusedTab | /**
* Sets the tab that has the focus currently.
*
* @param focusedTab
* the focused tab or {@code null} if no tab should be
* focused anymore
*/
private void setFocusedTab(Tab focusedTab) {
this.focusedTab = focusedTab;
} | 3.68 |
flink_BinarySegmentUtils_readBinary | /**
* Get binary, if len less than 8, will be include in variablePartOffsetAndLen.
*
* <p>Note: Need to consider the ByteOrder.
*
* @param baseOffset base offset of composite binary format.
* @param fieldOffset absolute start offset of 'variablePartOffsetAndLen'.
* @param variablePartOffsetAndLen a long value, r... | 3.68 |
framework_FocusableFlowPanel_addKeyPressHandler | /*
* (non-Javadoc)
*
* @see
* com.google.gwt.event.dom.client.HasKeyPressHandlers#addKeyPressHandler
* (com.google.gwt.event.dom.client.KeyPressHandler)
*/
@Override
public HandlerRegistration addKeyPressHandler(KeyPressHandler handler) {
return addDomHandler(handler, KeyPressEvent.getType());
} | 3.68 |
hadoop_TextView_echoWithoutEscapeHtml | /**
* Print strings as is (no newline, a la php echo).
* @param args the strings to print
*/
public void echoWithoutEscapeHtml(Object... args) {
PrintWriter out = writer();
for (Object s : args) {
out.print(s);
}
} | 3.68 |
hbase_SegmentFactory_createImmutableSegmentByCompaction | // create new flat immutable segment from compacting old immutable segments
// for compaction
public ImmutableSegment createImmutableSegmentByCompaction(final Configuration conf,
final CellComparator comparator, MemStoreSegmentsIterator iterator, int numOfCells,
CompactingMemStore.IndexType idxType, MemStoreCompact... | 3.68 |
flink_ProjectOperator_projectTuple8 | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5, T6, T7>
ProjectOperator<T, Tuple8<T0, T1, T2, T3, T4, T5, T6, T7>> projectTuple8() {
TypeInformation<?>[] fTypes = extract... | 3.68 |
hadoop_ExitUtil_getFirstExitException | /**
* @return the first {@code ExitException} thrown, null if none thrown yet.
*/
public static ExitException getFirstExitException() {
return FIRST_EXIT_EXCEPTION.get();
} | 3.68 |
flink_SplitFetcher_pauseOrResumeSplits | /**
* Called when some splits of this source instance progressed too much beyond the global
* watermark of all subtasks. If the split reader implements {@link SplitReader}, it will relay
* the information asynchronously through the split fetcher thread.
*
* @param splitsToPause the splits to pause
* @param splits... | 3.68 |
framework_UIEvents_getUI | /**
* Get the {@link UI} instance that received the poll request.
*
* @return the {@link UI} that received the poll request. Never
* <code>null</code>.
*/
public UI getUI() {
/*
* This cast is safe to make, since this class' constructor
* constrains the source to be a UI instance.
*/
... | 3.68 |
flink_FlinkContainersSettings_jobManagerHostname | /**
* Sets the job manager hostname and returns a reference to this Builder enabling method
* chaining.
*
* @param jobManagerHostname The job manager hostname to set.
* @return A reference to this Builder.
*/
public Builder jobManagerHostname(String jobManagerHostname) {
return setConfigOption(JobManagerOptio... | 3.68 |
framework_ComponentRendererConnector_createConnectorHierarchyChangeHandler | /**
* Adds a listener for grid hierarchy changes to find detached connectors
* previously handled by this renderer in order to detach from DOM their
* widgets before {@link AbstractComponentConnector#onUnregister()} is
* invoked otherwise an error message is logged.
*/
private void createConnectorHierarchyChangeHa... | 3.68 |
hadoop_StartupProgress_beginPhase | /**
* Begins execution of the specified phase.
*
* @param phase Phase to begin
*/
public void beginPhase(Phase phase) {
if (!isComplete()) {
phases.get(phase).beginTime = monotonicNow();
}
LOG.debug("Beginning of the phase: {}", phase);
} | 3.68 |
dubbo_SerializingExecutor_execute | /**
* Runs the given runnable strictly after all Runnables that were submitted
* before it, and using the {@code executor} passed to the constructor. .
*/
@Override
public void execute(Runnable r) {
runQueue.add(r);
schedule(r);
} | 3.68 |
hbase_TaskMonitor_getTasks | /**
* Produces a list containing copies of the current state of all non-expired MonitoredTasks
* handled by this TaskMonitor.
* @param filter type of wanted tasks
* @return A filtered list of MonitoredTasks.
*/
public synchronized List<MonitoredTask> getTasks(String filter) {
purgeExpiredTasks();
TaskFilter ta... | 3.68 |
flink_TieredStorageConfiguration_getMinReserveDiskSpaceFraction | /**
* Minimum reserved disk space fraction in disk tier.
*
* @return the fraction.
*/
public float getMinReserveDiskSpaceFraction() {
return minReserveDiskSpaceFraction;
} | 3.68 |
hadoop_RemoteSASKeyGeneratorImpl_makeRemoteRequest | /**
* Helper method to make a remote request.
*
* @param urls - Urls to use for the remote request
* @param path - hadoop.auth token for the remote request
* @param queryParams - queryParams to be used.
* @return RemoteSASKeyGenerationResponse
*/
private RemoteSASKeyGenerationResponse makeRemoteReq... | 3.68 |
flink_AbstractBlobCache_setBlobServerAddress | /**
* Sets the address of the {@link BlobServer}.
*
* @param blobServerAddress address of the {@link BlobServer}.
*/
public void setBlobServerAddress(InetSocketAddress blobServerAddress) {
serverAddress = checkNotNull(blobServerAddress);
} | 3.68 |
morf_SelectStatement_useImplicitJoinOrder | /**
* If supported by the dialect, hints to the database that joins should be applied in the order
* they are written in the SQL statement.
*
* <p>This is supported to greater or lesser extends on different SQL dialects. For instance,
* MySQL has no means to force ordering on anything except inner joins, but we d... | 3.68 |
morf_Criterion_like | /**
* Helper method to create a new "LIKE" expression.
*
* <blockquote><pre>
* Criterion.like(new Field("agreementnumber"), "A%");</pre></blockquote>
*
* <p>Note the escape character is set to '\' (backslash) by the underlying system.</p>
*
* @param field the field to evaluate in the expression (the left han... | 3.68 |
pulsar_SubscriptionStatsImpl_add | // if the stats are added for the 1st time, we will need to make a copy of these stats and add it to the current
// stats
public SubscriptionStatsImpl add(SubscriptionStatsImpl stats) {
Objects.requireNonNull(stats);
this.msgRateOut += stats.msgRateOut;
this.msgThroughputOut += stats.msgThroughputOut;
t... | 3.68 |
hadoop_DomainColumn_getColumnQualifier | /**
* @return the column name value
*/
private String getColumnQualifier() {
return columnQualifier;
} | 3.68 |
hbase_MasterRpcServices_checkMasterProcedureExecutor | /**
* @throws ServiceException If no MasterProcedureExecutor
*/
private void checkMasterProcedureExecutor() throws ServiceException {
if (this.server.getMasterProcedureExecutor() == null) {
throw new ServiceException("Master's ProcedureExecutor not initialized; retry later");
}
} | 3.68 |
flink_GatewayRetriever_getNow | /**
* Returns the currently retrieved gateway if there is such an object. Otherwise it returns an
* empty optional.
*
* @return Optional object to retrieve
*/
default Optional<T> getNow() {
CompletableFuture<T> leaderFuture = getFuture();
if (leaderFuture != null) {
if (leaderFuture.isCompletedExce... | 3.68 |
flink_SharedBufferAccessor_releaseNode | /**
* Decreases the reference counter for the given entry so that it can be removed once the
* reference counter reaches 0.
*
* @param node id of the entry
* @param version dewey number of the (potential) edge that locked the given node
* @throws Exception Thrown if the system cannot access the state.
*/
public ... | 3.68 |
querydsl_AbstractCollQuery_bind | /**
* Bind the given collection to an already existing query source
*
* @param <A> type of expression
* @param entity Path for the source
* @param col content of the source
* @return current object
*/
public <A> Q bind(Path<A> entity, Iterable<? extends A> col) {
iterables.put(entity, col);
return queryM... | 3.68 |
hbase_HRegionServer_getRegions | /**
* Gets the online regions of the specified table. This method looks at the in-memory
* onlineRegions. It does not go to <code>hbase:meta</code>. Only returns <em>online</em> regions.
* If a region on this table has been closed during a disable, etc., it will not be included in
* the returned list. So, the retur... | 3.68 |
hbase_MetricsHeapMemoryManager_setCurBlockCacheSizeGauge | /**
* Set the current blockcache size used gauge
* @param blockCacheSize the current memory usage in blockcache, in bytes.
*/
public void setCurBlockCacheSizeGauge(final long blockCacheSize) {
source.setCurBlockCacheSizeGauge(blockCacheSize);
} | 3.68 |
flink_Costs_addHeuristicCpuCost | /**
* Adds the given heuristic CPU cost to the current heuristic CPU cost for this Costs object.
*
* @param cost The heuristic CPU cost to add.
*/
public void addHeuristicCpuCost(double cost) {
if (cost <= 0) {
throw new IllegalArgumentException("Heuristic costs must be positive.");
}
this.heuri... | 3.68 |
framework_VDragAndDropManager_executeWhenReady | /**
* Method to execute commands when all existing dd related tasks are
* completed (some may require server visit).
* <p>
* Using this method may be handy if criterion that uses lazy initialization
* are used. Check
* <p>
* TODO Optimization: consider if we actually only need to keep the last
* command in queu... | 3.68 |
morf_OracleDialect_getSqlForNow | /**
* @see org.alfasoftware.morf.jdbc.SqlDialect#getSqlForNow(org.alfasoftware.morf.sql.element.Function)
*/
@Override
protected String getSqlForNow(Function function) {
return "SYSTIMESTAMP AT TIME ZONE 'UTC'";
} | 3.68 |
flink_FlinkDatabaseMetaData_storesMixedCaseIdentifiers | /** Flink sql is mixed case as sensitive. */
@Override
public boolean storesMixedCaseIdentifiers() throws SQLException {
return true;
} | 3.68 |
flink_KubernetesEntrypointUtils_loadConfiguration | /**
* For non-HA cluster, {@link JobManagerOptions#ADDRESS} has be set to Kubernetes service name
* on client side. See {@link KubernetesClusterDescriptor#deployClusterInternal}. So the
* TaskManager will use service address to contact with JobManager. For HA cluster, {@link
* JobManagerOptions#ADDRESS} will be set... | 3.68 |
querydsl_PathBuilder_getBoolean | /**
* Create a new Boolean typed path
*
* @param propertyName property name
* @return property path
*/
public BooleanPath getBoolean(String propertyName) {
validate(propertyName, Boolean.class);
return super.createBoolean(propertyName);
} | 3.68 |
pulsar_Commands_peekAndCopyMessageMetadata | /**
* Peek the message metadata from the buffer and return a deep copy of the metadata.
*
* If you want to hold multiple {@link MessageMetadata} instances from multiple buffers, you must call this method
* rather than {@link Commands#peekMessageMetadata(ByteBuf, String, long)}, which returns a thread local referenc... | 3.68 |
flink_UnorderedStreamElementQueue_emitCompleted | /**
* Pops one completed elements into the given output. Because an input element may produce
* an arbitrary number of output elements, there is no correlation between the size of the
* collection and the popped elements.
*
* @return the number of popped input elements.
*/
int emitCompleted(TimestampedCollector<O... | 3.68 |
open-banking-gateway_PathHeadersBodyMapperTemplate_forExecution | /**
* Converts context object into object that can be used for ASPSP API call.
* @param context Context to convert
* @return Object that can be used with {@code Xs2aAdapter} to perform ASPSP API calls
*/
public ValidatedPathHeadersBody<P, H, B> forExecution(C context) {
return new ValidatedPathHeadersBody<>(
... | 3.68 |
flink_RestServerEndpoint_shutDownInternal | /**
* Stops this REST server endpoint.
*
* @return Future which is completed once the shut down has been finished.
*/
protected CompletableFuture<Void> shutDownInternal() {
synchronized (lock) {
CompletableFuture<?> channelFuture = new CompletableFuture<>();
if (serverChannel != null) {
... | 3.68 |
hbase_RequestConverter_buildGetServerInfoRequest | /**
* Create a new GetServerInfoRequest
* @return a GetServerInfoRequest
*/
public static GetServerInfoRequest buildGetServerInfoRequest() {
return GetServerInfoRequest.getDefaultInstance();
} | 3.68 |
hbase_HBaseTestingUtility_getOtherRegionServer | /**
* Find any other region server which is different from the one identified by parameter
* @return another region server
*/
public HRegionServer getOtherRegionServer(HRegionServer rs) {
for (JVMClusterUtil.RegionServerThread rst : getMiniHBaseCluster().getRegionServerThreads()) {
if (!(rst.getRegionServer() ... | 3.68 |
hbase_CellCodec_write | /**
* Write int length followed by array bytes.
*/
private void write(final byte[] bytes, final int offset, final int length) throws IOException {
// TODO add BB backed os check and do for write. Pass Cell
this.out.write(Bytes.toBytes(length));
this.out.write(bytes, offset, length);
} | 3.68 |
hadoop_WriteOperationHelper_select | /**
* Execute an S3 Select operation.
* On a failure, the request is only logged at debug to avoid the
* select exception being printed.
*
* @param source source for selection
* @param request Select request to issue.
* @param action the action for use in exception creation
* @return response
* @throws IOExc... | 3.68 |
pulsar_Schema_generic | /**
* Returns a generic schema of existing schema info.
*
* <p>Only supports AVRO and JSON.
*
* @param schemaInfo schema info
* @return a generic schema instance
*/
static GenericSchema<GenericRecord> generic(SchemaInfo schemaInfo) {
return DefaultImplementation.getDefaultImplementation().getGenericSchema(sc... | 3.68 |
hbase_RegionLocations_mergeLocations | /**
* Merges this RegionLocations list with the given list assuming same range, and keeping the most
* up to date version of the HRegionLocation entries from either list according to seqNum. If
* seqNums are equal, the location from the argument (other) is taken.
* @param other the locations to merge with
* @retur... | 3.68 |
hibernate-validator_AnnotationTypeMemberCheck_checkMessageAttribute | /**
* Checks that the given type element
* <p/>
* <ul>
* <li>has a method with name "message",</li>
* <li>the return type of this method is {@link String}.</li>
* </ul>
*
* @param element The element of interest.
*
* @return A possibly non-empty set of constraint check errors, never null.
*/
private Set<Cons... | 3.68 |
dubbo_MetadataResolver_resolveConsumerServiceMetadata | /**
* for consumer
*
* @param targetClass target service class
* @param url consumer url
* @return rest metadata
* @throws CodeStyleNotSupportException not support type
*/
public static ServiceRestMetadata resolveConsumerServiceMetadata(
Class<?> targetClass, URL url, String contextPathFromUrl) {... | 3.68 |
framework_UIDL_getIntAttribute | /**
* Gets the named attribute as an int.
*
* @param name
* the name of the attribute to get
* @return the attribute value
*/
public int getIntAttribute(String name) {
return attr().getInt(name);
} | 3.68 |
hudi_S3EventsMetaSelector_createSourceSelector | /**
* Factory method for creating custom CloudObjectsMetaSelector. Default selector to use is {@link
* S3EventsMetaSelector}
*/
public static S3EventsMetaSelector createSourceSelector(TypedProperties props) {
String sourceSelectorClass = getStringWithAltKeys(
props, DFSPathSelectorConfig.SOURCE_INPUT_SELECTO... | 3.68 |
framework_UIDL_getStringArrayAttribute | /**
* Gets the named attribute as an array of Strings.
*
* @param name
* the name of the attribute to get
* @return the attribute value
*/
public String[] getStringArrayAttribute(String name) {
return attr().getStringArray(name);
} | 3.68 |
flink_CheckpointProperties_discardOnJobFailed | /**
* Returns whether the checkpoint should be discarded when the owning job reaches the {@link
* JobStatus#FAILED} state.
*
* @return <code>true</code> if the checkpoint should be discarded when the owning job reaches
* the {@link JobStatus#FAILED} state; <code>false</code> otherwise.
* @see CompletedCheckpo... | 3.68 |
flink_IOUtils_skipFully | /**
* Similar to readFully(). Skips bytes in a loop.
*
* @param in The InputStream to skip bytes from
* @param len number of bytes to skip
* @throws IOException if it could not skip requested number of bytes for any reason (including
* EOF)
*/
public static void skipFully(final InputStream in, long len) thro... | 3.68 |
graphhopper_VectorTile_addValues | /**
* <pre>
* Dictionary encoding for values
* </pre>
*
* <code>repeated .vector_tile.Tile.Value values = 4;</code>
*/
public Builder addValues(
int index, vector_tile.VectorTile.Tile.Value.Builder builderForValue) {
if (valuesBuilder_ == null) {
ensureValuesIsMutable();
values_.add(index, builderFo... | 3.68 |
querydsl_Expressions_list | /**
* Combine the given expressions into a list expression
*
* @param exprs list elements
* @return list expression
*/
public static Expression<Tuple> list(Expression<?>... exprs) {
return list(Tuple.class, exprs);
} | 3.68 |
hbase_PrivateCellUtil_writeRow | /**
* Writes the row from the given cell to the output stream
* @param out The outputstream to which the data has to be written
* @param cell The cell whose contents has to be written
* @param rlength the row length
*/
public static void writeRow(OutputStream out, Cell cell, short rlength) throws IOExceptio... | 3.68 |
hadoop_HdfsFileStatus_isEmptyLocalName | /**
* Check if the local name is empty.
* @return true if the name is empty
*/
default boolean isEmptyLocalName() {
return getLocalNameInBytes().length == 0;
} | 3.68 |
hudi_HoodieAvroUtils_generateProjectionSchema | /**
* Generate a reader schema off the provided writeSchema, to just project out the provided columns.
*/
public static Schema generateProjectionSchema(Schema originalSchema, List<String> fieldNames) {
Map<String, Field> schemaFieldsMap = originalSchema.getFields().stream()
.map(r -> Pair.of(r.name().toLowerC... | 3.68 |
hbase_Query_setReplicaId | /**
* Specify region replica id where Query will fetch data from. Use this together with
* {@link #setConsistency(Consistency)} passing {@link Consistency#TIMELINE} to read data from a
* specific replicaId. <br>
* <b> Expert: </b>This is an advanced API exposed. Only use it if you know what you are doing
*/
public... | 3.68 |
hbase_StorageClusterStatusModel_setMemStoreSizeMB | /**
* @param memstoreSizeMB memstore size, in MB
*/
public void setMemStoreSizeMB(int memstoreSizeMB) {
this.memstoreSizeMB = memstoreSizeMB;
} | 3.68 |
morf_WindowFunction_getPartitionBys | /**
* @return the fields to partition by.
*/
public ImmutableList<AliasedField> getPartitionBys() {
return partitionBys;
} | 3.68 |
flink_JaasModule_generateDefaultConfigFile | /** Generate the default JAAS config file. */
private static File generateDefaultConfigFile(String workingDir) {
checkArgument(workingDir != null, "working directory should not be null.");
final File jaasConfFile;
try {
Path path = Paths.get(workingDir);
if (Files.notExists(path)) {
... | 3.68 |
hbase_SimpleRpcServer_getNumOpenConnections | /**
* The number of open RPC conections
* @return the number of open rpc connections
*/
@Override
public int getNumOpenConnections() {
return connectionManager.size();
} | 3.68 |
pulsar_ReaderInterceptor_onPartitionsChange | /**
* This method is called when partitions of the topic (partitioned-topic) changes.
*
* @param topicName topic name
* @param partitions new updated number of partitions
*/
default void onPartitionsChange(String topicName, int partitions) {
} | 3.68 |
hbase_CompactingMemStore_snapshot | /**
* Push the current active memstore segment into the pipeline and create a snapshot of the tail of
* current compaction pipeline Snapshot must be cleared by call to {@link #clearSnapshot}.
* {@link #clearSnapshot(long)}.
* @return {@link MemStoreSnapshot}
*/
@Override
public MemStoreSnapshot snapshot() {
// I... | 3.68 |
flink_QueryableStateUtils_createKvStateServer | /**
* Initializes the {@link KvStateServer server} responsible for sending the requested internal
* state to the {@link KvStateClientProxy client proxy}.
*
* @param address the address to bind to.
* @param ports the range of ports the state server will attempt to listen to (see {@link
* org.apache.flink.confi... | 3.68 |
hbase_ReplicationSourceManager_releaseBufferQuota | /**
* To release the buffer quota which acquired by
* {@link ReplicationSourceManager#acquireBufferQuota}.
*/
void releaseBufferQuota(long size) {
if (size < 0) {
throw new IllegalArgumentException("size should not less than 0");
}
addTotalBufferUsed(-size);
} | 3.68 |
hadoop_RouterObserverReadProxyProvider_isRead | /**
* Check if a method is read-only.
*
* @return whether the 'method' is a read-only operation.
*/
private static boolean isRead(Method method) {
if (!method.isAnnotationPresent(ReadOnly.class)) {
return false;
}
return !method.getAnnotationsByType(ReadOnly.class)[0].activeOnly();
} | 3.68 |
hadoop_EntityTypeReader_getNextRowKey | /**
* Gets the possibly next row key prefix given current prefix and type.
*
* @param currRowKeyPrefix The current prefix that contains user, cluster,
* flow, run, and application id.
* @param entityType Current entity type.
* @return A new prefix for the possibly immediately next row key.... | 3.68 |
framework_AbstractOrderedLayoutConnector_needsExpand | /**
* Does the layout need to expand?
*/
private boolean needsExpand() {
return needsExpand;
} | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.