name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
hbase_PrivateCellUtil_createLastOnRow | /**
* Create a Cell that is larger than all other possible Cells for the given Cell's row.
* @return Last possible Cell on passed Cell's row.
*/
public static Cell createLastOnRow(final Cell cell) {
if (cell instanceof ByteBufferExtendedCell) {
return new LastOnRowByteBufferExtendedCell(((ByteBufferExtendedCel... | 3.68 |
hudi_HoodieTableConfig_isMetadataTableAvailable | /**
* @returns true if metadata table has been created and is being used for this dataset, else returns false.
*/
public boolean isMetadataTableAvailable() {
return isMetadataPartitionAvailable(MetadataPartitionType.FILES);
} | 3.68 |
hadoop_RBFMetrics_getNamespaceInfo | /**
* Build a set of unique values found in all namespaces.
*
* @param f Method reference of the appropriate FederationNamespaceInfo
* getter function
* @return Set of unique string values found in all discovered namespaces.
* @throws IOException if the query could not be executed.
*/
private Collection... | 3.68 |
hadoop_RouterFedBalance_getDefaultConf | /**
* Loads properties from hdfs-fedbalance-default.xml into configuration
* object.
*
* @return Configuration which includes properties from
* hdfs-fedbalance-default.xml and hdfs-fedbalance-site.xml
*/
@VisibleForTesting
static Configuration getDefaultConf() {
Configuration config = new HdfsConfigurat... | 3.68 |
hadoop_CachedDNSToSwitchMapping_getSwitchMap | /**
* Get the (host x switch) map.
* @return a copy of the cached map of hosts to rack
*/
@Override
public Map<String, String> getSwitchMap() {
return new HashMap<>(cache);
} | 3.68 |
dubbo_AccessLogData_setOutTime | /**
* Set the out date. As an argument it accept date string.
*
* @param outTime
*/
public void setOutTime(Date outTime) {
set(OUT_TIME, outTime);
} | 3.68 |
hadoop_JWTRedirectAuthenticationHandler_init | /**
* Initializes the authentication handler instance.
* <p>
* This method is invoked by the {@link AuthenticationFilter#init} method.
* </p>
* @param config
* configuration properties to initialize the handler.
*
* @throws ServletException
* thrown if the handler could not be initialized.
... | 3.68 |
graphhopper_PointList_shallowCopy | /**
* Create a shallow copy of this Pointlist from from to end, excluding end.
*
* @param makeImmutable makes this PointList immutable. If you don't ensure the consistency it might happen that due to changes of this
* object, the shallow copy might contain incorrect or corrupt data.
*/
public ... | 3.68 |
hbase_TableMapReduceUtil_findOrCreateJar | /**
* Finds the Jar for a class or creates it if it doesn't exist. If the class is in a directory in
* the classpath, it creates a Jar on the fly with the contents of the directory and returns the
* path to that Jar. If a Jar is created, it is created in the system temporary directory.
* Otherwise, returns an exist... | 3.68 |
hudi_BaseRollbackPlanActionExecutor_requestRollback | /**
* Creates a Rollback plan if there are files to be rolled back and stores them in instant file.
* Rollback Plan contains absolute file paths.
*
* @param startRollbackTime Rollback Instant Time
* @return Rollback Plan if generated
*/
protected Option<HoodieRollbackPlan> requestRollback(String startRollbackTime... | 3.68 |
framework_DropTargetExtensionConnector_onDragOver | /**
* Event handler for the {@code dragover} event.
* <p>
* Override this method in case custom handling for the dragover event is
* required. If the drop is allowed, the event should prevent default.
*
* @param event
* browser event to be handled
*/
protected void onDragOver(Event event) {
Nativ... | 3.68 |
morf_PortableSqlStatement_deepCopy | /**
* Deep copy.
*
* @return null.
*/
@Override
public Statement deepCopy() {
return null;
} | 3.68 |
framework_Table_setConverter | /**
* Sets a converter for a property id.
* <p>
* The converter is used to format the the data for the given property id
* before displaying it in the table.
* </p>
*
* @param propertyId
* The propertyId to format using the converter
* @param converter
* The converter to use for the prop... | 3.68 |
hadoop_JsonSerialization_fromResource | /**
* Convert from a JSON file.
* @param resource input file
* @return the parsed JSON
* @throws IOException IO problems
* @throws JsonParseException If the input is not well-formatted
* @throws JsonMappingException failure to map from the JSON to this class
*/
@SuppressWarnings({"IOResourceOpenedButNotSafelyClo... | 3.68 |
framework_TreeElement_getValue | /**
* Returns selected item of the tree. In multiselect mode returns first
* selected item. If there is no selected item returns empty string
*
* @return selected item of the tree
*/
public String getValue() {
List<WebElement> selectedElements = findElements(
By.className("v-tree-node-selected"));
... | 3.68 |
hbase_TableInputFormatBase_getRecordReader | /**
* Builds a TableRecordReader. If no TableRecordReader was provided, uses the default.
* @see InputFormat#getRecordReader(InputSplit, JobConf, Reporter)
*/
public RecordReader<ImmutableBytesWritable, Result> getRecordReader(InputSplit split, JobConf job,
Reporter reporter) throws IOException {
// In case a su... | 3.68 |
pulsar_LeastResourceUsageWithWeight_updateAndGetMaxResourceUsageWithWeight | /**
* Update and get the max resource usage with weight of broker according to the service configuration.
*
* @param broker the broker name.
* @param brokerData The broker load data.
* @param conf The service configuration.
* @return the max resource usage with weight of broker
*/
private double update... | 3.68 |
hadoop_BlockManager_requestPrefetch | /**
* Requests optional prefetching of the given block.
*
* @param blockNumber the id of the block to prefetch.
*
* @throws IllegalArgumentException if blockNumber is negative.
*/
public void requestPrefetch(int blockNumber) {
checkNotNegative(blockNumber, "blockNumber");
// Do nothing because we do not supp... | 3.68 |
hadoop_DecayRpcSchedulerDetailedMetrics_shutdown | /**
* Shutdown the instrumentation process.
*/
public void shutdown() {
DefaultMetricsSystem.instance().unregisterSource(name);
} | 3.68 |
hbase_HRegion_getRegionDir | /**
* Computes the Path of the HRegion
* @param tabledir qualified path for table
* @param name ENCODED region name
* @return Path of HRegion directory
* @deprecated For tests only; to be removed.
*/
@Deprecated
public static Path getRegionDir(final Path tabledir, final String name) {
return new Path(tabled... | 3.68 |
hadoop_BalanceProcedureScheduler_shutDown | /**
* Shutdown the scheduler.
*/
public synchronized void shutDown() {
if (!running.get()) {
return;
}
running.set(false);
readerThread.interrupt();
roosterThread.interrupt();
recoverThread.interrupt();
workersPool.shutdownNow();
} | 3.68 |
hbase_CoprocessorClassLoader_getClassLoader | /**
* Get a CoprocessorClassLoader for a coprocessor jar path from cache. If not in cache, create
* one.
* @param path the path to the coprocessor jar file to load classes from
* @param parent the parent class loader for exempted classes
* @param pathPrefix a prefix used in temp path name to store the ja... | 3.68 |
hibernate-validator_ClassVisitor_visitExecutableAsMethod | /**
* Checks whether the constraints of the given method are valid.
*
* @param element a method under investigation
* @param aVoid
*/
@Override
public Void visitExecutableAsMethod(ExecutableElement element, Void aVoid) {
processClassChecks( element );
return null;
} | 3.68 |
flink_LocalBufferPool_lazyDestroy | /** Destroy is called after the produce or consume phase of a task finishes. */
@Override
public void lazyDestroy() {
// NOTE: if you change this logic, be sure to update recycle() as well!
CompletableFuture<?> toNotify = null;
synchronized (availableMemorySegments) {
if (!isDestroyed) {
... | 3.68 |
hudi_BaseHoodieWriteClient_setWriteSchemaForDeletes | /**
* Sets write schema from last instant since deletes may not have schema set in the config.
*/
protected void setWriteSchemaForDeletes(HoodieTableMetaClient metaClient) {
try {
HoodieActiveTimeline activeTimeline = metaClient.getActiveTimeline();
Option<HoodieInstant> lastInstant =
activeTimeline... | 3.68 |
flink_Path_suffix | /**
* Adds a suffix to the final name in the path.
*
* @param suffix The suffix to be added
* @return the new path including the suffix
*/
public Path suffix(String suffix) {
return new Path(getParent(), getName() + suffix);
} | 3.68 |
flink_FlinkContainersSettings_flinkHome | /**
* Sets the path of the Flink distribution inside the container. Returns a reference to this
* Builder enabling method chaining.
*
* @param flinkHome The {@code flinkHome} to set.
* @return A reference to this Builder.
*/
public Builder flinkHome(String flinkHome) {
this.flinkHome = flinkHome;
return t... | 3.68 |
flink_DataStream_project | /**
* Initiates a Project transformation on a {@link Tuple} {@link DataStream}.<br>
* <b>Note: Only Tuple DataStreams can be projected.</b>
*
* <p>The transformation projects each Tuple of the DataSet onto a (sub)set of fields.
*
* @param fieldIndexes The field indexes of the input tuples that are retained. The o... | 3.68 |
morf_GraphBasedUpgradeSchemaChangeVisitor_create | /**
* Creates {@link GraphBasedUpgradeSchemaChangeVisitor} instance.
*
* @param sourceSchema schema prior to upgrade step
* @param sqlDialect dialect to generate statements for the target database
* @param idTable table for id generation
* @param upgradeNodes all the {@link GraphBasedUpgradeNode} instances... | 3.68 |
flink_ChannelStateWriteRequestExecutorImpl_waitAndTakeUnsafe | /**
* Retrieves and removes the head request of the {@link #deque}, waiting if necessary until an
* element becomes available.
*
* @return The head request, it can be null when the executor is closed.
*/
@Nullable
private ChannelStateWriteRequest waitAndTakeUnsafe() throws InterruptedException {
ChannelStateWr... | 3.68 |
rocketmq-connect_KafkaSinkAdaptorConnector_start | /**
* Start the component
*
* @param config component context
*/
@Override
public void start(KeyValue config) {
super.start(config);
sinkConnector.validate(taskConfig);
sinkConnector.initialize(new KafkaConnectorContext(connectorContext));
sinkConnector.start(taskConfig);
} | 3.68 |
hadoop_TFile_getMetaBlock | /**
* Stream access to a meta block.``
*
* @param name
* The name of the meta block.
* @return The input stream.
* @throws IOException
* on I/O error.
* @throws MetaBlockDoesNotExist
* If the meta block with the name does not exist.
*/
public DataInputStream getMetaBlock(String n... | 3.68 |
pulsar_ManagedLedgerConfig_setThrottleMarkDelete | /**
* Set the rate limiter on how many mark-delete calls per second are allowed. If the value is set to 0, the rate
* limiter is disabled. Default is 0.
*
* @param throttleMarkDelete
* the max number of mark-delete calls allowed per second
*/
public ManagedLedgerConfig setThrottleMarkDelete(double thro... | 3.68 |
hmily_HmilyTccTransactionExecutor_participantCancel | /**
* Participant cancel object.
*
* @param hmilyParticipants the hmily participants
* @param selfParticipantId the self participant id
* @return the object
*/
public Object participantCancel(final List<HmilyParticipant> hmilyParticipants, final Long selfParticipantId) {
LogUtil.debug(LOGGER, () -> "tcc cance... | 3.68 |
hadoop_OSSListResult_v1 | /**
* Restricted constructors to ensure v1 or v2, not both.
* @param result v1 result
* @return new list result container
*/
public static OSSListResult v1(ObjectListing result) {
return new OSSListResult(result, null);
} | 3.68 |
hadoop_HSAuditLogger_createSuccessLog | /**
* A helper api for creating an audit log for a successful event.
*/
static String createSuccessLog(String user, String operation, String target) {
StringBuilder b = new StringBuilder();
start(Keys.USER, user, b);
addRemoteIP(b);
add(Keys.OPERATION, operation, b);
add(Keys.TARGET, target, b);
add(Keys.... | 3.68 |
hbase_MobUtils_getTableNameString | /**
* Gets the table name from when this cell was written into a mob hfile as a string.
* @param cell to extract tag from
* @return table name as a string. empty if the tag is not found.
*/
public static Optional<String> getTableNameString(Cell cell) {
Optional<Tag> tag = getTableNameTag(cell);
Optional<String>... | 3.68 |
flink_KvStateLocationRegistry_notifyKvStateUnregistered | /**
* Notifies the registry about an unregistered KvState instance.
*
* @param jobVertexId JobVertexID the KvState instance belongs to
* @param keyGroupRange Key group index the KvState instance belongs to
* @param registrationName Name under which the KvState has been registered
* @throws IllegalArgumentExceptio... | 3.68 |
MagicPlugin_MagicConfigCommandExecutor_applySession | /**
* Note that this gets called asynchronously
*/
protected void applySession(String sessionId, Session session, CommandSender sender, String command, boolean load) {
String missingMessage = magic.getMessages().get("commands.mconfig." + command + ".missing");
String type = session.getType();
if (type == ... | 3.68 |
flink_DatadogHttpReporter_getTagsFromConfig | /** Get config tags from config 'metrics.reporter.dghttp.tags'. */
private List<String> getTagsFromConfig(String str) {
return Arrays.asList(str.split(","));
} | 3.68 |
framework_VTree_executeEventCommand | /*
* Must wait for Safari to focus before sending click and value change
* events (see #6373, #6374)
*/
private void executeEventCommand(ScheduledCommand command) {
if (BrowserInfo.get().isWebkit() && !treeHasFocus) {
Scheduler.get().scheduleDeferred(command);
} else {
command.execute();
... | 3.68 |
framework_VCalendar_isEventCaptionAsHtml | /**
* Checks whether event captions are rendered as HTML
* <p>
* The default is false, i.e. to render that caption as plain text.
*
* @return true if the captions are rendered as HTML, false if rendered as
* plain text
*/
public boolean isEventCaptionAsHtml() {
return eventCaptionAsHtml;
} | 3.68 |
hbase_AsyncTable_putAll | /**
* A simple version of batch put. It will fail if there are any failures.
* @param puts The list of mutations to apply.
* @return A {@link CompletableFuture} that always returns null when complete normally.
*/
default CompletableFuture<Void> putAll(List<Put> puts) {
return allOf(put(puts)).thenApply(r -> null)... | 3.68 |
hadoop_OBSObjectBucketUtils_renameFile | /**
* Implement rename file.
*
* @param owner OBS File System instance
* @param srcKey source object key
* @param dstKey destination object key
* @param srcStatus source object status
* @throws IOException any problem with rename operation
*/
private static void renameFile(final OBSFileSystem owner,
... | 3.68 |
flink_AbstractBinaryWriter_write | /**
* Writes the specified byte to this output stream. The general contract for <code>write
* </code> is that one byte is written to the output stream. The byte to be written is the
* eight low-order bits of the argument <code>b</code>. The 24 high-order bits of <code>b
* </code> are ignored.
*/
@Override
public v... | 3.68 |
dubbo_ScopeClusterInvoker_createInjvmInvoker | /**
* Creates a new Invoker for the current ScopeClusterInvoker and exports it to the local JVM.
*/
private void createInjvmInvoker(Exporter<?> exporter) {
if (injvmInvoker == null) {
synchronized (createLock) {
if (injvmInvoker == null) {
URL url = new ServiceConfigURL(
... | 3.68 |
framework_FieldBinder_bindField | /**
* Tries to bind the given {@link Component} instance to a member field of
* the bind target. The fields are matched based on localId, id and caption.
*
* @param instance
* the instance to be bound to a field
* @param localId
* the localId used for mapping the field to an instance field
... | 3.68 |
querydsl_ComparableExpression_gt | /**
* Create a {@code this > right} expression
*
* @param right rhs of the comparison
* @return this > right
* @see java.lang.Comparable#compareTo(Object)
*/
public BooleanExpression gt(Expression<T> right) {
return Expressions.booleanOperation(Ops.GT, mixin, right);
} | 3.68 |
pulsar_BindAddressValidator_migrateBindAddresses | /**
* Generates bind addresses based on legacy configuration properties.
*/
private static List<BindAddress> migrateBindAddresses(ServiceConfiguration config) {
List<BindAddress> addresses = new ArrayList<>(2);
if (config.getBrokerServicePort().isPresent()) {
addresses.add(new BindAddress(null, URI.cr... | 3.68 |
morf_DummyXmlOutputStreamProvider_getXmlString | /**
* @return Convert the output to a String
*/
public String getXmlString() {
return new String(testOutputStream.toByteArray());
} | 3.68 |
flink_SnapshotDirectory_permanent | /**
* Creates a permanent snapshot directory for the given path, which will not delete the
* underlying directory in {@link #cleanup()} after {@link #completeSnapshotAndGetHandle()} was
* called.
*/
public static SnapshotDirectory permanent(@Nonnull Path directory) throws IOException {
return new PermanentSnaps... | 3.68 |
flink_SliceAssigners_tumbling | /**
* Creates a tumbling window {@link SliceAssigner} that assigns elements to slices of tumbling
* windows.
*
* @param rowtimeIndex the index of rowtime field in the input row, {@code -1} if based on
* processing time.
* @param shiftTimeZone The shift timezone of the window, if the proctime or rowtime type i... | 3.68 |
hbase_PrivateCellUtil_createLastOnRowCol | /**
* Create a Cell that is larger than all other possible Cells for the given Cell's rk:cf:q. Used
* in creating "fake keys" for the multi-column Bloom filter optimization to skip the row/column
* we already know is not in the file.
* @return Last possible Cell on passed Cell's rk:cf:q.
*/
public static Cell crea... | 3.68 |
hbase_CreateStoreFileWriterParams_includeMVCCReadpoint | /**
* Whether to include MVCC or not
*/
public CreateStoreFileWriterParams includeMVCCReadpoint(boolean includeMVCCReadpoint) {
this.includeMVCCReadpoint = includeMVCCReadpoint;
return this;
} | 3.68 |
hadoop_HttpReferrerAuditHeader_addAttribute | /**
* Add a query parameter if not null/empty
* There's no need to escape here as it is done in the URI
* constructor.
* @param key query key
* @param value query value
*/
private void addAttribute(String key,
String value) {
if (StringUtils.isNotEmpty(value)) {
attributes.put(key, value);
}
} | 3.68 |
hbase_RSGroupInfo_addTable | /**
* Add a table
* @deprecated Since 3.0.0, will be removed in 4.0.0. The rsgroup information will be stored in
* the configuration of a table so this will be removed.
*/
@Deprecated
public void addTable(TableName table) {
tables.add(table);
} | 3.68 |
hadoop_LeveldbIterator_seekToFirst | /**
* Repositions the iterator so is is at the beginning of the Database.
*/
public void seekToFirst() throws DBException {
try {
iter.seekToFirst();
} catch (DBException e) {
throw e;
} catch (RuntimeException e) {
throw new DBException(e.getMessage(), e);
}
} | 3.68 |
morf_AuditRecordHelper_addAuditRecord | /**
* Add the audit record, writing out the SQL for the insert.
*
* @see org.alfasoftware.morf.upgrade.SchemaChangeVisitor#addAuditRecord(java.util.UUID, java.lang.String)
*
* @param visitor The schema change visitor adding the audit record.
* @param schema The schema to add the audit record to.
* @param uuid Th... | 3.68 |
hbase_LruAdaptiveBlockCache_evict | /**
* Eviction method. Evict items in order of use, allowing delete items which haven't been used for
* the longest amount of time.
* @return how many bytes were freed
*/
long evict() {
// Ensure only one eviction at a time
if (!evictionLock.tryLock()) {
return 0;
}
long bytesToFree = 0L;
try {
... | 3.68 |
morf_MathsField_getOperator | /**
* @return the operator
*/
public MathsOperator getOperator() {
return operator;
} | 3.68 |
pulsar_MessageAcknowledger_acknowledgeAsync | /**
* The asynchronous version of {@link #acknowledge(MessageId)}.
*/
default CompletableFuture<Void> acknowledgeAsync(MessageId messageId) {
return acknowledgeAsync(messageId, null);
} | 3.68 |
hmily_HmilyRepositoryStorage_releaseHmilyLocks | /**
* Release locks..
*
* @param hmilyLocks hmily locks
*/
public static void releaseHmilyLocks(final Collection<HmilyLock> hmilyLocks) {
if (!hmilyLocks.isEmpty()) {
PUBLISHER.syncPublishEvent(hmilyLocks, EventTypeEnum.RELEASE_HMILY_LOCKS.getCode());
}
} | 3.68 |
framework_VComboBox_dataReceived | /**
* Called by the connector when new data for the last requested filter
* is received from the server.
*/
public void dataReceived() {
if (initialData) {
suggestionPopup.menu.setSuggestions(currentSuggestions);
performSelection(serverSelectedKey, true, true);
updateSuggestionPopupMinWid... | 3.68 |
framework_VComboBox_setTotalSuggestions | /**
* Sets the total number of suggestions.
* <p>
* NOTE: this excluded the possible null selection item!
* <p>
* NOTE: this just updates the state, but doesn't update any UI.
*
* @since 8.0
* @param totalSuggestions
* total number of suggestions
*/
public void setTotalSuggestions(int totalSuggesti... | 3.68 |
framework_VCalendar_getDateClickListener | /**
* Gets the listener for listening to event clicks.
*
* @return
*/
public DateClickListener getDateClickListener() {
return dateClickListener;
} | 3.68 |
morf_AbstractSelectStatementBuilder_fullOuterJoin | /**
* Specifies an full outer join to a subselect:
*
* <blockquote><pre>
* TableReference sale = tableRef("Sale");
* TableReference customer = tableRef("Customer");
*
* // Define the subselect - a group by showing total sales by age in the
* // previous month.
* SelectStatement amountsByAgeLastMonth = select(f... | 3.68 |
hadoop_AzureNativeFileSystemStore_configureAzureStorageSession | /**
* Set the configuration parameters for this client storage session with
* Azure.
*
* @throws AzureException
*/
private void configureAzureStorageSession() throws AzureException {
// Assertion: Target session URI already should have been captured.
if (sessionUri == null) {
throw new AssertionError(
... | 3.68 |
flink_NFAStateNameHandler_getUniqueInternalName | /**
* Used to give a unique name to {@link org.apache.flink.cep.nfa.NFA} states created during the
* translation process. The name format will be {@code baseName:counter} , where the counter is
* increasing for states with the same {@code baseName}.
*
* @param baseName The base of the name.
* @return The (unique)... | 3.68 |
dubbo_PropertySourcesUtils_getSubProperties | /**
* Get prefixed {@link Properties}
*
* @param propertySources {@link PropertySources}
* @param propertyResolver {@link PropertyResolver} to resolve the placeholder if present
* @param prefix the prefix of property name
* @return Map
* @see Properties
*/
public static Map<String, Object> getSubProp... | 3.68 |
hadoop_AbstractTask_setTaskType | /**
* Set TaskType for a Task.
* @param type Simple or Composite Task
*/
public final void setTaskType(final TaskType type) {
this.taskType = type;
} | 3.68 |
dubbo_AbstractJSONImpl_getListOfStrings | /**
* Gets a list from an object for the given key, and verifies all entries are strings. If the key
* is not present, this returns null. If the value is not a List or an entry is not a string,
* throws an exception.
*/
@Override
public List<String> getListOfStrings(Map<String, ?> obj, String key) {
assert ob... | 3.68 |
hadoop_Validate_checkNotNullAndNotEmpty | /**
* Validates that the given buffer is not null and has non-zero capacity.
* @param <T> the type of iterable's elements.
* @param iter the argument reference to validate.
* @param argName the name of the argument being validated.
*/
public static <T> void checkNotNullAndNotEmpty(Iterable<T> iter,
String argN... | 3.68 |
hbase_CompactionProgress_getTotalCompactingKVs | /** Returns the total compacting key values in currently running compaction */
public long getTotalCompactingKVs() {
if (totalCompactingKVs < currentCompactedKVs) {
LOG.debug("totalCompactingKVs={} less than currentCompactedKVs={}", totalCompactingKVs,
currentCompactedKVs);
return currentCompactedKVs;
... | 3.68 |
flink_JoinOperator_projectTuple18 | /**
* Projects a pair of joined elements to a {@link Tuple} with the previously selected
* fields. Requires the classes of the fields of the resulting tuples.
*
* @return The projected data set.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17>... | 3.68 |
hadoop_FSEditLogAsync_tryRelease | // while draining, count the releases until release(int)
private void tryRelease(int permits) {
pendingReleases.getAndAdd(permits);
if (!draining.get()) {
super.release(pendingReleases.getAndSet(0));
}
} | 3.68 |
zxing_IntentIntegrator_startActivityForResult | /**
* Start an activity. This method is defined to allow different methods of activity starting for
* newer versions of Android and for compatibility library.
*
* @param intent Intent to start.
* @param code Request code for the activity
* @see Activity#startActivityForResult(Intent, int)
* @see Fragment#startAc... | 3.68 |
hbase_LruAdaptiveBlockCache_containsBlock | /**
* Whether the cache contains block with specified cacheKey
* @return true if contains the block
*/
@Override
public boolean containsBlock(BlockCacheKey cacheKey) {
return map.containsKey(cacheKey);
} | 3.68 |
flink_InputSelection_select | /**
* Selects an input identified by the given {@code inputId}.
*
* @param inputId the input id numbered starting from 1 to 64, and `1` indicates the first
* input. Specially, `-1` indicates all inputs.
* @return a reference to this object.
*/
public Builder select(int inputId) {
if (inputId > 0 && inputI... | 3.68 |
flink_TypeExtractionUtils_hasSuperclass | /**
* Returns true if the given class has a superclass of given name.
*
* @param clazz class to be analyzed
* @param superClassName class name of the super class
*/
public static boolean hasSuperclass(Class<?> clazz, String superClassName) {
List<Type> hierarchy = new ArrayList<>();
getTypeHierarchy(hierar... | 3.68 |
flink_FileInputFormat_getStatistics | /**
* Obtains basic file statistics containing only file size. If the input is a directory, then
* the size is the sum of all contained files.
*
* @see
* org.apache.flink.api.common.io.InputFormat#getStatistics(org.apache.flink.api.common.io.statistics.BaseStatistics)
*/
@Override
public FileBaseStatistics ge... | 3.68 |
framework_VScrollTable_ensureFocus | /**
* Ensure the component has a focus.
*
* TODO the current implementation simply always calls focus for the
* component. In case the Table at some point implements focus/blur
* listeners, this method needs to be evolved to conditionally call
* focus only if not currently focused.
*/
protected void ensureFocus(... | 3.68 |
flink_StreamTask_getCheckpointBarrierHandler | /**
* Acquires the optional {@link CheckpointBarrierHandler} associated with this stream task. The
* {@code CheckpointBarrierHandler} should exist if the task has data inputs and requires to
* align the barriers.
*/
protected Optional<CheckpointBarrierHandler> getCheckpointBarrierHandler() {
return Optional.emp... | 3.68 |
hadoop_ReadWriteDiskValidatorMetrics_sourceName | /**
* Get a source name by given directory name.
*
* @param dirName directory name
* @return the source name
*/
protected static String sourceName(String dirName) {
StringBuilder sb = new StringBuilder(RECORD_INFO.name());
sb.append(",dir=").append(dirName);
return sb.toString();
} | 3.68 |
AreaShop_RegionGroup_getName | /**
* Get the name of the group.
* @return The name of the group
*/
public String getName() {
return name;
} | 3.68 |
flink_CliFrontend_run | /**
* Executions the run action.
*
* @param args Command line arguments for the run action.
*/
protected void run(String[] args) throws Exception {
LOG.info("Running 'run' command.");
final Options commandOptions = CliFrontendParser.getRunCommandOptions();
final CommandLine commandLine = getCommandLine... | 3.68 |
rocketmq-connect_BrokerBasedLog_send | /**
* send data to all workers
*
* @param key
* @param value
* @param callback
*/
@Override
public void send(K key, V value, Callback callback) {
try {
Map.Entry<byte[], byte[]> encode = encode(key, value);
byte[] body = encode.getValue();
if (body.length > MAX_MESSAGE_SIZE) {
... | 3.68 |
hadoop_AzureNativeFileSystemStore_explicitFileExists | /**
* Checks whether an explicit file/folder exists.
* This is used by redo of atomic rename.
* There was a bug(apache jira HADOOP-12780) during atomic rename if
* process crashes after an inner directory has been renamed but still
* there are file under that directory to be renamed then after the
* process comes... | 3.68 |
querydsl_AbstractSQLQuery_as | /**
* Create an alias for the expression
*
* @param alias alias
* @return this as alias
*/
@SuppressWarnings("unchecked")
public SimpleExpression<T> as(Path<?> alias) {
return Expressions.as(this, (Path) alias);
} | 3.68 |
hadoop_FilePosition_setAbsolute | /**
* If the given {@code pos} lies within the current buffer, updates the current position to
* the specified value and returns true; otherwise returns false without changing the position.
*
* @param pos the absolute position to change the current position to if possible.
* @return true if the given current posit... | 3.68 |
framework_VComboBox_setPageLength | /**
* Sets the number of items to show per page, or 0 for showing all items.
*
* @param pageLength
* new page length or 0 for all items
*/
public void setPageLength(int pageLength) {
this.pageLength = pageLength;
} | 3.68 |
hadoop_DefaultStringifier_load | /**
* Restores the object from the configuration.
*
* @param <K> the class of the item
* @param conf the configuration to use
* @param keyName the name of the key to use
* @param itemClass the class of the item
* @return restored object
* @throws IOException : forwards Exceptions from the underlying
* {@link... | 3.68 |
hadoop_ManifestCommitter_cleanupJob | /**
* Execute the {@code CleanupJobStage} to remove the job attempt dir.
* This does
* @param jobContext Context of the job whose output is being written.
* @throws IOException failure during cleanup
*/
@SuppressWarnings("deprecation")
@Override
public void cleanupJob(final JobContext jobContext) throws IOExceptio... | 3.68 |
flink_StreamProjection_projectTuple11 | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10>
SingleOutputStreamOperator<Tuple11<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10>>
... | 3.68 |
hadoop_OBSPosixBucketUtils_fsRemoveSubdir | // Delete a sub dir.
private static int fsRemoveSubdir(final OBSFileSystem owner,
final String subdirKey,
final List<KeyAndVersion> subdirList)
throws IOException {
fsRecursivelyDeleteDir(owner, subdirKey, false);
subdirList.add(new KeyAndVersion(subdirKey));
if (subdirList.size() == owner.getMaxEntr... | 3.68 |
flink_FlinkCalciteSqlValidator_declaredDescriptorColumn | /**
* Returns whether the given column has been declared in a {@link SqlKind#DESCRIPTOR} next to a
* {@link SqlKind#EXPLICIT_TABLE} within TVF operands.
*/
private static boolean declaredDescriptorColumn(SelectScope scope, Column column) {
if (!(scope.getNode() instanceof ExplicitTableSqlSelect)) {
retur... | 3.68 |
hbase_DeleteNamespaceProcedure_removeNamespaceQuota | /**
* remove quota for the namespace
* @param env MasterProcedureEnv
* @param namespaceName name of the namespace in string format
**/
private static void removeNamespaceQuota(final MasterProcedureEnv env, final String namespaceName)
throws IOException {
env.getMasterServices().getMasterQuotaManager()... | 3.68 |
pulsar_TxnMetaImpl_addAckedPartitions | /**
* Remove the list partitions that the transaction acknowledges to.
*
* @param partitions the list of partitions that the txn acknowledges to
* @return the transaction itself.
* @throws InvalidTxnStatusException
*/
@Override
public synchronized TxnMetaImpl addAckedPartitions(List<TransactionSubscription> parti... | 3.68 |
framework_AbstractSplitPanel_setMinSplitPosition | /**
* Sets the minimum split position to the given position and unit. If the
* split position is reversed, maximum and minimum are also reversed.
*
* @param pos
* the minimum position of the split
* @param unit
* the unit (from {@link Sizeable}) in which the size is given.
* All... | 3.68 |
framework_VTwinColSelect_getSelectionsCaption | /**
* Gets the selections caption HTML widget.
*
* @return the selections caption widget
*/
protected HTML getSelectionsCaption() {
if (selectionsCaption == null) {
selectionsCaption = new HTML();
selectionsCaption.setStyleName(CLASSNAME + "-caption-right");
selectionsCaption.getElement(... | 3.68 |
flink_CatalogManager_getTable | /**
* Retrieves a fully qualified table with a specific time. If the path is not yet fully
* qualified, use {@link #qualifyIdentifier(UnresolvedIdentifier)} first.
*
* @param objectIdentifier full path of the table to retrieve
* @param timestamp Timestamp of the table snapshot, which is milliseconds since 1970-01-... | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.