name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
hadoop_DataJoinJob_main | /**
* @param args
*/
public static void main(String[] args) {
boolean success;
if (args.length < 8 || args.length > 10) {
System.out.println("usage: DataJoinJob " + "inputdirs outputdir map_input_file_format "
+ "numofParts " + "mapper_class " + "reducer_class "
+... | 3.68 |
hadoop_RoleModel_validate | /**
* Validation includes validating all statements.
*/
@Override
public void validate() {
requireNonNull(statement, "Statement");
checkState(VERSION.equals(version), "Invalid Version: %s", version);
statement.stream().forEach((a) -> a.validate());
} | 3.68 |
flink_BinaryRowDataSerializer_checkSkipReadForFixLengthPart | /**
* We need skip bytes to read when the remain bytes of current segment is not enough to write
* binary row fixed part. See {@link BinaryRowData}.
*/
public void checkSkipReadForFixLengthPart(AbstractPagedInputView source) throws IOException {
// skip if there is no enough size.
// Note: Use currentSegment... | 3.68 |
framework_GridRowDragger_getDropIndexCalculator | /**
* Gets the drop index calculator.
* <p>
* Default is {@code null} and the dropped items are placed on the drop
* location.
*
* @return the drop index calculator
*/
public DropIndexCalculator<T> getDropIndexCalculator() {
return dropTargetIndexCalculator;
} | 3.68 |
hudi_HoodieTable_getPendingCommitTimeline | /**
* Get only the inflights (no-completed) commit timeline.
*/
public HoodieTimeline getPendingCommitTimeline() {
return metaClient.getCommitsTimeline().filterPendingExcludingMajorAndMinorCompaction();
} | 3.68 |
framework_GridLayout_getColumns | /**
* Get the number of columns in the grid.
*
* @return the number of columns in the grid.
*/
public int getColumns() {
return getState(false).columns;
} | 3.68 |
hbase_ClusterMetricsBuilder_toOptions | /**
* Convert an enum set of ClusterMetrics.Option to a list of ClusterStatusProtos.Option
* @param options the ClusterMetrics options
* @return a list of ClusterStatusProtos.Option
*/
public static List<ClusterStatusProtos.Option> toOptions(EnumSet<ClusterMetrics.Option> options) {
return options.stream().map(Cl... | 3.68 |
flink_RocksDBNativeMetricMonitor_registerStatistics | /** Register gauges to pull native metrics for the database. */
private void registerStatistics() {
if (statistics != null) {
for (TickerType tickerType : options.getMonitorTickerTypes()) {
metricGroup.gauge(
String.format("rocksdb.%s", tickerType.name().toLowerCase()),
... | 3.68 |
hbase_BalancerClusterState_getRackForRegion | /**
* Maps region index to rack index
*/
public int getRackForRegion(int region) {
return serverIndexToRackIndex[regionIndexToServerIndex[region]];
} | 3.68 |
flink_FileRegionWriteReadUtils_readFixedSizeRegionFromFile | /**
* Read {@link FixedSizeRegion} from {@link FileChannel}.
*
* <p>Note that this type of region's length is fixed.
*
* @param channel the channel to read.
* @param regionBuffer the buffer to read {@link FixedSizeRegion}'s header.
* @param fileOffset the file offset to start read.
* @return the {@link FixedSiz... | 3.68 |
framework_VaadinSession_getUIProviders | /**
* Gets the UI providers configured for this session.
*
* @return an unmodifiable list of UI providers
*/
public List<UIProvider> getUIProviders() {
assert hasLock();
return Collections.unmodifiableList(uiProviders);
} | 3.68 |
hadoop_BlockManagerParameters_withBufferPoolSize | /**
* Sets the in-memory cache size as number of blocks.
*
* @param poolSize The buffer pool size as number of blocks.
* @return The builder.
*/
public BlockManagerParameters withBufferPoolSize(
final int poolSize) {
this.bufferPoolSize = poolSize;
return this;
} | 3.68 |
framework_FieldGroup_getPropertyType | /**
* Gets the type of the property with the given property id.
*
* @param propertyId
* The propertyId. Must be find
* @return The type of the property
*/
protected Class<?> getPropertyType(Object propertyId) throws BindException {
if (getItemDataSource() == null) {
throw new BindException(... | 3.68 |
hudi_BaseHoodieWriteClient_postWrite | /**
* Common method containing steps to be performed after write (upsert/insert/..) operations including auto-commit.
* @param result Commit Action Result
* @param instantTime Instant Time
* @param hoodieTable Hoodie Table
* @return Write Status
*/
public O postWrite(HoodieWriteMetadata<O> result, String instant... | 3.68 |
hadoop_LightWeightLinkedSet_getBookmark | /**
* Returns a new iterator starting at the bookmarked element.
*
* @return the iterator to the bookmarked element.
*/
public Iterator<T> getBookmark() {
LinkedSetIterator toRet = new LinkedSetIterator();
toRet.next = this.bookmark.next;
this.bookmark = toRet;
return toRet;
} | 3.68 |
dubbo_JValidatorNew_createMemberValue | // Copy from javassist.bytecode.annotation.Annotation.createMemberValue(ConstPool, CtClass);
private static MemberValue createMemberValue(ConstPool cp, CtClass type, Object value) throws NotFoundException {
MemberValue memberValue = javassist.bytecode.annotation.Annotation.createMemberValue(cp, type);
if (membe... | 3.68 |
framework_AbstractComponent_getHeight | /*
* (non-Javadoc)
*
* @see com.vaadin.Sizeable#getHeight()
*/
@Override
public float getHeight() {
return height;
} | 3.68 |
pulsar_ResourceGroupService_getRgTenantUnRegistersCount | // Visibility for testing.
protected static double getRgTenantUnRegistersCount (String rgName) {
return rgTenantUnRegisters.labels(rgName).get();
} | 3.68 |
hbase_FavoredNodeAssignmentHelper_getOneRandomServer | /**
* Gets a random server from the specified rack and skips anything specified.
* @param rack rack from a server is needed
* @param skipServerSet the server shouldn't belong to this set
*/
protected ServerName getOneRandomServer(String rack, Set<ServerName> skipServerSet) {
// Is the rack valid? Do we ... | 3.68 |
framework_VColorPickerGradient_setBGColor | /**
* Sets the given css color as the background.
*
* @param bgColor
* the color to set
*/
public void setBGColor(String bgColor) {
if (bgColor == null) {
background.getElement().getStyle().clearBackgroundColor();
} else {
background.getElement().getStyle().setBackgroundColor(bgC... | 3.68 |
framework_AbstractSelect_addListener | /**
* @deprecated As of 7.0, replaced by
* {@link #addItemSetChangeListener(Container.ItemSetChangeListener)}
*/
@Override
@Deprecated
public void addListener(Container.ItemSetChangeListener listener) {
addItemSetChangeListener(listener);
} | 3.68 |
framework_Panel_changeVariables | /**
* Called when one or more variables handled by the implementing class are
* changed.
*
* @see com.vaadin.server.VariableOwner#changeVariables(Object, Map)
*/
@Override
public void changeVariables(Object source, Map<String, Object> variables) {
// Get new size
final Integer newWidth = (Integer) variable... | 3.68 |
framework_GridElement_getFooterCell | /**
* Gets footer cell element with given row and column index.
*
* @param rowIndex
* Row index
* @param colIndex
* Column index
* @return Footer cell element with given indices.
*/
public GridCellElement getFooterCell(int rowIndex, int colIndex) {
return getSubPart("#footer[" + rowInd... | 3.68 |
hadoop_BufferPuller_close | /**
* Closes the iterator so that the underlying streams can be closed.
*/
@Override
public void close() throws IOException {
if (closed) {
return;
}
if (null != nativeReader) {
nativeReader.close();
}
closed = true;
} | 3.68 |
hadoop_LightWeightLinkedSet_pollFirst | /**
* Remove and return first element on the linked list of all elements.
*
* @return first element
*/
public T pollFirst() {
if (head == null) {
return null;
}
T first = head.element;
this.remove(first);
return first;
} | 3.68 |
hbase_HFileArchiver_archiveStoreFile | /**
* Archive the store file
* @param fs the filesystem where the store files live
* @param regionInfo region hosting the store files
* @param conf {@link Configuration} to examine to determine the archive directory
* @param tableDir {@link Path} to where the table is being stored (for building the... | 3.68 |
hbase_SnapshotScannerHDFSAclHelper_getUsersWithTableReadAction | /**
* Return users with table read permission
* @param tableName the table
* @param includeNamespace true if include users with namespace read action
* @param includeGlobal true if include users with global read action
* @return users with table read permission
* @throws IOException if an error occurred... | 3.68 |
dubbo_AccessLogData_setServiceName | /**
* Add service name.
*
* @param serviceName
*/
public void setServiceName(String serviceName) {
set(SERVICE, serviceName);
} | 3.68 |
hbase_MultiTableInputFormatBase_getScans | /**
* Allows subclasses to get the list of {@link Scan} objects.
*/
protected List<Scan> getScans() {
return this.scans;
} | 3.68 |
framework_VAccordion_replaceWidget | /**
* Replaces the existing wrapped widget (if any) with a new widget.
*
* @param newWidget
* the new widget to wrap
*/
public void replaceWidget(Widget newWidget) {
if (widget != null) {
widgets.remove(widget);
if (open) {
remove(widget);
}
}
widget = new... | 3.68 |
hbase_ChecksumType_codeToType | /**
* Cannot rely on enum ordinals . They change if item is removed or moved. Do our own codes.
* @return Type associated with passed code.
*/
public static ChecksumType codeToType(final byte b) {
for (ChecksumType t : ChecksumType.values()) {
if (t.getCode() == b) {
return t;
}
}
throw new Runti... | 3.68 |
hadoop_ClientToAMTokenSecretManagerInRM_registerMasterKey | // Only for RM recovery
public synchronized SecretKey registerMasterKey(
ApplicationAttemptId applicationAttemptID, byte[] keyData) {
SecretKey key = createSecretKey(keyData);
registerApplication(applicationAttemptID, key);
return key;
} | 3.68 |
hadoop_CachingGetSpaceUsed_getDirPath | /**
* @return The directory path being monitored.
*/
public String getDirPath() {
return dirPath;
} | 3.68 |
flink_PojoFieldUtils_readField | /**
* Reads a field from the given {@link DataInputView}.
*
* <p>This read methods avoids Java serialization, by reading the classname of the field's
* declaring class and dynamically loading it. The field is also read by field name and obtained
* via reflection.
*
* @param in the input view to read from.
* @pa... | 3.68 |
flink_BinarySegmentUtils_copyToView | /**
* Copy bytes of segments to output view.
*
* <p>Note: It just copies the data in, not include the length.
*
* @param segments source segments
* @param offset offset for segments
* @param sizeInBytes size in bytes
* @param target target output view
*/
public static void copyToView(
MemorySegment[] s... | 3.68 |
hbase_BackupManifest_getAllDependentListByTable | /**
* Get the full dependent image list in the whole dependency scope for a specific table of this
* backup in time order from old to new.
* @param table table
* @return the full backup image list for a table in time order in the whole scope of the
* dependency of this image
*/
public ArrayList<BackupImag... | 3.68 |
hbase_CommonFSUtils_getRootDir | /**
* Get the path for the root data directory
* @param c configuration
* @return {@link Path} to hbase root directory from configuration as a qualified Path.
* @throws IOException e
*/
public static Path getRootDir(final Configuration c) throws IOException {
Path p = new Path(c.get(HConstants.HBASE_DIR));
Fil... | 3.68 |
graphhopper_VectorTile_setVersion | /**
* <pre>
* Any compliant implementation must first read the version
* number encoded in this message and choose the correct
* implementation for this version number before proceeding to
* decode other parts of this message.
* </pre>
*
* <code>required uint32 version = 15 [default = 1];</code>
*/
public Buil... | 3.68 |
framework_VTabsheet_cancelNextBlurSchedule | /**
* Cancel the next scheduled execution. This method must be called only
* from an event occurring before the onBlur event. It's the case of IE
* which doesn't trigger the focus event, so we're using this approach
* to cancel the next blur event prior it's execution, calling the
* method from mouse down event.
... | 3.68 |
morf_DatabaseMetaDataProvider_createRealName | /**
* Creates {@link RealName}, which contractually remembers two versions of a
* database object name: the name as retrieved by the JDBC driver, and also
* the user-friendly camel-case name of that same object, often derived by
* looking at the comment of that object, or in schema descriptions.
*
* <p>
* Note: ... | 3.68 |
flink_CheckpointConfig_disableCheckpointing | /** Disables checkpointing. */
public void disableCheckpointing() {
configuration.removeConfig(ExecutionCheckpointingOptions.CHECKPOINTING_INTERVAL);
} | 3.68 |
framework_GridDropTargetConnector_isDroppingOnRowsPossible | /**
* Inspects whether the current drop would happen on the whole grid instead
* of specific row as the drop target. This is based on used drop mode,
* whether dropping on sorted grid rows is allowed (determined on server
* side and automatically updated to drop mode) and whether the grid is
* empty.
*
* @return... | 3.68 |
flink_DynamicConfiguration_addAppConfigurationEntry | /** Add entries for the given application name. */
public void addAppConfigurationEntry(String name, AppConfigurationEntry... entry) {
final AppConfigurationEntry[] existing = dynamicEntries.get(name);
final AppConfigurationEntry[] updated;
if (existing == null) {
updated = Arrays.copyOf(entry, entr... | 3.68 |
framework_StringToIntegerConverter_getModelType | /*
* (non-Javadoc)
*
* @see com.vaadin.data.util.converter.Converter#getModelType()
*/
@Override
public Class<Integer> getModelType() {
return Integer.class;
} | 3.68 |
flink_StreamElement_isLatencyMarker | /**
* Checks whether this element is a latency marker.
*
* @return True, if this element is a latency marker, false otherwise.
*/
public final boolean isLatencyMarker() {
return getClass() == LatencyMarker.class;
} | 3.68 |
morf_SqlScriptExecutorProvider_create | /**
* @param dataSource The database connection source to use
* @param sqlDialect The dialect to use for the dataSource
* @return new instance of {@link SqlScriptExecutorProvider}
*/
public SqlScriptExecutorProvider create(final DataSource dataSource, final SqlDialect sqlDialect) {
return new SqlScriptExecutorPro... | 3.68 |
hbase_TableMapReduceUtil_addDependencyJars | /**
* Add the jars containing the given classes to the job's configuration such that JobClient will
* ship them to the cluster and add them to the DistributedCache.
* @deprecated since 1.3.0 and will be removed in 3.0.0. Use {@link #addDependencyJars(Job)}
* instead.
* @see #addDependencyJars(Job)
* @... | 3.68 |
hadoop_StageConfig_getWriterQueueCapacity | /**
* Get writer queue capacity.
* @return the queue capacity
*/
public int getWriterQueueCapacity() {
return writerQueueCapacity;
} | 3.68 |
hbase_MasterObserver_preDisableTableAction | /**
* Called prior to disabling a table. Called as part of disable table procedure and it is asyn to
* the disable table RPC call.
* @param ctx the environment to interact with the framework and master
* @param tableName the name of the table
*/
default void preDisableTableAction(final ObserverContext<Master... | 3.68 |
hbase_MetricsMaster_setNumSpaceQuotas | /**
* Sets the number of space quotas defined.
* @see MetricsMasterQuotaSource#updateNumSpaceQuotas(long)
*/
public void setNumSpaceQuotas(final long numSpaceQuotas) {
masterQuotaSource.updateNumSpaceQuotas(numSpaceQuotas);
} | 3.68 |
morf_SqlUtils_nullLiteral | /**
* @return a {@link NullFieldLiteral}.
*/
public static FieldLiteral nullLiteral() {
return new NullFieldLiteral();
} | 3.68 |
flink_AbstractMetricGroup_getLogicalScope | /**
* Returns the logical scope of this group, for example {@code "taskmanager.job.task"}.
*
* @param filter character filter which is applied to the scope components
* @param delimiter delimiter to use for concatenating scope components
* @param reporterIndex index of the reporter
* @return logical scope
*/
Str... | 3.68 |
hbase_MutableRegionInfo_getStartKey | /** Returns the startKey */
@Override
public byte[] getStartKey() {
return startKey;
} | 3.68 |
hadoop_StageConfig_build | /**
* The build command makes the config immutable.
* Idempotent.
* @return the now-frozen config
*/
public StageConfig build() {
frozen = true;
return this;
} | 3.68 |
hadoop_BCFile_getCompressionName | /**
* Get the name of the compression algorithm used to compress the block.
*
* @return name of the compression algorithm.
*/
public String getCompressionName() {
return rBlkState.getCompressionName();
} | 3.68 |
framework_StreamResource_setMIMEType | /**
* Sets the mime type of the resource.
*
* @param mimeType
* the MIME type to be set.
*/
public void setMIMEType(String mimeType) {
this.mimeType = mimeType;
} | 3.68 |
querydsl_JPAExpressions_selectZero | /**
* Create a new detached JPQLQuery instance with the projection zero
*
* @return select(0)
*/
public static JPQLQuery<Integer> selectZero() {
return select(Expressions.ZERO);
} | 3.68 |
hbase_BufferedMutator_getWriteBufferPeriodicFlushTimerTickMs | /**
* Returns the current periodic flush timertick interval in milliseconds.
* @return The number of milliseconds between each check if the timeout has been exceeded. This
* value only has a real meaning if the timeout has been set to > 0
*/
default long getWriteBufferPeriodicFlushTimerTickMs() {
throw ne... | 3.68 |
pulsar_FileSystemManagedLedgerOffloader_offload | /*
* ledgerMetadata stored in an index of -1
* */
@Override
public CompletableFuture<Void> offload(ReadHandle readHandle, UUID uuid, Map<String, String> extraMetadata) {
CompletableFuture<Void> promise = new CompletableFuture<>();
scheduler.chooseThread(readHandle.getId()).execute(
new LedgerReader(... | 3.68 |
hudi_HoodieTableMetaClient_getTempFolderPath | /**
* @return Temp Folder path
*/
public String getTempFolderPath() {
return basePath + Path.SEPARATOR + TEMPFOLDER_NAME;
} | 3.68 |
hadoop_RouterClientRMService_finalize | /**
* Shutdown the chain of interceptors when the object is destroyed.
*/
@Override
protected void finalize() {
rootInterceptor.shutdown();
} | 3.68 |
hudi_LSMTimelineWriter_updateManifest | /**
* Updates a manifest file.
*
* <p>4 steps:
* <ol>
* <li>read the latest manifest version file;</li>
* <li>read the latest manifest file for valid files;</li>
* <li>remove files to the existing file list from step2;</li>
* <li>add this new file to the existing file list from step2.</li>
* </ol>
*
... | 3.68 |
pulsar_AuthorizationService_allowTopicOperation | /**
* @deprecated - will be removed after 2.12. Use async variant.
*/
@Deprecated
public Boolean allowTopicOperation(TopicName topicName,
TopicOperation operation,
String originalRole,
String role,
... | 3.68 |
morf_OracleDialect_dropStatements | /**
* @see org.alfasoftware.morf.jdbc.SqlDialect#dropStatements(org.alfasoftware.morf.metadata.View)
*/
@Override
public Collection<String> dropStatements(View view) {
return Arrays.asList("BEGIN FOR i IN (SELECT null FROM all_views WHERE OWNER='" + getSchemaName().toUpperCase() + "' AND VIEW_NAME='" + view.getName... | 3.68 |
hudi_HoodieTableMetaClient_getBootstrapIndexByFileIdFolderNameFolderPath | /**
* @return Bootstrap Index By Hudi File Id Folder
*/
public String getBootstrapIndexByFileIdFolderNameFolderPath() {
return basePath + Path.SEPARATOR + BOOTSTRAP_INDEX_BY_FILE_ID_FOLDER_PATH;
} | 3.68 |
open-banking-gateway_FintechSecureStorage_psuAspspKeyToPrivate | /**
* Sends PSU/Fintechs' to FinTechs' private storage.
* @param authSession Authorization session for this PSU/Fintech user
* @param fintech FinTech to store to
* @param psuKey Key to store
* @param password FinTechs Datasafe/Keystore password
*/
@SneakyThrows
public void psuAspspKeyToPrivate(AuthSession authSes... | 3.68 |
morf_AbstractDatabaseType_extractPath | /**
* Extracts the path part of the URL from the stack returned by
* {@link #splitJdbcUrl(String)}. Assumes that the stack has already
* had the preceding components (e.g. protocol, host and port) popped.
*
* @param splitURL A stack containing the remaining portions of a JDBC URL.
* @return The path.
*/
protect... | 3.68 |
hudi_AbstractHoodieLogRecordReader_getProgress | /**
* Return progress of scanning as a float between 0.0 to 1.0.
*/
public float getProgress() {
return progress;
} | 3.68 |
framework_VTextField_updateCursorPosition | /**
* Updates the cursor position variable if it has changed since the last
* update.
*
* @return true if the value was updated
*/
protected boolean updateCursorPosition() {
if (WidgetUtil.isAttachedAndDisplayed(this)) {
int cursorPos = prompting ? 0 : getCursorPos();
if (lastCursorPos != curso... | 3.68 |
framework_Tree_setNullSelectionItemId | /**
* Tree does not support <code>setNullSelectionItemId</code>.
*
* @see AbstractSelect#setNullSelectionItemId(java.lang.Object)
*/
@Override
public void setNullSelectionItemId(Object nullSelectionItemId)
throws UnsupportedOperationException {
if (nullSelectionItemId != null) {
throw new Unsupp... | 3.68 |
framework_Link_setTargetBorder | /**
* Sets the border of the target window.
*
* @param targetBorder
* the targetBorder to set.
*/
public void setTargetBorder(BorderStyle targetBorder) {
getState().targetBorder = targetBorder;
} | 3.68 |
framework_GridElement_getHeaderCellByCaption | /**
* Gets the header cell element with the given caption in the given header
* row. If there are multiple headers with the same name, the first one is
* returned.
*
* @param rowIndex
* The index of the header row
* @param caption
* The header caption
* @return The first header cell eleme... | 3.68 |
hadoop_LogParserUtil_stringToUnixTimestamp | /**
* Converts String date to unix timestamp. Note that we assume the time in the
* logs has the same time zone with the machine which runs the
* {@link RmSingleLineParser}.
*
* @param date The String date.
* @return Unix time stamp.
* @throws ParseException if data conversion from String to unix timestamp
* ... | 3.68 |
querydsl_MetaDataExporter_setNamingStrategy | /**
* Override the NamingStrategy (default: new DefaultNamingStrategy())
*
* @param namingStrategy naming strategy to override (default: new DefaultNamingStrategy())
*/
public void setNamingStrategy(NamingStrategy namingStrategy) {
module.bind(NamingStrategy.class, namingStrategy);
} | 3.68 |
querydsl_JTSGeometryExpression_disjoint | /**
* Returns 1 (TRUE) if this geometric object is “spatially disjoint” from anotherGeometry.
*
* @param geometry other geometry
* @return true, if disjoint
*/
public BooleanExpression disjoint(Expression<? extends Geometry> geometry) {
return Expressions.booleanOperation(SpatialOps.DISJOINT, mixin, geometry);... | 3.68 |
hadoop_StagingCommitter_validateContext | /**
* Validate the task attempt context; makes sure
* that the task attempt ID data is valid.
* @param context task context
*/
private static void validateContext(TaskAttemptContext context) {
requireNonNull(context, "null context");
requireNonNull(context.getTaskAttemptID(),
"null task attempt ID");
re... | 3.68 |
framework_AbstractComponentConnector_cancelTouchTimer | /**
* If a long touch event timer is running, cancel it.
*
* @since 7.6
*/
private void cancelTouchTimer() {
WidgetUtil.setTextSelectionEnabled(getWidget().getElement(), true);
if (longTouchTimer != null) {
// Re-enable text selection
longTouchTimer.cancel();
}
} | 3.68 |
flink_TaskExecutorMemoryConfiguration_getManagedMemoryTotal | /** Returns the total amount of memory reserved for by the MemoryManager. */
public Long getManagedMemoryTotal() {
return managedMemoryTotal;
} | 3.68 |
dubbo_AbstractConfigManager_getConfigById | /**
* Get config by id
*
* @param configType
* @param id
* @return
*/
protected <C extends AbstractConfig> C getConfigById(String configType, String id) {
return (C) getConfigsMap(configType).get(id);
} | 3.68 |
framework_WeekGrid_setCalendar | /**
* @param calendar
* the calendar to set
*/
public void setCalendar(VCalendar calendar) {
this.calendar = calendar;
} | 3.68 |
hadoop_Check_ge0 | /**
* Verifies an long is greater or equal to zero.
*
* @param value integer value.
* @param name the name to use in the exception message.
*
* @return the value.
*
* @throws IllegalArgumentException if the long is greater or equal to zero.
*/
public static long ge0(long value, String name) {
if (value < 0) ... | 3.68 |
hadoop_YarnClientUtils_generateToken | /**
* Generate SPNEGO challenge request token.
*
* @param server - hostname to contact
* @throws IOException thrown if doAs failed
* @throws InterruptedException thrown if doAs is interrupted
* @return SPNEGO token challenge
*/
public static String generateToken(String server) throws IOException,
Interrupted... | 3.68 |
pulsar_ManagedLedgerConfig_getMetadataAckQuorumSize | /**
* @return the metadataAckQuorumSize
*/
public int getMetadataAckQuorumSize() {
return metadataAckQuorumSize;
} | 3.68 |
dubbo_ReflectUtils_findMethodByMethodSignature | /**
* Find method from method signature
*
* @param clazz Target class to find method
* @param methodName Method signature, e.g.: method1(int, String). It is allowed to provide method name only, e.g.: method2
* @return target method
* @throws NoSuchMethodException
* @throws ClassNotFoundException
* @throws ... | 3.68 |
framework_Notification_show | /**
* Shows a notification message the current page. The position and behavior
* of the message depends on the type, which is one of the basic types
* defined in {@link Notification}, for instance
* Notification.TYPE_WARNING_MESSAGE.
*
* The caption is rendered as plain text with HTML automatically escaped.
*
*... | 3.68 |
hadoop_LogParserUtil_parseLog | /**
* Parse the log file/directory.
*
* @param logFile the file/directory of the log.
* @throws SkylineStoreException if fails to addHistory to
* {@link SkylineStore}.
* @throws IOException if fails to parse the log.
* @throws ResourceEstimatorException if t... | 3.68 |
hbase_IdentityTableMap_initJob | /**
* Use this before submitting a TableMap job. It will appropriately set up the JobConf.
* @param table table name
* @param columns columns to scan
* @param mapper mapper class
* @param job job configuration
*/
@SuppressWarnings("unchecked")
public static void initJob(String table, String columns, Class<... | 3.68 |
hudi_InLineFSUtils_getOuterFilePathFromInlinePath | /**
* InlineFS Path format:
* "inlinefs://path/to/outer/file/outer_file_scheme/?start_offset=start_offset>&length=<length>"
* <p>
* Outer File Path format:
* "outer_file_scheme://path/to/outer/file"
* <p>
* Example
* Input: "inlinefs://file1/s3a/?start_offset=20&length=40".
* Output: "s3a://file1"
*
* @param... | 3.68 |
hbase_HRegion_bulkLoadHFiles | /**
* Attempts to atomically load a group of hfiles. This is critical for loading rows with multiple
* column families atomically.
* @param familyPaths List of Pair<byte[] column family, String hfilePath>
* @param bulkLoadListener Internal hooks enabling massaging/preparation of a file about to be
* ... | 3.68 |
pulsar_StateChangeListeners_notifyOnCompletion | /**
* Notify all currently added listeners on completion of the future.
*
* @return future of a new completion stage
*/
public <T> CompletableFuture<T> notifyOnCompletion(CompletableFuture<T> future,
String serviceUnit,
... | 3.68 |
hbase_ProcedureExecutor_removeResult | /**
* Mark the specified completed procedure, as ready to remove.
* @param procId the ID of the procedure to remove
*/
public void removeResult(long procId) {
CompletedProcedureRetainer<TEnvironment> retainer = completed.get(procId);
if (retainer == null) {
assert !procedures.containsKey(procId) : "pid=" + p... | 3.68 |
dubbo_JValidationNew_createValidator | /**
* Return new instance of {@link JValidator}
* @param url Valid URL instance
* @return Instance of JValidator
*/
@Override
protected Validator createValidator(URL url) {
return new JValidatorNew(url);
} | 3.68 |
framework_DefaultConnectionStateHandler_getDialogTextGaveUp | /**
* Gets the text to show in the reconnect dialog after giving up (reconnect
* limit reached).
*
* @param reconnectAttempt
* The number of the current reconnection attempt
* @return The text to show in the reconnect dialog after giving up
*/
protected String getDialogTextGaveUp(int reconnectAttempt)... | 3.68 |
hbase_ResponseConverter_buildClearRegionBlockCacheResponse | /**
* Creates a protocol buffer ClearRegionBlockCacheResponse
* @return a ClearRegionBlockCacheResponse
*/
public static AdminProtos.ClearRegionBlockCacheResponse
buildClearRegionBlockCacheResponse(final HBaseProtos.CacheEvictionStats cacheEvictionStats) {
return AdminProtos.ClearRegionBlockCacheResponse.newBuil... | 3.68 |
hbase_RawDouble_decodeDouble | /**
* Read a {@code double} value from the buffer {@code buff}.
*/
public double decodeDouble(byte[] buff, int offset) {
double val = Bytes.toDouble(buff, offset);
return val;
} | 3.68 |
hbase_NewVersionBehaviorTracker_prepare | /**
* Reset the map if it is different with the last Cell. Save the cq array/offset/length for next
* Cell.
* @return If this put has duplicate ts with last cell, return the mvcc of last cell. Else return
* MAX_VALUE.
*/
protected long prepare(Cell cell) {
if (isColumnQualifierChanged(cell)) {
// The... | 3.68 |
hmily_JavaBeanBinder_getField | /**
* Gets field.
*
* @return the field
*/
public Field getField() {
return field;
} | 3.68 |
hbase_WALFactory_close | /**
* Shutdown all WALs and clean up any underlying storage. Use only when you will not need to
* replay and edits that have gone to any wals from this factory.
*/
public void close() throws IOException {
List<IOException> ioes = new ArrayList<>();
// these fields could be null if the WALFactory is created only ... | 3.68 |
hadoop_StageConfig_exitStage | /**
* Exit the stage; calls back to
* {@link #enterStageEventHandler} if non-null.
* @param stage stage entered
*/
public void exitStage(String stage) {
if (enterStageEventHandler != null) {
enterStageEventHandler.exitStage(stage);
}
} | 3.68 |
dubbo_ServiceInvokeRestFilter_getAcceptMediaType | /**
* return first match , if any multiple content-type
*
* @param request
* @return
*/
public static MediaType getAcceptMediaType(RequestFacade request, Class<?> returnType) {
String accept = request.getHeader(RestHeaderEnum.ACCEPT.getHeader());
accept = Objects.isNull(accept) ? MediaType.ALL_VALUE.value ... | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.