name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
hadoop_DataNodeFaultInjector_delayAckLastPacket | /**
* Used as a hook to delay sending the response of the last packet.
*/
public void delayAckLastPacket() throws IOException {
} | 3.68 |
hadoop_ManifestCommitter_setupTask | /**
* Set up a task through a {@link SetupTaskStage}.
* Classic FileOutputCommitter is a no-op here, relying
* on RecordWriters to create the dir implicitly on file
* create().
* FileOutputCommitter also uses the existence of that
* file as a flag to indicate task commit is needed.
* @param context task context.... | 3.68 |
hbase_HeapMemoryManager_getHeapOccupancyPercent | /** Returns heap occupancy percentage, 0 <= n <= 1. or -0.0 for error asking JVM */
public float getHeapOccupancyPercent() {
return this.heapOccupancyPercent == Float.MAX_VALUE
? HEAP_OCCUPANCY_ERROR_VALUE
: this.heapOccupancyPercent;
} | 3.68 |
framework_Design_createHtml | /**
* Generates an html tree representation of the component hierarchy having
* the root designContext.getRootComponent(). The hierarchy is stored under
* <body> in the tree. The generated tree represents a valid html
* document.
*
*
* @param designContext
* a DesignContext object specifying th... | 3.68 |
graphhopper_VectorTile_clearTags | /**
* <pre>
* Tags of this feature are encoded as repeated pairs of
* integers.
* A detailed description of tags is located in sections
* 4.2 and 4.4 of the specification
* </pre>
*
* <code>repeated uint32 tags = 2 [packed = true];</code>
*/
public Builder clearTags() {
tags_ = java.util.Collections.emptyLis... | 3.68 |
shardingsphere-elasticjob_TaskContext_getExecutorId | /**
* Get executor ID.
*
* @param appName application name
* @return executor ID
*/
public String getExecutorId(final String appName) {
return String.join(DELIMITER, appName, slaveId);
} | 3.68 |
hbase_MasterObserver_postUpdateReplicationPeerConfig | /**
* Called after update peerConfig for the specified peer
* @param ctx the environment to interact with the framework and master
* @param peerId a short name that identifies the peer
*/
default void postUpdateReplicationPeerConfig(
final ObserverContext<MasterCoprocessorEnvironment> ctx, String peerId,
Rep... | 3.68 |
framework_Button_removeClickShortcut | /**
* Removes the keyboard shortcut previously set with
* {@link #setClickShortcut(int, int...)}.
*/
public void removeClickShortcut() {
if (clickShortcut != null) {
removeShortcutListener(clickShortcut);
clickShortcut = null;
getState().clickShortcutKeyCode = 0;
}
} | 3.68 |
hbase_Procedure_isLockedWhenLoading | /**
* Can only be called when restarting, before the procedure actually being executed, as after we
* actually call the {@link #doAcquireLock(Object, ProcedureStore)} method, we will reset
* {@link #lockedWhenLoading} to false.
* <p/>
* Now it is only used in the ProcedureScheduler to determine whether we should p... | 3.68 |
morf_DataValueLookupMetadata_getIndexInArray | /**
* Get the array position of the specified column.
*
* @param columnName The column name.
* @return The array position, or null if not present.
*/
@Nullable
Integer getIndexInArray(CaseInsensitiveString columnName) {
return lookups.get(columnName);
} | 3.68 |
hbase_RingBufferEnvelope_getPayload | /**
* Retrieve current namedQueue payload {@link NamedQueuePayload} available on Envelope and free up
* the Envelope
* @return Retrieve rpc log details
*/
public NamedQueuePayload getPayload() {
final NamedQueuePayload namedQueuePayload = this.namedQueuePayload;
this.namedQueuePayload = null;
return namedQueu... | 3.68 |
flink_ReOpenableHashPartition_spillInMemoryPartition | /**
* Spills this partition to disk. This method is invoked once after the initial open() method
*
* @return Number of memorySegments in the writeBehindBuffers!
*/
int spillInMemoryPartition(
FileIOChannel.ID targetChannel,
IOManager ioManager,
LinkedBlockingQueue<MemorySegment> writeBehindB... | 3.68 |
hbase_HFileBlock_checkCallerProvidedOnDiskSizeWithHeader | /**
* Check that {@code value} provided by the calling context seems reasonable, within a large
* margin of error.
* @return {@code true} if the value is safe to proceed, {@code false} otherwise.
*/
private boolean checkCallerProvidedOnDiskSizeWithHeader(long value) {
// same validation logic as is used by Math.t... | 3.68 |
morf_AbstractSqlDialectTest_testInsertFromSelectIgnoresCase | /**
* Tests that when forming an insert/select from two table references, the {@link SqlDialect} acts in a case insensitive manner
* when determining if source fields are present in the destination table.
*/
@Test
public void testInsertFromSelectIgnoresCase() {
InsertStatement insertStatement = new InsertStatement... | 3.68 |
flink_AbstractStreamOperatorV2_getPartitionedState | /**
* Creates a partitioned state handle, using the state backend configured for this task.
*
* @throws IllegalStateException Thrown, if the key/value state was already initialized.
* @throws Exception Thrown, if the state backend cannot create the key/value state.
*/
protected <S extends State, N> S getPartitione... | 3.68 |
morf_OracleDialect_legacyFetchSizeForBulkSelects | /**
* @deprecated this method returns the legacy fetch size value for Oracle and is primarily for backwards compatibility.
* Please use {@link SqlDialect#fetchSizeForBulkSelects()} for the new recommended default value.
* @see SqlDialect#fetchSizeForBulkSelects()
*/
@Override
@Deprecated
public int legacyFetchSizeF... | 3.68 |
flink_FieldSet_toFieldList | /**
* Turns the FieldSet into an ordered FieldList.
*
* @return An ordered FieldList.
*/
public FieldList toFieldList() {
int[] pos = toArray();
Arrays.sort(pos);
return new FieldList(pos);
} | 3.68 |
pulsar_SecretsProvider_interpolateSecretForValue | /**
* If the passed value is formatted as a reference to a secret, as defined by the implementation, return the
* referenced secret. If the value is not formatted as a secret reference or the referenced secret does not exist,
* return null.
*
* @param value a config value that may be formatted as a reference to a ... | 3.68 |
pulsar_PulsarClientException_setPreviousExceptions | /**
* Add a list of previous exception which occurred for the same operation
* and have been retried.
*
* @param previous A collection of throwables that triggered retries
*/
public void setPreviousExceptions(Collection<Throwable> previous) {
this.previous = previous;
} | 3.68 |
hmily_PropertyName_getElementSize | /**
* Gets element size.
*
* @return the element size
*/
public int getElementSize() {
return this.elements.length;
} | 3.68 |
hbase_LeaseManager_getListener | /** Returns listener */
public LeaseListener getListener() {
return this.listener;
} | 3.68 |
shardingsphere-elasticjob_FailoverService_getAllFailoveringItems | /**
* Get all failovering items.
*
* @return all failovering items
*/
public Map<Integer, JobInstance> getAllFailoveringItems() {
int shardingTotalCount = configService.load(true).getShardingTotalCount();
Map<Integer, JobInstance> result = new LinkedHashMap<>(shardingTotalCount, 1);
for (int i = 0; i < ... | 3.68 |
graphhopper_VectorTile_getIntValue | /**
* <code>optional int64 int_value = 4;</code>
*/
public long getIntValue() {
return intValue_;
} | 3.68 |
hadoop_NativeTaskOutputFiles_getOutputIndexFileForWrite | /**
* Create a local map output index file name.
*
* @param size the size of the file
*/
public Path getOutputIndexFileForWrite(long size) throws IOException {
String path = String.format(OUTPUT_FILE_INDEX_FORMAT_STRING, TASKTRACKER_OUTPUT, id);
return lDirAlloc.getLocalPathForWrite(path, size, conf);
} | 3.68 |
flink_PekkoRpcServiceUtils_getRpcUrl | /**
* @param hostname The hostname or address where the target RPC service is listening.
* @param port The port where the target RPC service is listening.
* @param endpointName The name of the RPC endpoint.
* @param addressResolution Whether to try address resolution of the given hostname or not. This
* allows... | 3.68 |
flink_RowTimeMiniBatchAssginerOperator_getMiniBatchStart | // ------------------------------------------------------------------------
// Utilities
// ------------------------------------------------------------------------
/** Method to get the mini-batch start for a watermark. */
private static long getMiniBatchStart(long watermark, long interval) {
return watermark - (... | 3.68 |
dubbo_InvokerAndRestMethodMetadataPair_compareServiceMethod | /**
* same interface & same method desc
*
* @param beforeMetadata
* @return
*/
public boolean compareServiceMethod(InvokerAndRestMethodMetadataPair beforeMetadata) {
Class currentServiceInterface = this.invoker.getInterface();
Class<?> beforeServiceInterface = beforeMetadata.getInvoker().getInterface();... | 3.68 |
morf_AbstractSelectStatement_getJoins | /**
* Gets the list of joined tables in the order they are joined
*
* @return the joined tables
*/
public List<Join> getJoins() {
return joins;
} | 3.68 |
framework_VGridLayout_updateSpacingStyleName | /** For internal use only. May be removed or replaced in the future. */
public void updateSpacingStyleName(boolean spacingEnabled) {
String styleName = getStylePrimaryName();
if (spacingEnabled) {
spacingMeasureElement.addClassName(styleName + "-spacing-on");
spacingMeasureElement.removeClassNam... | 3.68 |
AreaShop_GeneralRegion_getLimitingGroup | /**
* Get the name of the group that is limiting the action, assuming actionAllowed() is false.
* @return The name of the group
*/
public String getLimitingGroup() {
return limitingGroup;
} | 3.68 |
framework_AtmospherePushConnection_onError | /**
* Called if the push connection fails. Atmosphere will automatically retry
* the connection until successful.
*
*/
protected void onError(AtmosphereResponse response) {
state = State.DISCONNECTED;
getConnectionStateHandler().pushError(this, response);
} | 3.68 |
MagicPlugin_ActionFactory_construct | /**
* Constructs a new action from a class name.
*
* @param actionClassName
* The class name of the action.
* @return The constructed action.
* @throws ActionFactoryException
* If no action could be constructed.
*/
public static BaseSpellAction construct(String actionClassName)
th... | 3.68 |
hadoop_RMDelegatedNodeLabelsUpdater_createRMNodeLabelsMappingProvider | /**
* Get the RMNodeLabelsMappingProvider which is used to provide node labels.
*/
private RMNodeLabelsMappingProvider createRMNodeLabelsMappingProvider(
Configuration conf) throws IOException {
RMNodeLabelsMappingProvider nodeLabelsMappingProvider = null;
try {
Class<? extends RMNodeLabelsMappingProvider... | 3.68 |
morf_InlineTableUpgrader_writeStatements | /**
* Write out SQL
*/
private void writeStatements(Collection<String> statements) {
sqlStatementWriter.writeSql(statements);
} | 3.68 |
pulsar_ClientConfiguration_setUseTcpNoDelay | /**
* Configure whether to use TCP no-delay flag on the connection, to disable Nagle algorithm.
* <p>
* No-delay features make sure packets are sent out on the network as soon as possible, and it's critical to achieve
* low latency publishes. On the other hand, sending out a huge number of small packets might limit... | 3.68 |
zxing_AztecCode_getSize | /**
* @return size in pixels (width and height)
*/
public int getSize() {
return size;
} | 3.68 |
framework_VTree_doRelationSelection | /**
* Selects a range of items which are in direct relation with each
* other.<br/>
* NOTE: The start node <b>MUST</b> be before the end node!
*
* @param startNode
*
* @param endNode
*/
private void doRelationSelection(TreeNode startNode, TreeNode endNode) {
TreeNode currentNode = endNode;
while (curren... | 3.68 |
hbase_CompoundConfiguration_freezeMutableConf | /**
* If set has been called, it will create a mutableConf. This converts the mutableConf to an
* immutable one and resets it to allow a new mutable conf. This is used when a new map or conf is
* added to the compound configuration to preserve proper override semantics.
*/
void freezeMutableConf() {
if (mutableCo... | 3.68 |
querydsl_GeometryExpression_contains | /**
* Returns 1 (TRUE) if this geometric object “spatially contains” anotherGeometry.
*
* @param geometry other geometry
* @return true, if contains
*/
public BooleanExpression contains(Expression<? extends Geometry> geometry) {
return Expressions.booleanOperation(SpatialOps.CONTAINS, mixin, geometry);
} | 3.68 |
framework_AbstractSplitPanel_readDesign | /*
* (non-Javadoc)
*
* @see com.vaadin.ui.AbstractComponent#readDesign(org.jsoup.nodes .Element,
* com.vaadin.ui.declarative.DesignContext)
*/
@Override
public void readDesign(Element design, DesignContext designContext) {
// handle default attributes
super.readDesign(design, designContext);
// handle ... | 3.68 |
hadoop_XMLUtils_newSecureSAXParserFactory | /**
* This method should be used if you need a {@link SAXParserFactory}. Use this method
* instead of {@link SAXParserFactory#newInstance()}. The factory that is returned has
* secure configuration enabled.
*
* @return a {@link SAXParserFactory} with secure configuration enabled
* @throws ParserConfigurationExcep... | 3.68 |
hbase_MetaTableAccessor_addSplitsToParent | /**
* Adds daughter region infos to hbase:meta row for the specified region.
* <p/>
* Note that this does not add its daughter's as different rows, but adds information about the
* daughters in the same row as the parent. Now only used in snapshot. Use
* {@link org.apache.hadoop.hbase.master.assignment.RegionState... | 3.68 |
hbase_CellUtil_makeColumn | /**
* Makes a column in family:qualifier form from separate byte arrays.
* <p>
* Not recommended for usage as this is old-style API.
* @return family:qualifier
*/
public static byte[] makeColumn(byte[] family, byte[] qualifier) {
return Bytes.add(family, COLUMN_FAMILY_DELIM_ARRAY, qualifier);
} | 3.68 |
hadoop_RegistryTypeUtils_requireAddressType | /**
* Require a specific address type on an endpoint
* @param required required type
* @param epr endpoint
* @throws InvalidRecordException if the type is wrong
*/
public static void requireAddressType(String required, Endpoint epr) throws
InvalidRecordException {
if (!required.equals(epr.addressType)) {
... | 3.68 |
framework_DeclarativeIconGenerator_setIcon | /**
* Sets an {@code icon} for the {@code item}.
*
* @param item
* a data item
* @param icon
* an icon for the {@code item}
*/
protected void setIcon(T item, Resource icon) {
captions.put(item, icon);
} | 3.68 |
querydsl_SQLExpressions_unionAll | /**
* Create a new UNION ALL clause
*
* @param sq subqueries
* @param <T>
* @return union
*/
public static <T> Union<T> unionAll(List<SubQueryExpression<T>> sq) {
return new SQLQuery<Void>().unionAll(sq);
} | 3.68 |
flink_RestServerEndpoint_getServerAddress | /**
* Returns the address on which this endpoint is accepting requests.
*
* @return address on which this endpoint is accepting requests or null if none
*/
@Nullable
public InetSocketAddress getServerAddress() {
synchronized (lock) {
assertRestServerHasBeenStarted();
Channel server = this.server... | 3.68 |
flink_KeyedStream_getKeyType | /**
* Gets the type of the key by which the stream is partitioned.
*
* @return The type of the key by which the stream is partitioned.
*/
@Internal
public TypeInformation<KEY> getKeyType() {
return keyType;
} | 3.68 |
hbase_NullComparator_toByteArray | /** Returns The comparator serialized using pb */
@Override
public byte[] toByteArray() {
ComparatorProtos.NullComparator.Builder builder = ComparatorProtos.NullComparator.newBuilder();
return builder.build().toByteArray();
} | 3.68 |
hadoop_MRJobConfUtil_getTaskProgressWaitDeltaTimeThreshold | /**
* Retrieves the min time required to log the task attempt current
* progress.
* @return the defined threshold in the conf.
* returns the default value if
* {@link #setTaskLogProgressDeltaThresholds} has not been called.
*/
public static long getTaskProgressWaitDeltaTimeThreshold() {
if (prog... | 3.68 |
hudi_HoodieSparkKeyGeneratorFactory_inferKeyGeneratorTypeFromWriteConfig | /**
* Infers the key generator type based on the record key and partition fields.
* If neither of the record key and partition fields are set, the default type is returned.
*
* @param props Properties from the write config.
* @return Inferred key generator type.
*/
public static KeyGeneratorType inferKeyGenerator... | 3.68 |
hbase_RequestConverter_toAssignRegionsRequest | // HBCK2
public static MasterProtos.AssignsRequest toAssignRegionsRequest(List<String> encodedRegionNames,
boolean override) {
MasterProtos.AssignsRequest.Builder b = MasterProtos.AssignsRequest.newBuilder();
return b.addAllRegion(toEncodedRegionNameRegionSpecifiers(encodedRegionNames))
.setOverride(override)... | 3.68 |
hibernate-validator_ValidationProviderHelper_getValidatorBeanClass | /**
* Determines the class of the {@link Validator} corresponding to the given configuration object.
*/
Class<? extends Validator> getValidatorBeanClass() {
return validatorClass;
} | 3.68 |
framework_ConnectorTracker_getLogger | /**
* Gets a logger for this class
*
* @return A logger instance for logging within this class
*
*/
private static Logger getLogger() {
return Logger.getLogger(ConnectorTracker.class.getName());
} | 3.68 |
flink_CopyOnWriteSkipListStateMap_compareSegmentAndNode | /**
* Compare the first skip list key in the given memory segment with the second skip list key in
* the given node.
*
* @param keySegment memory segment storing the first key.
* @param keyOffset offset of the first key in memory segment.
* @param keyLen length of the first key.
* @param targetNode the node stor... | 3.68 |
morf_TableDataHomology_copyAndSort | /**
* @return A sorted copy of the Record list
*/
private List<Record> copyAndSort(final Table table, Iterable<Record> records, Comparator<Record> comparator) {
// we need to transform the records to RecordBeans so we don't re-use the Record
return FluentIterable.from(records)
.transform(r -> RecordHelper.c... | 3.68 |
morf_SqlUtils_type | /**
* Specifies the data type for the parameter.
*
* @param dataType The data type
* @return the next phase of the parameter builder.
*/
public SqlParameterWidthBuilder type(DataType dataType) {
return new SqlParameterWidthBuilder(name, dataType);
} | 3.68 |
hadoop_CommonAuditContext_remove | /**
* Remove a context entry.
* @param key key
*/
public void remove(String key) {
if (LOG.isTraceEnabled()) {
LOG.trace("Remove context entry {}", key);
}
evaluatedEntries.remove(key);
} | 3.68 |
hbase_User_getGroupNames | /**
* Returns the list of groups of which this user is a member. On secure Hadoop this returns the
* group information for the user as resolved on the server. For 0.20 based Hadoop, the group
* names are passed from the client.
*/
public String[] getGroupNames() {
return ugi.getGroupNames();
} | 3.68 |
hudi_SparkBootstrapCommitActionExecutor_listAndProcessSourcePartitions | /**
* Return Bootstrap Mode selections for partitions listed and figure out bootstrap Schema.
* @return
* @throws IOException
*/
private Map<BootstrapMode, List<Pair<String, List<HoodieFileStatus>>>> listAndProcessSourcePartitions() throws IOException {
List<Pair<String, List<HoodieFileStatus>>> folders = Bootstr... | 3.68 |
morf_OracleDialect_createTrigger | /**
* Returns a list of SQL statement to create a trigger to populate a table's autonumber column
* from a sequence.
*
* @param table Table for which the trigger should be created.
* @param onColumn The autonumber column.
* @return SQL string list.
*/
private List<String> createTrigger(Table table, Column onColu... | 3.68 |
zxing_OneDReader_doDecode | /**
* We're going to examine rows from the middle outward, searching alternately above and below the
* middle, and farther out each time. rowStep is the number of rows between each successive
* attempt above and below the middle. So we'd scan row middle, then middle - rowStep, then
* middle + rowStep, then middle -... | 3.68 |
flink_RunLengthDecoder_readUnsignedVarInt | /** Reads the next varint encoded int. */
private int readUnsignedVarInt() throws IOException {
int value = 0;
int shift = 0;
int b;
do {
b = in.read();
value |= (b & 0x7F) << shift;
shift += 7;
} while ((b & 0x80) != 0);
return value;
} | 3.68 |
flink_TransientBlobCache_deleteInternal | /**
* Deletes the file associated with the blob key in this BLOB cache.
*
* @param jobId ID of the job this blob belongs to (or <tt>null</tt> if job-unrelated)
* @param key blob key associated with the file to be deleted
* @return <tt>true</tt> if the given blob is successfully deleted or non-existing;
* <tt>... | 3.68 |
hbase_HFileReaderImpl_next | /**
* Go to the next key/value in the block section. Loads the next block if necessary. If
* successful, {@link #getKey()} and {@link #getValue()} can be called.
* @return true if successfully navigated to the next key/value
*/
@Override
public boolean next() throws IOException {
// This is a hot method so extrem... | 3.68 |
flink_FlinkAssertions_assertThatChainOfCauses | /**
* Shorthand to assert chain of causes. Same as:
*
* <pre>{@code
* assertThat(throwable)
* .extracting(FlinkAssertions::chainOfCauses, FlinkAssertions.STREAM_THROWABLE)
* }</pre>
*/
public static ListAssert<Throwable> assertThatChainOfCauses(Throwable root) {
return assertThat(root).extracting(FlinkAs... | 3.68 |
flink_SortingThread_go | /** Entry point of the thread. */
@Override
public void go() throws InterruptedException {
boolean alive = true;
// loop as long as the thread is marked alive
while (isRunning() && alive) {
final CircularElement<E> element = this.dispatcher.take(SortStage.SORT);
if (element != EOF_MARKER &... | 3.68 |
framework_UIConnector_getPageState | /**
* Returns the state of the Page associated with the UI.
* <p>
* Note that state is considered an internal part of the connector. You
* should not rely on the state object outside of the connector who owns it.
* If you depend on the state of other connectors you should use their
* public API instead of their s... | 3.68 |
hbase_DeadServer_removeDeadServer | /**
* Called from rpc by operator cleaning up deadserver list.
* @param deadServerName the dead server name
* @return true if this server was removed
*/
public synchronized boolean removeDeadServer(final ServerName deadServerName) {
return this.deadServers.remove(deadServerName) != null;
} | 3.68 |
framework_VRadioButtonGroup_getItem | /**
* Returns the JsonObject used to populate the RadioButton widget that
* contains given Element.
*
* @since 8.2
* @param element
* the element to search for
* @return the related JsonObject; {@code null} if not found
*/
public JsonObject getItem(Element element) {
// The HTML populated in upda... | 3.68 |
framework_Escalator_reapplyColumnWidths | /**
* Reapplies all the cells' widths according to the calculated widths in
* the column configuration.
*/
public void reapplyColumnWidths() {
Element row = root.getFirstChildElement();
while (row != null) {
// Only handle non-spacer rows
if (!body.spacerContainer.isSpacer(row)) {
... | 3.68 |
hbase_VersionModel_getRESTVersion | /** Returns the REST gateway version */
@XmlAttribute(name = "REST")
public String getRESTVersion() {
return restVersion;
} | 3.68 |
framework_Calendar_fireEventMove | /**
* Fires an event move event to all server side move listeners.
*
* @param index
* The index of the event in the events list
* @param newFromDatetime
* The changed from date time
*/
protected void fireEventMove(int index, Date newFromDatetime) {
MoveEvent event = new MoveEvent(this, ... | 3.68 |
hadoop_FutureIO_raiseInnerCause | /**
* Extract the cause of a completion failure and rethrow it if an IOE
* or RTE.
* @param e exception.
* @param <T> type of return value.
* @return nothing, ever.
* @throws IOException either the inner IOException, or a wrapper around
* any non-Runtime-Exception
* @throws RuntimeException if that is the inner... | 3.68 |
zxing_MaskUtil_applyMaskPenaltyRule4 | /**
* Apply mask penalty rule 4 and return the penalty. Calculate the ratio of dark cells and give
* penalty if the ratio is far from 50%. It gives 10 penalty for 5% distance.
*/
static int applyMaskPenaltyRule4(ByteMatrix matrix) {
int numDarkCells = 0;
byte[][] array = matrix.getArray();
int width = matrix.g... | 3.68 |
hadoop_RBFMetrics_getJson | /**
* Get JSON for this record.
*
* @return Map representing the data for the JSON representation.
*/
private static Map<String, Object> getJson(BaseRecord record) {
Map<String, Object> json = new HashMap<>();
Map<String, Class<?>> fields = getFields(record);
for (String fieldName : fields.keySet()) {
if... | 3.68 |
hadoop_StageAllocatorLowCostAligned_canAllocate | // canAllocate() - boolean function, returns whether requestedResources
// can be allocated during the durationInterval without
// violating capacity constraints
public boolean canAllocate() {
return (gangsCanFit > 0);
} | 3.68 |
framework_VCalendarPanel_onCancel | /**
* Notifies submit-listeners of a cancel event
*/
private void onCancel() {
if (getSubmitListener() != null) {
getSubmitListener().onCancel();
}
} | 3.68 |
hbase_StorageClusterStatusModel_getLiveNodes | /** Returns the list of live nodes */
@XmlElement(name = "Node")
@XmlElementWrapper(name = "LiveNodes")
// workaround https://github.com/FasterXML/jackson-dataformat-xml/issues/192
@JsonProperty("LiveNodes")
public List<Node> getLiveNodes() {
return liveNodes;
} | 3.68 |
hbase_SnapshotManifest_convertToV2SingleManifest | /*
* In case of rolling-upgrade, we try to read all the formats and build the snapshot with the
* latest format.
*/
private void convertToV2SingleManifest() throws IOException {
// Try to load v1 and v2 regions
List<SnapshotRegionManifest> v1Regions, v2Regions;
ThreadPoolExecutor tpool = createExecutor("Snapsh... | 3.68 |
hadoop_BCFile_getAPIVersion | /**
* Get version of BCFile API.
*
* @return version of BCFile API.
*/
public Version getAPIVersion() {
return API_VERSION;
} | 3.68 |
flink_CheckpointStorageLocationReference_getReferenceBytes | /**
* Gets the reference bytes.
*
* <p><b>Important:</b> For efficiency, this method does not make a defensive copy, so the
* caller must not modify the bytes in the array.
*/
public byte[] getReferenceBytes() {
// return a non null object always
return encodedReference != null ? encodedReference : new byt... | 3.68 |
hbase_NamespaceAuditor_getRegionCountOfTable | /**
* Get region count for table
* @param tName - table name
* @return cached region count, or -1 if table status not found
* @throws IOException Signals that the namespace auditor has not been initialized
*/
public int getRegionCountOfTable(TableName tName) throws IOException {
if (stateManager.isInitialized())... | 3.68 |
hudi_HoodieBackedTableMetadata_closePartitionReaders | /**
* Close and clear all the partitions readers.
*/
private void closePartitionReaders() {
for (Pair<String, String> partitionFileSlicePair : partitionReaders.get().keySet()) {
close(partitionFileSlicePair);
}
partitionReaders.get().clear();
} | 3.68 |
flink_PlanReference_fromFile | /** Create a reference starting from a file path. */
public static PlanReference fromFile(File file) {
Objects.requireNonNull(file, "File cannot be null");
return new FilePlanReference(file);
} | 3.68 |
hadoop_FederationPolicyUtils_validateSubClusterAvailability | /**
* Validate if there is any active subcluster that is not blacklisted, it will
* throw an exception if there are no usable subclusters.
*
* @param activeSubClusters the list of subClusters as identified by
* {@link SubClusterId} currently active.
* @param blackListSubClusters the list of subClusters a... | 3.68 |
hbase_SampleUploader_run | /**
* Main entry point.
* @param otherArgs The command line parameters after ToolRunner handles standard.
* @throws Exception When running the job fails.
*/
@Override
public int run(String[] otherArgs) throws Exception {
if (otherArgs.length != 2) {
System.err.println("Wrong number of arguments: " + otherArgs... | 3.68 |
hbase_ProcedureCoordinator_getProcedureNames | /** Returns Return set of all procedure names. */
public Set<String> getProcedureNames() {
return new HashSet<>(procedures.keySet());
} | 3.68 |
graphhopper_PrepareContractionHierarchies_useFixedNodeOrdering | /**
* Instead of heuristically determining a node ordering for the graph contraction it is also possible
* to use a fixed ordering. For example this allows re-using a previously calculated node ordering.
* This will speed up CH preparation, but might lead to slower queries.
*/
public PrepareContractionHierarchies u... | 3.68 |
framework_CompositeValidator_removeValidator | /**
* Removes a validator from the composite.
*
* @param validator
* the Validator object which performs validation checks on this
* set of data field values.
*/
public void removeValidator(Validator validator) {
validators.remove(validator);
} | 3.68 |
flink_MetricListener_getCounter | /**
* Get registered {@link Counter} with identifier relative to the root metric group.
*
* @param identifier identifier relative to the root metric group
* @return Optional registered counter
*/
public Optional<Counter> getCounter(String... identifier) {
return getMetric(Counter.class, identifier);
} | 3.68 |
pulsar_PulsarAdminImpl_brokerStats | /**
* @return the broker statics
*/
public BrokerStats brokerStats() {
return brokerStats;
} | 3.68 |
querydsl_ComparableExpression_loe | /**
* Create a {@code this <= right} expression
*
* @param right rhs of the comparison
* @return this <= right
* @see java.lang.Comparable#compareTo(Object)
*/
public BooleanExpression loe(Expression<T> right) {
return Expressions.booleanOperation(Ops.LOE, mixin, right);
} | 3.68 |
hbase_Scan_getFamilyMap | /**
* Getting the familyMap
*/
public Map<byte[], NavigableSet<byte[]>> getFamilyMap() {
return this.familyMap;
} | 3.68 |
hadoop_AbfsHttpOperation_processStorageErrorResponse | /**
* When the request fails, this function is used to parse the responseAbfsHttpClient.LOG.debug("ExpectedError: ", ex);
* and extract the storageErrorCode and storageErrorMessage. Any errors
* encountered while attempting to process the error response are logged,
* but otherwise ignored.
*
* For storage errors... | 3.68 |
druid_DruidAbstractDataSource_setSocketTimeout | /**
* @since 1.2.12
*/
public void setSocketTimeout(int milliSeconds) {
this.socketTimeout = milliSeconds;
this.socketTimeoutSr = null;
} | 3.68 |
hbase_RootProcedureState_addRollbackStep | /**
* Called by the ProcedureExecutor after the procedure step is completed, to add the step to the
* rollback list (or procedure stack)
*/
protected synchronized void addRollbackStep(Procedure<TEnvironment> proc) {
if (proc.isFailed()) {
state = State.FAILED;
}
if (subprocStack == null) {
subprocStack... | 3.68 |
dubbo_PojoUtils_updatePropertyIfAbsent | /**
* Update the property if absent
*
* @param getterMethod the getter method
* @param setterMethod the setter method
* @param newValue the new value
* @param <T> the value type
* @since 2.7.8
*/
public static <T> void updatePropertyIfAbsent(Supplier<T> getterMethod, Consumer<T> setterMethod, T new... | 3.68 |
flink_ResourceCounter_getResourcesWithCount | /**
* Gets the stored resources and their counts. The counts are guaranteed to be positive (> 0).
*
* @return collection of {@link ResourceProfile} and count pairs
*/
public Collection<Map.Entry<ResourceProfile, Integer>> getResourcesWithCount() {
return resources.entrySet();
} | 3.68 |
flink_StreamSourceContexts_getSourceContext | /**
* Depending on the {@link TimeCharacteristic}, this method will return the adequate {@link
* org.apache.flink.streaming.api.functions.source.SourceFunction.SourceContext}. That is:
*
* <ul>
* <li>{@link TimeCharacteristic#IngestionTime} = {@code AutomaticWatermarkContext}
* <li>{@link TimeCharacteristic#P... | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.