name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
hbase_RegionCoprocessorHost_preDelete | /**
* Supports Coprocessor 'bypass'.
* @param delete The Delete object
* @param edit The WALEdit object.
* @return true if default processing should be bypassed
* @exception IOException Exception
*/
public boolean preDelete(final Delete delete, final WALEdit edit) throws IOException {
if (this.coprocEnvironme... | 3.68 |
hbase_BloomFilterUtil_createBySize | /**
* Creates a Bloom filter chunk of the given size.
* @param byteSizeHint the desired number of bytes for the Bloom filter bit array. Will be
* increased so that folding is possible.
* @param errorRate target false positive rate of the Bloom filter
* @param hashType Bloom filter hash f... | 3.68 |
dubbo_ClassSourceScanner_adaptiveClasses | /**
* The required adaptive class.
* For example: LoadBalance$Adaptive.class
* @return adaptive class
*/
public Map<String, Class<?>> adaptiveClasses() {
List<String> res = spiClassesWithAdaptive().stream()
.map((c) -> c.getName() + "$Adaptive")
.collect(Collectors.toList());
return ... | 3.68 |
flink_StreamExecutionEnvironment_getNumberOfExecutionRetries | /**
* Gets the number of times the system will try to re-execute failed tasks. A value of {@code
* -1} indicates that the system default value (as defined in the configuration) should be used.
*
* @return The number of times the system will try to re-execute failed tasks.
* @deprecated This method will be replaced... | 3.68 |
framework_BootstrapHandler_setupMainDiv | /**
* Method to write the div element into which that actual Vaadin application
* is rendered.
* <p>
* Override this method if you want to add some custom html around around
* the div element into which the actual Vaadin application will be
* rendered.
*
* @param context
*
* @throws IOException
*/
private vo... | 3.68 |
flink_TopNBuffer_getElement | /**
* Gets record which rank is given value.
*
* @param rank rank value to search
* @return the record which rank is given value
*/
public RowData getElement(int rank) {
int curRank = 0;
for (Map.Entry<RowData, Collection<RowData>> entry : treeMap.entrySet()) {
Collection<RowData> collection = entr... | 3.68 |
hbase_HFileLink_createPath | /**
* Create an HFileLink relative path for the table/region/family/hfile location
* @param table Table name
* @param region Region Name
* @param family Family Name
* @param hfile HFile Name
* @return the relative Path to open the specified table/region/family/hfile link
*/
public static Path createPath(final ... | 3.68 |
hadoop_YarnRegistryViewForProviders_getSelfRegistrationPath | /**
* Get the path to where the service has registered itself.
* Null until the service is registered
* @return the service registration path.
*/
public String getSelfRegistrationPath() {
return selfRegistrationPath;
} | 3.68 |
morf_AbstractSqlDialectTest_nullOrder | /**
* A database platform may need to specify the null order.
*
* <p>If a null order is not required for a SQL dialect descendant classes need to implement this method.</p>
*
* @return the null order for an SQL dialect
*/
protected String nullOrder() {
return StringUtils.EMPTY;
} | 3.68 |
hadoop_ManifestCommitterSupport_getEtag | /**
* Get an etag from a FileStatus which MUST BE
* an implementation of EtagSource and
* whose etag MUST NOT BE null/empty.
* @param status the status; may be null.
* @return the etag or null if not provided
*/
public static String getEtag(FileStatus status) {
if (status instanceof EtagSource) {
return ((E... | 3.68 |
hbase_ScanQueryMatcher_checkColumn | // Used only for testing purposes
static MatchCode checkColumn(ColumnTracker columnTracker, byte[] bytes, int offset, int length,
long ttl, byte type, boolean ignoreCount) throws IOException {
KeyValue kv = KeyValueUtil.createFirstOnRow(HConstants.EMPTY_BYTE_ARRAY, 0, 0,
HConstants.EMPTY_BYTE_ARRAY, 0, 0, bytes... | 3.68 |
hbase_Bytes_hashCode | /**
* Calculate the hash code for a given range of bytes.
* @param bytes array to hash
* @param offset offset to start from
* @param length length to hash
*/
public static int hashCode(byte[] bytes, int offset, int length) {
int hash = 1;
for (int i = offset; i < offset + length; i++)
hash = (31 * hash) +... | 3.68 |
framework_Notification_getStyle | /**
* @since 7.2
*
* @return the style name for this notification type.
*/
public String getStyle() {
return style;
} | 3.68 |
flink_CliFrontend_loadCustomCommandLine | /**
* Loads a class from the classpath that implements the CustomCommandLine interface.
*
* @param className The fully-qualified class name to load.
* @param params The constructor parameters
*/
private static CustomCommandLine loadCustomCommandLine(String className, Object... params)
throws Exception {
... | 3.68 |
hbase_WALStreamReader_next | /**
* Read the next entry in WAL.
* <p/>
* In most cases you should just use this method, especially when reading a closed wal file for
* splitting or printing.
*/
default WAL.Entry next() throws IOException {
return next(null);
} | 3.68 |
framework_InMemoryDataProvider_setFilterByValue | /**
* Sets a filter that requires an item property to have a specific value.
* The property value and the provided value are compared using
* {@link Object#equals(Object)}. The filter replaces any filter that has
* been set or added previously.
*
* @see #setFilter(SerializablePredicate)
* @see #setFilter(ValuePr... | 3.68 |
framework_VComboBox_getItemOffsetWidth | /*
* Gets the width of one menu item.
*/
int getItemOffsetWidth() {
List<MenuItem> items = getItems();
return items != null && !items.isEmpty()
? items.get(0).getOffsetWidth()
: 0;
} | 3.68 |
hadoop_CommitContext_commitOrFail | /**
* Commit the operation, throwing an exception on any failure.
* See {@code CommitOperations#commitOrFail(SinglePendingCommit)}.
* @param commit commit to execute
* @throws IOException on a failure
*/
public void commitOrFail(SinglePendingCommit commit) throws IOException {
commitOperations.commitOrFail(commi... | 3.68 |
hadoop_StorageStatistics_getName | /**
* Get the name of this StorageStatistics object.
* @return name of this StorageStatistics object
*/
public String getName() {
return name;
} | 3.68 |
framework_VaadinService_getDependencyFilters | /**
* Gets the filters which all resource dependencies are passed through
* before being sent to the client for loading.
*
* @see #initDependencyFilters(List)
*
* @since 8.1
* @return the dependency filters to pass resources dependencies through
* before loading
*/
public Iterable<DependencyFilter> get... | 3.68 |
flink_RocksDBStateBackend_getNumberOfTransferingThreads | /** @deprecated Typo in method name. Use {@link #getNumberOfTransferThreads} instead. */
@Deprecated
public int getNumberOfTransferingThreads() {
return getNumberOfTransferThreads();
} | 3.68 |
flink_SideOutputDataStream_cache | /**
* Caches the intermediate result of the transformation. Only support bounded streams and
* currently only block mode is supported. The cache is generated lazily at the first time the
* intermediate result is computed. The cache will be clear when {@link
* CachedDataStream#invalidate()} called or the {@link Stre... | 3.68 |
hadoop_FTPInputStream_seek | // We don't support seek.
@Override
public void seek(long pos) throws IOException {
throw new IOException("Seek not supported");
} | 3.68 |
flink_SubtaskStateStats_getProcessedData | /** @return the total number of processed bytes during the checkpoint. */
public long getProcessedData() {
return processedData;
} | 3.68 |
hbase_Compression_createPlainCompressionStream | /**
* Creates a compression stream without any additional wrapping into buffering streams.
*/
public CompressionOutputStream createPlainCompressionStream(OutputStream downStream,
Compressor compressor) throws IOException {
CompressionCodec codec = getCodec(conf);
((Configurable) codec).getConf().setInt("io.file... | 3.68 |
flink_FileInputFormat_registerInflaterInputStreamFactory | /**
* Registers a decompression algorithm through a {@link
* org.apache.flink.api.common.io.compression.InflaterInputStreamFactory} with a file extension
* for transparent decompression.
*
* @param fileExtension of the compressed files
* @param factory to create an {@link java.util.zip.InflaterInputStream} that h... | 3.68 |
hudi_SparkBootstrapCommitActionExecutor_metadataBootstrap | /**
* Perform Metadata Bootstrap.
* @param partitionFilesList List of partitions and files within that partitions
*/
protected Option<HoodieWriteMetadata<HoodieData<WriteStatus>>> metadataBootstrap(List<Pair<String, List<HoodieFileStatus>>> partitionFilesList) {
if (null == partitionFilesList || partitionFilesList... | 3.68 |
hadoop_BlockResolver_resolve | /**
* @param s the external reference.
* @return sequence of blocks that make up the reference.
*/
public Iterable<BlockProto> resolve(FileStatus s) {
List<Long> lengths = blockLengths(s);
ArrayList<BlockProto> ret = new ArrayList<>(lengths.size());
long tot = 0;
for (long l : lengths) {
tot += l;
re... | 3.68 |
hudi_HoodieTableMetaClient_getSchemaFolderName | /**
* @return schema folder path
*/
public String getSchemaFolderName() {
return new Path(metaPath.get(), SCHEMA_FOLDER_NAME).toString();
} | 3.68 |
flink_NumericSummaryAggregator_combine | /** combine two aggregations. */
@Override
public void combine(Aggregator<T, NumericColumnSummary<T>> otherSameType) {
NumericSummaryAggregator<T> other = (NumericSummaryAggregator<T>) otherSameType;
nullCount += other.nullCount;
nanCount += other.nanCount;
infinityCount += other.infinityCount;
if... | 3.68 |
hibernate-validator_Contracts_assertValueNotNull | /**
* Asserts that the given object is not {@code null}.
*
* @param o The object to check.
* @param name The name of the value to check. A message of the form
* "<name> must not be null" will be used as message of
* the resulting exception if the given object is {@code null}.
*
* @throws IllegalArgumentEx... | 3.68 |
hbase_HFileReaderImpl_getCurCellSerializedSize | // Returns the #bytes in HFile for the current cell. Used to skip these many bytes in current
// HFile block's buffer so as to position to the next cell.
private int getCurCellSerializedSize() {
int curCellSize = KEY_VALUE_LEN_SIZE + currKeyLen + currValueLen + currMemstoreTSLen;
if (this.reader.getFileContext().is... | 3.68 |
framework_RpcDataSourceConnector_updateRowData | /**
* Updates row data based on row key.
*
* @since 7.6
* @param row
* new row object
*/
protected void updateRowData(JsonObject row) {
int index = indexOfKey(getRowKey(row));
if (index >= 0) {
setRowData(index, Collections.singletonList(row));
}
} | 3.68 |
rocketmq-connect_FieldsMetadata_extractRecordValuePk | /**
* record value
*
* @param tableName
* @param configuredPkFields
* @param valueSchema
* @param headers
* @param allFields
* @param keyFieldNames
*/
private static void extractRecordValuePk(
final String tableName,
final List<String> configuredPkFields,
final Schema valueSchema,
... | 3.68 |
framework_DragSourceExtensionConnector_removeDragListeners | /**
* Removes dragstart and dragend event listeners from the given DOM element.
*
* @param element
* DOM element to remove event listeners from.
*/
protected void removeDragListeners(Element element) {
EventTarget target = element.cast();
target.removeEventListener(Event.DRAGSTART, dragStartLis... | 3.68 |
hadoop_UriUtils_generateUniqueTestPath | /**
* Generate unique test path for multiple user tests.
*
* @return root test path
*/
public static String generateUniqueTestPath() {
String testUniqueForkId = System.getProperty("test.unique.fork.id");
return testUniqueForkId == null ? "/test" : "/" + testUniqueForkId + "/test";
} | 3.68 |
hudi_BootstrapOperator_waitForBootstrapReady | /**
* Wait for other bootstrap tasks to finish the index bootstrap.
*/
private void waitForBootstrapReady(int taskID) {
int taskNum = getRuntimeContext().getNumberOfParallelSubtasks();
int readyTaskNum = 1;
while (taskNum != readyTaskNum) {
try {
readyTaskNum = aggregateManager.updateGlobalAggregate(B... | 3.68 |
hbase_AssignmentVerificationReport_getTotalFavoredAssignments | /**
* Return the number of regions assigned to their favored nodes
* @return number of regions assigned to their favored nodes
*/
int getTotalFavoredAssignments() {
return totalFavoredAssignments;
} | 3.68 |
hadoop_GetAllResourceTypeInfoResponsePBImpl_initResourceTypeInfosList | // Once this is called. containerList will never be null - until a getProto
// is called.
private void initResourceTypeInfosList() {
if (this.resourceTypeInfo != null) {
return;
}
GetAllResourceTypeInfoResponseProtoOrBuilder p = viaProto ? proto : builder;
List<ResourceTypeInfoProto> list = p.getResourceTyp... | 3.68 |
flink_SegmentPartitionFileWriter_writeSegmentFinishFile | /**
* Writing a segment-finish file when the current segment is complete. The downstream can
* determine if the current segment is complete by checking for the existence of the
* segment-finish file.
*
* <p>Note that the method is only called by the flushing thread.
*/
private void writeSegmentFinishFile(
... | 3.68 |
framework_VTabsheet_scheduleBlur | /**
* Schedule a new blur event for a deferred execution.
*
* @param blurSource
* the source tab
*/
private void scheduleBlur(Tab blurSource) {
if (nextBlurScheduleCancelled) {
// This will set the stopNextBlurCommand back to false as well.
cancelLastBlurSchedule();
// Res... | 3.68 |
flink_HardwareDescription_getSizeOfPhysicalMemory | /**
* Returns the size of physical memory in bytes available on the compute node.
*
* @return the size of physical memory in bytes available on the compute node
*/
public long getSizeOfPhysicalMemory() {
return this.sizeOfPhysicalMemory;
} | 3.68 |
framework_ApplicationConnection_getConnector | /**
* Get either an existing ComponentConnector or create a new
* ComponentConnector with the given type and id.
*
* If a ComponentConnector with the given id already exists, returns it.
* Otherwise creates and registers a new ComponentConnector of the given
* type.
*
* @param connectorId
* Id of th... | 3.68 |
framework_EventCellReference_isHeader | /**
* Is the cell reference for a cell in the header of the Grid.
*
* @since 7.5
* @return <code>true</code> if referenced cell is in the header,
* <code>false</code> if not
*/
public boolean isHeader() {
return section == Section.HEADER;
} | 3.68 |
flink_NettyPartitionRequestClient_requestSubpartition | /**
* Requests a remote intermediate result partition queue.
*
* <p>The request goes to the remote producer, for which this partition request client instance
* has been created.
*/
@Override
public void requestSubpartition(
final ResultPartitionID partitionId,
final int subpartitionIndex,
f... | 3.68 |
hudi_PreferWriterConflictResolutionStrategy_getCandidateInstantsForTableServicesCommits | /**
* To find which instants are conflicting, we apply the following logic
* Get both completed instants and ingestion inflight commits that have happened since the last successful write.
* We need to check for write conflicts since they may have mutated the same files
* that are being newly created by the current ... | 3.68 |
flink_RocksDBProperty_getRocksDBProperty | /**
* @return property string that can be used to query {@link
* RocksDB#getLongProperty(ColumnFamilyHandle, String)}.
*/
public String getRocksDBProperty() {
return String.format(ROCKS_DB_PROPERTY_FORMAT, property);
} | 3.68 |
morf_BaseDataSetReader_open | /**
* @see org.alfasoftware.morf.xml.XmlStreamProvider#open()
*/
@Override
public void open() {
// no-op
} | 3.68 |
hadoop_FederationStateStoreFacade_incrementCurrentKeyId | /**
* stateStore provides CurrentKeyId increase.
*
* @return currentKeyId.
*/
public int incrementCurrentKeyId() {
return stateStore.incrementCurrentKeyId();
} | 3.68 |
framework_ColorPickerGradient_setBackgroundColor | /**
* Sets the background color.
*
* @param color
* the new background color
*/
public void setBackgroundColor(Color color) {
getState().bgColor = color.getCSS();
} | 3.68 |
hbase_ServerCrashProcedure_getRegionsOnCrashedServer | /** Returns List of Regions on crashed server. */
List<RegionInfo> getRegionsOnCrashedServer(MasterProcedureEnv env) {
return env.getMasterServices().getAssignmentManager().getRegionsOnServer(serverName);
} | 3.68 |
rocketmq-connect_AvroData_defaultValueFromConnect | // Convert default values from Connect data format to Avro's format, which is an
// org.codehaus.jackson.JsonNode. The default value is provided as an argument because even
// though you can get a default value from the schema, default values for complex structures need
// to perform the same translation but those defa... | 3.68 |
hbase_UserProvider_create | /**
* Wraps an underlying {@code UserGroupInformation} instance.
* @param ugi The base Hadoop user
*/
public User create(UserGroupInformation ugi) {
if (ugi == null) {
return null;
}
return new User.SecureHadoopUser(ugi, groupCache);
} | 3.68 |
flink_OperatorChain_close | /**
* This method releases all resources of the record writer output. It stops the output flushing
* thread (if there is one) and releases all buffers currently held by the output serializers.
*
* <p>This method should never fail.
*/
public void close() throws IOException {
closer.close();
} | 3.68 |
hbase_ExceptionUtil_rethrowIfInterrupt | /** Throw InterruptedIOException if t was an interruption, nothing otherwise. */
public static void rethrowIfInterrupt(Throwable t) throws InterruptedIOException {
InterruptedIOException iie = asInterrupt(t);
if (iie != null) {
throw iie;
}
} | 3.68 |
morf_UnionSetOperator_validateOrderBy | /**
* Don't allow sub-select statements to have ORDER BY statements, as this is
* an invalid construct in in SQL-92.
*
* @param selectStatement the select statement to be validated.
*/
private void validateOrderBy(SelectStatement selectStatement) throws IllegalArgumentException {
if (!selectStatement.getOrderBys... | 3.68 |
hbase_TableDescriptorBuilder_getNormalizerTargetRegionCount | /**
* Check if there is the target region count. If so, the normalize plan will be calculated based
* on the target region count.
* @return target region count after normalize done
*/
@Override
public int getNormalizerTargetRegionCount() {
return getOrDefault(NORMALIZER_TARGET_REGION_COUNT_KEY, Integer::valueOf,
... | 3.68 |
MagicPlugin_SpellAction_load | /**
* This mechanism never worked properly and is no longer called.
* Actions that need to store data should interact with CastContext.getVariables instead.
*/
@Deprecated
default void load(Mage mage, ConfigurationSection data) { } | 3.68 |
hadoop_TFile_getBlockContainsKey | /**
* if greater is true then returns the beginning location of the block
* containing the key strictly greater than input key. if greater is false
* then returns the beginning location of the block greater than equal to
* the input key
*
* @param key
* the input key
* @param greater
* boole... | 3.68 |
flink_SpillingThread_disposeSortBuffers | /** Releases the memory that is registered for in-memory sorted run generation. */
private void disposeSortBuffers(boolean releaseMemory) {
CircularElement<E> element;
while ((element = this.dispatcher.poll(SortStage.READ)) != null) {
element.getBuffer().dispose();
if (releaseMemory) {
... | 3.68 |
flink_QueryableStateConfiguration_getStateServerPortRange | /**
* Returns the port range where the queryable state server can listen. See {@link
* org.apache.flink.configuration.QueryableStateOptions#SERVER_PORT_RANGE
* QueryableStateOptions.SERVER_PORT_RANGE}.
*/
public Iterator<Integer> getStateServerPortRange() {
return qserverPortRange;
} | 3.68 |
hadoop_IOStatisticsStoreImpl_getMaximumReference | /**
* Get a reference to the atomic instance providing the
* value for a specific maximum. This is useful if
* the value is passed around.
* @param key statistic name
* @return the reference
* @throws NullPointerException if there is no entry of that name
*/
@Override
public AtomicLong getMaximumReference(String... | 3.68 |
hbase_RegionCoprocessorHost_postClose | /**
* Invoked after a region is closed
* @param abortRequested true if the server is aborting
*/
public void postClose(final boolean abortRequested) {
try {
execOperation(new RegionObserverOperationWithoutResult() {
@Override
public void call(RegionObserver observer) throws IOException {
ob... | 3.68 |
hbase_OrderedInt64_decodeLong | /**
* Read a {@code long} value from the buffer {@code src}.
* @param src the {@link PositionedByteRange} to read the {@code long} from
* @return the {@code long} read from the buffer
*/
public long decodeLong(PositionedByteRange src) {
return OrderedBytes.decodeInt64(src);
} | 3.68 |
framework_NotificationElement_getCaption | /**
* Returns the caption of the Notification element.
*
* @since 8.0
* @return the caption of the Notification element
*/
public String getCaption() {
WebElement popup = findElement(By.className("popupContent"));
WebElement caption = popup.findElement(By.tagName("h1"));
return caption.getText();
} | 3.68 |
hbase_MetricsSource_refreshAgeOfLastShippedOp | /**
* Convenience method to use the last given timestamp to refresh the age of the last edit. Used
* when replication fails and need to keep that metric accurate.
* @param walGroupId id of the group to update
*/
public void refreshAgeOfLastShippedOp(String walGroupId) {
Long lastTimestamp = this.lastShippedTimeSt... | 3.68 |
hudi_Triple_compareTo | /**
* <p>
* Compares the triple based on the left element, followed by the middle element, finally the right element. The types
* must be {@code Comparable}.
* </p>
*
* @param other the other triple, not null
* @return negative if this is less, zero if equal, positive if greater
*/
@Override
public int compareT... | 3.68 |
flink_S3TestCredentials_assumeCredentialsAvailable | /**
* Checks whether credentials are available in the environment variables of this JVM. If not,
* throws an {@link AssumptionViolatedException} which causes JUnit tests to be skipped.
*/
public static void assumeCredentialsAvailable() {
Assume.assumeTrue(
"No S3 credentials available in this test's ... | 3.68 |
hbase_PrivateCellUtil_writeValue | /**
* Writes the value from the given cell to the output stream
* @param out The outputstream to which the data has to be written
* @param cell The cell whose contents has to be written
* @param vlength the value length
*/
public static void writeValue(OutputStream out, Cell cell, int vlength) throws IOExce... | 3.68 |
hadoop_AzureFileSystemInstrumentation_getCurrentWebResponses | /**
* Gets the current number of web responses obtained from Azure Storage.
* @return The number of web responses.
*/
public long getCurrentWebResponses() {
return inMemoryNumberOfWebResponses.get();
} | 3.68 |
morf_InsertStatement_useDirectPath | /**
* If supported by the dialect, hints to the database that an {@code APPEND} query hint should be used in the insert statement.
*
* <p>In general, as with all query plan modification, <strong>do not use this unless you know
* exactly what you are doing</strong>.</p>
*
* <p>These directives are applied in the S... | 3.68 |
hadoop_RouterAuditLogger_logFailure | /**
* Create a readable and parsable audit log string for a failed event.
*
* @param user User who made the service request.
* @param operation Operation requested by the user.
* @param perm Target permissions.
* @param target The target on which the operation is being performed.
* @param description Some additi... | 3.68 |
rocketmq-connect_Worker_stopConnector | /**
* Stop a connector managed by this worker.
*
* @param connName the connector name.
*/
private void stopConnector(String connName) {
WorkerConnector workerConnector = connectors.get(connName);
log.info("Stopping connector {}", connName);
if (workerConnector == null) {
log.warn("Ignoring stop ... | 3.68 |
flink_GenericDataSinkBase_setInput | /**
* Sets the given operator as the input to this operator.
*
* @param input The operator to use as the input.
*/
public void setInput(Operator<IN> input) {
this.input = checkNotNull(input, "The input may not be null.");
} | 3.68 |
hadoop_YarnServiceConf_getLong | /**
* Get long value for the property. First get from the userConf, if not
* present, get from systemConf.
*
* @param name name of the property
* @param defaultValue default value of the property, if it is not defined in
* userConf and systemConf.
* @param userConf Configuration provided by c... | 3.68 |
framework_Potus_getLastName | /**
* @return the lastName
*/
public String getLastName() {
return lastName;
} | 3.68 |
flink_ArrayListConverter_createObjectArrayKind | /** Creates the kind of array for {@link List#toArray(Object[])}. */
private static Object[] createObjectArrayKind(Class<?> elementClazz) {
// e.g. int[] is not a Object[]
if (elementClazz.isPrimitive()) {
return (Object[]) Array.newInstance(primitiveToWrapper(elementClazz), 0);
}
// e.g. int[][... | 3.68 |
hbase_StoreFileInfo_getCreatedTimestamp | /** Returns timestamp when this file was created (as returned by filesystem) */
public long getCreatedTimestamp() {
return createdTimestamp;
} | 3.68 |
zxing_HistoryManager_buildHistory | /**
* <p>Builds a text representation of the scanning history. Each scan is encoded on one
* line, terminated by a line break (\r\n). The values in each line are comma-separated,
* and double-quoted. Double-quotes within values are escaped with a sequence of two
* double-quotes. The fields output are:</p>
*
* <ol... | 3.68 |
hbase_WALSplitter_createOutputSinkAndEntryBuffers | /**
* Setup the output sinks and entry buffers ahead of splitting WAL.
*/
private void createOutputSinkAndEntryBuffers() {
PipelineController controller = new PipelineController();
if (this.hfile) {
this.entryBuffers = new BoundedEntryBuffers(controller, this.bufferSize);
this.outputSink = new BoundedReco... | 3.68 |
hadoop_AzureBlobFileSystem_checkException | /**
* Given a path and exception, choose which IOException subclass
* to create.
* Will return if and only iff the error code is in the list of allowed
* error codes.
* @param path path of operation triggering exception; may be null
* @param exception the exception caught
* @param allowedErrorCodesList varargs l... | 3.68 |
hbase_DisableTableProcedure_setTableStateToDisabling | /**
* Mark table state to Disabling
* @param env MasterProcedureEnv
*/
private static void setTableStateToDisabling(final MasterProcedureEnv env,
final TableName tableName) throws IOException {
// Set table disabling flag up in zk.
env.getMasterServices().getTableStateManager().setTableState(tableName,
Tab... | 3.68 |
hbase_RegionServerAccounting_getGlobalMemStoreDataSize | /** Returns the global Memstore data size in the RegionServer */
public long getGlobalMemStoreDataSize() {
return globalMemStoreDataSize.sum();
} | 3.68 |
morf_ViewDeploymentValidator_validateExistingView | /**
*
* @param view View being examined. This is an existing schema view.
* @return boolean true if view is valid.
* @deprecated kept to ensure backwards compatibility.
*/
@Override
@Deprecated
public boolean validateExistingView(View view) {
return true; // the given existing view is okay as is
} | 3.68 |
querydsl_StringExpression_containsIgnoreCase | /**
* Create a {@code this.containsIgnoreCase(str)} expression
*
* <p>Returns true if the given String is contained, compare case insensitively</p>
*
* @param str string
* @return this.containsIgnoreCase(str)
*/
public BooleanExpression containsIgnoreCase(String str) {
return containsIgnoreCase(ConstantImpl.... | 3.68 |
flink_SingleOutputStreamOperator_setChainingStrategy | /**
* Sets the {@link ChainingStrategy} for the given operator affecting the way operators will
* possibly be co-located on the same thread for increased performance.
*
* @param strategy The selected {@link ChainingStrategy}
* @return The operator with the modified chaining strategy
*/
@PublicEvolving
private Sin... | 3.68 |
hadoop_WritableName_setName | /**
* Set the name that a class should be known as to something other than the
* class name.
*
* @param writableClass input writableClass.
* @param name input name.
*/
public static synchronized void setName(Class<?> writableClass, String name) {
CLASS_TO_NAME.put(writableClass, name);
NAME_TO_CLASS.put(name,... | 3.68 |
hadoop_BlockManagerParameters_getPrefetchingStatistics | /**
* @return The prefetching statistics for the stream.
*/
public PrefetchingStatistics getPrefetchingStatistics() {
return prefetchingStatistics;
} | 3.68 |
hudi_HoodieTable_getCompletedSavepointTimeline | /**
* Get only the completed (no-inflights) savepoint timeline.
*/
public HoodieTimeline getCompletedSavepointTimeline() {
return getActiveTimeline().getSavePointTimeline().filterCompletedInstants();
} | 3.68 |
hadoop_AuditContextUpdater_resetCurrentAuditContext | /**
* Remove job/task info from the current audit context.
*/
public void resetCurrentAuditContext() {
currentAuditContext().remove(AuditConstants.PARAM_JOB_ID);
currentAuditContext().remove(CommitConstants.PARAM_TASK_ATTEMPT_ID);
} | 3.68 |
hadoop_StripedBlockChecksumReconstructor_clearBuffers | /**
* Clear all associated buffers.
*/
private void clearBuffers() {
getStripedReader().clearBuffers();
targetBuffer.clear();
} | 3.68 |
framework_ProgressBar_setIndeterminate | /**
* Sets whether or not this progress indicator is indeterminate. In
* indeterminate mode there is an animation indicating that the task is
* running but without providing any information about the current progress.
*
* @param indeterminate
* <code>true</code> to set to indeterminate mode; otherwise
... | 3.68 |
hibernate-validator_ExecutableHelper_instanceMethodParametersResolveToSameTypes | /**
* Whether the parameters of the two given instance methods resolve to the same types or not. Takes type parameters into account.
*
* @param subTypeMethod a method on a supertype
* @param superTypeMethod a method on a subtype
*
* @return {@code true} if the parameters of the two methods resolve to the same typ... | 3.68 |
framework_MultiSelect_isSelected | /**
* Returns whether the given item is currently selected.
*
* @param item
* the item to check, not null
* @return {@code true} if the item is selected, {@code false} otherwise
*/
public default boolean isSelected(T item) {
return getSelectedItems().contains(item);
} | 3.68 |
hbase_HRegionServer_getReplicationSinkService | /** Returns Return the object that implements the replication sink executorService. */
public ReplicationSinkService getReplicationSinkService() {
return replicationSinkHandler;
} | 3.68 |
flink_ActiveResourceManager_requestNewWorker | /**
* Allocates a resource using the worker resource specification.
*
* @param workerResourceSpec workerResourceSpec specifies the size of the to be allocated
* resource
*/
@VisibleForTesting
public void requestNewWorker(WorkerResourceSpec workerResourceSpec) {
final TaskExecutorProcessSpec taskExecutorPro... | 3.68 |
hadoop_LocalityMulticastAMRMProxyPolicy_getHeadroomWeighting | /**
* Compute the weighting based on available headroom. This is proportional to
* the available headroom memory announced by RM, or to 1/N for RMs we have
* not seen yet. If all RMs report zero headroom, we fallback to 1/N again.
*/
private float getHeadroomWeighting(SubClusterId targetId,
AllocationBookkeeper... | 3.68 |
framework_VScrollTable_setRowFocus | /**
* Moves the selection head to a specific row.
*
* @param row
* The row to where the selection head should move
* @return Returns true if focus was moved successfully, else false
*/
public boolean setRowFocus(VScrollTableRow row) {
if (!isSelectable()) {
return false;
}
// Remo... | 3.68 |
hudi_HoodieIngestionService_onIngestionCompletes | /**
* A callback method to be invoked after ingestion completes.
* <p>
* For continuous mode, this is invoked once after exiting the ingestion loop.
*/
protected boolean onIngestionCompletes(boolean hasError) {
return true;
} | 3.68 |
flink_AvailabilityProvider_resetUnavailable | /** Judges to reset the current available state as unavailable. */
public void resetUnavailable() {
if (isAvailable()) {
availableFuture = new CompletableFuture<>();
}
} | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.