name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
hadoop_AuditingIntegration_enterStageWorker | /**
* Update the thread context with the stage name and
* job ID.
* This MUST be invoked at the start of methods invoked in helper threads,
* to ensure that they are all annotated with job and stage.
* @param jobId job ID.
* @param stage stage name.
*/
public static void enterStageWorker(String jobId, String sta... | 3.68 |
streampipes_IgnoreBlocksAfterContentFilter_getDefaultInstance | /**
* Returns the singleton instance for DeleteBlocksAfterContentFilter.
*/
public static IgnoreBlocksAfterContentFilter getDefaultInstance() {
return DEFAULT_INSTANCE;
} | 3.68 |
framework_AbsoluteLayout_getTopValue | /**
* Gets the 'top' attributes value in current units.
*
* @see #getTopUnits()
* @return The value of the 'top' attribute, null if not set
*/
public Float getTopValue() {
return topValue;
} | 3.68 |
AreaShop_FileManager_postUpdateFiles | /**
* Checks for old file formats and converts them to the latest format.
* This is to be triggered after the load of the region files.
*/
private void postUpdateFiles() {
Integer fileStatus = versions.get(AreaShop.versionFiles);
// If the the files are already the current version
if(fileStatus != null && fileSt... | 3.68 |
flink_CopyOnWriteStateMap_removeEntry | /** Helper method that is the basis for operations that remove mappings. */
private StateMapEntry<K, N, S> removeEntry(K key, N namespace) {
final int hash = computeHashForOperationAndDoIncrementalRehash(key, namespace);
final StateMapEntry<K, N, S>[] tab = selectActiveTable(hash);
int index = hash & (tab.... | 3.68 |
morf_DatabaseSchemaManager_mutateToSupportSchema | /**
* Mutates the current database schema so that it supports the one requested.
*
* <p>When this method returns, it guarantees that all the tables in {code schema} are
* present in the database and also empty.</p>
*
* <p>Note it does not guarantee that no other tables exist.</p>
*
* @param schema The schema wh... | 3.68 |
dubbo_Converter_getTargetType | /**
* Get the target type
*
* @return non-null
*/
default Class<T> getTargetType() {
return findActualTypeArgument(getClass(), Converter.class, 1);
} | 3.68 |
framework_TreeTable_addExpandListener | /**
* Adds an expand listener.
*
* @param listener
* the Listener to be added.
*/
public void addExpandListener(ExpandListener listener) {
addListener(ExpandEvent.class, listener, ExpandListener.EXPAND_METHOD);
} | 3.68 |
hbase_PrivateCellUtil_getValueAsBigDecimal | /**
* Converts the value bytes of the given cell into a BigDecimal
* @return value as BigDecimal
*/
public static BigDecimal getValueAsBigDecimal(Cell cell) {
if (cell instanceof ByteBufferExtendedCell) {
return ByteBufferUtils.toBigDecimal(((ByteBufferExtendedCell) cell).getValueByteBuffer(),
((ByteBuff... | 3.68 |
framework_ScrollbarBundle_truncate | /**
* Truncates a double such that no decimal places are retained.
* <p>
* E.g. {@code trunc(2.3d) == 2.0d} and {@code trunc(-2.3d) == -2.0d}.
*
* @param num
* the double value to be truncated
* @return the {@code num} value without any decimal digits
*/
private static double truncate(double num) {
... | 3.68 |
hbase_RollingStatCalculator_insertDataValue | /**
* Inserts given data value to array of data values to be considered for statistics calculation
*/
public void insertDataValue(long data) {
// if current number of data points already equals rolling period and rolling period is
// non-zero then remove one data and update the statistics
if (numberOfDataValues... | 3.68 |
framework_VaadinPortletRequest_getPortletPreference | /**
* Reads a portlet preference from the portlet of the request.
*
* @param name
* The name of the portlet preference. Cannot be
* <code>null</code>.
*
* @return The value of the portlet preference, <code>null</code> if the
* preference is not defined.
*/
public String getPortlet... | 3.68 |
hadoop_S3ListRequest_v1 | /**
* Restricted constructors to ensure v1 or v2, not both.
* @param request v1 request
* @return new list request container
*/
public static S3ListRequest v1(ListObjectsRequest request) {
return new S3ListRequest(request, null);
} | 3.68 |
framework_VFilterSelect_setPromptingOff | /**
* Turns prompting off. When prompting is turned on a command prompt is
* shown in the text box if nothing has been entered.
* <p>
* For internal use only. May be removed or replaced in the future.
*
* @param text
* The text the text box should contain.
*/
public void setPromptingOff(String text) ... | 3.68 |
hbase_CloseChecker_isTimeLimit | /**
* Check periodically to see if a system stop is requested every time.
* @return if true, system stop.
*/
public boolean isTimeLimit(Store store, long now) {
if (closeCheckTimeLimit <= 0) {
return false;
}
final long elapsedMillis = now - lastCloseCheckMillis;
if (elapsedMillis <= closeCheckTimeLimit... | 3.68 |
hbase_StorageClusterStatusModel_setDeadNodes | /**
* @param nodes the list of dead node names
*/
public void setDeadNodes(List<String> nodes) {
this.deadNodes = nodes;
} | 3.68 |
hadoop_AbstractS3ACommitter_warnOnActiveUploads | /**
* Scan for active uploads and list them along with a warning message.
* Errors are ignored.
* @param path output path of job.
*/
protected void warnOnActiveUploads(final Path path) {
List<MultipartUpload> pending;
try {
pending = getCommitOperations()
.listPendingUploadsUnderPath(path);
} catc... | 3.68 |
flink_CompactingHashTable_compactPartition | /**
* Compacts (garbage collects) partition with copy-compact strategy using compaction partition
*
* @param partitionNumber partition to compact
* @throws IOException
*/
private void compactPartition(final int partitionNumber) throws IOException {
// do nothing if table was closed, parameter is invalid or no ... | 3.68 |
dubbo_FileSystemDynamicConfiguration_detectPoolingBasedWatchService | /**
* Detect the argument of {@link WatchService} is based on {@linkplain sun.nio.fs.PollingWatchService}
* or not.
* <p>
* Some platforms do not provide the native implementation of {@link WatchService}, just use
* {@linkplain sun.nio.fs.PollingWatchService} in periodic poll file modifications.
*
* @param watch... | 3.68 |
hadoop_EntityTableRW_createTable | /*
* (non-Javadoc)
*
* @see
* org.apache.hadoop.yarn.server.timelineservice.storage.BaseTableRW#
* createTable(org.apache.hadoop.hbase.client.Admin,
* org.apache.hadoop.conf.Configuration)
*/
public void createTable(Admin admin, Configuration hbaseConf)
throws IOException {
TableName table = getTableName(... | 3.68 |
hbase_AbstractFSWAL_getNewPath | /**
* retrieve the next path to use for writing. Increments the internal filenum.
*/
private Path getNewPath() throws IOException {
this.filenum.set(Math.max(getFilenum() + 1, EnvironmentEdgeManager.currentTime()));
Path newPath = getCurrentFileName();
return newPath;
} | 3.68 |
hadoop_ConfigurationUtils_resolve | /**
* Returns a new ConfigurationUtils instance with all inline values resolved.
*
* @return a new ConfigurationUtils instance with all inline values resolved.
*/
public static Configuration resolve(Configuration conf) {
Configuration resolved = new Configuration(false);
for (Map.Entry<String, String> entry : c... | 3.68 |
hmily_MotanServerConfig_baseServiceConfig | /**
* Base service config basic service config bean.
*
* @return the basic service config bean
*/
@Bean
@ConfigurationProperties(prefix = "hmily.motan.server")
public BasicServiceConfigBean baseServiceConfig() {
return new BasicServiceConfigBean();
} | 3.68 |
hadoop_RouterWebServices_createRequestInterceptorChain | /**
* This method creates and returns reference of the first interceptor in the
* chain of request interceptor instances.
*
* @return the reference of the first interceptor in the chain
*/
@VisibleForTesting
protected RESTRequestInterceptor createRequestInterceptorChain() {
return RouterServerUtil.createRequestI... | 3.68 |
flink_SegmentsUtil_bitGet | /**
* read bit from segments.
*
* @param segments target segments.
* @param baseOffset bits base offset.
* @param index bit index from base offset.
*/
public static boolean bitGet(MemorySegment[] segments, int baseOffset, int index) {
int offset = baseOffset + byteIndex(index);
byte current = getByte(segm... | 3.68 |
flink_RocksDBStateBackend_setDbStoragePath | /**
* Sets the path where the RocksDB local database files should be stored on the local file
* system. Setting this path overrides the default behavior, where the files are stored across
* the configured temp directories.
*
* <p>Passing {@code null} to this function restores the default behavior, where the config... | 3.68 |
framework_ColumnVisibilityChangeEvent_isHidden | /**
* Was the column set hidden or visible.
*
* @return <code>true</code> if the column was hidden <code>false</code> if
* it was set visible
*/
public boolean isHidden() {
return hidden;
} | 3.68 |
hadoop_KerberosAuthException_getUser | /** @return The user, or null if not set. */
public String getUser() {
return user;
} | 3.68 |
open-banking-gateway_AccountInformationRequestCommon_fintech_calls_list_accounts_for_max_musterman | // Note that max.musterman is typically used for EMBEDDED (real EMBEDDED that is returned by bank, and not EMBEDDED approach in table)
public SELF fintech_calls_list_accounts_for_max_musterman(String bankProfileId) {
ExtractableResponse<Response> response = withAccountsHeaders(MAX_MUSTERMAN, bankProfileId)
... | 3.68 |
hudi_HoodieTableMetaClient_getTableType | /**
* @return Hoodie Table Type
*/
public HoodieTableType getTableType() {
return tableType;
} | 3.68 |
hbase_TimestampsFilter_getNextCellHint | /**
* Pick the next cell that the scanner should seek to. Since this can skip any number of cells any
* of which can be a delete this can resurect old data. The method will only be used if canHint
* was set to true while creating the filter.
* @throws IOException This will never happen.
*/
@Override
public Cell ge... | 3.68 |
hadoop_BoundedByteArrayOutputStream_getBuffer | /**
* Returns the underlying buffer.
* Data is only valid to {@link #size()}.
* @return the underlying buffer.
*/
public byte[] getBuffer() {
return buffer;
} | 3.68 |
hadoop_MySQLDataDrivenDBRecordReader_executeQuery | // Execute statements for mysql in unbuffered mode.
protected ResultSet executeQuery(String query) throws SQLException {
statement = getConnection().prepareStatement(query,
ResultSet.TYPE_FORWARD_ONLY, ResultSet.CONCUR_READ_ONLY);
statement.setFetchSize(Integer.MIN_VALUE); // MySQL: read row-at-a-time.
return... | 3.68 |
hbase_TableDescriptorBuilder_newBuilder | /**
* Copy all values, families, and name from the input.
* @param desc The desciptor to copy
* @return A clone of input
*/
public static TableDescriptorBuilder newBuilder(final TableDescriptor desc) {
return new TableDescriptorBuilder(desc);
} | 3.68 |
flink_FileMergingSnapshotManagerBase_createLogicalFile | /**
* Create a logical file on a physical file.
*
* @param physicalFile the underlying physical file.
* @param startOffset the offset of the physical file that the logical file start from.
* @param length the length of the logical file.
* @param subtaskKey the id of the subtask that the logical file belongs to.
... | 3.68 |
hbase_TableDescriptorBuilder_toStringTableAttributes | /** Returns map of all table attributes formatted into string. */
public String toStringTableAttributes() {
return getValues(true).toString();
} | 3.68 |
framework_PropertyFilterDefinition_getIgnorePackageNamesStartingWith | /**
* Returns a list of package name prefixes to ignore.
*
* @return list of strings that
*/
public List<String> getIgnorePackageNamesStartingWith() {
return ignorePackageNamesStartingWith;
} | 3.68 |
framework_PropertyFormatter_readOnlyStatusChange | /**
* Listens for changes in the datasource.
*
* This should not be called directly.
*/
@Override
public void readOnlyStatusChange(Property.ReadOnlyStatusChangeEvent event) {
fireReadOnlyStatusChange();
} | 3.68 |
hadoop_IOStatisticsSnapshot_serializer | /**
* Get a JSON serializer for this class.
* @return a serializer.
*/
public static JsonSerialization<IOStatisticsSnapshot> serializer() {
return new JsonSerialization<>(IOStatisticsSnapshot.class, false, true);
} | 3.68 |
pulsar_LinuxBrokerHostUsageImpl_getTotalCpuUsageForEntireHost | /**
* Reads first line of /proc/stat to get total cpu usage.
*
* <pre>
* cpu user nice system idle iowait irq softirq steal guest guest_nice
* cpu 317808 128 58637 2503692 7634 0 13472 0 0 0
* </pre>
*
* Line is split in "words", filtering the first. The sum of all numbers give th... | 3.68 |
hbase_EnvironmentEdgeManager_getDelegate | /**
* Retrieves the singleton instance of the {@link EnvironmentEdge} that is being managed.
* @return the edge.
*/
public static EnvironmentEdge getDelegate() {
return delegate;
} | 3.68 |
hbase_DateTieredCompactionPolicy_selectMinorCompaction | /**
* We receive store files sorted in ascending order by seqId then scan the list of files. If the
* current file has a maxTimestamp older than last known maximum, treat this file as it carries
* the last known maximum. This way both seqId and timestamp are in the same order. If files carry
* the same maxTimestamp... | 3.68 |
hbase_WALProcedureMap_isIncreasing | /**
* @return True if this new procedure is 'richer' than the current one else false and we log this
* incidence where it appears that the WAL has older entries appended after newer ones.
* See HBASE-18152.
*/
private static boolean isIncreasing(ProcedureProtos.Procedure current,
ProcedureProtos.P... | 3.68 |
hbase_HBaseTestingUtility_setupDataTestDir | /**
* Home our data in a dir under {@link #DEFAULT_BASE_TEST_DIRECTORY}. Give it a random name so can
* have many concurrent tests running if we need to. It needs to amend the
* {@link #TEST_DIRECTORY_KEY} System property, as it's what minidfscluster bases it data dir on.
* Moding a System property is not the way t... | 3.68 |
hadoop_NamenodeStatusReport_getNumLiveDatanodes | /**
* Get the number of live blocks.
*
* @return The number of dead nodes.
*/
public int getNumLiveDatanodes() {
return this.liveDatanodes;
} | 3.68 |
framework_JsonPaintTarget_flush | /**
* Method flush.
*/
private void flush() {
uidlBuffer.flush();
} | 3.68 |
hbase_RegionRemoteProcedureBase_persistAndWake | // A bit strange but the procedure store will throw RuntimeException if we can not persist the
// state, so upper layer should take care of this...
private void persistAndWake(MasterProcedureEnv env, RegionStateNode regionNode) {
env.getMasterServices().getMasterProcedureExecutor().getStore().update(this);
regionNo... | 3.68 |
querydsl_PathBuilder_getSimple | /**
* Create a new Simple path
*
* @param <A>
* @param property property name
* @param type property type
* @return property path
*/
@SuppressWarnings("unchecked")
public <A> SimplePath<A> getSimple(String property, Class<A> type) {
Class<? extends A> vtype = validate(property, type);
return super.create... | 3.68 |
hudi_HoodieFlinkCopyOnWriteTable_upsert | /**
* Upsert a batch of new records into Hoodie table at the supplied instantTime.
*
* <p>Specifies the write handle explicitly in order to have fine-grained control with
* the underneath file.
*
* @param context HoodieEngineContext
* @param writeHandle The write handle
* @param instantTime Instant Time for... | 3.68 |
framework_ConnectorTracker_getDirtyVisibleConnectors | /**
* Returns a collection of those {@link #getDirtyConnectors() dirty
* connectors} that are actually visible to the client.
*
* @return A list of dirty and visible connectors.
*/
public ArrayList<ClientConnector> getDirtyVisibleConnectors() {
Collection<ClientConnector> dirtyConnectors = getDirtyConnectors()... | 3.68 |
framework_Buffered_getCause | /**
* Gets the cause of the exception.
*
* @return The (first) cause for the exception, null if no cause.
*/
@Override
public final Throwable getCause() {
if (causes.length == 0) {
return null;
}
return causes[0];
} | 3.68 |
flink_FlinkRexBuilder_toComparable | /** Copied from the {@link RexBuilder} to fix the {@link RexBuilder#makeIn}. */
@SuppressWarnings("rawtypes")
private static <C extends Comparable<C>> C toComparable(Class<C> clazz, RexNode point) {
switch (point.getKind()) {
case LITERAL:
final RexLiteral literal = (RexLiteral) point;
... | 3.68 |
hmily_JavaBeanBinder_getProperties | /**
* Gets properties.
*
* @return the properties
*/
Map<String, BeanProperty> getProperties() {
return this.properties;
} | 3.68 |
hudi_HoodieAdbJdbcClient_getPartitionClause | /**
* Generate Hive Partition from partition values.
*
* @param partition Partition path
* @return partition clause
*/
private String getPartitionClause(String partition) {
List<String> partitionValues = partitionValueExtractor.extractPartitionValuesInPath(partition);
ValidationUtils.checkArgument(config.getSp... | 3.68 |
flink_KvStateService_fromConfiguration | /**
* Creates and returns the KvState service.
*
* @param taskManagerServicesConfiguration task manager configuration
* @return service for kvState related components
*/
public static KvStateService fromConfiguration(
TaskManagerServicesConfiguration taskManagerServicesConfiguration) {
KvStateRegistry ... | 3.68 |
flink_FlinkContainerTestEnvironment_getFlinkContainers | /**
* Get instance of Flink containers for cluster controlling.
*
* @return Flink cluster on Testcontainers
*/
public FlinkContainers getFlinkContainers() {
return this.flinkContainers;
} | 3.68 |
flink_SourceTestSuiteBase_getTestDataSize | /**
* Get the size of test data.
*
* @param collections test data
* @return the size of test data
*/
protected int getTestDataSize(List<List<T>> collections) {
int sumSize = 0;
for (Collection<T> collection : collections) {
sumSize += collection.size();
}
return sumSize;
} | 3.68 |
hbase_WALEntryStream_checkAllBytesParsed | // HBASE-15984 check to see we have in fact parsed all data in a cleanly closed file
private boolean checkAllBytesParsed() {
// -1 means the wal wasn't closed cleanly.
final long trailerSize = currentTrailerSize();
FileStatus stat = null;
try {
stat = getCurrentPathFileStatus();
} catch (IOException e) {
... | 3.68 |
morf_OracleDialect_defaultNullOrder | /**
* {@inheritDoc}
* @see org.alfasoftware.morf.jdbc.SqlDialect#defaultNullOrder()
*/
@Override
protected String defaultNullOrder() {
return DEFAULT_NULL_ORDER;
} | 3.68 |
hadoop_OBSFileSystem_getBucket | /**
* Return the bucket of this filesystem.
*
* @return the bucket
*/
String getBucket() {
return bucket;
} | 3.68 |
hudi_AbstractTableFileSystemView_getAllLogsMergedFileSliceBeforeOrOn | /**
* Stream all "merged" file-slices before on an instant time
* for a MERGE_ON_READ table with index that can index log files(which means it writes pure logs first).
*
* <p>In streaming read scenario, in order for better reading efficiency, the user can choose to skip the
* base files that are produced by compac... | 3.68 |
flink_PendingCheckpoint_discard | /**
* Discard state. Must be called after {@link #dispose(boolean, CheckpointsCleaner,
* Runnable, Executor) dispose}.
*/
@Override
public void discard() {
synchronized (lock) {
if (discarded) {
Preconditions.checkState(
disposed, "Checkpoint should be disposed before bein... | 3.68 |
shardingsphere-elasticjob_JobRegistry_getJobScheduleController | /**
* Get job schedule controller.
*
* @param jobName job name
* @return job schedule controller
*/
public JobScheduleController getJobScheduleController(final String jobName) {
return schedulerMap.get(jobName);
} | 3.68 |
framework_DefaultItemSorter_setSortProperties | /*
* (non-Javadoc)
*
* @see com.vaadin.data.util.ItemSorter#setSortProperties(com.vaadin.data.
* Container .Sortable, java.lang.Object[], boolean[])
*/
@Override
public void setSortProperties(Container.Sortable container,
Object[] propertyId, boolean[] ascending) {
this.container = container;
// R... | 3.68 |
hbase_RegionNormalizerWorkQueue_put | /**
* Inserts the specified element at the tail of the queue, if it's not already present.
* @param e the element to add
*/
public void put(E e) {
if (e == null) {
throw new NullPointerException();
}
lock.writeLock().lock();
try {
delegate.add(e);
if (!delegate.isEmpty()) {
notEmpty.signal(... | 3.68 |
hadoop_StartupProgress_setTotal | /**
* Sets the total associated with the specified phase and step. For example,
* this can be used while loading edits to indicate the number of operations to
* be applied.
*
* @param phase Phase to set
* @param step Step to set
* @param total long to set
*/
public void setTotal(Phase phase, Step step, long t... | 3.68 |
flink_SnapshotDirectory_listDirectory | /**
* List the files in the snapshot directory.
*
* @return the files in the snapshot directory.
* @throws IOException if there is a problem creating the file statuses.
*/
public Path[] listDirectory() throws IOException {
return FileUtils.listDirectory(directory);
} | 3.68 |
flink_MetricGroup_counter | /**
* Registers a {@link org.apache.flink.metrics.Counter} with Flink.
*
* @param name name of the counter
* @param counter counter to register
* @param <C> counter type
* @return the given counter
*/
default <C extends Counter> C counter(int name, C counter) {
return counter(String.valueOf(name), counter);
... | 3.68 |
hudi_CollectionUtils_reduce | /**
* Reduces provided {@link Collection} using provided {@code reducer} applied to
* every element of the collection like following
*
* {@code reduce(reduce(reduce(identity, e1), e2), ...)}
*
* @param c target collection to be reduced
* @param identity element for reducing to start from
* @param reducer actual... | 3.68 |
flink_CrossOperator_projectTuple14 | /**
* Projects a pair of crossed elements to a {@link Tuple} with the previously selected
* fields.
*
* @return The projected data set.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13>
ProjectCross<
I1,
I2,
... | 3.68 |
hadoop_AbfsManifestStoreOperations_storePreservesEtagsThroughRenames | /**
* Etags are preserved through Gen2 stores, but not wasb stores.
* @param path path to probe.
* @return true if this store preserves etags.
*/
@Override
public boolean storePreservesEtagsThroughRenames(final Path path) {
return etagsPreserved;
} | 3.68 |
zxing_UPCEANExtension5Support_parseExtensionString | /**
* @param raw raw content of extension
* @return formatted interpretation of raw content as a {@link Map} mapping
* one {@link ResultMetadataType} to appropriate value, or {@code null} if not known
*/
private static Map<ResultMetadataType,Object> parseExtensionString(String raw) {
if (raw.length() != 5) {
... | 3.68 |
druid_IPAddress_getIPAddress | /**
* Return the integer representation of the IP address.
*
* @return The IP address.
*/
public final int getIPAddress() {
return ipAddress;
} | 3.68 |
flink_ResultPartition_setup | /**
* Registers a buffer pool with this result partition.
*
* <p>There is one pool for each result partition, which is shared by all its sub partitions.
*
* <p>The pool is registered with the partition *after* it as been constructed in order to
* conform to the life-cycle of task registrations in the {@link TaskE... | 3.68 |
hbase_MobFileCache_getAccessCount | /**
* Gets the count of accesses to the mob file cache.
* @return The count of accesses to the mob file cache.
*/
public long getAccessCount() {
return count.get();
} | 3.68 |
hudi_HoodieBackedTableMetadataWriter_updateFunctionalIndexIfPresent | /**
* Update functional index from {@link HoodieCommitMetadata}.
*/
private void updateFunctionalIndexIfPresent(HoodieCommitMetadata commitMetadata, String instantTime, Map<MetadataPartitionType, HoodieData<HoodieRecord>> partitionToRecordMap) {
dataMetaClient.getTableConfig().getMetadataPartitions()
.stream(... | 3.68 |
hbase_RequestConverter_buildModifyColumnRequest | /**
* Create a protocol buffer ModifyColumnRequest
* @return an ModifyColumnRequest
*/
public static ModifyColumnRequest buildModifyColumnRequest(final TableName tableName,
final ColumnFamilyDescriptor column, final long nonceGroup, final long nonce) {
ModifyColumnRequest.Builder builder = ModifyColumnRequest.ne... | 3.68 |
hadoop_ManifestCommitter_isCommitJobRepeatable | /**
* Failure during Job Commit is not recoverable from.
*
* @param jobContext
* Context of the job whose output is being written.
* @return false, always
* @throws IOException never
*/
@Override
public boolean isCommitJobRepeatable(final JobContext jobContext)
throws IOException {
LOG.info("Probe... | 3.68 |
hadoop_WasbTokenRenewer_renew | /**
* Renew the delegation token.
* @param token token to renew.
* @param conf configuration object.
* @return extended expiry time of the token.
* @throws IOException thrown when trying get current user.
* @throws InterruptedException thrown when thread is interrupted
*/
@Override
public long renew(final Token<... | 3.68 |
flink_SingleLogicalSlot_release | /**
* A release of the payload by the {@link AllocatedSlot} triggers a release of the payload of
* the logical slot.
*
* @param cause of the payload release
*/
@Override
public void release(Throwable cause) {
if (STATE_UPDATER.compareAndSet(this, State.ALIVE, State.RELEASING)) {
signalPayloadRelease(ca... | 3.68 |
flink_CatalogManager_createDatabase | /**
* Create a database.
*
* @param catalogName Name of the catalog for database
* @param databaseName Name of the database to be created
* @param database The database definition
* @param ignoreIfExists Flag to specify behavior when a database with the given name already
* exists: if set to false, throw a D... | 3.68 |
hadoop_ContainerAllocationHistory_addAllocationEntry | /**
* Record the allocation history for the container.
*
* @param container to add record for
* @param requestSet resource request ask set
* @param fulfillTimeStamp time at which allocation happened
* @param fulfillLatency time elapsed in allocating since asked
*/
public synchronized void addAllocationEntry(Cont... | 3.68 |
hadoop_Paths_getAppAttemptId | /**
* Get the Application Attempt ID for this job.
* @param conf the config to look in
* @return the Application Attempt ID for a given job.
*/
private static int getAppAttemptId(Configuration conf) {
return conf.getInt(
MRJobConfig.APPLICATION_ATTEMPT_ID, 0);
} | 3.68 |
dubbo_RpcServiceContext_getLocalHost | /**
* get local host.
*
* @return local host
*/
@Override
public String getLocalHost() {
String host = localAddress == null
? null
: localAddress.getAddress() == null
? localAddress.getHostName()
: NetUtils.filterLocalHost(localAddress.getAddress()... | 3.68 |
morf_DatabaseMetaDataProviderUtils_shouldIgnoreIndex | /**
* Indexes which contain the suffix _PRF and a digit are to be ignored:
* this allows performance testing of new index to verify their effect,
* without breaking the schema checking.
*
* eg. Schedule_PRF1
*
* @param indexName The name of an index
* @return Whether it should be ignored
*/
public static boole... | 3.68 |
flink_MiniCluster_isRunning | /** Checks if the mini cluster was started and is running. */
public boolean isRunning() {
return running;
} | 3.68 |
hibernate-validator_ClassLoadingHelper_run | /**
* Runs the given privileged action, using a privileged block if required.
* <p>
* <b>NOTE:</b> This must never be changed into a publicly available method to avoid execution of arbitrary
* privileged actions within HV's protection domain.
*/
@IgnoreForbiddenApisErrors(reason = "SecurityManager is deprecated in... | 3.68 |
pulsar_TransactionBufferProvider_newProvider | /**
* Construct a provider from the provided class.
*
* @param providerClassName the provider class name.
* @return an instance of transaction buffer provider.
*/
static TransactionBufferProvider newProvider(String providerClassName) throws IOException {
try {
TransactionBufferProvider transactionBuffe... | 3.68 |
flink_ManuallyTriggeredScheduledExecutorService_triggerNonPeriodicScheduledTask | /**
* Triggers a single non-periodically scheduled task.
*
* @throws NoSuchElementException If there is no such task.
*/
public void triggerNonPeriodicScheduledTask() {
final ScheduledTask<?> poll = nonPeriodicScheduledTasks.remove();
if (poll != null) {
poll.execute();
}
} | 3.68 |
hadoop_ValidationFailure_verify | /**
* Verify that a condition holds.
* @param expression expression which must be true
* @param message message to raise on a failure
* @param args arguments for the message formatting
* @throws ValidationFailure on a failure
*/
public static void verify(boolean expression,
String message,
Object... args)... | 3.68 |
hbase_CoprocessorHost_checkAndLoadInstance | /**
* @param implClass Implementation class
* @param priority priority
* @param conf configuration
* @throws java.io.IOException Exception
*/
public E checkAndLoadInstance(Class<?> implClass, int priority, Configuration conf)
throws IOException {
// create the instance
C impl;
try {
impl = checkA... | 3.68 |
streampipes_AssetLinkBuilder_withEditingDisabled | /**
* Sets whether editing is disabled for the AssetLink being built.
*
* @param editingDisabled Whether editing is disabled.
* @return The AssetLinkBuilder instance for method chaining.
*/
public AssetLinkBuilder withEditingDisabled(boolean editingDisabled) {
this.assetLink.setEditingDisabled(editingDisabled);
... | 3.68 |
flink_SqlJsonUtils_getNodeFactory | /** Returns the {@link JsonNodeFactory} for creating nodes. */
public static JsonNodeFactory getNodeFactory() {
return MAPPER.getNodeFactory();
} | 3.68 |
pulsar_FieldParser_stringToList | /**
* Converts comma separated string to List.
*
* @param <T>
* type of list
* @param val
* comma separated values.
* @return The converted list with type {@code <T>}.
*/
public static <T> List<T> stringToList(String val, Class<T> type) {
if (val == null) {
return null;
}
... | 3.68 |
framework_FilesystemContainer_equals | /**
* Tests if the given object is the same as the this object. Two
* Properties got from an Item with the same ID are equal.
*
* @param obj
* an object to compare with this object.
* @return <code>true</code> if the given object is the same as this
* object, <code>false</code> if not
*/
@Ove... | 3.68 |
dubbo_URLStrParser_parseDecodedStr | /**
* @param decodedURLStr : after {@link URL#decode} string
* decodedURLStr format: protocol://username:password@host:port/path?k1=v1&k2=v2
* [protocol://][username:password@][host:port]/[path][?k1=v1&k2=v2]
*/
public static URL parseDecodedStr(String decodedURLStr) {
... | 3.68 |
MagicPlugin_Base64Coder_encodeString | /**
* Encodes a string into Base64 format.
* No blanks or line breaks are inserted.
*
* @param s A String to be encoded.
* @return A String containing the Base64 encoded data.
*/
public static String encodeString(String s) {
return new String(encode(s.getBytes(StandardCharsets.UTF_8)));
} | 3.68 |
framework_Notification_setHtmlContentAllowed | /**
* Sets whether html is allowed in the caption and description. If set to
* true, the texts are passed to the browser as html and the developer is
* responsible for ensuring no harmful html is used. If set to false, the
* texts are passed to the browser as plain text.
*
* @param htmlContentAllowed
* ... | 3.68 |
flink_RocksDBIncrementalRestoreOperation_restoreWithRescaling | /**
* Recovery from multi incremental states with rescaling. For rescaling, this method creates a
* temporary RocksDB instance for a key-groups shard. All contents from the temporary instance
* are copied into the real restore instance and then the temporary instance is discarded.
*/
private void restoreWithRescali... | 3.68 |
flink_CopyOnWriteSkipListStateMap_helpGetBytesForKeyAndNamespace | /**
* Returns the byte arrays of serialized key and namespace.
*
* @param node the node.
* @return a tuple of byte arrays of serialized key and namespace
*/
Tuple2<byte[], byte[]> helpGetBytesForKeyAndNamespace(long node) {
Node nodeStorage = getNodeSegmentAndOffset(node);
MemorySegment segment = nodeStora... | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.