name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
hadoop_AuditingIntegration_enterStageWorker
/** * Update the thread context with the stage name and * job ID. * This MUST be invoked at the start of methods invoked in helper threads, * to ensure that they are all annotated with job and stage. * @param jobId job ID. * @param stage stage name. */ public static void enterStageWorker(String jobId, String sta...
3.68
streampipes_IgnoreBlocksAfterContentFilter_getDefaultInstance
/** * Returns the singleton instance for DeleteBlocksAfterContentFilter. */ public static IgnoreBlocksAfterContentFilter getDefaultInstance() { return DEFAULT_INSTANCE; }
3.68
framework_AbsoluteLayout_getTopValue
/** * Gets the 'top' attributes value in current units. * * @see #getTopUnits() * @return The value of the 'top' attribute, null if not set */ public Float getTopValue() { return topValue; }
3.68
AreaShop_FileManager_postUpdateFiles
/** * Checks for old file formats and converts them to the latest format. * This is to be triggered after the load of the region files. */ private void postUpdateFiles() { Integer fileStatus = versions.get(AreaShop.versionFiles); // If the the files are already the current version if(fileStatus != null && fileSt...
3.68
flink_CopyOnWriteStateMap_removeEntry
/** Helper method that is the basis for operations that remove mappings. */ private StateMapEntry<K, N, S> removeEntry(K key, N namespace) { final int hash = computeHashForOperationAndDoIncrementalRehash(key, namespace); final StateMapEntry<K, N, S>[] tab = selectActiveTable(hash); int index = hash & (tab....
3.68
morf_DatabaseSchemaManager_mutateToSupportSchema
/** * Mutates the current database schema so that it supports the one requested. * * <p>When this method returns, it guarantees that all the tables in {code schema} are * present in the database and also empty.</p> * * <p>Note it does not guarantee that no other tables exist.</p> * * @param schema The schema wh...
3.68
dubbo_Converter_getTargetType
/** * Get the target type * * @return non-null */ default Class<T> getTargetType() { return findActualTypeArgument(getClass(), Converter.class, 1); }
3.68
framework_TreeTable_addExpandListener
/** * Adds an expand listener. * * @param listener * the Listener to be added. */ public void addExpandListener(ExpandListener listener) { addListener(ExpandEvent.class, listener, ExpandListener.EXPAND_METHOD); }
3.68
hbase_PrivateCellUtil_getValueAsBigDecimal
/** * Converts the value bytes of the given cell into a BigDecimal * @return value as BigDecimal */ public static BigDecimal getValueAsBigDecimal(Cell cell) { if (cell instanceof ByteBufferExtendedCell) { return ByteBufferUtils.toBigDecimal(((ByteBufferExtendedCell) cell).getValueByteBuffer(), ((ByteBuff...
3.68
framework_ScrollbarBundle_truncate
/** * Truncates a double such that no decimal places are retained. * <p> * E.g. {@code trunc(2.3d) == 2.0d} and {@code trunc(-2.3d) == -2.0d}. * * @param num * the double value to be truncated * @return the {@code num} value without any decimal digits */ private static double truncate(double num) { ...
3.68
hbase_RollingStatCalculator_insertDataValue
/** * Inserts given data value to array of data values to be considered for statistics calculation */ public void insertDataValue(long data) { // if current number of data points already equals rolling period and rolling period is // non-zero then remove one data and update the statistics if (numberOfDataValues...
3.68
framework_VaadinPortletRequest_getPortletPreference
/** * Reads a portlet preference from the portlet of the request. * * @param name * The name of the portlet preference. Cannot be * <code>null</code>. * * @return The value of the portlet preference, <code>null</code> if the * preference is not defined. */ public String getPortlet...
3.68
hadoop_S3ListRequest_v1
/** * Restricted constructors to ensure v1 or v2, not both. * @param request v1 request * @return new list request container */ public static S3ListRequest v1(ListObjectsRequest request) { return new S3ListRequest(request, null); }
3.68
framework_VFilterSelect_setPromptingOff
/** * Turns prompting off. When prompting is turned on a command prompt is * shown in the text box if nothing has been entered. * <p> * For internal use only. May be removed or replaced in the future. * * @param text * The text the text box should contain. */ public void setPromptingOff(String text) ...
3.68
hbase_CloseChecker_isTimeLimit
/** * Check periodically to see if a system stop is requested every time. * @return if true, system stop. */ public boolean isTimeLimit(Store store, long now) { if (closeCheckTimeLimit <= 0) { return false; } final long elapsedMillis = now - lastCloseCheckMillis; if (elapsedMillis <= closeCheckTimeLimit...
3.68
hbase_StorageClusterStatusModel_setDeadNodes
/** * @param nodes the list of dead node names */ public void setDeadNodes(List<String> nodes) { this.deadNodes = nodes; }
3.68
hadoop_AbstractS3ACommitter_warnOnActiveUploads
/** * Scan for active uploads and list them along with a warning message. * Errors are ignored. * @param path output path of job. */ protected void warnOnActiveUploads(final Path path) { List<MultipartUpload> pending; try { pending = getCommitOperations() .listPendingUploadsUnderPath(path); } catc...
3.68
flink_CompactingHashTable_compactPartition
/** * Compacts (garbage collects) partition with copy-compact strategy using compaction partition * * @param partitionNumber partition to compact * @throws IOException */ private void compactPartition(final int partitionNumber) throws IOException { // do nothing if table was closed, parameter is invalid or no ...
3.68
dubbo_FileSystemDynamicConfiguration_detectPoolingBasedWatchService
/** * Detect the argument of {@link WatchService} is based on {@linkplain sun.nio.fs.PollingWatchService} * or not. * <p> * Some platforms do not provide the native implementation of {@link WatchService}, just use * {@linkplain sun.nio.fs.PollingWatchService} in periodic poll file modifications. * * @param watch...
3.68
hadoop_EntityTableRW_createTable
/* * (non-Javadoc) * * @see * org.apache.hadoop.yarn.server.timelineservice.storage.BaseTableRW# * createTable(org.apache.hadoop.hbase.client.Admin, * org.apache.hadoop.conf.Configuration) */ public void createTable(Admin admin, Configuration hbaseConf) throws IOException { TableName table = getTableName(...
3.68
hbase_AbstractFSWAL_getNewPath
/** * retrieve the next path to use for writing. Increments the internal filenum. */ private Path getNewPath() throws IOException { this.filenum.set(Math.max(getFilenum() + 1, EnvironmentEdgeManager.currentTime())); Path newPath = getCurrentFileName(); return newPath; }
3.68
hadoop_ConfigurationUtils_resolve
/** * Returns a new ConfigurationUtils instance with all inline values resolved. * * @return a new ConfigurationUtils instance with all inline values resolved. */ public static Configuration resolve(Configuration conf) { Configuration resolved = new Configuration(false); for (Map.Entry<String, String> entry : c...
3.68
hmily_MotanServerConfig_baseServiceConfig
/** * Base service config basic service config bean. * * @return the basic service config bean */ @Bean @ConfigurationProperties(prefix = "hmily.motan.server") public BasicServiceConfigBean baseServiceConfig() { return new BasicServiceConfigBean(); }
3.68
hadoop_RouterWebServices_createRequestInterceptorChain
/** * This method creates and returns reference of the first interceptor in the * chain of request interceptor instances. * * @return the reference of the first interceptor in the chain */ @VisibleForTesting protected RESTRequestInterceptor createRequestInterceptorChain() { return RouterServerUtil.createRequestI...
3.68
flink_SegmentsUtil_bitGet
/** * read bit from segments. * * @param segments target segments. * @param baseOffset bits base offset. * @param index bit index from base offset. */ public static boolean bitGet(MemorySegment[] segments, int baseOffset, int index) { int offset = baseOffset + byteIndex(index); byte current = getByte(segm...
3.68
flink_RocksDBStateBackend_setDbStoragePath
/** * Sets the path where the RocksDB local database files should be stored on the local file * system. Setting this path overrides the default behavior, where the files are stored across * the configured temp directories. * * <p>Passing {@code null} to this function restores the default behavior, where the config...
3.68
framework_ColumnVisibilityChangeEvent_isHidden
/** * Was the column set hidden or visible. * * @return <code>true</code> if the column was hidden <code>false</code> if * it was set visible */ public boolean isHidden() { return hidden; }
3.68
hadoop_KerberosAuthException_getUser
/** @return The user, or null if not set. */ public String getUser() { return user; }
3.68
open-banking-gateway_AccountInformationRequestCommon_fintech_calls_list_accounts_for_max_musterman
// Note that max.musterman is typically used for EMBEDDED (real EMBEDDED that is returned by bank, and not EMBEDDED approach in table) public SELF fintech_calls_list_accounts_for_max_musterman(String bankProfileId) { ExtractableResponse<Response> response = withAccountsHeaders(MAX_MUSTERMAN, bankProfileId) ...
3.68
hudi_HoodieTableMetaClient_getTableType
/** * @return Hoodie Table Type */ public HoodieTableType getTableType() { return tableType; }
3.68
hbase_TimestampsFilter_getNextCellHint
/** * Pick the next cell that the scanner should seek to. Since this can skip any number of cells any * of which can be a delete this can resurect old data. The method will only be used if canHint * was set to true while creating the filter. * @throws IOException This will never happen. */ @Override public Cell ge...
3.68
hadoop_BoundedByteArrayOutputStream_getBuffer
/** * Returns the underlying buffer. * Data is only valid to {@link #size()}. * @return the underlying buffer. */ public byte[] getBuffer() { return buffer; }
3.68
hadoop_MySQLDataDrivenDBRecordReader_executeQuery
// Execute statements for mysql in unbuffered mode. protected ResultSet executeQuery(String query) throws SQLException { statement = getConnection().prepareStatement(query, ResultSet.TYPE_FORWARD_ONLY, ResultSet.CONCUR_READ_ONLY); statement.setFetchSize(Integer.MIN_VALUE); // MySQL: read row-at-a-time. return...
3.68
hbase_TableDescriptorBuilder_newBuilder
/** * Copy all values, families, and name from the input. * @param desc The desciptor to copy * @return A clone of input */ public static TableDescriptorBuilder newBuilder(final TableDescriptor desc) { return new TableDescriptorBuilder(desc); }
3.68
flink_FileMergingSnapshotManagerBase_createLogicalFile
/** * Create a logical file on a physical file. * * @param physicalFile the underlying physical file. * @param startOffset the offset of the physical file that the logical file start from. * @param length the length of the logical file. * @param subtaskKey the id of the subtask that the logical file belongs to. ...
3.68
hbase_TableDescriptorBuilder_toStringTableAttributes
/** Returns map of all table attributes formatted into string. */ public String toStringTableAttributes() { return getValues(true).toString(); }
3.68
framework_PropertyFilterDefinition_getIgnorePackageNamesStartingWith
/** * Returns a list of package name prefixes to ignore. * * @return list of strings that */ public List<String> getIgnorePackageNamesStartingWith() { return ignorePackageNamesStartingWith; }
3.68
framework_PropertyFormatter_readOnlyStatusChange
/** * Listens for changes in the datasource. * * This should not be called directly. */ @Override public void readOnlyStatusChange(Property.ReadOnlyStatusChangeEvent event) { fireReadOnlyStatusChange(); }
3.68
hadoop_IOStatisticsSnapshot_serializer
/** * Get a JSON serializer for this class. * @return a serializer. */ public static JsonSerialization<IOStatisticsSnapshot> serializer() { return new JsonSerialization<>(IOStatisticsSnapshot.class, false, true); }
3.68
pulsar_LinuxBrokerHostUsageImpl_getTotalCpuUsageForEntireHost
/** * Reads first line of /proc/stat to get total cpu usage. * * <pre> * cpu user nice system idle iowait irq softirq steal guest guest_nice * cpu 317808 128 58637 2503692 7634 0 13472 0 0 0 * </pre> * * Line is split in "words", filtering the first. The sum of all numbers give th...
3.68
hbase_EnvironmentEdgeManager_getDelegate
/** * Retrieves the singleton instance of the {@link EnvironmentEdge} that is being managed. * @return the edge. */ public static EnvironmentEdge getDelegate() { return delegate; }
3.68
hbase_DateTieredCompactionPolicy_selectMinorCompaction
/** * We receive store files sorted in ascending order by seqId then scan the list of files. If the * current file has a maxTimestamp older than last known maximum, treat this file as it carries * the last known maximum. This way both seqId and timestamp are in the same order. If files carry * the same maxTimestamp...
3.68
hbase_WALProcedureMap_isIncreasing
/** * @return True if this new procedure is 'richer' than the current one else false and we log this * incidence where it appears that the WAL has older entries appended after newer ones. * See HBASE-18152. */ private static boolean isIncreasing(ProcedureProtos.Procedure current, ProcedureProtos.P...
3.68
hbase_HBaseTestingUtility_setupDataTestDir
/** * Home our data in a dir under {@link #DEFAULT_BASE_TEST_DIRECTORY}. Give it a random name so can * have many concurrent tests running if we need to. It needs to amend the * {@link #TEST_DIRECTORY_KEY} System property, as it's what minidfscluster bases it data dir on. * Moding a System property is not the way t...
3.68
hadoop_NamenodeStatusReport_getNumLiveDatanodes
/** * Get the number of live blocks. * * @return The number of dead nodes. */ public int getNumLiveDatanodes() { return this.liveDatanodes; }
3.68
framework_JsonPaintTarget_flush
/** * Method flush. */ private void flush() { uidlBuffer.flush(); }
3.68
hbase_RegionRemoteProcedureBase_persistAndWake
// A bit strange but the procedure store will throw RuntimeException if we can not persist the // state, so upper layer should take care of this... private void persistAndWake(MasterProcedureEnv env, RegionStateNode regionNode) { env.getMasterServices().getMasterProcedureExecutor().getStore().update(this); regionNo...
3.68
querydsl_PathBuilder_getSimple
/** * Create a new Simple path * * @param <A> * @param property property name * @param type property type * @return property path */ @SuppressWarnings("unchecked") public <A> SimplePath<A> getSimple(String property, Class<A> type) { Class<? extends A> vtype = validate(property, type); return super.create...
3.68
hudi_HoodieFlinkCopyOnWriteTable_upsert
/** * Upsert a batch of new records into Hoodie table at the supplied instantTime. * * <p>Specifies the write handle explicitly in order to have fine-grained control with * the underneath file. * * @param context HoodieEngineContext * @param writeHandle The write handle * @param instantTime Instant Time for...
3.68
framework_ConnectorTracker_getDirtyVisibleConnectors
/** * Returns a collection of those {@link #getDirtyConnectors() dirty * connectors} that are actually visible to the client. * * @return A list of dirty and visible connectors. */ public ArrayList<ClientConnector> getDirtyVisibleConnectors() { Collection<ClientConnector> dirtyConnectors = getDirtyConnectors()...
3.68
framework_Buffered_getCause
/** * Gets the cause of the exception. * * @return The (first) cause for the exception, null if no cause. */ @Override public final Throwable getCause() { if (causes.length == 0) { return null; } return causes[0]; }
3.68
flink_FlinkRexBuilder_toComparable
/** Copied from the {@link RexBuilder} to fix the {@link RexBuilder#makeIn}. */ @SuppressWarnings("rawtypes") private static <C extends Comparable<C>> C toComparable(Class<C> clazz, RexNode point) { switch (point.getKind()) { case LITERAL: final RexLiteral literal = (RexLiteral) point; ...
3.68
hmily_JavaBeanBinder_getProperties
/** * Gets properties. * * @return the properties */ Map<String, BeanProperty> getProperties() { return this.properties; }
3.68
hudi_HoodieAdbJdbcClient_getPartitionClause
/** * Generate Hive Partition from partition values. * * @param partition Partition path * @return partition clause */ private String getPartitionClause(String partition) { List<String> partitionValues = partitionValueExtractor.extractPartitionValuesInPath(partition); ValidationUtils.checkArgument(config.getSp...
3.68
flink_KvStateService_fromConfiguration
/** * Creates and returns the KvState service. * * @param taskManagerServicesConfiguration task manager configuration * @return service for kvState related components */ public static KvStateService fromConfiguration( TaskManagerServicesConfiguration taskManagerServicesConfiguration) { KvStateRegistry ...
3.68
flink_FlinkContainerTestEnvironment_getFlinkContainers
/** * Get instance of Flink containers for cluster controlling. * * @return Flink cluster on Testcontainers */ public FlinkContainers getFlinkContainers() { return this.flinkContainers; }
3.68
flink_SourceTestSuiteBase_getTestDataSize
/** * Get the size of test data. * * @param collections test data * @return the size of test data */ protected int getTestDataSize(List<List<T>> collections) { int sumSize = 0; for (Collection<T> collection : collections) { sumSize += collection.size(); } return sumSize; }
3.68
hbase_WALEntryStream_checkAllBytesParsed
// HBASE-15984 check to see we have in fact parsed all data in a cleanly closed file private boolean checkAllBytesParsed() { // -1 means the wal wasn't closed cleanly. final long trailerSize = currentTrailerSize(); FileStatus stat = null; try { stat = getCurrentPathFileStatus(); } catch (IOException e) { ...
3.68
morf_OracleDialect_defaultNullOrder
/** * {@inheritDoc} * @see org.alfasoftware.morf.jdbc.SqlDialect#defaultNullOrder() */ @Override protected String defaultNullOrder() { return DEFAULT_NULL_ORDER; }
3.68
hadoop_OBSFileSystem_getBucket
/** * Return the bucket of this filesystem. * * @return the bucket */ String getBucket() { return bucket; }
3.68
hudi_AbstractTableFileSystemView_getAllLogsMergedFileSliceBeforeOrOn
/** * Stream all "merged" file-slices before on an instant time * for a MERGE_ON_READ table with index that can index log files(which means it writes pure logs first). * * <p>In streaming read scenario, in order for better reading efficiency, the user can choose to skip the * base files that are produced by compac...
3.68
flink_PendingCheckpoint_discard
/** * Discard state. Must be called after {@link #dispose(boolean, CheckpointsCleaner, * Runnable, Executor) dispose}. */ @Override public void discard() { synchronized (lock) { if (discarded) { Preconditions.checkState( disposed, "Checkpoint should be disposed before bein...
3.68
shardingsphere-elasticjob_JobRegistry_getJobScheduleController
/** * Get job schedule controller. * * @param jobName job name * @return job schedule controller */ public JobScheduleController getJobScheduleController(final String jobName) { return schedulerMap.get(jobName); }
3.68
framework_DefaultItemSorter_setSortProperties
/* * (non-Javadoc) * * @see com.vaadin.data.util.ItemSorter#setSortProperties(com.vaadin.data. * Container .Sortable, java.lang.Object[], boolean[]) */ @Override public void setSortProperties(Container.Sortable container, Object[] propertyId, boolean[] ascending) { this.container = container; // R...
3.68
hbase_RegionNormalizerWorkQueue_put
/** * Inserts the specified element at the tail of the queue, if it's not already present. * @param e the element to add */ public void put(E e) { if (e == null) { throw new NullPointerException(); } lock.writeLock().lock(); try { delegate.add(e); if (!delegate.isEmpty()) { notEmpty.signal(...
3.68
hadoop_StartupProgress_setTotal
/** * Sets the total associated with the specified phase and step. For example, * this can be used while loading edits to indicate the number of operations to * be applied. * * @param phase Phase to set * @param step Step to set * @param total long to set */ public void setTotal(Phase phase, Step step, long t...
3.68
flink_SnapshotDirectory_listDirectory
/** * List the files in the snapshot directory. * * @return the files in the snapshot directory. * @throws IOException if there is a problem creating the file statuses. */ public Path[] listDirectory() throws IOException { return FileUtils.listDirectory(directory); }
3.68
flink_MetricGroup_counter
/** * Registers a {@link org.apache.flink.metrics.Counter} with Flink. * * @param name name of the counter * @param counter counter to register * @param <C> counter type * @return the given counter */ default <C extends Counter> C counter(int name, C counter) { return counter(String.valueOf(name), counter); ...
3.68
hudi_CollectionUtils_reduce
/** * Reduces provided {@link Collection} using provided {@code reducer} applied to * every element of the collection like following * * {@code reduce(reduce(reduce(identity, e1), e2), ...)} * * @param c target collection to be reduced * @param identity element for reducing to start from * @param reducer actual...
3.68
flink_CrossOperator_projectTuple14
/** * Projects a pair of crossed elements to a {@link Tuple} with the previously selected * fields. * * @return The projected data set. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13> ProjectCross< I1, I2, ...
3.68
hadoop_AbfsManifestStoreOperations_storePreservesEtagsThroughRenames
/** * Etags are preserved through Gen2 stores, but not wasb stores. * @param path path to probe. * @return true if this store preserves etags. */ @Override public boolean storePreservesEtagsThroughRenames(final Path path) { return etagsPreserved; }
3.68
zxing_UPCEANExtension5Support_parseExtensionString
/** * @param raw raw content of extension * @return formatted interpretation of raw content as a {@link Map} mapping * one {@link ResultMetadataType} to appropriate value, or {@code null} if not known */ private static Map<ResultMetadataType,Object> parseExtensionString(String raw) { if (raw.length() != 5) { ...
3.68
druid_IPAddress_getIPAddress
/** * Return the integer representation of the IP address. * * @return The IP address. */ public final int getIPAddress() { return ipAddress; }
3.68
flink_ResultPartition_setup
/** * Registers a buffer pool with this result partition. * * <p>There is one pool for each result partition, which is shared by all its sub partitions. * * <p>The pool is registered with the partition *after* it as been constructed in order to * conform to the life-cycle of task registrations in the {@link TaskE...
3.68
hbase_MobFileCache_getAccessCount
/** * Gets the count of accesses to the mob file cache. * @return The count of accesses to the mob file cache. */ public long getAccessCount() { return count.get(); }
3.68
hudi_HoodieBackedTableMetadataWriter_updateFunctionalIndexIfPresent
/** * Update functional index from {@link HoodieCommitMetadata}. */ private void updateFunctionalIndexIfPresent(HoodieCommitMetadata commitMetadata, String instantTime, Map<MetadataPartitionType, HoodieData<HoodieRecord>> partitionToRecordMap) { dataMetaClient.getTableConfig().getMetadataPartitions() .stream(...
3.68
hbase_RequestConverter_buildModifyColumnRequest
/** * Create a protocol buffer ModifyColumnRequest * @return an ModifyColumnRequest */ public static ModifyColumnRequest buildModifyColumnRequest(final TableName tableName, final ColumnFamilyDescriptor column, final long nonceGroup, final long nonce) { ModifyColumnRequest.Builder builder = ModifyColumnRequest.ne...
3.68
hadoop_ManifestCommitter_isCommitJobRepeatable
/** * Failure during Job Commit is not recoverable from. * * @param jobContext * Context of the job whose output is being written. * @return false, always * @throws IOException never */ @Override public boolean isCommitJobRepeatable(final JobContext jobContext) throws IOException { LOG.info("Probe...
3.68
hadoop_WasbTokenRenewer_renew
/** * Renew the delegation token. * @param token token to renew. * @param conf configuration object. * @return extended expiry time of the token. * @throws IOException thrown when trying get current user. * @throws InterruptedException thrown when thread is interrupted */ @Override public long renew(final Token<...
3.68
flink_SingleLogicalSlot_release
/** * A release of the payload by the {@link AllocatedSlot} triggers a release of the payload of * the logical slot. * * @param cause of the payload release */ @Override public void release(Throwable cause) { if (STATE_UPDATER.compareAndSet(this, State.ALIVE, State.RELEASING)) { signalPayloadRelease(ca...
3.68
flink_CatalogManager_createDatabase
/** * Create a database. * * @param catalogName Name of the catalog for database * @param databaseName Name of the database to be created * @param database The database definition * @param ignoreIfExists Flag to specify behavior when a database with the given name already * exists: if set to false, throw a D...
3.68
hadoop_ContainerAllocationHistory_addAllocationEntry
/** * Record the allocation history for the container. * * @param container to add record for * @param requestSet resource request ask set * @param fulfillTimeStamp time at which allocation happened * @param fulfillLatency time elapsed in allocating since asked */ public synchronized void addAllocationEntry(Cont...
3.68
hadoop_Paths_getAppAttemptId
/** * Get the Application Attempt ID for this job. * @param conf the config to look in * @return the Application Attempt ID for a given job. */ private static int getAppAttemptId(Configuration conf) { return conf.getInt( MRJobConfig.APPLICATION_ATTEMPT_ID, 0); }
3.68
dubbo_RpcServiceContext_getLocalHost
/** * get local host. * * @return local host */ @Override public String getLocalHost() { String host = localAddress == null ? null : localAddress.getAddress() == null ? localAddress.getHostName() : NetUtils.filterLocalHost(localAddress.getAddress()...
3.68
morf_DatabaseMetaDataProviderUtils_shouldIgnoreIndex
/** * Indexes which contain the suffix _PRF and a digit are to be ignored: * this allows performance testing of new index to verify their effect, * without breaking the schema checking. * * eg. Schedule_PRF1 * * @param indexName The name of an index * @return Whether it should be ignored */ public static boole...
3.68
flink_MiniCluster_isRunning
/** Checks if the mini cluster was started and is running. */ public boolean isRunning() { return running; }
3.68
hibernate-validator_ClassLoadingHelper_run
/** * Runs the given privileged action, using a privileged block if required. * <p> * <b>NOTE:</b> This must never be changed into a publicly available method to avoid execution of arbitrary * privileged actions within HV's protection domain. */ @IgnoreForbiddenApisErrors(reason = "SecurityManager is deprecated in...
3.68
pulsar_TransactionBufferProvider_newProvider
/** * Construct a provider from the provided class. * * @param providerClassName the provider class name. * @return an instance of transaction buffer provider. */ static TransactionBufferProvider newProvider(String providerClassName) throws IOException { try { TransactionBufferProvider transactionBuffe...
3.68
flink_ManuallyTriggeredScheduledExecutorService_triggerNonPeriodicScheduledTask
/** * Triggers a single non-periodically scheduled task. * * @throws NoSuchElementException If there is no such task. */ public void triggerNonPeriodicScheduledTask() { final ScheduledTask<?> poll = nonPeriodicScheduledTasks.remove(); if (poll != null) { poll.execute(); } }
3.68
hadoop_ValidationFailure_verify
/** * Verify that a condition holds. * @param expression expression which must be true * @param message message to raise on a failure * @param args arguments for the message formatting * @throws ValidationFailure on a failure */ public static void verify(boolean expression, String message, Object... args)...
3.68
hbase_CoprocessorHost_checkAndLoadInstance
/** * @param implClass Implementation class * @param priority priority * @param conf configuration * @throws java.io.IOException Exception */ public E checkAndLoadInstance(Class<?> implClass, int priority, Configuration conf) throws IOException { // create the instance C impl; try { impl = checkA...
3.68
streampipes_AssetLinkBuilder_withEditingDisabled
/** * Sets whether editing is disabled for the AssetLink being built. * * @param editingDisabled Whether editing is disabled. * @return The AssetLinkBuilder instance for method chaining. */ public AssetLinkBuilder withEditingDisabled(boolean editingDisabled) { this.assetLink.setEditingDisabled(editingDisabled); ...
3.68
flink_SqlJsonUtils_getNodeFactory
/** Returns the {@link JsonNodeFactory} for creating nodes. */ public static JsonNodeFactory getNodeFactory() { return MAPPER.getNodeFactory(); }
3.68
pulsar_FieldParser_stringToList
/** * Converts comma separated string to List. * * @param <T> * type of list * @param val * comma separated values. * @return The converted list with type {@code <T>}. */ public static <T> List<T> stringToList(String val, Class<T> type) { if (val == null) { return null; } ...
3.68
framework_FilesystemContainer_equals
/** * Tests if the given object is the same as the this object. Two * Properties got from an Item with the same ID are equal. * * @param obj * an object to compare with this object. * @return <code>true</code> if the given object is the same as this * object, <code>false</code> if not */ @Ove...
3.68
dubbo_URLStrParser_parseDecodedStr
/** * @param decodedURLStr : after {@link URL#decode} string * decodedURLStr format: protocol://username:password@host:port/path?k1=v1&k2=v2 * [protocol://][username:password@][host:port]/[path][?k1=v1&k2=v2] */ public static URL parseDecodedStr(String decodedURLStr) { ...
3.68
MagicPlugin_Base64Coder_encodeString
/** * Encodes a string into Base64 format. * No blanks or line breaks are inserted. * * @param s A String to be encoded. * @return A String containing the Base64 encoded data. */ public static String encodeString(String s) { return new String(encode(s.getBytes(StandardCharsets.UTF_8))); }
3.68
framework_Notification_setHtmlContentAllowed
/** * Sets whether html is allowed in the caption and description. If set to * true, the texts are passed to the browser as html and the developer is * responsible for ensuring no harmful html is used. If set to false, the * texts are passed to the browser as plain text. * * @param htmlContentAllowed * ...
3.68
flink_RocksDBIncrementalRestoreOperation_restoreWithRescaling
/** * Recovery from multi incremental states with rescaling. For rescaling, this method creates a * temporary RocksDB instance for a key-groups shard. All contents from the temporary instance * are copied into the real restore instance and then the temporary instance is discarded. */ private void restoreWithRescali...
3.68
flink_CopyOnWriteSkipListStateMap_helpGetBytesForKeyAndNamespace
/** * Returns the byte arrays of serialized key and namespace. * * @param node the node. * @return a tuple of byte arrays of serialized key and namespace */ Tuple2<byte[], byte[]> helpGetBytesForKeyAndNamespace(long node) { Node nodeStorage = getNodeSegmentAndOffset(node); MemorySegment segment = nodeStora...
3.68