name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
flink_HiveParserSqlFunctionConverter_getName
// TODO: this is not valid. Function names for built-in UDFs are specified in // FunctionRegistry, and only happen to match annotations. For user UDFs, the // name is what user specifies at creation time (annotation can be absent, // different, or duplicate some other function). private static String getName(GenericUDF...
3.68
hbase_TableDescriptorBuilder_getValues
/** * Getter for fetching an unmodifiable {@link #values} map. * @return unmodifiable map {@link #values}. * @see #values */ @Override public Map<Bytes, Bytes> getValues() { // shallow pointer copy return Collections.unmodifiableMap(values); }
3.68
druid_SQLMethodInvokeExpr_getParameters
/** * instead of getArguments * * @deprecated */ public List<SQLExpr> getParameters() { return this.arguments; }
3.68
framework_AbstractComponentConnector_onDragSourceDetached
/** * Invoked when a {@link DragSourceExtensionConnector} has been removed from * this component. * <p> * By default, does nothing. * <p> * This is a framework internal method, and should not be invoked manually. * * @since 8.1 * @see #onDragSourceAttached() */ public void onDragSourceDetached() { }
3.68
flink_CheckpointConfig_setTolerableCheckpointFailureNumber
/** * This defines how many consecutive checkpoint failures will be tolerated, before the whole job * is failed over. The default value is `0`, which means no checkpoint failures will be * tolerated, and the job will fail on first reported checkpoint failure. */ public void setTolerableCheckpointFailureNumber(int t...
3.68
flink_FlinkContainersSettings_build
/** * Returns a {@code FlinkContainersConfig} built from the parameters previously set. * * @return A {@code FlinkContainersConfig} built with parameters of this {@code * FlinkContainersConfig.Builder}. */ public FlinkContainersSettings build() { return new FlinkContainersSettings(this); }
3.68
hadoop_HdfsConfiguration_init
/** * This method is here so that when invoked, HdfsConfiguration is class-loaded * if it hasn't already been previously loaded. Upon loading the class, the * static initializer block above will be executed to add the deprecated keys * and to add the default resources. It is safe for this method to be called * mu...
3.68
framework_SharedUtil_propertyIdToHumanFriendly
/** * Converts a property id to a human friendly format. Handles nested * properties by only considering the last part, e.g. "address.streetName" * is equal to "streetName" for this method. * * @since 7.4 * @param propertyId * The propertyId to format * @return A human friendly version of the propert...
3.68
flink_JavaFieldPredicates_isStatic
/** * Match the static modifier of the {@link JavaField}. * * @return A {@link DescribedPredicate} returning true, if and only if the tested {@link * JavaField} has the static modifier. */ public static DescribedPredicate<JavaField> isStatic() { return DescribedPredicate.describe( "static", fie...
3.68
framework_ColorPickerPopup_setHSVTabVisible
/** * Sets the HSV tab visibility. * * @param visible * The visibility of the HSV tab */ public void setHSVTabVisible(boolean visible) { if (visible && !isTabVisible(hsvTab)) { tabs.addTab(hsvTab, "HSV", null); checkIfTabsNeeded(); } else if (!visible && isTabVisible(hsvTab)) { ...
3.68
hadoop_AbstractS3ACommitter_getDestinationFS
/** * Get the destination filesystem from the output path and the configuration. * @param out output path * @param config job/task config * @return the associated FS * @throws PathCommitException output path isn't to an S3A FS instance. * @throws IOException failure to instantiate the FS. */ protected FileSystem...
3.68
flink_ArchivedExecutionGraph_createSparseArchivedExecutionGraph
/** * Create a sparse ArchivedExecutionGraph for a job. Most fields will be empty, only job status * and error-related fields are set. */ public static ArchivedExecutionGraph createSparseArchivedExecutionGraph( JobID jobId, String jobName, JobStatus jobStatus, @Nullable Throwable thro...
3.68
flink_DataStream_printToErr
/** * Writes a DataStream to the standard error stream (stderr). * * <p>For each element of the DataStream the result of {@link Object#toString()} is written. * * <p>NOTE: This will print to stderr on the machine where the code is executed, i.e. the Flink * worker. * * @param sinkIdentifier The string to prefix...
3.68
hadoop_ActiveAuditManagerS3A_prune
/** * Prune all null weak references, calling the referenceLost * callback for each one. * * non-atomic and non-blocking. * @return the number of entries pruned. */ @VisibleForTesting int prune() { return activeSpanMap.prune(); }
3.68
flink_SplitEnumeratorContext_registeredReadersOfAttempts
/** * Get the currently registered readers of all the subtask attempts. The mapping is from subtask * id to a map which maps an attempt to its reader info. * * @return the currently registered readers. */ default Map<Integer, Map<Integer, ReaderInfo>> registeredReadersOfAttempts() { throw new UnsupportedOperat...
3.68
rocketmq-connect_ClusterManagementServiceImpl_prepare
/** * Preparation before startup * * @param connectConfig */ private void prepare(WorkerConfig connectConfig) { String consumerGroup = this.defaultMQPullConsumer.getConsumerGroup(); Set<String> consumerGroupSet = ConnectUtil.fetchAllConsumerGroupList(connectConfig); if (!consumerGroupSet.contains(consum...
3.68
framework_Table_setColumnHeader
/** * Sets the column header for the specified column. * * @param propertyId * the propertyId identifying the column. * @param header * the header to set. */ public void setColumnHeader(Object propertyId, String header) { if (header == null) { columnHeaders.remove(propertyId); ...
3.68
framework_VaadinService_getCurrentRequest
/** * Gets the currently processed Vaadin request. The current request is * automatically defined when the request is started. The current request * can not be used in e.g. background threads because of the way server * implementations reuse request instances. * * @return the current Vaadin request instance if av...
3.68
framework_ExternalResource_setMIMEType
/** * Sets the MIME type of the resource. */ public void setMIMEType(String mimeType) { this.mimeType = mimeType; }
3.68
hadoop_PartitionResourcesInfo_getUserAmLimit
/** * @return the userAmLimit */ public ResourceInfo getUserAmLimit() { return userAmLimit; }
3.68
hadoop_AMRMClientAsyncImpl_serviceStop
/** * Tells the heartbeat and handler threads to stop and waits for them to * terminate. */ @Override protected void serviceStop() throws Exception { keepRunning = false; heartbeatThread.interrupt(); try { heartbeatThread.join(); } catch (InterruptedException ex) { LOG.error("Error joining with heart...
3.68
morf_AbstractSqlDialectTest_expectedLeftTrim
/** * @return The expected SQL for a Left Trim */ protected String expectedLeftTrim() { return "SELECT LTRIM(field1) FROM " + tableName("schedule"); }
3.68
hudi_FlinkOptions_getPropertiesWithPrefix
/** * Collects the config options that start with specified prefix {@code prefix} into a 'key'='value' list. */ public static Map<String, String> getPropertiesWithPrefix(Map<String, String> options, String prefix) { final Map<String, String> hoodieProperties = new HashMap<>(); if (hasPropertyOptions(options, pref...
3.68
flink_CheckpointConfig_enableApproximateLocalRecovery
/** * Enables the approximate local recovery mode. * * <p>In this recovery mode, when a task fails, the entire downstream of the tasks (including * the failed task) restart. * * <p>Notice that 1. Approximate recovery may lead to data loss. The amount of data which leads * the failed task from the state of the la...
3.68
pulsar_FunctionRuntimeManager_getFunctionStats
/** * Get stats of all function instances. * * @param tenant the tenant the function belongs to * @param namespace the namespace the function belongs to * @param functionName the function name * @return a list of function statuses * @throws PulsarAdminException */ public FunctionStatsImpl getFunctionSt...
3.68
hbase_MetaTableAccessor_updateLocation
/** * Updates the location of the specified region to be the specified server. * <p> * Connects to the specified server which should be hosting the specified catalog region name to * perform the edit. * @param connection connection we're using * @param regionInfo region to update location of * @param...
3.68
druid_SQLCommitStatement_getChain
// mysql public Boolean getChain() { return chain; }
3.68
flink_ProcessorArchitecture_getAddressSize
/** Gets the address size of the memory (32 bit, 64 bit). */ public MemoryAddressSize getAddressSize() { return addressSize; }
3.68
dubbo_StringUtils_decodeHexByte
/** * Decode a 2-digit hex byte from within a string. */ public static byte decodeHexByte(CharSequence s, int pos) { int hi = decodeHexNibble(s.charAt(pos)); int lo = decodeHexNibble(s.charAt(pos + 1)); if (hi == -1 || lo == -1) { throw new IllegalArgumentException( String.format("...
3.68
hbase_ZKUtil_getPath
/** Returns path to znode where the ZKOp will occur */ public String getPath() { return path; }
3.68
hbase_AdvancedScanResultConsumer_onHeartbeat
/** * Indicate that there is a heartbeat message but we have not cumulated enough cells to call * {@link #onNext(Result[], ScanController)}. * <p> * Note that this method will always be called when RS returns something to us but we do not have * enough cells to call {@link #onNext(Result[], ScanController)}. Somet...
3.68
hadoop_AssumedRoleCredentialProvider_buildSessionName
/** * Build the session name from the current user's shortname. * @return a string for the session name. * @throws IOException failure to get the current user */ static String buildSessionName() throws IOException { return sanitize(UserGroupInformation.getCurrentUser() .getShortUserName()); }
3.68
hmily_NacosConfig_fileName
/** * File name string. * * @return the string */ public String fileName() { return dataId + "." + fileExtension; }
3.68
hadoop_ZStandardCompressor_getBytesWritten
/** * Returns the total number of compressed bytes output so far. * * @return the total (non-negative) number of compressed bytes output so far */ @Override public long getBytesWritten() { checkStream(); return bytesWritten; }
3.68
flink_RocksDBNativeMetricOptions_enableNumRunningFlushes
/** Returns the number of currently running flushes. */ public void enableNumRunningFlushes() { this.properties.add(RocksDBProperty.NumRunningFlushes.getRocksDBProperty()); }
3.68
hadoop_TimelineReaderWebServicesUtils_parseEventFilters
/** * Parse a delimited string and convert it into a set of strings. For * instance, if delimiter is ",", then the string should be represented as * "value1,value2,value3". * @param str delimited string. * @param delimiter string is delimited by this delimiter. * @return set of strings. */ static TimelineFilterL...
3.68
hadoop_OBSFileSystem_isEnableTrash
/** * Return a flag that indicates if fast delete is enabled. * * @return the flag */ boolean isEnableTrash() { return enableTrash; }
3.68
open-banking-gateway_FintechConsentAccessImpl_delete
/** * Deletes consent from the database. */ @Override public void delete(ProtocolFacingConsent consent) { consents.delete(((ProtocolFacingConsentImpl) consent).getConsent()); }
3.68
hbase_FavoredNodeAssignmentHelper_generateMissingFavoredNode
/* * Generates a missing favored node based on the input favored nodes. This helps to generate new * FN when there is already 2 FN and we need a third one. For eg, while generating new FN for * split daughters after inheriting 2 FN from the parent. If the cluster has only one rack it * generates from the same rack....
3.68
flink_CoGroupOperatorBase_getGroupOrderForInputOne
/** * Gets the order of elements within a group for the first input. If no such order has been set, * this method returns null. * * @return The group order for the first input. */ public Ordering getGroupOrderForInputOne() { return getGroupOrder(0); }
3.68
flink_MemorySize_parse
/** * Parses the given string with a default unit. * * @param text The string to parse. * @param defaultUnit specify the default unit. * @return The parsed MemorySize. * @throws IllegalArgumentException Thrown, if the expression cannot be parsed. */ public static MemorySize parse(String text, MemoryUnit defaultU...
3.68
pulsar_WebServer_addRestResource
/** * Add a REST resource to the servlet context. * * @param basePath The base path for the resource. * @param attribute An attribute associated with the resource. * @param attributeValue The value of the attribute. * @param resourceClass The class representing the resource. *...
3.68
streampipes_HTMLFetcher_fetch
/** * Fetches the document at the given URL, using {@link URLConnection}. * * @param url * @return * @throws IOException */ public static HTMLDocument fetch(final URL url) throws IOException { final URLConnection conn = url.openConnection(); final String ct = conn.getContentType(); if (ct == null || !(ct.e...
3.68
hadoop_SnappyDecompressor_setInput
/** * Sets input data for decompression. * This should be called if and only if {@link #needsInput()} returns * <code>true</code> indicating that more input data is required. * (Both native and non-native versions of various Decompressors require * that the data passed in via <code>b[]</code> remain unmodified unt...
3.68
morf_SqlDialect_columns
/** * @see org.alfasoftware.morf.metadata.Table#columns() */ @Override public List<Column> columns() { List<Column> columns = new ArrayList<>(); columns.add(SchemaUtils.column(ID_INCREMENTOR_TABLE_COLUMN_NAME, DataType.STRING, 132).primaryKey()); columns.add(SchemaUtils.column(ID_INCREMENTOR_TABLE_COLUMN_VALUE,...
3.68
dubbo_ClassUtils_resolveClass
/** * Resolve the {@link Class} by the specified name and {@link ClassLoader} * * @param className the name of {@link Class} * @param classLoader {@link ClassLoader} * @return If can't be resolved , return <code>null</code> * @since 2.7.6 */ public static Class<?> resolveClass(String className, ClassLoader cla...
3.68
framework_VScrollTable_cancel
/** * Cancels the current context touch timeout. */ public void cancel() { if (contextTouchTimeout != null) { contextTouchTimeout.cancel(); contextTouchTimeout = null; } touchStart = null; }
3.68
hudi_BaseHoodieWriteClient_updateColumnComment
/** * update col comment for hudi table. * * @param colName col name to be changed. if we want to change col from a nested filed, the fullName should be specified * @param doc . */ public void updateColumnComment(String colName, String doc) { Pair<InternalSchema, HoodieTableMetaClient> pair = getInternalSche...
3.68
hbase_PrivateCellUtil_compareQualifier
/** * Compare cell's qualifier against given comparator * @param cell the cell to use for comparison * @param comparator the {@link CellComparator} to use for comparison * @return result comparing cell's qualifier */ public static int compareQualifier(Cell cell, ByteArrayComparable comparator) { if (cell i...
3.68
flink_DeletePushDownUtils_prepareFilter
/** Prepare the filter with reducing && simplifying. */ private static Filter prepareFilter(Filter filter) { // we try to reduce and simplify the filter ReduceExpressionsRuleProxy reduceExpressionsRuleProxy = ReduceExpressionsRuleProxy.INSTANCE; SimplifyFilterConditionRule simplifyFilterConditionRule = ...
3.68
hbase_AbstractProcedureScheduler_schedLock
// ========================================================================== // Internal helpers // ========================================================================== protected void schedLock() { schedulerLock.lock(); }
3.68
hadoop_KMSAudit_op
/** * Logs to the audit service a single operation on the KMS or on a key. * * @param opStatus * The outcome of the audited event * @param op * The operation being audited (either {@link KMS.KMSOp} or * {@link Type} N.B this is passed as an {@link Object} to allow * either en...
3.68
hbase_WALKeyImpl_getSequenceId
/** * SequenceId is only available post WAL-assign. Calls before this will get you a * {@link SequenceId#NO_SEQUENCE_ID}. See the comment on FSHLog#append and #getWriteNumber in this * method for more on when this sequenceId comes available. * @return long the new assigned sequence number */ @Override public long ...
3.68
hbase_EncodedDataBlock_getIterator
/** * Provides access to compressed value. * @param headerSize header size of the block. * @return Forwards sequential iterator. */ public Iterator<Cell> getIterator(int headerSize) { final int rawSize = rawKVs.length; byte[] encodedDataWithHeader = getEncodedData(); int bytesToSkip = headerSize + Bytes.SIZEO...
3.68
framework_CalendarEvent_getCalendarEvent
/** * @return the {@link com.vaadin.addon.calendar.event.CalendarEvent * CalendarEvent} that has changed */ public CalendarEvent getCalendarEvent() { return source; }
3.68
rocketmq-connect_Worker_stopConnectors
/** * stop connectors * * @param ids */ private void stopConnectors(Collection<String> ids) { for (String connectorName : ids) { stopConnector(connectorName); } }
3.68
open-banking-gateway_ProcessResultEventHandler_add
/** * Adds the subscriber to the BPMN process. If any already exists - old one will be removed. * * @param processId BPMN process id to subscribe to * @param subscriber Internal BPMN event handling function */ void add(String processId, Consumer<InternalProcessResult> subscriber) { InternalProcessResult delay...
3.68
flink_RequestedLocalProperties_filterBySemanticProperties
/** * Filters these properties by what can be preserved by the given SemanticProperties when * propagated down to the given input. * * @param props The SemanticProperties which define which fields are preserved. * @param input The index of the operator's input. * @return The filtered RequestedLocalProperties */ ...
3.68
dubbo_ServiceDiscoveryRegistryDirectory_destroyAllInvokers
/** * Close all invokers */ @Override protected void destroyAllInvokers() { Map<ProtocolServiceKeyWithAddress, Invoker<T>> localUrlInvokerMap = this.urlInvokerMap; // local reference if (localUrlInvokerMap != null) { for (Invoker<T> invoker : new ArrayList<>(localUrlInvokerMap.values())) { ...
3.68
flink_HiveStatsUtil_createTableColumnStats
/** Create Flink ColumnStats from Hive ColumnStatisticsData. */ private static CatalogColumnStatisticsDataBase createTableColumnStats( DataType colType, ColumnStatisticsData stats, String hiveVersion) { HiveShim hiveShim = HiveShimLoader.loadHiveShim(hiveVersion); if (stats.isSetBinaryStats()) { ...
3.68
hadoop_StageConfig_withProgressable
/** * Optional progress callback. * @param value new value * @return this */ public StageConfig withProgressable(final Progressable value) { checkOpen(); progressable = value; return this; }
3.68
flink_UploadThrottle_releaseCapacity
/** * Release previously {@link #seizeCapacity(long) seized} capacity. Called by {@link * BatchingStateChangeUploadScheduler} (IO thread). */ public void releaseCapacity(long bytes) { inFlightBytesCounter -= bytes; }
3.68
flink_StreamExecutionEnvironment_executeAsync
/** * Triggers the program execution asynchronously. The environment will execute all parts of the * program that have resulted in a "sink" operation. Sink operations are for example printing * results or forwarding them to a message queue. * * @param streamGraph the stream graph representing the transformations ...
3.68
hbase_HBaseRpcServicesBase_checkOOME
/** * Check if an OOME and, if so, abort immediately to avoid creating more objects. * @return True if we OOME'd and are aborting. */ @Override public boolean checkOOME(Throwable e) { return OOMEChecker.exitIfOOME(e, getClass().getSimpleName()); }
3.68
hbase_Mutation_getRow
/** * Method for retrieving the delete's row */ @Override public byte[] getRow() { return this.row; }
3.68
hbase_MemoryBoundedLogMessageBuffer_dumpTo
/** * Dump the contents of the buffer to the given stream. */ public synchronized void dumpTo(PrintWriter out) { SimpleDateFormat df = new SimpleDateFormat("yyyy-MM-dd'T'HH:mm:ss"); for (LogMessage msg : messages) { out.write(df.format(new Date(msg.timestamp))); out.write(" "); out.println(new String...
3.68
querydsl_JPAExpressions_min
/** * Create a min(col) expression * * @param left collection * @return min(col) */ public static <A extends Comparable<? super A>> ComparableExpression<A> min(CollectionExpression<?,A> left) { return Expressions.comparableOperation((Class) left.getParameter(0), Ops.QuantOps.MIN_IN_COL, (Expression<?>) left); ...
3.68
hadoop_Quota_getEachQuotaUsage
/** * Get quota usage for the federation path. * @param path Federation path. * @return quota usage for each remote location. * @throws IOException If the quota system is disabled. */ Map<RemoteLocation, QuotaUsage> getEachQuotaUsage(String path) throws IOException { rpcServer.checkOperation(OperationCategor...
3.68
hudi_JdbcSource_validatePropsAndGetDataFrameReader
/** * Validates all user properties and prepares the {@link DataFrameReader} to read from RDBMS. * * @param session The {@link SparkSession}. * @param properties The JDBC connection properties and data source options. * @return The {@link DataFrameReader} to read from RDBMS * @throws HoodieException */ privat...
3.68
flink_FunctionContext_getCachedFile
/** * Gets the local temporary file copy of a distributed cache files. * * @param name distributed cache file name * @return local temporary file copy of a distributed cache file. */ public File getCachedFile(String name) { if (context == null) { throw new TableException( "Calls to Func...
3.68
flink_PojoSerializerSnapshot_restoreSerializers
/** * Transforms a {@link LinkedHashMap} with {@link TypeSerializerSnapshot}s as the value to * {@link TypeSerializer} as the value by restoring the snapshot. */ private static <K> LinkedHashMap<K, TypeSerializer<?>> restoreSerializers( LinkedHashMap<K, TypeSerializerSnapshot<?>> snapshotsMap) { final Li...
3.68
dubbo_ThreadLocalCache_get
/** * API to return stored value using a key against the calling thread specific store. * @param key Unique identifier for cache lookup * @return Return stored object against key */ @Override public Object get(Object key) { return store.get().get(key); }
3.68
hadoop_OBSFileSystem_getDefaultPort
/** * Return the default port for this FileSystem. * * @return -1 to indicate the port is undefined, which agrees with the * contract of {@link URI#getPort()} */ @Override public int getDefaultPort() { return OBSConstants.OBS_DEFAULT_PORT; }
3.68
hadoop_MawoConfiguration_getJobQueueStorageEnabled
/** * Check if Job Queue Storage is Enabled. * @return True if Job queue storage is enabled otherwise False */ public boolean getJobQueueStorageEnabled() { return Boolean.parseBoolean(configsMap.get(JOB_QUEUE_STORAGE_ENABLED)); }
3.68
querydsl_StringExpression_notLike
/** * Create a {@code this not like str} expression * * @param str string * @return this not like str */ public BooleanExpression notLike(Expression<String> str, char escape) { return like(str, escape).not(); }
3.68
AreaShop_RegionGroup_getSettings
/** * Get the configurationsection with the settings of this group. * @return The ConfigurationSection with the settings of the group */ public ConfigurationSection getSettings() { ConfigurationSection result = plugin.getFileManager().getGroupSettings(name); if(result != null) { return result; } else { retur...
3.68
flink_KvStateRegistry_createTaskRegistry
/** * Creates a {@link TaskKvStateRegistry} facade for the {@link Task} identified by the given * JobID and JobVertexID instance. * * @param jobId JobID of the task * @param jobVertexId JobVertexID of the task * @return A {@link TaskKvStateRegistry} facade for the task */ public TaskKvStateRegistry createTaskReg...
3.68
hadoop_InputSplit_getLocationInfo
/** * Gets info about which nodes the input split is stored on and how it is * stored at each location. * * @return list of <code>SplitLocationInfo</code>s describing how the split * data is stored at each location. A null value indicates that all the * locations have the data stored on disk. * @throws IO...
3.68
flink_FineGrainedSlotManager_declareNeededResources
/** DO NOT call this method directly. Use {@link #declareNeededResourcesWithDelay()} instead. */ private void declareNeededResources() { Map<InstanceID, WorkerResourceSpec> unWantedTaskManagers = taskManagerTracker.getUnWantedTaskManager(); Map<WorkerResourceSpec, Set<InstanceID>> unWantedTaskManage...
3.68
hadoop_ContainerReapContext_build
/** * Builds the context with the attributes set. * * @return the context. */ public ContainerReapContext build() { return new ContainerReapContext(this); }
3.68
flink_LinkedOptionalMap_absentKeysOrValues
/** Returns the key names of any keys or values that are absent. */ public Set<String> absentKeysOrValues() { return underlyingMap.entrySet().stream() .filter(LinkedOptionalMap::keyOrValueIsAbsent) .map(Entry::getKey) .collect(Collectors.toCollection(LinkedHashSet::new)); }
3.68
pulsar_AbstractHierarchicalLedgerManager_isLedgerParentNode
/** * whether the child of ledgersRootPath is a top level parent znode for * ledgers (in HierarchicalLedgerManager) or znode of a ledger (in * FlatLedgerManager). */ public boolean isLedgerParentNode(String path) { return path.matches(getLedgerParentNodeRegex()); }
3.68
hudi_HoodieWriteCommitPulsarCallback_createProducer
/** * Method helps to create {@link Producer}. * * @param hoodieConfig Pulsar configs * @return A {@link Producer} */ public Producer<String> createProducer(HoodieConfig hoodieConfig) throws PulsarClientException { MessageRoutingMode routeMode = Enum.valueOf(MessageRoutingMode.class, PRODUCER_ROUTE_MODE.de...
3.68
morf_AbstractConnectionResources_getConnection
/** * @see javax.sql.DataSource#getConnection(java.lang.String, java.lang.String) */ @Override public Connection getConnection(String username, String password) throws SQLException { log.info("Opening new database connection to [" + AbstractConnectionResources.this.getJdbcUrl() + "] with username [" + username + "]...
3.68
framework_Profiler_getOwnTime
/** * Gets the total time spent in this node, excluding time spent in sub * nodes. * * @return the total time spent, in milliseconds */ public double getOwnTime() { double time = getTimeSpent(); for (Node node : children.values()) { time -= node.getTimeSpent(); } return time; }
3.68
flink_PartitionedFileWriter_releaseQuietly
/** Used to close and delete the failed {@link PartitionedFile} when any exception occurs. */ public void releaseQuietly() { IOUtils.closeQuietly(this); IOUtils.deleteFileQuietly(dataFilePath); IOUtils.deleteFileQuietly(indexFilePath); }
3.68
hibernate-validator_ReflectionHelper_getIndexedValue
/** * Tries to retrieve the indexed value from the specified object. * * @param value The object from which to retrieve the indexed value. The object has to be non {@code null} and * either a collection or array. * @param index The index. * * @return The indexed value or {@code null} if {@code value} is {@code n...
3.68
hbase_ArrayBackedTag_getValueArray
/** Returns The byte array backing this Tag. */ @Override public byte[] getValueArray() { return this.bytes; }
3.68
hadoop_ReplicaInfo_setVolume
/** * Set the volume where this replica is located on disk. */ void setVolume(FsVolumeSpi vol) { this.volume = vol; }
3.68
hibernate-validator_MetaConstraints_getWrappedValueType
/** * Returns the sub-types binding for the single type parameter of the super-type. E.g. for {@code IntegerProperty} * and {@code Property<T>}, {@code Integer} would be returned. */ private static Class<?> getWrappedValueType(TypeResolutionHelper typeResolutionHelper, Type declaredType, ValueExtractorDescriptor val...
3.68
flink_PrioritizedDeque_clear
/** Removes all priority and non-priority elements. */ public void clear() { deque.clear(); numPriorityElements = 0; }
3.68
pulsar_MultiTopicsConsumerImpl_removeConsumerAsync
// Remove a consumer for a topic public CompletableFuture<Void> removeConsumerAsync(String topicName) { checkArgument(TopicName.isValid(topicName), "Invalid topic name:" + topicName); if (getState() == State.Closing || getState() == State.Closed) { return FutureUtil.failedFuture( new Pulsar...
3.68
hadoop_Container_setAllocationRequestId
/** * Set the optional <em>ID</em> corresponding to the original {@code * ResourceRequest{@link #setAllocationRequestId(long)} * etAllocationRequestId()}}s which is satisfied by this allocated {@code * Container}. * <p> * The scheduler may return multiple {@code AllocateResponse}s corresponding * to the same ID ...
3.68
hbase_HMaster_getNamespace
/** * Get a Namespace * @param name Name of the Namespace * @return Namespace descriptor for <code>name</code> */ NamespaceDescriptor getNamespace(String name) throws IOException { checkInitialized(); if (this.cpHost != null) this.cpHost.preGetNamespaceDescriptor(name); NamespaceDescriptor nsd = this.clusterS...
3.68
framework_VAbstractSplitPanel_setLocked
/** * For internal use only. May be removed or replaced in the future. * * @param newValue * {@code true} if split position should be locked, {@code false} * otherwise */ public void setLocked(boolean newValue) { if (locked != newValue) { locked = newValue; splitterSize =...
3.68
hadoop_IOStatisticsStoreImpl_lookupQuietly
/** * Get a reference to the map type providing the * value for a specific key, returning null if it not found. * @param <T> type of map/return type. * @param map map to look up * @param key statistic name * @return the value */ private static <T> T lookupQuietly(final Map<String, T> map, String key) { return ...
3.68
framework_FileUploadHandler_streamToReceiver
/** * @param in * @param streamVariable * @param filename * @param type * @param contentLength * @return true if the streamvariable has informed that the terminal can * forget this variable * @throws UploadException */ protected final boolean streamToReceiver(VaadinSession session, final InputS...
3.68
flink_TypeTransformation_transform
/** * Transforms the given data type to a different data type. * * <p>This method provides a {@link DataTypeFactory} if available. */ default DataType transform(@Nullable DataTypeFactory factory, DataType typeToTransform) { return transform(typeToTransform); }
3.68
dubbo_DubboConfigBeanDefinitionConflictApplicationListener_resolveUniqueApplicationConfigBean
/** * Resolve the unique {@link ApplicationConfig} Bean * * @param registry {@link BeanDefinitionRegistry} instance * @param beanFactory {@link ConfigurableListableBeanFactory} instance * @see EnableDubboConfig */ private void resolveUniqueApplicationConfigBean(BeanDefinitionRegistry registry, ListableBeanFact...
3.68
dubbo_ServiceInvokeRestFilter_executeResponseIntercepts
/** * execute response Intercepts * * @param restFilterContext * @throws Exception */ public void executeResponseIntercepts(RestInterceptContext restFilterContext) throws Exception { for (RestResponseInterceptor restResponseInterceptor : restResponseInterceptors) { restResponseInterceptor.intercept(r...
3.68