name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_HiveParserSqlFunctionConverter_getName | // TODO: this is not valid. Function names for built-in UDFs are specified in
// FunctionRegistry, and only happen to match annotations. For user UDFs, the
// name is what user specifies at creation time (annotation can be absent,
// different, or duplicate some other function).
private static String getName(GenericUDF... | 3.68 |
hbase_TableDescriptorBuilder_getValues | /**
* Getter for fetching an unmodifiable {@link #values} map.
* @return unmodifiable map {@link #values}.
* @see #values
*/
@Override
public Map<Bytes, Bytes> getValues() {
// shallow pointer copy
return Collections.unmodifiableMap(values);
} | 3.68 |
druid_SQLMethodInvokeExpr_getParameters | /**
* instead of getArguments
*
* @deprecated
*/
public List<SQLExpr> getParameters() {
return this.arguments;
} | 3.68 |
framework_AbstractComponentConnector_onDragSourceDetached | /**
* Invoked when a {@link DragSourceExtensionConnector} has been removed from
* this component.
* <p>
* By default, does nothing.
* <p>
* This is a framework internal method, and should not be invoked manually.
*
* @since 8.1
* @see #onDragSourceAttached()
*/
public void onDragSourceDetached() {
} | 3.68 |
flink_CheckpointConfig_setTolerableCheckpointFailureNumber | /**
* This defines how many consecutive checkpoint failures will be tolerated, before the whole job
* is failed over. The default value is `0`, which means no checkpoint failures will be
* tolerated, and the job will fail on first reported checkpoint failure.
*/
public void setTolerableCheckpointFailureNumber(int t... | 3.68 |
flink_FlinkContainersSettings_build | /**
* Returns a {@code FlinkContainersConfig} built from the parameters previously set.
*
* @return A {@code FlinkContainersConfig} built with parameters of this {@code
* FlinkContainersConfig.Builder}.
*/
public FlinkContainersSettings build() {
return new FlinkContainersSettings(this);
} | 3.68 |
hadoop_HdfsConfiguration_init | /**
* This method is here so that when invoked, HdfsConfiguration is class-loaded
* if it hasn't already been previously loaded. Upon loading the class, the
* static initializer block above will be executed to add the deprecated keys
* and to add the default resources. It is safe for this method to be called
* mu... | 3.68 |
framework_SharedUtil_propertyIdToHumanFriendly | /**
* Converts a property id to a human friendly format. Handles nested
* properties by only considering the last part, e.g. "address.streetName"
* is equal to "streetName" for this method.
*
* @since 7.4
* @param propertyId
* The propertyId to format
* @return A human friendly version of the propert... | 3.68 |
flink_JavaFieldPredicates_isStatic | /**
* Match the static modifier of the {@link JavaField}.
*
* @return A {@link DescribedPredicate} returning true, if and only if the tested {@link
* JavaField} has the static modifier.
*/
public static DescribedPredicate<JavaField> isStatic() {
return DescribedPredicate.describe(
"static", fie... | 3.68 |
framework_ColorPickerPopup_setHSVTabVisible | /**
* Sets the HSV tab visibility.
*
* @param visible
* The visibility of the HSV tab
*/
public void setHSVTabVisible(boolean visible) {
if (visible && !isTabVisible(hsvTab)) {
tabs.addTab(hsvTab, "HSV", null);
checkIfTabsNeeded();
} else if (!visible && isTabVisible(hsvTab)) {
... | 3.68 |
hadoop_AbstractS3ACommitter_getDestinationFS | /**
* Get the destination filesystem from the output path and the configuration.
* @param out output path
* @param config job/task config
* @return the associated FS
* @throws PathCommitException output path isn't to an S3A FS instance.
* @throws IOException failure to instantiate the FS.
*/
protected FileSystem... | 3.68 |
flink_ArchivedExecutionGraph_createSparseArchivedExecutionGraph | /**
* Create a sparse ArchivedExecutionGraph for a job. Most fields will be empty, only job status
* and error-related fields are set.
*/
public static ArchivedExecutionGraph createSparseArchivedExecutionGraph(
JobID jobId,
String jobName,
JobStatus jobStatus,
@Nullable Throwable thro... | 3.68 |
flink_DataStream_printToErr | /**
* Writes a DataStream to the standard error stream (stderr).
*
* <p>For each element of the DataStream the result of {@link Object#toString()} is written.
*
* <p>NOTE: This will print to stderr on the machine where the code is executed, i.e. the Flink
* worker.
*
* @param sinkIdentifier The string to prefix... | 3.68 |
hadoop_ActiveAuditManagerS3A_prune | /**
* Prune all null weak references, calling the referenceLost
* callback for each one.
*
* non-atomic and non-blocking.
* @return the number of entries pruned.
*/
@VisibleForTesting
int prune() {
return activeSpanMap.prune();
} | 3.68 |
flink_SplitEnumeratorContext_registeredReadersOfAttempts | /**
* Get the currently registered readers of all the subtask attempts. The mapping is from subtask
* id to a map which maps an attempt to its reader info.
*
* @return the currently registered readers.
*/
default Map<Integer, Map<Integer, ReaderInfo>> registeredReadersOfAttempts() {
throw new UnsupportedOperat... | 3.68 |
rocketmq-connect_ClusterManagementServiceImpl_prepare | /**
* Preparation before startup
*
* @param connectConfig
*/
private void prepare(WorkerConfig connectConfig) {
String consumerGroup = this.defaultMQPullConsumer.getConsumerGroup();
Set<String> consumerGroupSet = ConnectUtil.fetchAllConsumerGroupList(connectConfig);
if (!consumerGroupSet.contains(consum... | 3.68 |
framework_Table_setColumnHeader | /**
* Sets the column header for the specified column.
*
* @param propertyId
* the propertyId identifying the column.
* @param header
* the header to set.
*/
public void setColumnHeader(Object propertyId, String header) {
if (header == null) {
columnHeaders.remove(propertyId);
... | 3.68 |
framework_VaadinService_getCurrentRequest | /**
* Gets the currently processed Vaadin request. The current request is
* automatically defined when the request is started. The current request
* can not be used in e.g. background threads because of the way server
* implementations reuse request instances.
*
* @return the current Vaadin request instance if av... | 3.68 |
framework_ExternalResource_setMIMEType | /**
* Sets the MIME type of the resource.
*/
public void setMIMEType(String mimeType) {
this.mimeType = mimeType;
} | 3.68 |
hadoop_PartitionResourcesInfo_getUserAmLimit | /**
* @return the userAmLimit
*/
public ResourceInfo getUserAmLimit() {
return userAmLimit;
} | 3.68 |
hadoop_AMRMClientAsyncImpl_serviceStop | /**
* Tells the heartbeat and handler threads to stop and waits for them to
* terminate.
*/
@Override
protected void serviceStop() throws Exception {
keepRunning = false;
heartbeatThread.interrupt();
try {
heartbeatThread.join();
} catch (InterruptedException ex) {
LOG.error("Error joining with heart... | 3.68 |
morf_AbstractSqlDialectTest_expectedLeftTrim | /**
* @return The expected SQL for a Left Trim
*/
protected String expectedLeftTrim() {
return "SELECT LTRIM(field1) FROM " + tableName("schedule");
} | 3.68 |
hudi_FlinkOptions_getPropertiesWithPrefix | /**
* Collects the config options that start with specified prefix {@code prefix} into a 'key'='value' list.
*/
public static Map<String, String> getPropertiesWithPrefix(Map<String, String> options, String prefix) {
final Map<String, String> hoodieProperties = new HashMap<>();
if (hasPropertyOptions(options, pref... | 3.68 |
flink_CheckpointConfig_enableApproximateLocalRecovery | /**
* Enables the approximate local recovery mode.
*
* <p>In this recovery mode, when a task fails, the entire downstream of the tasks (including
* the failed task) restart.
*
* <p>Notice that 1. Approximate recovery may lead to data loss. The amount of data which leads
* the failed task from the state of the la... | 3.68 |
pulsar_FunctionRuntimeManager_getFunctionStats | /**
* Get stats of all function instances.
*
* @param tenant the tenant the function belongs to
* @param namespace the namespace the function belongs to
* @param functionName the function name
* @return a list of function statuses
* @throws PulsarAdminException
*/
public FunctionStatsImpl getFunctionSt... | 3.68 |
hbase_MetaTableAccessor_updateLocation | /**
* Updates the location of the specified region to be the specified server.
* <p>
* Connects to the specified server which should be hosting the specified catalog region name to
* perform the edit.
* @param connection connection we're using
* @param regionInfo region to update location of
* @param... | 3.68 |
druid_SQLCommitStatement_getChain | // mysql
public Boolean getChain() {
return chain;
} | 3.68 |
flink_ProcessorArchitecture_getAddressSize | /** Gets the address size of the memory (32 bit, 64 bit). */
public MemoryAddressSize getAddressSize() {
return addressSize;
} | 3.68 |
dubbo_StringUtils_decodeHexByte | /**
* Decode a 2-digit hex byte from within a string.
*/
public static byte decodeHexByte(CharSequence s, int pos) {
int hi = decodeHexNibble(s.charAt(pos));
int lo = decodeHexNibble(s.charAt(pos + 1));
if (hi == -1 || lo == -1) {
throw new IllegalArgumentException(
String.format("... | 3.68 |
hbase_ZKUtil_getPath | /** Returns path to znode where the ZKOp will occur */
public String getPath() {
return path;
} | 3.68 |
hbase_AdvancedScanResultConsumer_onHeartbeat | /**
* Indicate that there is a heartbeat message but we have not cumulated enough cells to call
* {@link #onNext(Result[], ScanController)}.
* <p>
* Note that this method will always be called when RS returns something to us but we do not have
* enough cells to call {@link #onNext(Result[], ScanController)}. Somet... | 3.68 |
hadoop_AssumedRoleCredentialProvider_buildSessionName | /**
* Build the session name from the current user's shortname.
* @return a string for the session name.
* @throws IOException failure to get the current user
*/
static String buildSessionName() throws IOException {
return sanitize(UserGroupInformation.getCurrentUser()
.getShortUserName());
} | 3.68 |
hmily_NacosConfig_fileName | /**
* File name string.
*
* @return the string
*/
public String fileName() {
return dataId + "." + fileExtension;
} | 3.68 |
hadoop_ZStandardCompressor_getBytesWritten | /**
* Returns the total number of compressed bytes output so far.
*
* @return the total (non-negative) number of compressed bytes output so far
*/
@Override
public long getBytesWritten() {
checkStream();
return bytesWritten;
} | 3.68 |
flink_RocksDBNativeMetricOptions_enableNumRunningFlushes | /** Returns the number of currently running flushes. */
public void enableNumRunningFlushes() {
this.properties.add(RocksDBProperty.NumRunningFlushes.getRocksDBProperty());
} | 3.68 |
hadoop_TimelineReaderWebServicesUtils_parseEventFilters | /**
* Parse a delimited string and convert it into a set of strings. For
* instance, if delimiter is ",", then the string should be represented as
* "value1,value2,value3".
* @param str delimited string.
* @param delimiter string is delimited by this delimiter.
* @return set of strings.
*/
static TimelineFilterL... | 3.68 |
hadoop_OBSFileSystem_isEnableTrash | /**
* Return a flag that indicates if fast delete is enabled.
*
* @return the flag
*/
boolean isEnableTrash() {
return enableTrash;
} | 3.68 |
open-banking-gateway_FintechConsentAccessImpl_delete | /**
* Deletes consent from the database.
*/
@Override
public void delete(ProtocolFacingConsent consent) {
consents.delete(((ProtocolFacingConsentImpl) consent).getConsent());
} | 3.68 |
hbase_FavoredNodeAssignmentHelper_generateMissingFavoredNode | /*
* Generates a missing favored node based on the input favored nodes. This helps to generate new
* FN when there is already 2 FN and we need a third one. For eg, while generating new FN for
* split daughters after inheriting 2 FN from the parent. If the cluster has only one rack it
* generates from the same rack.... | 3.68 |
flink_CoGroupOperatorBase_getGroupOrderForInputOne | /**
* Gets the order of elements within a group for the first input. If no such order has been set,
* this method returns null.
*
* @return The group order for the first input.
*/
public Ordering getGroupOrderForInputOne() {
return getGroupOrder(0);
} | 3.68 |
flink_MemorySize_parse | /**
* Parses the given string with a default unit.
*
* @param text The string to parse.
* @param defaultUnit specify the default unit.
* @return The parsed MemorySize.
* @throws IllegalArgumentException Thrown, if the expression cannot be parsed.
*/
public static MemorySize parse(String text, MemoryUnit defaultU... | 3.68 |
pulsar_WebServer_addRestResource | /**
* Add a REST resource to the servlet context.
*
* @param basePath The base path for the resource.
* @param attribute An attribute associated with the resource.
* @param attributeValue The value of the attribute.
* @param resourceClass The class representing the resource.
*... | 3.68 |
streampipes_HTMLFetcher_fetch | /**
* Fetches the document at the given URL, using {@link URLConnection}.
*
* @param url
* @return
* @throws IOException
*/
public static HTMLDocument fetch(final URL url) throws IOException {
final URLConnection conn = url.openConnection();
final String ct = conn.getContentType();
if (ct == null || !(ct.e... | 3.68 |
hadoop_SnappyDecompressor_setInput | /**
* Sets input data for decompression.
* This should be called if and only if {@link #needsInput()} returns
* <code>true</code> indicating that more input data is required.
* (Both native and non-native versions of various Decompressors require
* that the data passed in via <code>b[]</code> remain unmodified unt... | 3.68 |
morf_SqlDialect_columns | /**
* @see org.alfasoftware.morf.metadata.Table#columns()
*/
@Override
public List<Column> columns() {
List<Column> columns = new ArrayList<>();
columns.add(SchemaUtils.column(ID_INCREMENTOR_TABLE_COLUMN_NAME, DataType.STRING, 132).primaryKey());
columns.add(SchemaUtils.column(ID_INCREMENTOR_TABLE_COLUMN_VALUE,... | 3.68 |
dubbo_ClassUtils_resolveClass | /**
* Resolve the {@link Class} by the specified name and {@link ClassLoader}
*
* @param className the name of {@link Class}
* @param classLoader {@link ClassLoader}
* @return If can't be resolved , return <code>null</code>
* @since 2.7.6
*/
public static Class<?> resolveClass(String className, ClassLoader cla... | 3.68 |
framework_VScrollTable_cancel | /**
* Cancels the current context touch timeout.
*/
public void cancel() {
if (contextTouchTimeout != null) {
contextTouchTimeout.cancel();
contextTouchTimeout = null;
}
touchStart = null;
} | 3.68 |
hudi_BaseHoodieWriteClient_updateColumnComment | /**
* update col comment for hudi table.
*
* @param colName col name to be changed. if we want to change col from a nested filed, the fullName should be specified
* @param doc .
*/
public void updateColumnComment(String colName, String doc) {
Pair<InternalSchema, HoodieTableMetaClient> pair = getInternalSche... | 3.68 |
hbase_PrivateCellUtil_compareQualifier | /**
* Compare cell's qualifier against given comparator
* @param cell the cell to use for comparison
* @param comparator the {@link CellComparator} to use for comparison
* @return result comparing cell's qualifier
*/
public static int compareQualifier(Cell cell, ByteArrayComparable comparator) {
if (cell i... | 3.68 |
flink_DeletePushDownUtils_prepareFilter | /** Prepare the filter with reducing && simplifying. */
private static Filter prepareFilter(Filter filter) {
// we try to reduce and simplify the filter
ReduceExpressionsRuleProxy reduceExpressionsRuleProxy = ReduceExpressionsRuleProxy.INSTANCE;
SimplifyFilterConditionRule simplifyFilterConditionRule =
... | 3.68 |
hbase_AbstractProcedureScheduler_schedLock | // ==========================================================================
// Internal helpers
// ==========================================================================
protected void schedLock() {
schedulerLock.lock();
} | 3.68 |
hadoop_KMSAudit_op | /**
* Logs to the audit service a single operation on the KMS or on a key.
*
* @param opStatus
* The outcome of the audited event
* @param op
* The operation being audited (either {@link KMS.KMSOp} or
* {@link Type} N.B this is passed as an {@link Object} to allow
* either en... | 3.68 |
hbase_WALKeyImpl_getSequenceId | /**
* SequenceId is only available post WAL-assign. Calls before this will get you a
* {@link SequenceId#NO_SEQUENCE_ID}. See the comment on FSHLog#append and #getWriteNumber in this
* method for more on when this sequenceId comes available.
* @return long the new assigned sequence number
*/
@Override
public long ... | 3.68 |
hbase_EncodedDataBlock_getIterator | /**
* Provides access to compressed value.
* @param headerSize header size of the block.
* @return Forwards sequential iterator.
*/
public Iterator<Cell> getIterator(int headerSize) {
final int rawSize = rawKVs.length;
byte[] encodedDataWithHeader = getEncodedData();
int bytesToSkip = headerSize + Bytes.SIZEO... | 3.68 |
framework_CalendarEvent_getCalendarEvent | /**
* @return the {@link com.vaadin.addon.calendar.event.CalendarEvent
* CalendarEvent} that has changed
*/
public CalendarEvent getCalendarEvent() {
return source;
} | 3.68 |
rocketmq-connect_Worker_stopConnectors | /**
* stop connectors
*
* @param ids
*/
private void stopConnectors(Collection<String> ids) {
for (String connectorName : ids) {
stopConnector(connectorName);
}
} | 3.68 |
open-banking-gateway_ProcessResultEventHandler_add | /**
* Adds the subscriber to the BPMN process. If any already exists - old one will be removed.
*
* @param processId BPMN process id to subscribe to
* @param subscriber Internal BPMN event handling function
*/
void add(String processId, Consumer<InternalProcessResult> subscriber) {
InternalProcessResult delay... | 3.68 |
flink_RequestedLocalProperties_filterBySemanticProperties | /**
* Filters these properties by what can be preserved by the given SemanticProperties when
* propagated down to the given input.
*
* @param props The SemanticProperties which define which fields are preserved.
* @param input The index of the operator's input.
* @return The filtered RequestedLocalProperties
*/
... | 3.68 |
dubbo_ServiceDiscoveryRegistryDirectory_destroyAllInvokers | /**
* Close all invokers
*/
@Override
protected void destroyAllInvokers() {
Map<ProtocolServiceKeyWithAddress, Invoker<T>> localUrlInvokerMap = this.urlInvokerMap; // local reference
if (localUrlInvokerMap != null) {
for (Invoker<T> invoker : new ArrayList<>(localUrlInvokerMap.values())) {
... | 3.68 |
flink_HiveStatsUtil_createTableColumnStats | /** Create Flink ColumnStats from Hive ColumnStatisticsData. */
private static CatalogColumnStatisticsDataBase createTableColumnStats(
DataType colType, ColumnStatisticsData stats, String hiveVersion) {
HiveShim hiveShim = HiveShimLoader.loadHiveShim(hiveVersion);
if (stats.isSetBinaryStats()) {
... | 3.68 |
hadoop_StageConfig_withProgressable | /**
* Optional progress callback.
* @param value new value
* @return this
*/
public StageConfig withProgressable(final Progressable value) {
checkOpen();
progressable = value;
return this;
} | 3.68 |
flink_UploadThrottle_releaseCapacity | /**
* Release previously {@link #seizeCapacity(long) seized} capacity. Called by {@link
* BatchingStateChangeUploadScheduler} (IO thread).
*/
public void releaseCapacity(long bytes) {
inFlightBytesCounter -= bytes;
} | 3.68 |
flink_StreamExecutionEnvironment_executeAsync | /**
* Triggers the program execution asynchronously. The environment will execute all parts of the
* program that have resulted in a "sink" operation. Sink operations are for example printing
* results or forwarding them to a message queue.
*
* @param streamGraph the stream graph representing the transformations
... | 3.68 |
hbase_HBaseRpcServicesBase_checkOOME | /**
* Check if an OOME and, if so, abort immediately to avoid creating more objects.
* @return True if we OOME'd and are aborting.
*/
@Override
public boolean checkOOME(Throwable e) {
return OOMEChecker.exitIfOOME(e, getClass().getSimpleName());
} | 3.68 |
hbase_Mutation_getRow | /**
* Method for retrieving the delete's row
*/
@Override
public byte[] getRow() {
return this.row;
} | 3.68 |
hbase_MemoryBoundedLogMessageBuffer_dumpTo | /**
* Dump the contents of the buffer to the given stream.
*/
public synchronized void dumpTo(PrintWriter out) {
SimpleDateFormat df = new SimpleDateFormat("yyyy-MM-dd'T'HH:mm:ss");
for (LogMessage msg : messages) {
out.write(df.format(new Date(msg.timestamp)));
out.write(" ");
out.println(new String... | 3.68 |
querydsl_JPAExpressions_min | /**
* Create a min(col) expression
*
* @param left collection
* @return min(col)
*/
public static <A extends Comparable<? super A>> ComparableExpression<A> min(CollectionExpression<?,A> left) {
return Expressions.comparableOperation((Class) left.getParameter(0), Ops.QuantOps.MIN_IN_COL, (Expression<?>) left);
... | 3.68 |
hadoop_Quota_getEachQuotaUsage | /**
* Get quota usage for the federation path.
* @param path Federation path.
* @return quota usage for each remote location.
* @throws IOException If the quota system is disabled.
*/
Map<RemoteLocation, QuotaUsage> getEachQuotaUsage(String path)
throws IOException {
rpcServer.checkOperation(OperationCategor... | 3.68 |
hudi_JdbcSource_validatePropsAndGetDataFrameReader | /**
* Validates all user properties and prepares the {@link DataFrameReader} to read from RDBMS.
*
* @param session The {@link SparkSession}.
* @param properties The JDBC connection properties and data source options.
* @return The {@link DataFrameReader} to read from RDBMS
* @throws HoodieException
*/
privat... | 3.68 |
flink_FunctionContext_getCachedFile | /**
* Gets the local temporary file copy of a distributed cache files.
*
* @param name distributed cache file name
* @return local temporary file copy of a distributed cache file.
*/
public File getCachedFile(String name) {
if (context == null) {
throw new TableException(
"Calls to Func... | 3.68 |
flink_PojoSerializerSnapshot_restoreSerializers | /**
* Transforms a {@link LinkedHashMap} with {@link TypeSerializerSnapshot}s as the value to
* {@link TypeSerializer} as the value by restoring the snapshot.
*/
private static <K> LinkedHashMap<K, TypeSerializer<?>> restoreSerializers(
LinkedHashMap<K, TypeSerializerSnapshot<?>> snapshotsMap) {
final Li... | 3.68 |
dubbo_ThreadLocalCache_get | /**
* API to return stored value using a key against the calling thread specific store.
* @param key Unique identifier for cache lookup
* @return Return stored object against key
*/
@Override
public Object get(Object key) {
return store.get().get(key);
} | 3.68 |
hadoop_OBSFileSystem_getDefaultPort | /**
* Return the default port for this FileSystem.
*
* @return -1 to indicate the port is undefined, which agrees with the
* contract of {@link URI#getPort()}
*/
@Override
public int getDefaultPort() {
return OBSConstants.OBS_DEFAULT_PORT;
} | 3.68 |
hadoop_MawoConfiguration_getJobQueueStorageEnabled | /**
* Check if Job Queue Storage is Enabled.
* @return True if Job queue storage is enabled otherwise False
*/
public boolean getJobQueueStorageEnabled() {
return Boolean.parseBoolean(configsMap.get(JOB_QUEUE_STORAGE_ENABLED));
} | 3.68 |
querydsl_StringExpression_notLike | /**
* Create a {@code this not like str} expression
*
* @param str string
* @return this not like str
*/
public BooleanExpression notLike(Expression<String> str, char escape) {
return like(str, escape).not();
} | 3.68 |
AreaShop_RegionGroup_getSettings | /**
* Get the configurationsection with the settings of this group.
* @return The ConfigurationSection with the settings of the group
*/
public ConfigurationSection getSettings() {
ConfigurationSection result = plugin.getFileManager().getGroupSettings(name);
if(result != null) {
return result;
} else {
retur... | 3.68 |
flink_KvStateRegistry_createTaskRegistry | /**
* Creates a {@link TaskKvStateRegistry} facade for the {@link Task} identified by the given
* JobID and JobVertexID instance.
*
* @param jobId JobID of the task
* @param jobVertexId JobVertexID of the task
* @return A {@link TaskKvStateRegistry} facade for the task
*/
public TaskKvStateRegistry createTaskReg... | 3.68 |
hadoop_InputSplit_getLocationInfo | /**
* Gets info about which nodes the input split is stored on and how it is
* stored at each location.
*
* @return list of <code>SplitLocationInfo</code>s describing how the split
* data is stored at each location. A null value indicates that all the
* locations have the data stored on disk.
* @throws IO... | 3.68 |
flink_FineGrainedSlotManager_declareNeededResources | /** DO NOT call this method directly. Use {@link #declareNeededResourcesWithDelay()} instead. */
private void declareNeededResources() {
Map<InstanceID, WorkerResourceSpec> unWantedTaskManagers =
taskManagerTracker.getUnWantedTaskManager();
Map<WorkerResourceSpec, Set<InstanceID>> unWantedTaskManage... | 3.68 |
hadoop_ContainerReapContext_build | /**
* Builds the context with the attributes set.
*
* @return the context.
*/
public ContainerReapContext build() {
return new ContainerReapContext(this);
} | 3.68 |
flink_LinkedOptionalMap_absentKeysOrValues | /** Returns the key names of any keys or values that are absent. */
public Set<String> absentKeysOrValues() {
return underlyingMap.entrySet().stream()
.filter(LinkedOptionalMap::keyOrValueIsAbsent)
.map(Entry::getKey)
.collect(Collectors.toCollection(LinkedHashSet::new));
} | 3.68 |
pulsar_AbstractHierarchicalLedgerManager_isLedgerParentNode | /**
* whether the child of ledgersRootPath is a top level parent znode for
* ledgers (in HierarchicalLedgerManager) or znode of a ledger (in
* FlatLedgerManager).
*/
public boolean isLedgerParentNode(String path) {
return path.matches(getLedgerParentNodeRegex());
} | 3.68 |
hudi_HoodieWriteCommitPulsarCallback_createProducer | /**
* Method helps to create {@link Producer}.
*
* @param hoodieConfig Pulsar configs
* @return A {@link Producer}
*/
public Producer<String> createProducer(HoodieConfig hoodieConfig) throws PulsarClientException {
MessageRoutingMode routeMode = Enum.valueOf(MessageRoutingMode.class,
PRODUCER_ROUTE_MODE.de... | 3.68 |
morf_AbstractConnectionResources_getConnection | /**
* @see javax.sql.DataSource#getConnection(java.lang.String, java.lang.String)
*/
@Override
public Connection getConnection(String username, String password) throws SQLException {
log.info("Opening new database connection to [" + AbstractConnectionResources.this.getJdbcUrl() + "] with username [" + username + "]... | 3.68 |
framework_Profiler_getOwnTime | /**
* Gets the total time spent in this node, excluding time spent in sub
* nodes.
*
* @return the total time spent, in milliseconds
*/
public double getOwnTime() {
double time = getTimeSpent();
for (Node node : children.values()) {
time -= node.getTimeSpent();
}
return time;
} | 3.68 |
flink_PartitionedFileWriter_releaseQuietly | /** Used to close and delete the failed {@link PartitionedFile} when any exception occurs. */
public void releaseQuietly() {
IOUtils.closeQuietly(this);
IOUtils.deleteFileQuietly(dataFilePath);
IOUtils.deleteFileQuietly(indexFilePath);
} | 3.68 |
hibernate-validator_ReflectionHelper_getIndexedValue | /**
* Tries to retrieve the indexed value from the specified object.
*
* @param value The object from which to retrieve the indexed value. The object has to be non {@code null} and
* either a collection or array.
* @param index The index.
*
* @return The indexed value or {@code null} if {@code value} is {@code n... | 3.68 |
hbase_ArrayBackedTag_getValueArray | /** Returns The byte array backing this Tag. */
@Override
public byte[] getValueArray() {
return this.bytes;
} | 3.68 |
hadoop_ReplicaInfo_setVolume | /**
* Set the volume where this replica is located on disk.
*/
void setVolume(FsVolumeSpi vol) {
this.volume = vol;
} | 3.68 |
hibernate-validator_MetaConstraints_getWrappedValueType | /**
* Returns the sub-types binding for the single type parameter of the super-type. E.g. for {@code IntegerProperty}
* and {@code Property<T>}, {@code Integer} would be returned.
*/
private static Class<?> getWrappedValueType(TypeResolutionHelper typeResolutionHelper, Type declaredType, ValueExtractorDescriptor val... | 3.68 |
flink_PrioritizedDeque_clear | /** Removes all priority and non-priority elements. */
public void clear() {
deque.clear();
numPriorityElements = 0;
} | 3.68 |
pulsar_MultiTopicsConsumerImpl_removeConsumerAsync | // Remove a consumer for a topic
public CompletableFuture<Void> removeConsumerAsync(String topicName) {
checkArgument(TopicName.isValid(topicName), "Invalid topic name:" + topicName);
if (getState() == State.Closing || getState() == State.Closed) {
return FutureUtil.failedFuture(
new Pulsar... | 3.68 |
hadoop_Container_setAllocationRequestId | /**
* Set the optional <em>ID</em> corresponding to the original {@code
* ResourceRequest{@link #setAllocationRequestId(long)}
* etAllocationRequestId()}}s which is satisfied by this allocated {@code
* Container}.
* <p>
* The scheduler may return multiple {@code AllocateResponse}s corresponding
* to the same ID ... | 3.68 |
hbase_HMaster_getNamespace | /**
* Get a Namespace
* @param name Name of the Namespace
* @return Namespace descriptor for <code>name</code>
*/
NamespaceDescriptor getNamespace(String name) throws IOException {
checkInitialized();
if (this.cpHost != null) this.cpHost.preGetNamespaceDescriptor(name);
NamespaceDescriptor nsd = this.clusterS... | 3.68 |
framework_VAbstractSplitPanel_setLocked | /**
* For internal use only. May be removed or replaced in the future.
*
* @param newValue
* {@code true} if split position should be locked, {@code false}
* otherwise
*/
public void setLocked(boolean newValue) {
if (locked != newValue) {
locked = newValue;
splitterSize =... | 3.68 |
hadoop_IOStatisticsStoreImpl_lookupQuietly | /**
* Get a reference to the map type providing the
* value for a specific key, returning null if it not found.
* @param <T> type of map/return type.
* @param map map to look up
* @param key statistic name
* @return the value
*/
private static <T> T lookupQuietly(final Map<String, T> map, String key) {
return ... | 3.68 |
framework_FileUploadHandler_streamToReceiver | /**
* @param in
* @param streamVariable
* @param filename
* @param type
* @param contentLength
* @return true if the streamvariable has informed that the terminal can
* forget this variable
* @throws UploadException
*/
protected final boolean streamToReceiver(VaadinSession session,
final InputS... | 3.68 |
flink_TypeTransformation_transform | /**
* Transforms the given data type to a different data type.
*
* <p>This method provides a {@link DataTypeFactory} if available.
*/
default DataType transform(@Nullable DataTypeFactory factory, DataType typeToTransform) {
return transform(typeToTransform);
} | 3.68 |
dubbo_DubboConfigBeanDefinitionConflictApplicationListener_resolveUniqueApplicationConfigBean | /**
* Resolve the unique {@link ApplicationConfig} Bean
*
* @param registry {@link BeanDefinitionRegistry} instance
* @param beanFactory {@link ConfigurableListableBeanFactory} instance
* @see EnableDubboConfig
*/
private void resolveUniqueApplicationConfigBean(BeanDefinitionRegistry registry, ListableBeanFact... | 3.68 |
dubbo_ServiceInvokeRestFilter_executeResponseIntercepts | /**
* execute response Intercepts
*
* @param restFilterContext
* @throws Exception
*/
public void executeResponseIntercepts(RestInterceptContext restFilterContext) throws Exception {
for (RestResponseInterceptor restResponseInterceptor : restResponseInterceptors) {
restResponseInterceptor.intercept(r... | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.