name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
dubbo_JCacheFactory_createCache | /**
* Takes url as an method argument and return new instance of cache store implemented by JCache.
* @param url url of the method
* @return JCache instance of cache
*/
@Override
protected Cache createCache(URL url) {
return new JCache(url);
} | 3.68 |
framework_Escalator_scrollToColumn | /**
* Scrolls the body horizontally so that the column at the given index is
* visible and there is at least {@code padding} pixels in the direction of
* the given scroll destination.
*
* @param columnIndex
* the index of the column to scroll to
* @param destination
* where the column shou... | 3.68 |
hbase_TableRecordReader_getCurrentKey | /**
* Returns the current key.
* @return The current key.
* @throws InterruptedException When the job is aborted.
* @see org.apache.hadoop.mapreduce.RecordReader#getCurrentKey()
*/
@Override
public ImmutableBytesWritable getCurrentKey() throws IOException, InterruptedException {
return this.recordReaderImpl.getC... | 3.68 |
flink_ConfigurationUtils_getSystemResourceMetricsProbingInterval | /**
* @return extracted {@link MetricOptions#SYSTEM_RESOURCE_METRICS_PROBING_INTERVAL} or {@code
* Optional.empty()} if {@link MetricOptions#SYSTEM_RESOURCE_METRICS} are disabled.
*/
public static Optional<Time> getSystemResourceMetricsProbingInterval(
Configuration configuration) {
if (!configuratio... | 3.68 |
flink_FlinkPreparingTableBase_getCollationList | /**
* Returns a description of the physical ordering (or orderings) of the rows returned from this
* table.
*
* @see org.apache.calcite.rel.metadata.RelMetadataQuery#collations(RelNode)
*/
public List<RelCollation> getCollationList() {
return ImmutableList.of();
} | 3.68 |
hadoop_FileIoProvider_mkdirsWithExistsCheck | /**
* Create the target directory using {@link File#mkdirs()} only if
* it doesn't exist already.
*
* @param volume target volume. null if unavailable.
* @param dir directory to be created.
* @throws IOException if the directory could not created
*/
public void mkdirsWithExistsCheck(
@Nullable FsVolumeSpi... | 3.68 |
hudi_Pair_getKey | /**
* <p>
* Gets the key from this pair.
* </p>
*
* <p>
* This method implements the {@code Map.Entry} interface returning the left element as the key.
* </p>
*
* @return the left element as the key, may be null
*/
@Override
public final L getKey() {
return getLeft();
} | 3.68 |
hadoop_S3LogParser_q | /**
* Quoted entry using the {@link #QUOTED} pattern.
* @param name name of the element (for code clarity only)
* @return the pattern for the regexp
*/
private static String q(String name) {
return e(name, QUOTED);
} | 3.68 |
framework_VDebugWindow_addSection | /**
* Adds the given {@link Section} as a tab in the {@link VDebugWindow} UI.
* {@link Section#getTabButton()} is called to obtain a button which is used
* tab.
*
* @param section
*/
public void addSection(final Section section) {
Button b = section.getTabButton();
b.addClickHandler(event -> {
act... | 3.68 |
framework_TabSheet_readTabFromDesign | /**
* Reads the given tab element from design
*
* @since 7.4
*
* @param tabElement
* the element to be read
* @param designContext
* the design context
*/
private void readTabFromDesign(Element tabElement,
DesignContext designContext) {
Attributes attr = tabElement.attributes(... | 3.68 |
hbase_HRegionFileSystem_openRegionFromFileSystem | /**
* Open Region from file-system.
* @param conf the {@link Configuration} to use
* @param fs {@link FileSystem} from which to add the region
* @param tableDir {@link Path} to where the table is being stored
* @param regionInfo {@link RegionInfo} for region to be added
* @param readOnly True if... | 3.68 |
hbase_FutureUtils_rethrow | /**
* If we could propagate the given {@code error} directly, we will fill the stack trace with the
* current thread's stack trace so it is easier to trace where is the exception thrown. If not, we
* will just create a new IOException and then throw it.
*/
public static IOException rethrow(Throwable error) throws I... | 3.68 |
querydsl_SQLExpressions_covarSamp | /**
* CORR returns the coefficient of correlation of a set of number pairs.
*
* @param expr1 first arg
* @param expr2 second arg
* @return corr(expr1, expr2)
*/
public static WindowOver<Double> covarSamp(Expression<? extends Number> expr1, Expression<? extends Number> expr2) {
return new WindowOver<Double>(Do... | 3.68 |
framework_PasswordField_readDesign | /*
* (non-Javadoc)
*
* @see com.vaadin.ui.AbstractField#readDesign(org.jsoup.nodes.Element ,
* com.vaadin.ui.declarative.DesignContext)
*/
@Override
public void readDesign(Element design, DesignContext designContext) {
super.readDesign(design, designContext);
Attributes attr = design.attributes();
if (... | 3.68 |
hbase_RegionCoprocessorHost_postCompactSelection | /**
* Called after the {@link HStoreFile}s to be compacted have been selected from the available
* candidates.
* @param store The store where compaction is being requested
* @param selected The store files selected to compact
* @param tracker used to track the life cycle of a compaction
* @param request the ... | 3.68 |
druid_MySqlStatementParser_parseCase | /**
* parse case statement
*
* @return MySqlCaseStatement
*/
public MySqlCaseStatement parseCase() {
MySqlCaseStatement stmt = new MySqlCaseStatement();
accept(Token.CASE);
if (lexer.token() == Token.WHEN)// grammar 1
{
while (lexer.token() == Token.WHEN) {
MySqlWhenStatement wh... | 3.68 |
hudi_BaseHoodieTableServiceClient_cluster | /**
* Ensures clustering instant is in expected state and performs clustering for the plan stored in metadata.
*
* @param clusteringInstant Clustering Instant Time
* @return Collection of Write Status
*/
public HoodieWriteMetadata<O> cluster(String clusteringInstant, boolean shouldComplete) {
HoodieTable<?, I, ?... | 3.68 |
hbase_HBaseTestingUtility_generateColumnDescriptors | /**
* Create a set of column descriptors with the combination of compression, encoding, bloom codecs
* available.
* @param prefix family names prefix
* @return the list of column descriptors
*/
public static List<ColumnFamilyDescriptor> generateColumnDescriptors(final String prefix) {
List<ColumnFamilyDescriptor... | 3.68 |
graphhopper_Distributions_logExponentialDistribution | /**
* Use this function instead of Math.log(exponentialDistribution(beta, x)) to avoid an
* arithmetic underflow for very small probabilities.
*
* @param beta =1/lambda with lambda being the standard exponential distribution rate parameter
*/
static double logExponentialDistribution(double beta, double x) {
re... | 3.68 |
hadoop_IOStatisticsBinding_emptyStatisticsStore | /**
* Get the shared instance of the immutable empty statistics
* store.
* @return an empty statistics object.
*/
public static IOStatisticsStore emptyStatisticsStore() {
return EmptyIOStatisticsStore.getInstance();
} | 3.68 |
hbase_PersistentIOEngine_calculateChecksum | /**
* Using an encryption algorithm to calculate a checksum, the default encryption algorithm is MD5
* @return the checksum which is convert to HexString
* @throws IOException something happened like file not exists
* @throws NoSuchAlgorithmException no such algorithm
*/
protected byte[] calculateChec... | 3.68 |
flink_NetUtils_socketAddressToUrlString | /**
* Encodes an IP address and port to be included in URL. in particular, this method makes sure
* that IPv6 addresses have the proper formatting to be included in URLs.
*
* @param address The socket address with the IP address and port.
* @return The proper URL string encoded IP address and port.
*/
public stat... | 3.68 |
hbase_HRegion_registerService | /**
* Registers a new protocol buffer {@link Service} subclass as a coprocessor endpoint to be
* available for handling {@link #execService(RpcController, CoprocessorServiceCall)} calls.
* <p/>
* Only a single instance may be registered per region for a given {@link Service} subclass (the
* instances are keyed on ... | 3.68 |
flink_SqlNodeConvertUtils_validateAlterView | /**
* Validate the view to alter is valid and existed and return the {@link CatalogView} to alter.
*/
static CatalogView validateAlterView(SqlAlterView alterView, ConvertContext context) {
UnresolvedIdentifier unresolvedIdentifier =
UnresolvedIdentifier.of(alterView.fullViewName());
ObjectIdentifi... | 3.68 |
hbase_ProcedureExecutor_isRunning | // ==========================================================================
// Accessors
// ==========================================================================
public boolean isRunning() {
return running.get();
} | 3.68 |
framework_ContainerHierarchicalWrapper_getItemIds | /*
* Gets the ID's of all Items stored in the Container Don't add a JavaDoc
* comment here, we use the default documentation from implemented
* interface.
*/
@Override
public Collection<?> getItemIds() {
return container.getItemIds();
} | 3.68 |
framework_PropertysetItem_hashCode | /*
* (non-Javadoc)
*
* @see java.lang.Object#hashCode()
*/
@Override
public int hashCode() {
return (list == null ? 0 : list.hashCode())
^ (map == null ? 0 : map.hashCode())
^ ((propertySetChangeListeners == null
|| propertySetChangeListeners.isEmpty()) ? 0
... | 3.68 |
flink_DeltaIterationBase_getSolutionSet | /**
* Gets the contract that represents the solution set for the step function.
*
* @return The solution set for the step function.
*/
public Operator getSolutionSet() {
return this.solutionSetPlaceholder;
} | 3.68 |
hadoop_AMRMProxyService_recover | /**
* Recover from NM state store. Called after serviceInit before serviceStart.
*
* @throws IOException if recover fails
*/
public void recover() throws IOException {
LOG.info("Recovering AMRMProxyService.");
RecoveredAMRMProxyState state =
this.nmContext.getNMStateStore().loadAMRMProxyState();
this.... | 3.68 |
open-banking-gateway_AuthSessionHandler_createNewAuthSessionAndEnhanceResult | /**
* Creates new authorization session associated with the request.
* @param request Request to associate session with.
* @param sessionKey Authorization session encryption key.
* @param context Service context for the request
* @param result Protocol response that required to open the session
* @param <O> Outco... | 3.68 |
flink_RocksDBStateDownloader_transferAllStateDataToDirectory | /**
* Transfer all state data to the target directory, as specified in the download requests.
*
* @param downloadRequests the list of downloads.
* @throws Exception If anything about the download goes wrong.
*/
public void transferAllStateDataToDirectory(
Collection<StateHandleDownloadSpec> downloadRequest... | 3.68 |
hbase_MetricsConnection_getServerStats | /** serverStats metric */
public ConcurrentHashMap<ServerName, ConcurrentMap<byte[], RegionStats>> getServerStats() {
return serverStats;
} | 3.68 |
hbase_ServerNonceManager_endOperation | /**
* Ends the operation started by startOperation.
* @param group Nonce group.
* @param nonce Nonce.
* @param success Whether the operation has succeeded.
*/
public void endOperation(long group, long nonce, boolean success) {
if (nonce == HConstants.NO_NONCE) return;
NonceKey nk = new NonceKey(group, nonc... | 3.68 |
hadoop_ComponentContainers_containers | /**
* Sets the containers.
* @param compContainers containers of the component.
*/
public ComponentContainers containers(List<Container> compContainers) {
this.containers = compContainers;
return this;
} | 3.68 |
dubbo_AdaptiveClassCodeGenerator_generateUrlAssignmentIndirectly | /**
* get parameter with type <code>URL</code> from method parameter:
* <p>
* test if parameter has method which returns type <code>URL</code>
* <p>
* if not found, throws IllegalStateException
*/
private String generateUrlAssignmentIndirectly(Method method) {
Class<?>[] pts = method.getParameterTypes();
... | 3.68 |
hbase_DependentColumnFilter_parseFrom | /**
* Parse a seralized representation of {@link DependentColumnFilter}
* @param pbBytes A pb serialized {@link DependentColumnFilter} instance
* @return An instance of {@link DependentColumnFilter} made from <code>bytes</code>
* @throws DeserializationException if an error occurred
* @see #toByteArray
*/
public ... | 3.68 |
hudi_Triple_equals | /**
* <p>
* Compares this triple to another based on the three elements.
* </p>
*
* @param obj the object to compare to, null returns false
* @return true if the elements of the triple are equal
*/
@SuppressWarnings("deprecation") // ObjectUtils.equals(Object, Object) has been deprecated in 3.2
@Override
public ... | 3.68 |
flink_JobGraph_getVertices | /**
* Returns an Iterable to iterate all vertices registered with the job graph.
*
* @return an Iterable to iterate all vertices registered with the job graph
*/
public Iterable<JobVertex> getVertices() {
return this.taskVertices.values();
} | 3.68 |
hadoop_ByteArray_buffer | /**
* @return the underlying buffer.
*/
@Override
public byte[] buffer() {
return buffer;
} | 3.68 |
hbase_SingleColumnValueExcludeFilter_areSerializedFieldsEqual | /**
* Returns true if and only if the fields of the filter that are serialized are equal to the
* corresponding fields in other. Used for testing.
*/
@Override
boolean areSerializedFieldsEqual(Filter o) {
if (o == this) {
return true;
}
if (!(o instanceof SingleColumnValueExcludeFilter)) {
return false... | 3.68 |
hadoop_FedBalanceContext_build | /**
* Build the FedBalanceContext.
*
* @return the FedBalanceContext obj.
*/
public FedBalanceContext build() {
FedBalanceContext context = new FedBalanceContext();
context.src = this.src;
context.dst = this.dst;
context.mount = this.mount;
context.conf = this.conf;
context.forceCloseOpenFiles = this.fo... | 3.68 |
hbase_KeyValue_compareIgnoringPrefix | /**
* Overridden
* @param commonPrefix location of expected common prefix
* @param left the left kv serialized byte[] to be compared with
* @param loffset the offset in the left byte[]
* @param llength the length in the left byte[]
* @param right the right kv serialized byte[] to be compa... | 3.68 |
hudi_CleanActionExecutor_clean | /**
* Performs cleaning of partition paths according to cleaning policy and returns the number of files cleaned. Handles
* skews in partitions to clean by making files to clean as the unit of task distribution.
*
* @throws IllegalArgumentException if unknown cleaning policy is provided
*/
List<HoodieCleanStat> cle... | 3.68 |
hadoop_ServiceLauncher_getServiceName | /**
* Get the service name via {@link Service#getName()}.
*
* If the service is not instantiated, the classname is returned instead.
* @return the service name
*/
public String getServiceName() {
Service s = service;
String name = null;
if (s != null) {
try {
name = s.getName();
} catch (Except... | 3.68 |
querydsl_MathExpressions_round | /**
* Round to s decimal places
*
* @param num numeric expression
* @param s decimal places
* @return round(num, s)
*/
public static <A extends Number & Comparable<?>> NumberExpression<A> round(Expression<A> num, int s) {
return Expressions.numberOperation(num.getType(), MathOps.ROUND2, num, ConstantImpl.crea... | 3.68 |
hadoop_IdentityTransformer_transformAclEntriesForSetRequest | /**
* Perform Identity transformation when calling setAcl(),removeAclEntries() and modifyAclEntries()
* If the AclEntry type is a user or group, and its name is one of the following:
* 1.short name; 2.$superuser; 3.Fully qualified name; 4. principal id.
* <pre>
* Short name could be transformed to:
* - A servi... | 3.68 |
pulsar_AbstractMetrics_createMetricsByDimension | /**
* Creates a dimension key for replication metrics.
*
* @param namespace
* @param fromClusterName
* @param toClusterName
* @return
*/
protected Metrics createMetricsByDimension(String namespace, String fromClusterName, String toClusterName) {
Map<String, String> dimensionMap = new HashMap<>();
dimens... | 3.68 |
zxing_FinderPatternFinder_foundPatternDiagonal | /**
* @param stateCount count of black/white/black/white/black pixels just read
* @return true iff the proportions of the counts is close enough to the 1/1/3/1/1 ratios
* used by finder patterns to be considered a match
*/
protected static boolean foundPatternDiagonal(int[] stateCount) {
int totalModuleSi... | 3.68 |
querydsl_SQLExpressions_datediff | /**
* Get a datediff(unit, start, end) expression
*
* @param unit date part
* @param start start
* @param end end
* @return difference in units
*/
public static <D extends Comparable> NumberExpression<Integer> datediff(DatePart unit,
DateTimeExpression<D> start, D end) {
return Expressions.numberOper... | 3.68 |
flink_TableConfig_getConfiguration | /**
* Gives direct access to the underlying application-specific key-value map for advanced
* configuration.
*/
public Configuration getConfiguration() {
return configuration;
} | 3.68 |
hadoop_IdentifierResolver_setOutputReaderClass | /**
* Sets the {@link OutputReader} class.
*/
protected void setOutputReaderClass(Class<? extends OutputReader>
outputReaderClass) {
this.outputReaderClass = outputReaderClass;
} | 3.68 |
flink_ExecutionConfig_getRegisteredTypesWithKryoSerializers | /** Returns the registered types with Kryo Serializers. */
public LinkedHashMap<Class<?>, SerializableSerializer<?>>
getRegisteredTypesWithKryoSerializers() {
return registeredTypesWithKryoSerializers;
} | 3.68 |
framework_GridLayoutElement_getColumnCount | /**
* Gets the total number of columns in the layout.
*
* @return the number of columns in the layout
* @since 8.0.6
*/
public long getColumnCount() {
Long res = (Long) getCommandExecutor()
.executeScript("return arguments[0].getColumnCount()", this);
if (res == null) {
throw new Illega... | 3.68 |
pulsar_AdminResource_isRedirectException | /**
* Check current exception whether is redirect exception.
*
* @param ex The throwable.
* @return Whether is redirect exception
*/
protected static boolean isRedirectException(Throwable ex) {
Throwable realCause = FutureUtil.unwrapCompletionException(ex);
return realCause instanceof WebApplicationExcepti... | 3.68 |
flink_PekkoInvocationHandler_tell | /**
* Sends the message to the RPC endpoint.
*
* @param message to send to the RPC endpoint.
*/
protected void tell(Object message) {
rpcEndpoint.tell(message, ActorRef.noSender());
} | 3.68 |
hadoop_SchedulingResponse_isSuccess | /**
* Returns true if Scheduler was able to accept and commit this request.
* @return isSuccessful.
*/
public boolean isSuccess() {
return this.isSuccess;
} | 3.68 |
hadoop_DistributedCache_getTimestamp | /**
* Returns mtime of a given cache file on hdfs. Internal to MapReduce.
* @param conf configuration
* @param cache cache file
* @return mtime of a given cache file on hdfs
* @throws IOException
*/
@Deprecated
public static long getTimestamp(Configuration conf, URI cache)
throws IOException {
return getFileS... | 3.68 |
flink_ServiceType_buildUpExternalRestService | /**
* Build up the external rest service template, according to the jobManager parameters.
*
* @param kubernetesJobManagerParameters the parameters of jobManager.
* @return the external rest service
*/
public Service buildUpExternalRestService(
KubernetesJobManagerParameters kubernetesJobManagerParameters)... | 3.68 |
framework_Form_bindPropertyToField | /**
* Binds an item property to a field. The default behavior is to bind
* property straight to Field. If Property.Viewer type property (e.g.
* PropertyFormatter) is already set for field, the property is bound to
* that Property.Viewer.
*
* @param propertyId
* @param property
* @param field
* @since 6.7.3
*/... | 3.68 |
pulsar_FunctionMetaDataManager_acquireLeadership | /**
* Called by the leader service when this worker becomes the leader.
* We first get exclusive producer on the metadata topic. Next we drain the tailer
* to ensure that we have caught up to metadata topic. After which we close the tailer.
* Note that this method cannot be syncrhonized because the tailer might sti... | 3.68 |
framework_VaadinUriResolver_resolveVaadinUri | /**
* Translates a Vaadin URI to a URL that can be loaded by the browser. The
* following URI schemes are supported:
* <ul>
* <li><code>theme://</code> - resolves to the URL of the currently active
* theme.</li>
* <li><code>published://</code> - resolves to resources on the classpath
* published by {@link com.va... | 3.68 |
pulsar_ResourceUnitRanking_getEstimatedMessageRate | /**
* Get the estimated message rate.
*/
public double getEstimatedMessageRate() {
return this.estimatedMessageRate;
} | 3.68 |
hadoop_JavaCommandLineBuilder_setJVMOpts | /**
* Set JVM opts.
* @param jvmOpts JVM opts
*/
public void setJVMOpts(String jvmOpts) {
if (ServiceUtils.isSet(jvmOpts)) {
add(jvmOpts);
}
} | 3.68 |
flink_LogicalScopeProvider_castFrom | /**
* Casts the given metric group to a {@link LogicalScopeProvider}, if it implements the
* interface.
*
* @param metricGroup metric group to cast
* @return cast metric group
* @throws IllegalStateException if the metric group did not implement the LogicalScopeProvider
* interface
*/
static LogicalScopePro... | 3.68 |
flink_GSChecksumWriteChannel_write | /**
* Writes bytes to the underlying channel and updates checksum.
*
* @param content The content to write
* @param start The start position
* @param length The number of bytes to write
* @return The number of bytes written
* @throws IOException On underlying failure
*/
public int write(byte[] content, int star... | 3.68 |
hadoop_SampleQuantiles_insertBatch | /**
* Merges items from buffer into the samples array in one pass.
* This is more efficient than doing an insert on every item.
*/
private void insertBatch() {
if (bufferCount == 0) {
return;
}
Arrays.sort(buffer, 0, bufferCount);
// Base case: no samples
int start = 0;
if (samples.size() == 0) {
... | 3.68 |
framework_ImmediateUpload_setup | /*
* (non-Javadoc)
*
* @see com.vaadin.tests.components.AbstractTestUI#setup(com.vaadin.server.
* VaadinRequest)
*/
@Override
protected void setup(VaadinRequest request) {
// by default is in immediate mode (since 8.0)
Upload immediateUpload = new Upload();
immediateUpload.setId("immediateupload");
... | 3.68 |
zxing_AddressBookParsedResult_getBirthday | /**
* @return birthday formatted as yyyyMMdd (e.g. 19780917)
*/
public String getBirthday() {
return birthday;
} | 3.68 |
flink_CompositeType_createComparator | /**
* Generic implementation of the comparator creation. Composite types are supplying the
* infrastructure to create the actual comparators
*
* @return The comparator
*/
@PublicEvolving
public TypeComparator<T> createComparator(
int[] logicalKeyFields,
boolean[] orders,
int logicalFieldOff... | 3.68 |
hbase_HRegion_close | /**
* Close down this HRegion. Flush the cache unless abort parameter is true, Shut down each HStore,
* don't service any more calls. This method could take some time to execute, so don't call it
* from a time-sensitive thread.
* @param abort true if server is aborting (only during testing)
* @param ignor... | 3.68 |
morf_SqlUtils_asString | /**
* Returns a SQL DSL expression to return the field CASTed to
* a string of the specified length
*
* @param length length of the string cast
* @return {@link Cast} as string of given length
*/
public Cast asString(int length) {
return asType(DataType.STRING, length);
} | 3.68 |
hadoop_SinglePendingCommit_getLength | /**
* Destination file size.
* @return size of destination object
*/
public long getLength() {
return length;
} | 3.68 |
dubbo_NacosRegistry_getServiceNames | /**
* Get the service names from the specified {@link URL url}
*
* @param url {@link URL}
* @param listener {@link NotifyListener}
* @return non-null
*/
private Set<String> getServiceNames(URL url, NacosAggregateListener listener) {
if (isAdminProtocol(url)) {
scheduleServiceNamesLookup(url, list... | 3.68 |
hibernate-validator_SizeValidatorForArraysOfFloat_isValid | /**
* Checks the number of entries in an array.
*
* @param array The array to validate.
* @param constraintValidatorContext context in which the constraint is evaluated.
*
* @return Returns {@code true} if the array is {@code null} or the number of entries in
* {@code array} is between the specified {@co... | 3.68 |
flink_SegmentsUtil_byteIndex | /**
* Given a bit index, return the byte index containing it.
*
* @param bitIndex the bit index.
* @return the byte index.
*/
private static int byteIndex(int bitIndex) {
return bitIndex >>> ADDRESS_BITS_PER_WORD;
} | 3.68 |
druid_ZookeeperNodeRegister_register | /**
* Register a Node which has a Properties as the payload.
* <pre>
* CAUTION: only one node can be registered,
* if you want to register another one,
* call deregister first
* </pre>
*
* @param payload The information used to generate the payload Properties
* @return true, register successf... | 3.68 |
hadoop_TemporaryAWSCredentialsProvider_createCredentials | /**
* The credentials here must include a session token, else this operation
* will raise an exception.
* @param config the configuration
* @return temporary credentials.
* @throws IOException on any failure to load the credentials.
* @throws NoAuthWithAWSException validation failure
* @throws NoAwsCredentialsEx... | 3.68 |
morf_DatabaseMetaDataProvider_loadTable | /**
* Loads a table.
*
* @param tableName Name of the table.
* @return The table metadata.
*/
protected Table loadTable(AName tableName) {
final RealName realTableName = tableNames.get().get(tableName);
if (realTableName == null) {
throw new IllegalArgumentException("Table [" + tableName + "] not found.")... | 3.68 |
framework_Slot_getSpacingResizeListener | /**
* Returns the spacing element resize listener for this slot if one has been
* set.
*
* @return the listener or {@code null} if not set
*/
public ElementResizeListener getSpacingResizeListener() {
return spacingResizeListener;
} | 3.68 |
framework_DDEventHandleStrategy_restoreDragImage | /**
* Restores drag image after temporary update by
* {@link #updateDragImage(NativePreviewEvent, DDManagerMediator)}.
*
* @param originalImageDisplay
* original "display" CSS style property of drag image element
* @param mediator
* VDragAndDropManager data accessor
* @param event
* ... | 3.68 |
flink_ResolvedCatalogBaseTable_getSchema | /**
* @deprecated This method returns the deprecated {@link TableSchema} class. The old class was a
* hybrid of resolved and unresolved schema information. It has been replaced by the new
* {@link ResolvedSchema} which is resolved by the framework and accessible via {@link
* #getResolvedSchema()}.
*/
@... | 3.68 |
hbase_SplitTableRegionProcedure_rollbackState | /**
* To rollback {@link SplitTableRegionProcedure}, an AssignProcedure is asynchronously submitted
* for parent region to be split (rollback doesn't wait on the completion of the AssignProcedure)
* . This can be improved by changing rollback() to support sub-procedures. See HBASE-19851 for
* details.
*/
@Override... | 3.68 |
dubbo_AbstractServiceDiscovery_doUpdate | /**
* Update Service Instance. Unregister and then register by default.
* Can be override if registry support update instance directly.
* <br/>
* NOTICE: Remind to update {@link AbstractServiceDiscovery#serviceInstance}'s reference if updated
* and report metadata by {@link AbstractServiceDiscovery#reportMetadata(... | 3.68 |
hadoop_MutableInverseQuantiles_getQuantiles | /**
* Returns the array of Inverse Quantiles declared in MutableInverseQuantiles.
*
* @return array of Inverse Quantiles
*/
public synchronized Quantile[] getQuantiles() {
return INVERSE_QUANTILES;
} | 3.68 |
hibernate-validator_NotEmptyValidatorForArraysOfBoolean_isValid | /**
* Checks the array is not {@code null} and not empty.
*
* @param array the array to validate
* @param constraintValidatorContext context in which the constraint is evaluated
* @return returns {@code true} if the array is not {@code null} and the array is not empty
*/
@Override
public boolean isValid(boolean[]... | 3.68 |
hbase_CommonFSUtils_getNamespaceDir | /**
* Returns the {@link org.apache.hadoop.fs.Path} object representing the namespace directory under
* path rootdir
* @param rootdir qualified path of HBase root directory
* @param namespace namespace name
* @return {@link org.apache.hadoop.fs.Path} for table
*/
public static Path getNamespaceDir(Path rootdir,... | 3.68 |
rocketmq-connect_WorkerSinkTask_execute | /**
* execute poll and send record
*/
@Override
protected void execute() {
while (isRunning()) {
try {
long startTimeStamp = System.currentTimeMillis();
log.info("START pullMessageFromQueues, time started : {}", startTimeStamp);
if (messageQueues.size() == 0) {
... | 3.68 |
pulsar_FileUtils_deleteFilesInDirectory | /**
* Deletes all files (not directories) in the given directory (recursive)
* that match the given filename filter. If any file cannot be deleted then
* this is printed at warn to the given logger.
*
* @param directory to delete contents of
* @param filter if null then no filter is used
* @param logger to notif... | 3.68 |
flink_OperationExecutorFactory_getDecimalDigits | /**
* The number of fractional digits for this type. Null is returned for data types where this is
* not applicable.
*/
private static @Nullable Integer getDecimalDigits(LogicalType columnType) {
switch (columnType.getTypeRoot()) {
case BOOLEAN:
case TINYINT:
case SMALLINT:
case I... | 3.68 |
hadoop_ApplicationMaster_addAsLocalResourceFromEnv | /**
* Add the given resource into the map of resources, using information from
* the supplied environment variables.
*
* @param resource The resource to add.
* @param localResources Map of local resources to insert into.
* @param env Map of environment variables.
*/
public void addAsLocalResourceFromEnv(DynoReso... | 3.68 |
framework_Window_addCloseListener | /**
* Adds a CloseListener to the window.
*
* For a window the CloseListener is fired when the user closes it (clicks
* on the close button).
*
* For a browser level window the CloseListener is fired when the browser
* level window is closed. Note that closing a browser level window does not
* mean it will be d... | 3.68 |
flink_FromClasspathEntryClassInformationProvider_getJarFile | /**
* Always returns an empty {@code Optional} because this implementation relies on the JAR
* archive being available on either the user or the system classpath.
*
* @return An empty {@code Optional}.
*/
@Override
public Optional<File> getJarFile() {
return Optional.empty();
} | 3.68 |
pulsar_ReplicatedSubscriptionSnapshotCache_advancedMarkDeletePosition | /**
* Signal that the mark-delete position on the subscription has been advanced. If there is a snapshot that
* correspond to this position, it will returned, other it will return null.
*/
public synchronized ReplicatedSubscriptionsSnapshot advancedMarkDeletePosition(PositionImpl pos) {
ReplicatedSubscriptionsSn... | 3.68 |
framework_LayoutManager_getInnerHeightDouble | /**
* Gets the inner height (excluding margins, paddings and borders) of the
* given element, provided that it has been measured. These elements are
* guaranteed to be measured:
* <ul>
* <li>ManagedLayouts and their child Connectors
* <li>Elements for which there is at least one ElementResizeListener
* <li>Eleme... | 3.68 |
hadoop_ManifestCommitterSupport_getAppAttemptId | /**
* Get the Application Attempt Id for this job
* by looking for {@link MRJobConfig#APPLICATION_ATTEMPT_ID}
* in the configuration, falling back to 0 if unset.
* For spark it will always be 0, for MR it will be set in the AM
* to the {@code ApplicationAttemptId} the AM is launched with.
* @param conf job config... | 3.68 |
morf_SelectStatement_except | /**
* Perform a EXCEPT set operation with another {@code selectStatement},
* eliminating any rows from the top select statement which exist in the bottom
* select statement.
* <p>
* If an except operation is performed then all participating select statements
* require the same selected column list.
* </p>
*
* ... | 3.68 |
dubbo_Converter_getSourceType | /**
* Get the source type
*
* @return non-null
*/
default Class<S> getSourceType() {
return findActualTypeArgument(getClass(), Converter.class, 0);
} | 3.68 |
flink_HadoopOutputFormatBase_open | /**
* create the temporary output file for hadoop RecordWriter.
*
* @param taskNumber The number of the parallel instance.
* @param numTasks The number of parallel tasks.
* @throws java.io.IOException
*/
@Override
public void open(int taskNumber, int numTasks) throws IOException {
// enforce sequential open(... | 3.68 |
framework_VaadinService_addSessionInitListener | /**
* Adds a listener that gets notified when a new Vaadin service session is
* initialized for this service.
* <p>
* Because of the way different service instances share the same session,
* the listener is not necessarily notified immediately when the session is
* created but only when the first request for that... | 3.68 |
framework_DragSourceExtension_addDragStartListener | /**
* Attaches dragstart listener for the current drag source.
* {@link DragStartListener#dragStart(DragStartEvent)} is called when
* dragstart event happens on the client side.
*
* @param listener
* Listener to handle dragstart event.
* @return Handle to be used to remove this listener.
*/
public Re... | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.