name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
hbase_RegionServerObserver_postCreateReplicationEndPoint | /**
* This will be called after the replication endpoint is instantiated.
* @param ctx the environment to interact with the framework and region server.
* @param endpoint - the base endpoint for replication
* @return the endpoint to use during replication.
*/
default ReplicationEndpoint postCreateReplicationE... | 3.68 |
framework_VScrollTable_resizeSortedColumnForSortIndicator | /** For internal use only. May be removed or replaced in the future. */
public void resizeSortedColumnForSortIndicator() {
// Force recalculation of the captionContainer element inside the header
// cell to accommodate for the size of the sort arrow.
HeaderCell sortedHeader = tHead.getHeaderCell(sortColumn)... | 3.68 |
hadoop_AbfsClientThrottlingIntercept_initializeSingleton | /**
* Creates a singleton object of the AbfsClientThrottlingIntercept.
* which is shared across all filesystem instances.
* @param abfsConfiguration configuration set.
* @return singleton object of intercept.
*/
static AbfsClientThrottlingIntercept initializeSingleton(AbfsConfiguration abfsConfiguration) {
if (s... | 3.68 |
pulsar_BrokersBase_closeAndReCheck | /**
* Close producer and reader and then to re-check if this operation is success.
*
* Re-check
* - Producer: If close fails we will print error log to notify user.
* - Consumer: If close fails we will force delete subscription.
*
* @param producer Producer
* @param reader Reader
* @param topic Topic
* @para... | 3.68 |
framework_Overlay_getOverlayContainer | /**
* Gets the 'overlay container' element.
*
* @return the overlay container element
*/
public com.google.gwt.user.client.Element getOverlayContainer() {
return RootPanel.get().getElement();
} | 3.68 |
hbase_HFileBlock_startWriting | /**
* Starts writing into the block. The previous block's data is discarded.
* @return the stream the user can write their data into
*/
DataOutputStream startWriting(BlockType newBlockType) throws IOException {
if (state == State.BLOCK_READY && startOffset != -1) {
// We had a previous block that was written t... | 3.68 |
flink_Conditions_haveLeafArgumentTypes | /**
* Tests leaf argument types of a method against the given predicate.
*
* <p>See {@link #haveLeafTypes(DescribedPredicate)} for details.
*/
public static ArchCondition<JavaMethod> haveLeafArgumentTypes(
DescribedPredicate<JavaClass> typePredicate) {
return new ArchCondition<JavaMethod>(
"... | 3.68 |
hadoop_ConnectionContext_isClosed | /**
* Check if the connection is closed.
*
* @return If the connection is closed.
*/
public synchronized boolean isClosed() {
return this.closed;
} | 3.68 |
hbase_HRegion_compact | /**
* We are trying to remove / relax the region read lock for compaction. Let's see what are the
* potential race conditions among the operations (user scan, region split, region close and
* region bulk load). user scan ---> region read lock region split --> region close first -->
* region write lock region close ... | 3.68 |
hbase_TableMapReduceUtil_updateMap | /**
* Add entries to <code>packagedClasses</code> corresponding to class files contained in
* <code>jar</code>.
* @param jar The jar who's content to list.
* @param packagedClasses map[class -> jar]
*/
private static void updateMap(String jar, Map<String, String> packagedClasses)
throws IOException {... | 3.68 |
framework_ContainerHierarchicalWrapper_getType | /*
* Gets the data type of all Properties identified by the given Property ID.
* Don't add a JavaDoc comment here, we use the default documentation from
* implemented interface.
*/
@Override
public Class<?> getType(Object propertyId) {
return container.getType(propertyId);
} | 3.68 |
pulsar_InetAddressUtils_isIPv6Address | /**
* Checks whether the parameter is a valid IPv6 address (including compressed).
*
* @param input the address string to check for validity
* @return true if the input parameter is a valid standard or compressed IPv6 address
*/
public static boolean isIPv6Address(final String input) {
return isIPv6StdAddress(... | 3.68 |
morf_DeleteStatement_getWhereCriterion | /**
* Gets the where criteria.
*
* @return the where criteria
*/
public Criterion getWhereCriterion() {
return whereCriterion;
} | 3.68 |
flink_BlobOutputStream_sendPutHeader | /**
* Constructs and writes the header data for a PUT request to the given output stream.
*
* @param outputStream the output stream to write the PUT header data to
* @param jobId the ID of job the BLOB belongs to (or <tt>null</tt> if job-unrelated)
* @param blobType whether the BLOB should become permanent or tran... | 3.68 |
framework_Embedded_getAlternateText | /**
* Gets this component's "alt-text".
*
* @see #setAlternateText(String)
*/
public String getAlternateText() {
return getState(false).altText;
} | 3.68 |
hadoop_NodePlan_setURI | /**
* Sets the DataNodeURI.
*
* @param dataNodeName - String
*/
public void setURI(String dataNodeName) {
this.nodeName = dataNodeName;
} | 3.68 |
hudi_HoodieMetrics_reportMetrics | /**
* Given a commit action, metrics name and value this method reports custom metrics.
*/
public void reportMetrics(String commitAction, String metricName, long value) {
metrics.registerGauge(getMetricsName(commitAction, metricName), value);
} | 3.68 |
hbase_RegionNormalizerWorkQueue_putFirst | /**
* Inserts the specified element at the head of the queue.
* @param e the element to add
*/
public void putFirst(E e) {
if (e == null) {
throw new NullPointerException();
}
putAllFirst(Collections.singleton(e));
} | 3.68 |
hadoop_AuditingIntegration_updateCommonContextOnCommitterExit | /**
* Remove commit info at the end of the task or job.
*/
public static void updateCommonContextOnCommitterExit() {
currentAuditContext().remove(PARAM_JOB_ID);
currentAuditContext().remove(CONTEXT_ATTR_TASK_ATTEMPT_ID);
} | 3.68 |
morf_HumanReadableStatementProducer_addIndex | /** @see org.alfasoftware.morf.upgrade.SchemaEditor#addIndex(java.lang.String, org.alfasoftware.morf.metadata.Index) **/
@Override
public void addIndex(String tableName, Index index) {
consumer.schemaChange(HumanReadableStatementHelper.generateAddIndexString(tableName, index));
} | 3.68 |
pulsar_Transactions_getPendingAckStats | /**
* Get transaction pending ack stats.
*
* @param topic the topic of this transaction pending ack stats
* @param subName the subscription name of this transaction pending ack stats
* @return the stats of transaction pending ack.
*/
default TransactionPendingAckStats getPendingAckStats(String topic, String subNa... | 3.68 |
flink_StreamProjection_projectTuple16 | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15>
SingleOutputStreamOperator<
Tuple16<
... | 3.68 |
querydsl_Expressions_numberOperation | /**
* Create a new Operation expression
*
* @param type type of expression
* @param operator operator
* @param args operation arguments
* @return operation expression
*/
public static <T extends Number & Comparable<?>> NumberOperation<T> numberOperation(Class<? extends T> type,
Operator operator, Express... | 3.68 |
hbase_BucketAllocator_freeBytes | /**
* How many more bytes can be allocated from the currently claimed blocks of this bucket size
*/
public long freeBytes() {
return freeCount * itemSize;
} | 3.68 |
flink_CopyOnWriteStateMap_incrementalRehash | /** Runs a number of steps for incremental rehashing. */
@SuppressWarnings("unchecked")
private void incrementalRehash() {
StateMapEntry<K, N, S>[] oldMap = primaryTable;
StateMapEntry<K, N, S>[] newMap = incrementalRehashTable;
int oldCapacity = oldMap.length;
int newMask = newMap.length - 1;
int... | 3.68 |
pulsar_ConcurrentOpenHashMap_forEach | /**
* Iterate over all the entries in the map and apply the processor function to each of them.
* <p>
* <b>Warning: Do Not Guarantee Thread-Safety.</b>
* @param processor the function to apply to each entry
*/
public void forEach(BiConsumer<? super K, ? super V> processor) {
for (int i = 0; i < sections.length... | 3.68 |
flink_OuterJoinPaddingUtil_padLeft | /**
* Returns a padding result with the given left row.
*
* @param leftRow the left row to pad
* @return the reusable null padding result
*/
public final RowData padLeft(RowData leftRow) {
return joinedRow.replace(leftRow, rightNullPaddingRow);
} | 3.68 |
hadoop_JWTRedirectAuthenticationHandler_validateSignature | /**
* Verify the signature of the JWT token in this method. This method depends
* on the public key that was established during init based upon the
* provisioned public key. Override this method in subclasses in order to
* customize the signature verification behavior.
*
* @param jwtToken the token that contains ... | 3.68 |
framework_AbstractListing_deserializeDeclarativeRepresentation | /**
* Deserializes a string to a data item.
* <p>
* Default implementation is able to handle only {@link String} as an item
* type. There will be a {@link ClassCastException} if {@code T } is not a
* {@link String}.
*
* @see #serializeDeclarativeRepresentation(Object)
*
* @param item
* string to de... | 3.68 |
hbase_HFileBlock_allocateBufferForUnpacking | /**
* Always allocates a new buffer of the correct size. Copies header bytes from the existing
* buffer. Does not change header fields. Reserve room to keep checksum bytes too.
*/
private ByteBuff allocateBufferForUnpacking() {
int headerSize = headerSize();
int capacityNeeded = headerSize + uncompressedSizeWith... | 3.68 |
hmily_HmilyLockCacheManager_removeByKey | /**
* remove guava cache by key.
*
* @param lockId guava cache key.
*/
public void removeByKey(final String lockId) {
if (Objects.nonNull(lockId)) {
loadingCache.invalidate(lockId);
}
} | 3.68 |
hadoop_FilterFileSystem_listLocatedStatus | /** List files and its block locations in a directory. */
@Override
public RemoteIterator<LocatedFileStatus> listLocatedStatus(Path f)
throws IOException {
return fs.listLocatedStatus(f);
} | 3.68 |
framework_CustomizedSystemMessages_setCommunicationErrorNotificationEnabled | /**
* Enables or disables the notification. If disabled, the set URL (or
* current) is loaded directly.
*
* @param communicationErrorNotificationEnabled
* true = enabled, false = disabled
*/
public void setCommunicationErrorNotificationEnabled(
boolean communicationErrorNotificationEnabled) {
... | 3.68 |
framework_RenderInformation_updateSize | /**
* Update the size of the widget.
*
* @param widget
*
* @return true if the size has changed since last update
*
* @since 7.2
*/
public boolean updateSize(Element element) {
return updateSize(DOM.asOld(element));
} | 3.68 |
hbase_RestoreSnapshotProcedure_updateTableDescriptor | /**
* Update descriptor
* @param env MasterProcedureEnv
**/
private void updateTableDescriptor(final MasterProcedureEnv env) throws IOException {
env.getMasterServices().getTableDescriptors().update(modifiedTableDescriptor);
} | 3.68 |
pulsar_MLTransactionSequenceIdGenerator_onManagedLedgerLastLedgerInitialize | // When we don't roll over ledger, we can init sequenceId from the getLastAddConfirmed transaction metadata entry
@Override
public CompletableFuture<Void> onManagedLedgerLastLedgerInitialize(String name, LedgerHandle lh) {
CompletableFuture<Void> promise = new CompletableFuture<>();
if (lh.getLastAddConfirmed()... | 3.68 |
hadoop_StageConfig_getConf | /**
* Get configuration.
* @return the configuration
*/
public Configuration getConf() {
return conf;
} | 3.68 |
framework_BaseLayoutTestUI_getTicketNumber | /*
* (non-Javadoc)
*
* @see com.vaadin.tests.components.AbstractTestUI#getTicketNumber()
*/
@Override
protected Integer getTicketNumber() {
// TODO Auto-generated method stub
return null;
} | 3.68 |
hbase_KeyValueTestUtil_containsIgnoreMvccVersion | /**
* Checks whether KeyValues from kvCollection2 are contained in kvCollection1. The comparison is
* made without distinguishing MVCC version of the KeyValues
* @return true if KeyValues from kvCollection2 are contained in kvCollection1
*/
public static boolean containsIgnoreMvccVersion(Collection<? extends Cell> ... | 3.68 |
hibernate-validator_AnnotationProxy_equals | /**
* Performs an equality check as described in {@link Annotation#equals(Object)}.
*
* @param obj The object to compare
*
* @return Whether the given object is equal to this annotation proxy or not
*
* @see Annotation#equals(Object)
*/
@Override
public boolean equals(Object obj) {
if ( this == obj ) {
retur... | 3.68 |
flink_TopologyGraph_unlink | /**
* Remove the edge from `from` node to `to` node. If there is no edge between them then do
* nothing.
*/
void unlink(ExecNode<?> from, ExecNode<?> to) {
TopologyNode fromNode = getOrCreateTopologyNode(from);
TopologyNode toNode = getOrCreateTopologyNode(to);
fromNode.outputs.remove(toNode);
toNod... | 3.68 |
morf_TableNameDecorator_getName | /**
* @see org.alfasoftware.morf.metadata.Table#getName()
*/
@Override
public String getName() {
return name;
} | 3.68 |
flink_LogicalType_is | /**
* Returns whether the family type of the type equals to the {@code family} or not.
*
* @param family The family type to check against for equality
*/
public boolean is(LogicalTypeFamily family) {
return typeRoot.getFamilies().contains(family);
} | 3.68 |
flink_SourceCoordinatorContext_unregisterSourceReader | /**
* Unregister a source reader.
*
* @param subtaskId the subtask id of the source reader.
* @param attemptNumber the attempt number of the source reader.
*/
void unregisterSourceReader(int subtaskId, int attemptNumber) {
final Map<Integer, ReaderInfo> attemptReaders = registeredReaders.get(subtaskId);
if... | 3.68 |
pulsar_PulsarConfigurationLoader_create | /**
* Creates PulsarConfiguration and loads it with populated attribute values from provided Properties object.
*
* @param properties The properties to populate the attributed from
* @throws IOException
* @throws IllegalArgumentException
*/
@SuppressWarnings({ "rawtypes", "unchecked" })
public static <T extends P... | 3.68 |
streampipes_ProcessingElementBuilder_category | /**
* Assigns a category to the element which later serves to categorize data processors in the UI.
*
* @param epaCategory The {@link org.apache.streampipes.model.DataProcessorType} of the element.
* @return {@link ProcessingElementBuilder}
*/
public ProcessingElementBuilder category(DataProcessorType... epaCatego... | 3.68 |
flink_CheckpointConfig_isForceCheckpointing | /**
* Checks whether checkpointing is forced, despite currently non-checkpointable iteration
* feedback.
*
* @return True, if checkpointing is forced, false otherwise.
* @deprecated This will be removed once iterations properly participate in checkpointing.
*/
@Deprecated
@PublicEvolving
public boolean isForceChe... | 3.68 |
flink_HashJoinOperator_fallbackSMJProcessPartition | /**
* If here also exists partitions which spilled to disk more than three time when hash join end,
* means that the key in these partitions is very skewed, so fallback to sort merge join
* algorithm to process it.
*/
private void fallbackSMJProcessPartition() throws Exception {
if (!table.getPartitionsPendingF... | 3.68 |
framework_Sort_by | /**
* Start building a Sort order by sorting a provided column.
*
* @param column
* a grid column object reference
* @param direction
* indicator of sort direction - either ascending or descending
* @return a sort instance, typed to the grid data type
*/
public static Sort by(Grid.Column<?... | 3.68 |
graphhopper_VectorTile_clearExtent | /**
* <pre>
* Although this is an "optional" field it is required by the specification.
* See https://github.com/mapbox/vector-tile-spec/issues/47
* </pre>
*
* <code>optional uint32 extent = 5 [default = 4096];</code>
*/
public Builder clearExtent() {
bitField0_ = (bitField0_ & ~0x00000020);
extent_ = 4096;
... | 3.68 |
hadoop_HttpFSAuthenticationFilter_getConfiguration | /**
* Returns the hadoop-auth configuration from HttpFSServer's configuration.
* <p>
* It returns all HttpFSServer's configuration properties prefixed with
* <code>hadoop.http.authentication</code>. The
* <code>hadoop.http.authentication</code> prefix is removed from the
* returned property names.
*
* @param co... | 3.68 |
hadoop_PlacementConstraints_nodeAttribute | /**
* Constructs a target expression on a node attribute. It is satisfied if
* the specified node attribute has one of the specified values.
*
* @param attributeKey the name of the node attribute
* @param attributeValues the set of values that the attribute should take
* values from
* @return the result... | 3.68 |
hbase_HFileSystem_getNoChecksumFs | /**
* Returns the filesystem that is specially setup for doing reads from storage. This object avoids
* doing checksum verifications for reads.
* @return The FileSystem object that can be used to read data from files.
*/
public FileSystem getNoChecksumFs() {
return noChecksumFs;
} | 3.68 |
hadoop_ResourceUsageMetrics_setPhysicalMemoryUsage | /**
* Set the physical memory usage.
*/
public void setPhysicalMemoryUsage(long usage) {
physicalMemoryUsage = usage;
} | 3.68 |
hadoop_AbstractMultipartUploader_checkUploadId | /**
* Utility method to validate uploadIDs.
* @param uploadId Upload ID
* @throws IllegalArgumentException invalid ID
*/
protected void checkUploadId(byte[] uploadId)
throws IllegalArgumentException {
checkArgument(uploadId != null, "null uploadId");
checkArgument(uploadId.length > 0,
"Empty UploadId ... | 3.68 |
hbase_SnapshotManifest_getRegionManifestsMap | /**
* Get all the Region Manifest from the snapshot. This is an helper to get a map with the region
* encoded name
*/
public Map<String, SnapshotRegionManifest> getRegionManifestsMap() {
if (regionManifests == null || regionManifests.isEmpty()) return null;
HashMap<String, SnapshotRegionManifest> regionsMap = n... | 3.68 |
framework_VaadinPortlet_invokeStaticLiferayMethod | /**
* Simplified version of what Liferay PortalClassInvoker did. This is
* used because the API of PortalClassInvoker has changed in Liferay
* 6.2.
*
* This simply uses reflection with Liferay class loader. Parameters are
* Strings to avoid static dependencies and to load all classes with
* Liferay's own class l... | 3.68 |
dubbo_MediaTypeUtil_convertMediaType | /**
* return first match , if any multiple content-type ,acquire mediaType by targetClass type .if contentTypes is empty
*
* @param contentTypes
* @return
*/
public static MediaType convertMediaType(Class<?> targetType, String... contentTypes) {
if (contentTypes == null || contentTypes.length == 0) {
... | 3.68 |
framework_ComboBox_getItemStyleGenerator | /**
* Gets the currently used item style generator.
*
* @return the itemStyleGenerator the currently used item style generator,
* or <code>null</code> if no generator is used
* @since 7.5.6
*/
public ItemStyleGenerator getItemStyleGenerator() {
return itemStyleGenerator;
} | 3.68 |
hadoop_FedBalance_setTrashOpt | /**
* Specify the trash behaviour of the source path.
* @param value the trash option.
*/
public Builder setTrashOpt(TrashOption value) {
this.trashOpt = value;
return this;
} | 3.68 |
flink_Pattern_getAfterMatchSkipStrategy | /** @return the pattern's {@link AfterMatchSkipStrategy.SkipStrategy} after match. */
public AfterMatchSkipStrategy getAfterMatchSkipStrategy() {
return afterMatchSkipStrategy;
} | 3.68 |
hadoop_VirtualInputFormat_getSplits | // Number of splits = Number of mappers. Creates fakeSplits to launch
// the required number of mappers
@Override
public List<InputSplit> getSplits(JobContext job) throws IOException {
Configuration conf = job.getConfiguration();
int numMappers = conf.getInt(CreateFileMapper.NUM_MAPPERS_KEY, -1);
if (numMappers =... | 3.68 |
flink_ResultInfo_getResultSchema | /** Get the schemas of the results. */
public ResolvedSchema getResultSchema() {
return ResolvedSchema.of(
columnInfos.stream().map(ColumnInfo::toColumn).collect(Collectors.toList()));
} | 3.68 |
hbase_FileChangeWatcher_waitForState | /**
* Blocks until the current state becomes <code>desiredState</code>. Currently only used by tests,
* thus package-private.
* @param desiredState the desired state.
* @throws InterruptedException if the current thread gets interrupted.
*/
synchronized void waitForState(State desiredState) throws InterruptedExcep... | 3.68 |
framework_VaadinPortlet_getRequestType | /**
* @param vaadinRequest
* @return
*
* @deprecated As of 7.0. This is no longer used and only provided for
* backwards compatibility. Each {@link RequestHandler} can
* individually decide whether it wants to handle a request or
* not.
*/
@Deprecated
protected RequestType ge... | 3.68 |
flink_DataStream_getTransformation | /**
* Returns the {@link Transformation} that represents the operation that logically creates this
* {@link DataStream}.
*
* @return The Transformation
*/
@Internal
public Transformation<T> getTransformation() {
return transformation;
} | 3.68 |
flink_TypeInferenceUtil_inferOutputType | /**
* Infers an output type using the given {@link TypeStrategy}. It assumes that input arguments
* have been adapted before if necessary.
*/
public static DataType inferOutputType(
CallContext callContext, TypeStrategy outputTypeStrategy) {
final Optional<DataType> potentialOutputType = outputTypeStrate... | 3.68 |
flink_RpcEndpoint_validateRunsInMainThread | /**
* Validates that the method call happens in the RPC endpoint's main thread.
*
* <p><b>IMPORTANT:</b> This check only happens when assertions are enabled, such as when
* running tests.
*
* <p>This can be used for additional checks, like
*
* <pre>{@code
* protected void concurrencyCriticalMethod() {
* v... | 3.68 |
framework_VaadinService_writeToHttpSession | /**
* Performs the actual write of the VaadinSession to the underlying HTTP
* session after sanity checks have been performed.
* <p>
* Called by {@link #storeSession(VaadinSession, WrappedSession)}
*
* @since 7.6
* @param wrappedSession
* the underlying HTTP session
* @param session
* th... | 3.68 |
framework_NestedMethodProperty_initialize | /**
* Initializes most of the internal fields based on the top-level bean
* instance and property name (dot-separated string).
*
* @param beanClass
* class of the top-level bean to which the property applies
* @param propertyName
* dot separated nested property name
* @throws IllegalArgume... | 3.68 |
morf_DataValueLookupMetadata_getColumnNames | /**
* @return The column names stored, where the array index corresponds
* with the {@link DataValueLookupBuilderImpl} internal array position.
*/
List<CaseInsensitiveString> getColumnNames() {
return keys;
} | 3.68 |
querydsl_BooleanBuilder_orNot | /**
* Create the union of this and the negation of the given predicate
*
* @param right predicate to be negated
* @return the current object
*/
public BooleanBuilder orNot(Predicate right) {
return or(right.not());
} | 3.68 |
flink_FlinkContainersSettings_checkpointPath | /**
* Sets the {@code checkpointPath} and returns a reference to this Builder enabling method
* chaining.
*
* @param checkpointPath The checkpoint path to set.
* @return A reference to this Builder.
*/
public Builder checkpointPath(String checkpointPath) {
this.checkpointPath = checkpointPath;
return setC... | 3.68 |
hadoop_SchedulerAppReport_getReservedContainers | /**
* Get the list of reserved containers
* @return All of the reserved containers.
*/
public Collection<RMContainer> getReservedContainers() {
return reserved;
} | 3.68 |
hadoop_TokenIdentifier_getTrackingId | /**
* Returns a tracking identifier that can be used to associate usages of a
* token across multiple client sessions.
*
* Currently, this function just returns an MD5 of {{@link #getBytes()}.
*
* @return tracking identifier
*/
public String getTrackingId() {
if (trackingId == null) {
trackingId = DigestUt... | 3.68 |
morf_DatabaseSchemaManager_dropTablesIfPresent | /**
* Drop the specified tables from the schema if they are present.
*
* @param tablesToDrop The tables to delete if they are present in the database.
*/
public void dropTablesIfPresent(Set<String> tablesToDrop) {
ProducerCache producerCache = new ProducerCache();
try {
Collection<String> sql = Lists.newLin... | 3.68 |
flink_DeweyNumber_increase | /**
* Creates a new dewey number from this such that its last digit is increased by the supplied
* number.
*
* @param times how many times to increase the Dewey number
* @return A new dewey number derived from this whose last digit is increased by given number
*/
public DeweyNumber increase(int times) {
int[]... | 3.68 |
framework_VRadioButtonGroup_isHtmlContentAllowed | /**
* Returns whether HTML is allowed in the item captions.
*
* @return {@code true} if the captions are used as HTML, {@code false} if
* used as plain text
*/
public boolean isHtmlContentAllowed() {
return htmlContentAllowed;
} | 3.68 |
flink_FlinkAggregateExpandDistinctAggregatesRule_doRewrite | /**
* Converts all distinct aggregate calls to a given set of arguments.
*
* <p>This method is called several times, one for each set of arguments. Each time it is
* called, it generates a JOIN to a new SELECT DISTINCT relational expression, and modifies the
* set of top-level calls.
*
* @param aggregate Origina... | 3.68 |
framework_VTooltip_setOpenDelay | /**
* Sets the time (in ms) that should elapse after an event triggering
* tooltip showing has occurred (e.g. mouse over) before the tooltip is
* shown. If a tooltip has recently been shown, then
* {@link #getQuickOpenDelay()} is used instead of this.
*
* @param openDelay
* The open delay (in ms)
*/
... | 3.68 |
framework_Navigator_addProvider | /**
* Registers a view provider (factory).
* <p>
* Providers are called in order of registration until one that can handle
* the requested view name is found.
*
* @param provider
* provider to register, not <code>null</code>
* @throws IllegalArgumentException
* if the provided view provi... | 3.68 |
hbase_RegionCoprocessorHost_postMemStoreCompaction | /**
* Invoked after in memory compaction.
*/
public void postMemStoreCompaction(HStore store) throws IOException {
execOperation(coprocEnvironments.isEmpty() ? null : new RegionObserverOperationWithoutResult() {
@Override
public void call(RegionObserver observer) throws IOException {
observer.postMemS... | 3.68 |
pulsar_ClientCnxIdleState_tryMarkIdleAndInitIdleTime | /**
* Try to transform the state of the connection to #{@link State#IDLE}, state should only be
* transformed to #{@link State#IDLE} from state #{@link State#USING}. if the state
* is successfully transformed, "idleMarkTime" will be assigned to current time.
*/
public void tryMarkIdleAndInitIdleTime() {
if (c... | 3.68 |
pulsar_ReaderConfiguration_setSubscriptionRolePrefix | /**
* Set the subscription role prefix for subscription auth. The default prefix is "reader".
*
* @param subscriptionRolePrefix
*/
public ReaderConfiguration setSubscriptionRolePrefix(String subscriptionRolePrefix) {
checkArgument(StringUtils.isNotBlank(subscriptionRolePrefix));
conf.setSubscriptionRolePref... | 3.68 |
hadoop_DatanodeAdminProperties_getPort | /**
* Get the port number of the datanode.
* @return the port number of the datanode.
*/
public int getPort() {
return port;
} | 3.68 |
flink_CheckpointProperties_forCheckpoint | /**
* Creates the checkpoint properties for a checkpoint.
*
* <p>Checkpoints may be queued in case too many other checkpoints are currently happening. They
* are garbage collected automatically, except when the owning job terminates in state {@link
* JobStatus#FAILED}. The user is required to configure the clean u... | 3.68 |
flink_ArrowFieldWriter_getCount | /** Returns the current count of elements written. */
public int getCount() {
return count;
} | 3.68 |
hadoop_GangliaConf_setSlope | /**
* @param slope the slope to set
*/
void setSlope(GangliaSlope slope) {
this.slope = slope;
} | 3.68 |
hbase_RawBytesTerminated_encode | /**
* Write {@code val} into {@code dst}, respecting {@code offset} and {@code length}.
* @return number of bytes written.
*/
public int encode(PositionedByteRange dst, byte[] val, int voff, int vlen) {
return ((RawBytes) wrapped).encode(dst, val, voff, vlen);
} | 3.68 |
hadoop_ProbeStatus_markAsSuccessful | /**
* Flip the success bit on while the real outcome bit is kept false
*/
public void markAsSuccessful() {
success = true;
} | 3.68 |
hadoop_CipherSuite_getAlgorithmBlockSize | /**
* @return size of an algorithm block in bytes
*/
public int getAlgorithmBlockSize() {
return algoBlockSize;
} | 3.68 |
hadoop_HttpFSReleaseFilter_getFileSystemAccess | /**
* Returns the {@link FileSystemAccess} service to return the FileSystemAccess filesystem
* instance to.
*
* @return the FileSystemAccess service.
*/
@Override
protected FileSystemAccess getFileSystemAccess() {
return HttpFSServerWebApp.get().get(FileSystemAccess.class);
} | 3.68 |
flink_RestartStrategies_fixedDelayRestart | /**
* Generates a FixedDelayRestartStrategyConfiguration.
*
* @param restartAttempts Number of restart attempts for the FixedDelayRestartStrategy
* @param delayInterval Delay in-between restart attempts for the FixedDelayRestartStrategy
* @return FixedDelayRestartStrategy
*/
public static RestartStrategyConfigura... | 3.68 |
hudi_HoodieTableMetadataUtil_deleteMetadataPartition | /**
* Deletes the metadata partition from the file system.
*
* @param basePath - base path of the dataset
* @param context - instance of {@link HoodieEngineContext}
* @param partitionType - {@link MetadataPartitionType} of the partition to delete
*/
public static void deleteMetadataPartition(String bas... | 3.68 |
hadoop_FederationProtocolPBTranslator_getProtoOrBuilder | /**
* Returns an interface to access data stored within this object. The object
* may have been initialized either via a builder or by an existing protobuf
* byte stream.
*
* @return MessageOrBuilder protobuf interface for the requested class.
*/
@SuppressWarnings("unchecked")
public T getProtoOrBuilder() {
if ... | 3.68 |
hadoop_AMRMProxyService_initializePipeline | /**
* Initializes the request interceptor pipeline for the specified application.
*
* @param applicationAttemptId attempt id
* @param user user name
* @param amrmToken amrmToken issued by RM
* @param localToken amrmToken issued by AMRMProxy
* @param recoveredDataMap the recovered states for AMRMProxy from NMSS
... | 3.68 |
hadoop_ConfigurationBasicValidator_validate | /**
* This method handles the base case where the configValue is null, based on the throwIfInvalid it either throws or returns the defaultVal,
* otherwise it returns null indicating that the configValue needs to be validated further.
* @param configValue the configuration value set by the user
* @return the default... | 3.68 |
rocketmq-connect_ServiceProviderUtil_getPositionManagementService | /**
* Get position management service by class name
*
* @param positionManagementServiceClazz
* @return
*/
@NotNull
public static PositionManagementService getPositionManagementService(String positionManagementServiceClazz) {
if (StringUtils.isEmpty(positionManagementServiceClazz)) {
positionManagement... | 3.68 |
hibernate-validator_ModUtil_calculateMod11Check | /**
* Calculate Modulo 11 checksum assuming that the threshold is Integer.MAX_VALUE
*
* @param digits the digits for which to calculate the checksum
*
* @return the result of the mod11 checksum calculation
*/
public static int calculateMod11Check(final List<Integer> digits) {
return calculateMod11Check( digits, ... | 3.68 |
hbase_RowModel_getCells | /** Returns the cells */
public List<CellModel> getCells() {
return cells;
} | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.