name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
morf_HumanReadableStatementHelper_generatePortableSqlStatementString | /**
* Generates a human-readable description of a raw SQL operation.
*
* @param statement the data upgrade statement to describe.
* @param preferredSQLDialect the SQL dialect to use, if available. If this is {@code null} or the preferred
* dialect is not available in the statement bundle then an arbitrary... | 3.68 |
hadoop_BondedS3AStatisticsContext_newInputStreamStatistics | /**
* Create a stream input statistics instance.
* The FileSystem.Statistics instance of the {@link #statisticsSource}
* is used as the reference to FileSystem statistics to update
* @return the new instance
*/
@Override
public S3AInputStreamStatistics newInputStreamStatistics() {
return getInstrumentation().new... | 3.68 |
framework_Design_read | /**
* Loads a design from the given input stream.
*
* @param design
* The stream to read the design from
* @return The root component of the design
*/
public static Component read(InputStream design) {
DesignContext context = read(design, null);
return context.getRootComponent();
} | 3.68 |
framework_EventRouter_removeListener | /*
* Removes the event listener method matching the given given parameters.
* Don't add a JavaDoc comment here, we use the default documentation from
* implemented interface.
*/
@Override
@Deprecated
public void removeListener(Class<?> eventType, Object target,
String methodName) {
// Find the correct ... | 3.68 |
flink_GettingStartedExample_eval | // the 'eval()' method defines input and output types (reflectively extracted)
// and contains the runtime logic
public String eval(String street, String zipCode, String city) {
return normalize(street) + ", " + normalize(zipCode) + ", " + normalize(city);
} | 3.68 |
hudi_TableServicePipeline_execute | /**
* Run all table services task sequentially.
*/
public void execute() {
tableServiceTasks.forEach(TableServiceTask::run);
} | 3.68 |
hbase_ServerSideScanMetrics_createCounter | /**
* Create a new counter with the specified name
* @return {@link AtomicLong} instance for the counter with counterName
*/
protected AtomicLong createCounter(String counterName) {
AtomicLong c = new AtomicLong(0);
counters.put(counterName, c);
return c;
} | 3.68 |
dubbo_BaseServiceMetadata_getDisplayServiceKey | /**
* Format : interface:version
*
* @return
*/
public String getDisplayServiceKey() {
StringBuilder serviceNameBuilder = new StringBuilder();
serviceNameBuilder.append(serviceInterfaceName);
serviceNameBuilder.append(COLON_SEPARATOR).append(version);
return serviceNameBuilder.toString();
} | 3.68 |
flink_ProjectOperator_types | /** @deprecated Deprecated method only kept for compatibility. */
@SuppressWarnings("unchecked")
@Deprecated
@PublicEvolving
public <R extends Tuple> ProjectOperator<IN, R> types(Class<?>... types) {
TupleTypeInfo<R> typeInfo = (TupleTypeInfo<R>) this.getResultType();
if (types.length != typeInfo.getArity()) {... | 3.68 |
hbase_InclusiveStopFilter_toByteArray | /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.InclusiveStopFilter.Builder builder =
FilterProtos.InclusiveStopFilter.newBuilder();
if (this.stopRowKey != null)
builder.setStopRowKey(UnsafeByteOperations.unsafeWrap(this.stopRowKey));
return builder.build... | 3.68 |
framework_AriaHelper_ensureHasId | /**
* Makes sure that the provided element has an id attribute. Adds a new
* unique id if not.
*
* @param element
* Element to check
* @return String with the id of the element
*/
public static String ensureHasId(Element element) {
assert element != null : "Valid Element required";
String id ... | 3.68 |
hbase_HMaster_getSnapshotManager | /** Returns the underlying snapshot manager */
@Override
public SnapshotManager getSnapshotManager() {
return this.snapshotManager;
} | 3.68 |
shardingsphere-elasticjob_TransactionOperation_opCheckExists | /**
* Operation check exists.
*
* @param key key
* @return TransactionOperation
*/
public static TransactionOperation opCheckExists(final String key) {
return new TransactionOperation(Type.CHECK_EXISTS, key, null);
} | 3.68 |
framework_AbstractMedia_setPreload | /**
* Sets the preload attribute that is intended to provide a hint to the
* browser how the media should be preloaded. Valid values are 'none',
* 'metadata', 'preload', see the <a href=
* "https://developer.mozilla.org/en/docs/Web/HTML/Element/video#attr-preload">
* Mozilla Developer Network</a> for details.
*
... | 3.68 |
framework_PointerDownEvent_getType | /**
* Gets the event type associated with PointerDownEvent events.
*
* @return the handler type
*/
public static Type<PointerDownHandler> getType() {
return TYPE;
} | 3.68 |
framework_ComboBox_setPopupWidth | /**
* Sets the suggestion pop-up's width as a CSS string. By using relative
* units (e.g. "50%") it's possible to set the popup's width relative to the
* ComboBox itself.
*
* @see #getPopupWidth()
* @since 7.7
* @param width
* the width
*/
public void setPopupWidth(String width) {
suggestionPopu... | 3.68 |
hbase_MasterObserver_preDeleteTable | /**
* Called before {@link org.apache.hadoop.hbase.master.HMaster} deletes a table. Called as part of
* delete table RPC call.
* @param ctx the environment to interact with the framework and master
* @param tableName the name of the table
*/
default void preDeleteTable(final ObserverContext<MasterCoprocessor... | 3.68 |
flink_SlotProfile_getPreferredLocations | /** Returns the preferred locations for the slot. */
public Collection<TaskManagerLocation> getPreferredLocations() {
return preferredLocations;
} | 3.68 |
hbase_RegionCoprocessorHost_preBulkLoadHFile | /**
* @param familyPaths pairs of { CF, file path } submitted for bulk load
*/
public void preBulkLoadHFile(final List<Pair<byte[], String>> familyPaths) throws IOException {
execOperation(coprocEnvironments.isEmpty() ? null : new RegionObserverOperationWithoutResult() {
@Override
public void call(RegionObs... | 3.68 |
dubbo_CtClassBuilder_getQualifiedClassName | /**
* get full qualified class name
*
* @param className super class name, maybe qualified or not
*/
protected String getQualifiedClassName(String className) {
if (className.contains(".")) {
return className;
}
if (fullNames.containsKey(className)) {
return fullNames.get(className);
... | 3.68 |
hadoop_ExecutorServiceFuturePool_shutdown | /**
* Utility to shutdown the {@link ExecutorService} used by this class. Will wait up to a
* certain timeout for the ExecutorService to gracefully shutdown.
*
* @param logger Logger
* @param timeout the maximum time to wait
* @param unit the time unit of the timeout argument
*/
public void shutdown(Logger logge... | 3.68 |
flink_DataTypeTemplate_isForceAnyPattern | /** Returns whether the given class must be treated as RAW type. */
boolean isForceAnyPattern(@Nullable Class<?> clazz) {
if (forceRawPattern == null || clazz == null) {
return false;
}
final String className = clazz.getName();
for (String pattern : forceRawPattern) {
if (className.start... | 3.68 |
hudi_HoodieFlinkWriteClient_waitForCleaningFinish | /**
* Blocks and wait for the async cleaning service to finish.
*
* <p>The Flink write client is designed to write data set as buckets
* but cleaning action should trigger after all the write actions within a
* checkpoint finish.
*/
public void waitForCleaningFinish() {
if (tableServiceClient.asyncCleanerServic... | 3.68 |
hbase_StorageClusterStatusModel_getRootIndexSizeKB | /** Returns The current total size of root-level indexes for the region, in KB. */
@XmlAttribute
public int getRootIndexSizeKB() {
return rootIndexSizeKB;
} | 3.68 |
hudi_HoodieTableMetadataUtil_deleteMetadataTablePartition | /**
* Delete a partition within the metadata table.
* <p>
* This can be used to delete a partition so that it can be re-bootstrapped.
*
* @param dataMetaClient {@code HoodieTableMetaClient} of the dataset for which metadata table is to be deleted
* @param context instance of {@code HoodieEngineContext}.
*... | 3.68 |
hadoop_ShortWritable_equals | /** Returns true iff <code>o</code> is a ShortWritable with the same value. */
@Override
public boolean equals(Object o) {
if (!(o instanceof ShortWritable))
return false;
ShortWritable other = (ShortWritable) o;
return this.value == other.value;
} | 3.68 |
flink_HighAvailabilityServices_getWebMonitorLeaderRetriever | /**
* This retriever should no longer be used on the cluster side. The web monitor retriever is
* only required on the client-side and we have a dedicated high-availability services for the
* client, named {@link ClientHighAvailabilityServices}. See also FLINK-13750.
*
* @return the leader retriever for web monito... | 3.68 |
hbase_SaslAuthMethod_getAuthMethod | /**
* Returns the Hadoop {@link AuthenticationMethod} for this method.
*/
public AuthenticationMethod getAuthMethod() {
return method;
} | 3.68 |
pulsar_TxnMetaImpl_checkTxnStatus | /**
* Check if the transaction is in an expected status.
*
* @param expectedStatus
*/
private synchronized void checkTxnStatus(TxnStatus expectedStatus) throws InvalidTxnStatusException {
if (this.txnStatus != expectedStatus) {
throw new InvalidTxnStatusException(
txnID, expectedStatus, txnS... | 3.68 |
AreaShop_Utils_getImportantWorldEditRegions | /**
* Get a list of regions around a location.
* - Returns highest priority, child instead of parent regions
* @param location The location to check for regions
* @return empty list if no regions found, 1 member if 1 region is a priority, more if regions with the same priority
*/
public static List<ProtectedRegion... | 3.68 |
hbase_ClientMetaTableAccessor_getResults | /** Returns Collected results; wait till visits complete to collect all possible results */
List<T> getResults() {
return this.results;
} | 3.68 |
hudi_CleanPlanner_getSavepointedDataFiles | /**
* Get the list of data file names savepointed.
*/
public Stream<String> getSavepointedDataFiles(String savepointTime) {
if (!hoodieTable.getSavepointTimestamps().contains(savepointTime)) {
throw new HoodieSavepointException(
"Could not get data files for savepoint " + savepointTime + ". No such save... | 3.68 |
AreaShop_AreaShop_onEnable | /**
* Called on start or reload of the server.
*/
@Override
public void onEnable() {
AreaShop.instance = this;
Do.init(this);
managers = new HashSet<>();
boolean error = false;
// Find WorldEdit integration version to load
String weVersion = null;
String rawWeVersion = null;
String weBeta = null;
Plugin plu... | 3.68 |
flink_JoinedStreams_apply | /**
* Completes the join operation with the user function that is executed for each combination
* of elements with the same key in a window.
*
* <p>Note: This method's return type does not support setting an operator-specific
* parallelism. Due to binary backwards compatibility, this cannot be altered. Use the
* ... | 3.68 |
framework_VComboBox_updateSelectedIconPosition | /**
* Positions the icon vertically in the middle. Should be called after the
* icon has loaded
*/
private void updateSelectedIconPosition() {
// Position icon vertically to middle
int availableHeight = 0;
availableHeight = getOffsetHeight();
int iconHeight = WidgetUtil.getRequiredHeight(selectedIte... | 3.68 |
rocketmq-connect_Worker_stop | /**
* We can choose to persist in-memory task status
* so we can view history tasks
*/
public void stop() {
// stop and await connectors
if (!connectors.isEmpty()) {
log.warn("Shutting down connectors {} uncleanly; herder should have shut down connectors before the Worker is stopped", connectors.key... | 3.68 |
flink_EnrichedRowData_from | /**
* Creates a new {@link EnrichedRowData} with the provided {@code fixedRow} as the immutable
* static row, and uses the {@code producedRowFields}, {@code fixedRowFields} and {@code
* mutableRowFields} arguments to compute the indexes mapping.
*
* <p>The {@code producedRowFields} should include the name of field... | 3.68 |
flink_S3TestCredentials_getS3SecretKey | /**
* Gets the S3 Secret Key.
*
* <p>This method throws an exception if the key is not available. Tests should use {@link
* #assumeCredentialsAvailable()} to skip tests when credentials are not available.
*/
public static String getS3SecretKey() {
if (S3_TEST_SECRET_KEY != null) {
return S3_TEST_SECRET... | 3.68 |
morf_Cast_as | /**
* @see org.alfasoftware.morf.sql.element.AliasedField#as(java.lang.String)
*/
@Override
public Cast as(String aliasName) {
return (Cast) super.as(aliasName);
} | 3.68 |
flink_RoundRobinOperatorStateRepartitioner_repartitionBroadcastState | /** Repartition BROADCAST state. */
private void repartitionBroadcastState(
Map<String, List<Tuple2<StreamStateHandle, OperatorStateHandle.StateMetaInfo>>>
broadcastState,
List<Map<StreamStateHandle, OperatorStateHandle>> mergeMapList) {
int newParallelism = mergeMapList.size();
... | 3.68 |
flink_CrossOperator_projectTuple6 | /**
* Projects a pair of crossed elements to a {@link Tuple} with the previously selected
* fields.
*
* @return The projected data set.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5>
ProjectCross<I1, I2, Tuple6<T0, T1, T2, T3, T4, T5>> projectTuple6() {
TypeInformation<?>[] fTypes = e... | 3.68 |
morf_SqlUtils_substring | /**
* Returns a SQL DSL expression to return the substring of <strong>field</strong>
* of <strong>length</strong> characters, starting from <strong>from</strong>.
*
* @param field The source expression.
* @param from The start character offset.
* @param length The length of the substring.
* @return The SQL DSL e... | 3.68 |
hadoop_ManifestStoreOperationsThroughFileSystem_storePreservesEtagsThroughRenames | /**
* Probe filesystem capabilities.
* @param path path to probe.
* @return true if the FS declares its renames work.
*/
@Override
public boolean storePreservesEtagsThroughRenames(Path path) {
try {
return fileSystem.hasPathCapability(path,
CommonPathCapabilities.ETAGS_PRESERVED_IN_RENAME);
} catch ... | 3.68 |
framework_Tree_setItemDescriptionGenerator | /**
* Set the item description generator which generates tooltips for the tree
* items.
*
* @param generator
* The generator to use or null to disable
*/
public void setItemDescriptionGenerator(
ItemDescriptionGenerator generator) {
if (generator != itemDescriptionGenerator) {
itemD... | 3.68 |
MagicPlugin_MagicController_getNPCSuppliers | /**
* @return The supplier set that is used.
*/
public NPCSupplierSet getNPCSuppliers() {
return npcSuppliers;
} | 3.68 |
morf_AliasedField_getImpliedName | /**
* Returns the name of the field either implied by its source or by its alias.
* @return The implied name of the field
*/
public String getImpliedName() {
return getAlias();
} | 3.68 |
hudi_HoodieTableFactory_inferAvroSchema | /**
* Inferences the deserialization Avro schema from the table schema (e.g. the DDL)
* if both options {@link FlinkOptions#SOURCE_AVRO_SCHEMA_PATH} and
* {@link FlinkOptions#SOURCE_AVRO_SCHEMA} are not specified.
*
* @param conf The configuration
* @param rowType The specified table row type
*/
private stati... | 3.68 |
hudi_DFSHoodieDatasetInputReader_iteratorSize | /**
* Returns the number of elements remaining in {@code iterator}. The iterator will be left exhausted: its {@code hasNext()} method will return {@code false}.
*/
private static int iteratorSize(Iterator<?> iterator) {
int count = 0;
while (iterator.hasNext()) {
iterator.next();
count++;
}
return cou... | 3.68 |
querydsl_BeanMap_reinitialise | /**
* Reinitializes this bean. Called during {@link #setBean(Object)}.
* Does introspection to find properties.
*/
protected void reinitialise() {
readMethods.clear();
writeMethods.clear();
types.clear();
initialise();
} | 3.68 |
hadoop_ProxyCombiner_combine | /**
* Combine two or more proxies which together comprise a single proxy
* interface. This can be used for a protocol interface which {@code extends}
* multiple other protocol interfaces. The returned proxy will implement
* all of the methods of the combined proxy interface, delegating calls
* to which proxy imple... | 3.68 |
hadoop_AbfsClient_checkUserError | /**
* Returns true if the status code lies in the range of user error.
* @param responseStatusCode http response status code.
* @return True or False.
*/
private boolean checkUserError(int responseStatusCode) {
return (responseStatusCode >= HttpURLConnection.HTTP_BAD_REQUEST
&& responseStatusCode < HttpURLC... | 3.68 |
morf_SqlDialect_getSqlFrom | /**
* Convert a {@link WindowFunction} into standards compliant SQL.
* @param windowFunctionField The field to convert
* @return The resulting SQL
**/
protected String getSqlFrom(WindowFunction windowFunctionField) {
if (requiresOrderByForWindowFunction(windowFunctionField.getFunction()) && windowFunctionField.g... | 3.68 |
flink_AbstractParameterTool_getShort | /**
* Returns the Short value for the given key. If the key does not exists it will return the
* default value given. The method fails if the value is not a Short.
*/
public short getShort(String key, short defaultValue) {
addToDefaults(key, Short.toString(defaultValue));
String value = get(key);
if (val... | 3.68 |
morf_ConnectionResourcesBean_setHostName | /**
* @see org.alfasoftware.morf.jdbc.AbstractConnectionResources#setHostName(java.lang.String)
*/
@Override
public void setHostName(String hostName) {
this.hostName = hostName;
} | 3.68 |
morf_AbstractSelectStatementBuilder_getJoins | /**
* Gets the list of joined tables in the order they are joined
*
* @return the joined tables
*/
List<Join> getJoins() {
return joins;
} | 3.68 |
AreaShop_BuyRegion_getPlayerName | /**
* Get the name of the player that owns this region.
* @return The name of the player that owns this region, if unavailable by UUID it will return the old cached name, if that is unavailable it will return <UNKNOWN>
*/
public String getPlayerName() {
String result = Utils.toName(getBuyer());
if(result == ... | 3.68 |
hadoop_StoragePolicySatisfyManager_removeAllPathIds | /**
* Clean up all sps path ids.
*/
public void removeAllPathIds() {
synchronized (pathsToBeTraversed) {
pathsToBeTraversed.clear();
}
} | 3.68 |
flink_ResourceManager_deregisterApplication | /**
* Cleanup application and shut down cluster.
*
* @param finalStatus of the Flink application
* @param diagnostics diagnostics message for the Flink application or {@code null}
*/
@Override
public CompletableFuture<Acknowledge> deregisterApplication(
final ApplicationStatus finalStatus, @Nullable final ... | 3.68 |
hadoop_OuterJoinRecordReader_combine | /**
* Emit everything from the collector.
*/
protected boolean combine(Object[] srcs, TupleWritable dst) {
assert srcs.length == dst.size();
return true;
} | 3.68 |
hadoop_CSQueueStore_add | /**
* Method for adding a queue to the store.
* @param queue Queue to be added
*/
public void add(CSQueue queue) {
String fullName = queue.getQueuePath();
String shortName = queue.getQueueShortName();
try {
modificationLock.writeLock().lock();
fullNameQueues.put(fullName, queue);
getMap.put(fullN... | 3.68 |
dubbo_ReferenceBean_getVersion | /**
* The version of the service
*/
public String getVersion() {
// Compatible with seata-1.4.0: io.seata.rm.tcc.remoting.parser.DubboRemotingParser#getServiceDesc()
return referenceConfig.getVersion();
} | 3.68 |
hadoop_BoundedResourcePool_release | /**
* Releases a previously acquired resource.
*
* @throws IllegalArgumentException if item is null.
*/
@Override
public void release(T item) {
checkNotNull(item, "item");
synchronized (createdItems) {
if (!createdItems.contains(item)) {
throw new IllegalArgumentException("This item is not a part of ... | 3.68 |
framework_DesignContext_getComponentLocalId | /**
* Returns the local id for a component.
*
* @since 7.5.0
*
* @param component
* The component whose local id to get.
* @return the local id of the component, or null if the component has no
* local id assigned
*/
public String getComponentLocalId(Component component) {
return compone... | 3.68 |
hadoop_SnappyCompressor_getBytesRead | /**
* Return number of bytes given to this compressor since last reset.
*/
@Override
public long getBytesRead() {
return bytesRead;
} | 3.68 |
flink_SqlAlterTable_getPartitionKVs | /** Get partition spec as key-value strings. */
public LinkedHashMap<String, String> getPartitionKVs() {
return SqlPartitionUtils.getPartitionKVs(getPartitionSpec());
} | 3.68 |
querydsl_SQLExpressions_lead | /**
* expr evaluated at the row that is one row after the current row within the partition;
*
* @param expr expression
* @return lead(expr)
*/
public static <T> WindowOver<T> lead(Expression<T> expr) {
return new WindowOver<T>(expr.getType(), SQLOps.LEAD, expr);
} | 3.68 |
hbase_LockAndQueue_releaseExclusiveLock | /** Returns whether we should wake the procedures waiting on the lock here. */
public boolean releaseExclusiveLock(Procedure<?> proc) {
if (
exclusiveLockOwnerProcedure == null
|| exclusiveLockOwnerProcedure.getProcId() != proc.getProcId()
) {
// We are not the lock owner, it is probably inherited fro... | 3.68 |
hbase_SnapshotScannerHDFSAclHelper_revokeAcl | /**
* Remove acl when grant or revoke user permission
* @param userPermission the user and permission
* @param skipNamespaces the namespace set to skip remove acl
* @param skipTables the table set to skip remove acl
* @return false if an error occurred, otherwise true
*/
public boolean revokeAcl(UserPermissio... | 3.68 |
morf_OracleDialect_selectStatementPreFieldDirectives | /**
* @see org.alfasoftware.morf.jdbc.SqlDialect#selectStatementPreFieldDirectives(org.alfasoftware.morf.sql.SelectStatement)
*/
@Override
protected String selectStatementPreFieldDirectives(SelectStatement selectStatement) {
StringBuilder builder = new StringBuilder();
for (Hint hint : selectStatement.getHints()... | 3.68 |
hbase_BlockIOUtils_readFullyWithHeapBuffer | /**
* Copying bytes from InputStream to {@link ByteBuff} by using an temporary heap byte[] (default
* size is 1024 now).
* @param in the InputStream to read
* @param out the destination {@link ByteBuff}
* @param length to read
* @throws IOException if any io error encountered.
*/
public static void readFu... | 3.68 |
framework_View_getViewComponent | /**
* Gets the component to show when navigating to the view.
*
* By default casts this View to a {@link Component} if possible, otherwise
* throws an IllegalStateException.
*
* @since 8.1
* @return the component to show, by default the view instance itself
*/
public default Component getViewComponent() {
i... | 3.68 |
hadoop_NamenodeStatusReport_getNumDecommissioningDatanodes | /**
* Get the number of decommissionining nodes.
*
* @return The number of decommissionining nodes.
*/
public int getNumDecommissioningDatanodes() {
return this.decomDatanodes;
} | 3.68 |
hmily_MetricsReporter_counterIncrement | /**
* Counter increment by count.
*
* @param name name
* @param labelValues label values
* @param count count
*/
public static void counterIncrement(final String name, final String[] labelValues, final long count) {
Optional.ofNullable(metricsRegister).ifPresent(register -> register.counterIncrement(name, lab... | 3.68 |
flink_ExceptionUtils_assertThrowableWithMessage | /**
* The same as {@link #findThrowableWithMessage(Throwable, String)}, but rethrows original
* exception if the expected exception was not found.
*/
public static <T extends Throwable> void assertThrowableWithMessage(
Throwable throwable, String searchMessage) throws T {
if (!findThrowableWithMessage(th... | 3.68 |
hadoop_S3AReadOpContext_withAuditSpan | /**
* Set builder value.
* @param value new value
* @return the builder
*/
public S3AReadOpContext withAuditSpan(final AuditSpan value) {
auditSpan = value;
return this;
} | 3.68 |
flink_SlotSharingGroup_setManagedMemoryMB | /** Set the task managed memory for this SlotSharingGroup in MB. */
public Builder setManagedMemoryMB(int managedMemoryMB) {
this.managedMemory = MemorySize.ofMebiBytes(managedMemoryMB);
return this;
} | 3.68 |
pulsar_TestRetrySupport_incrementSetupNumber | /**
* This method should be called in the setup method of the concrete class.
*
* This increases an internal counter and resets the failure state which are used to determine
* whether cleanup is needed before a test method is called.
*
*/
protected final void incrementSetupNumber() {
currentSetupNumber++;
... | 3.68 |
flink_HybridSource_build | /** Build the source. */
public HybridSource<T> build() {
return new HybridSource(sources);
} | 3.68 |
hbase_RegionReplicaUtil_addReplicas | /**
* Create any replicas for the regions (the default replicas that was already created is passed to
* the method)
* @param regions existing regions
* @param oldReplicaCount existing replica count
* @param newReplicaCount updated replica count due to modify table
* @return the combined list of default an... | 3.68 |
framework_VCustomLayout_clear | /** Clear all widgets from the layout. */
@Override
public void clear() {
super.clear();
locationToWidget.clear();
childWidgetToCaptionWrapper.clear();
} | 3.68 |
hbase_HMaster_isOnline | /**
* Report whether this master is started This method is used for testing.
* @return true if master is ready to go, false if not.
*/
public boolean isOnline() {
return serviceStarted;
} | 3.68 |
hudi_SparkInternalSchemaConverter_collectColNamesFromSparkStruct | /**
* Collect all the leaf nodes names.
*
* @param sparkSchema a spark schema
* @return leaf nodes full names.
*/
public static List<String> collectColNamesFromSparkStruct(StructType sparkSchema) {
List<String> result = new ArrayList<>();
collectColNamesFromStructType(sparkSchema, new LinkedList<>(), result);... | 3.68 |
hudi_AvroSchemaUtils_isNullable | /**
* Returns true in case provided {@link Schema} is nullable (ie accepting null values),
* returns false otherwise
*/
public static boolean isNullable(Schema schema) {
if (schema.getType() != Schema.Type.UNION) {
return false;
}
List<Schema> innerTypes = schema.getTypes();
return innerTypes.size() > 1... | 3.68 |
hbase_UnsafeAccess_toShort | /**
* Reads a short value at the given Object's offset considering it was written in big-endian
* format.
* @return short value at offset
*/
public static short toShort(Object ref, long offset) {
if (LITTLE_ENDIAN) {
return Short.reverseBytes(HBasePlatformDependent.getShort(ref, offset));
}
return HBasePl... | 3.68 |
framework_StateChangeEvent_getChangedPropertiesFastSet | /**
* Gets the properties that have changed.
*
* @return a set of names of the changed properties
*
* @deprecated As of 7.0.1, use {@link #hasPropertyChanged(String)} instead
* for improved performance.
*/
@Deprecated
public FastStringSet getChangedPropertiesFastSet() {
if (changedProperties == n... | 3.68 |
flink_AsynchronousFileIOChannel_registerAllRequestsProcessedListener | /**
* Registers a listener to be notified when all outstanding requests have been processed.
*
* <p>New requests can arrive right after the listener got notified. Therefore, it is not safe
* to assume that the number of outstanding requests is still zero after a notification unless
* there was a close right before... | 3.68 |
hbase_SingleColumnValueFilter_parseFrom | /**
* Parse a serialized representation of {@link SingleColumnValueFilter}
* @param pbBytes A pb serialized {@link SingleColumnValueFilter} instance
* @return An instance of {@link SingleColumnValueFilter} made from <code>bytes</code>
* @throws DeserializationException if an error occurred
* @see #toByteArray
*/
... | 3.68 |
hbase_MetricsTableRequests_updateCheckAndDelete | /**
* Update the CheckAndDelete time histogram.
* @param time time it took
*/
public void updateCheckAndDelete(long time) {
if (isEnableTableLatenciesMetrics()) {
checkAndDeleteTimeHistogram.update(time);
}
} | 3.68 |
flink_AbstractCheckpointStats_getTaskStateStats | /**
* Returns the task state stats for the given job vertex ID or <code>null</code> if no task with
* such an ID is available.
*
* @param jobVertexId Job vertex ID of the task stats to look up.
* @return The task state stats instance for the given ID or <code>null</code>.
*/
public TaskStateStats getTaskStateStat... | 3.68 |
hadoop_RpcProgramPortmap_set | /**
* When a program first becomes available on a machine, it registers itself
* with the port mapper program on the same machine. The program passes its
* program number "prog", version number "vers", transport protocol number
* "prot", and the port "port" on which it awaits service request. The
* procedure retur... | 3.68 |
hbase_CacheConfig_setEvictOnClose | /**
* Only used for testing.
* @param evictOnClose whether blocks should be evicted from the cache when an HFile reader is
* closed
*/
public void setEvictOnClose(boolean evictOnClose) {
this.evictOnClose = evictOnClose;
} | 3.68 |
hudi_CleanActionExecutor_runPendingClean | /**
* Executes the Cleaner plan stored in the instant metadata.
*/
HoodieCleanMetadata runPendingClean(HoodieTable<T, I, K, O> table, HoodieInstant cleanInstant) {
try {
HoodieCleanerPlan cleanerPlan = CleanerUtils.getCleanerPlan(table.getMetaClient(), cleanInstant);
return runClean(table, cleanInstant, cle... | 3.68 |
druid_IPRange_computeMaskFromNetworkPrefix | /**
* Convert a extended network prefix integer into an IP number.
*
* @param prefix The network prefix number.
* @return Return the IP number corresponding to the extended network prefix.
*/
private IPAddress computeMaskFromNetworkPrefix(int prefix) {
/*
* int subnet = 0; for (int i=0; i<prefix; i++) { s... | 3.68 |
hadoop_ApplicationMaster_main | /**
* @param args Command line args
*/
public static void main(String[] args) {
boolean result = false;
try {
ApplicationMaster appMaster = new ApplicationMaster();
LOG.info("Initializing ApplicationMaster");
boolean doRun = appMaster.init(args);
if (!doRun) {
System.exit(0);
}
resul... | 3.68 |
hadoop_ManifestCommitter_abortTask | /**
* Abort a task.
* @param context task context
* @throws IOException failure during the delete
*/
@Override
public void abortTask(final TaskAttemptContext context)
throws IOException {
ManifestCommitterConfig committerConfig = enterCommitter(true,
context);
try {
new AbortTaskStage(
com... | 3.68 |
framework_ServerRpcManager_applyInvocation | /**
* Invoke a method in a server side RPC target class. This method is to be
* used by the RPC framework and unit testing tools only.
*
* @param invocation
* method invocation to perform
*/
public void applyInvocation(ServerRpcMethodInvocation invocation)
throws RpcInvocationException {
Me... | 3.68 |
morf_OracleDialect_getSqlForAddDays | /**
* @see org.alfasoftware.morf.jdbc.SqlDialect#getSqlForAddDays(org.alfasoftware.morf.sql.element.Function)
*/
@Override
protected String getSqlForAddDays(Function function) {
return String.format(
"(%s) + (%s)",
getSqlFrom(function.getArguments().get(0)),
getSqlFrom(function.getArguments().get(1))
... | 3.68 |
framework_AbstractLegacyComponent_isImmediate | /**
* Returns the immediate mode of the component.
* <p>
* Since Vaadin 8, the default mode is immediate.
*
* @return true if the component is in immediate mode (explicitly or
* implicitly set), false if the component if not in immediate mode
*/
public boolean isImmediate() {
if (explicitImmediateVal... | 3.68 |
hadoop_ResourceRequest_executionType | /**
* Set the <code>executionTypeRequest</code> of the request with 'ensure
* execution type' flag set to true.
* @see ResourceRequest#setExecutionTypeRequest(
* ExecutionTypeRequest)
* @param executionType <code>executionType</code> of the request.
* @return {@link ResourceRequestBuilder}
*/
@Public
@Evolving
p... | 3.68 |
querydsl_Expressions_stringOperation | /**
* Create a new Operation expression
*
* @param operator operator
* @param args operation arguments
* @return operation expression
*/
public static StringOperation stringOperation(Operator operator, Expression<?>... args) {
return new StringOperation(operator, args);
} | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.