name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
morf_GraphBasedUpgradeTraversalService_allNodesCompleted | /**
* @return true if all nodes of the upgrade have been executed
*/
public boolean allNodesCompleted() {
lock.lock();
try {
return allNodesCompletedNoLock();
} finally {
lock.unlock();
}
} | 3.68 |
flink_ExecutionEnvironment_getJobName | /**
* Gets the job name. If user defined job name is not found in the configuration, the default
* name based on the timestamp when this method is invoked will return.
*
* @return A job name.
*/
private String getJobName() {
return configuration.getString(
PipelineOptions.NAME, "Flink Java Job at "... | 3.68 |
dubbo_DubboAbstractTDigest_weightedAverage | /**
* Same as {@link #weightedAverageSorted(double, double, double, double)} but flips
* the order of the variables if <code>x2</code> is greater than
* <code>x1</code>.
*/
static double weightedAverage(double x1, double w1, double x2, double w2) {
if (x1 <= x2) {
return weightedAverageSorted(x1, w1, x2... | 3.68 |
hbase_AccessController_preCleanupBulkLoad | /**
* Authorization security check for SecureBulkLoadProtocol.cleanupBulkLoad()
* @param ctx the context
*/
@Override
public void preCleanupBulkLoad(ObserverContext<RegionCoprocessorEnvironment> ctx)
throws IOException {
requireAccess(ctx, "preCleanupBulkLoad",
ctx.getEnvironment().getRegion().getTableDescri... | 3.68 |
pulsar_TokenClient_exchangeClientCredentials | /**
* Performs a token exchange using client credentials.
* @param req the client credentials request details.
* @return a token result
* @throws TokenExchangeException
*/
public TokenResult exchangeClientCredentials(ClientCredentialsExchangeRequest req)
throws TokenExchangeException, IOException {
Str... | 3.68 |
hbase_MasterProcedureScheduler_getGlobalQueue | // ============================================================================
// Global Queue Lookup Helpers
// ============================================================================
private GlobalQueue getGlobalQueue(String globalId) {
GlobalQueue node = AvlTree.get(globalMap, globalId, GLOBAL_QUEUE_KEY_COMP... | 3.68 |
framework_ApplicationConfiguration_getServiceUrlParameterName | /**
* Return the name of the parameter used to to send data to the service url.
* This method should only be called if {@link #useServiceUrlPathParam()} is
* true.
*
* @since 7.1.6
* @return The parameter name, by default <code>v-resourcePath</code>
*/
public String getServiceUrlParameterName() {
return getJ... | 3.68 |
rocketmq-connect_ColumnDefinition_isSignedNumber | /**
* Indicates whether values in the column are signed numbers.
*
* @return <code>true</code> if so; <code>false</code> otherwise
*/
public boolean isSignedNumber() {
return signedNumbers;
} | 3.68 |
querydsl_CollQueryFactory_from | /**
* Create a new query
*
* @param path source expression
* @param col source collection
* @return query
*/
public static <A> CollQuery<A> from(Path<A> path, Iterable<A> col) {
return new CollQuery<Void>().from(path, col).select(path);
} | 3.68 |
morf_AbstractSqlDialectTest_assertSQLEquals | /**
* Helper to allow lists of SQL strings to be compared in Eclipse.
*
* @param message The message to show on failure.
* @param expected The expected list of strings.
* @param actual The actual list of strings.
*/
private void assertSQLEquals(String message, List<String> expected, List<String> actual) {
Asser... | 3.68 |
morf_BaseDataSetReader_clear | /**
* Clear the local table lists.
*/
protected void clear() {
tableNameToFileNameMap.clear();
tableNames.clear();
} | 3.68 |
flink_TypeInformation_getGenericParameters | /**
* Optional method for giving Flink's type extraction system information about the mapping of a
* generic type parameter to the type information of a subtype. This information is necessary in
* cases where type information should be deduced from an input type.
*
* <p>For instance, a method for a {@link Tuple2} ... | 3.68 |
framework_StaticSection_isVisible | /**
* Gets the visibility of this section.
*
* @return {@code true} if visible; {@code false} if not
*
* @since 8.1.1
*/
public boolean isVisible() {
return getState(false).visible;
} | 3.68 |
hadoop_S3ALocatedFileStatus_toS3AFileStatus | /**
* Generate an S3AFileStatus instance, including etag and
* version ID, if present.
* @return the S3A status.
*/
public S3AFileStatus toS3AFileStatus() {
return new S3AFileStatus(
getPath(),
isDirectory(),
isEmptyDirectory,
getLen(),
getModificationTime(),
getBlockSize(),
... | 3.68 |
hbase_VersionModel_setJerseyVersion | /**
* @param version the Jersey framework version string
*/
public void setJerseyVersion(String version) {
this.jerseyVersion = version;
} | 3.68 |
druid_DruidAbstractDataSource_isInitExceptionThrow | /**
* @since 1.1.11
*/
public boolean isInitExceptionThrow() {
return initExceptionThrow;
} | 3.68 |
framework_ApplicationConfiguration_getRootElement | /**
* Get the root element instance used for this application.
*
* @return registered root element
* @since 8.4
*/
public Element getRootElement() {
return rootElement;
} | 3.68 |
hadoop_DynoInfraUtils_waitForNameNodeJMXValue | /**
* Poll the launched NameNode's JMX for a specific value, waiting for it to
* cross some threshold. Continues until the threshold has been crossed or
* {@code shouldExit} returns true. Periodically logs the current value.
*
* @param valueName The human-readable name of the value which is being
* ... | 3.68 |
hbase_ImplType_toString | /** Returns <code>-option</code> */
@Override
public String toString() {
return "-" + option;
} | 3.68 |
framework_FocusableComplexPanel_addKeyPressHandler | /*
* (non-Javadoc)
*
* @see
* com.google.gwt.event.dom.client.HasKeyPressHandlers#addKeyPressHandler
* (com.google.gwt.event.dom.client.KeyPressHandler)
*/
@Override
public HandlerRegistration addKeyPressHandler(KeyPressHandler handler) {
return addDomHandler(handler, KeyPressEvent.getType());
} | 3.68 |
hudi_GenericRecordFullPayloadGenerator_getNewPayload | /**
* Create a new {@link GenericRecord} with random value according to given schema.
*
* Long fields which are specified within partitionPathFieldNames are constrained to the value of the partition for which records are being generated.
*
* @return {@link GenericRecord} with random value
*/
public GenericRecord ... | 3.68 |
hudi_CatalogOptions_allOptions | /**
* Returns all the config options.
*/
public static List<ConfigOption<?>> allOptions() {
return OptionsResolver.allOptions(CatalogOptions.class);
} | 3.68 |
hadoop_ClusterMetrics_getDecommissionedTaskTrackerCount | /**
* Get the number of decommissioned trackers in the cluster.
*
* @return decommissioned tracker count
*/
public int getDecommissionedTaskTrackerCount() {
return numDecommissionedTrackers;
} | 3.68 |
morf_SqlDialect_indexes | /**
* @see org.alfasoftware.morf.metadata.Table#indexes()
*/
@Override
public List<Index> indexes() {
return new ArrayList<>();
} | 3.68 |
flink_PropertiesUtil_getBoolean | /**
* Get boolean from properties. This method returns {@code true} iff the parsed value is "true".
*
* @param config Properties
* @param key key in Properties
* @param defaultValue default value if value is not set
* @return default or value of key
*/
public static boolean getBoolean(Properties config, String k... | 3.68 |
flink_ExpressionResolver_resolve | /**
* Resolves given expressions with configured set of rules. All expressions of an operation
* should be given at once as some rules might assume the order of expressions.
*
* <p>After this method is applied the returned expressions should be ready to be converted to
* planner specific expressions.
*
* @param ... | 3.68 |
hadoop_NMStateStoreService_serviceStop | /** Shutdown the state storage. */
@Override
public void serviceStop() throws IOException {
closeStorage();
} | 3.68 |
framework_VaadinService_writeStringResponse | /**
* Writes the given string as a response using the given content type.
*
* @param response
* The response reference
* @param contentType
* The content type of the response
* @param responseString
* The actual response
* @throws IOException
* If an error occurred... | 3.68 |
hbase_InputStreamBlockDistribution_isEnabled | /**
* True if we should derive StoreFile HDFSBlockDistribution from the underlying input stream
*/
public static boolean isEnabled(Configuration conf) {
return conf.getBoolean(HBASE_LOCALITY_INPUTSTREAM_DERIVE_ENABLED,
DEFAULT_HBASE_LOCALITY_INPUTSTREAM_DERIVE_ENABLED);
} | 3.68 |
hadoop_StoreContext_getRequestFactory | /**
* Get the request factory.
* @return the factory for requests.
*/
public RequestFactory getRequestFactory() {
return contextAccessors.getRequestFactory();
} | 3.68 |
flink_SkipListUtils_helpGetNextNode | /**
* Return the next of the given node at the given level.
*
* @param node the node to find the next node for.
* @param level the level to find the next node.
* @param levelIndexHeader the header of the level index.
* @param spaceAllocator the space allocator.
* @return the pointer to the next node of the given... | 3.68 |
framework_ComponentRootSetter_setRoot | /**
* Sets the composition root for the given custom component or composite.
* <p>
* For internal use only.
*
* @param customComponent
* the custom component or composite
* @param component
* the component to assign as composition root
*/
public static void setRoot(Component customCompone... | 3.68 |
hbase_HRegionServer_finishRegionProcedure | /**
* See {@link #submitRegionProcedure(long)}.
* @param procId the id of the open/close region procedure
*/
public void finishRegionProcedure(long procId) {
executedRegionProcedures.put(procId, procId);
submittedRegionProcedures.remove(procId);
} | 3.68 |
flink_DeltaIterationBase_getInitialWorkset | /**
* Returns the initial workset input, or null, if none is set.
*
* @return The iteration's workset input.
*/
public Operator<WT> getInitialWorkset() {
return getSecondInput();
} | 3.68 |
hbase_ByteBufferKeyValue_hashCode | /**
* In line with {@link #equals(Object)}, only uses the key portion, not the value.
*/
@Override
public int hashCode() {
return calculateHashForKey(this);
} | 3.68 |
hadoop_StoreContext_isCSEEnabled | /**
* return if the store context have client side encryption enabled.
* @return boolean indicating if CSE is enabled or not.
*/
public boolean isCSEEnabled() {
return isCSEEnabled;
} | 3.68 |
framework_Embedded_setCodebase | /**
* This attribute specifies the base path used to resolve relative URIs
* specified by the classid, data, and archive attributes. When absent, its
* default value is the base URI of the current document.
*
* @param codebase
* The base path
*/
public void setCodebase(String codebase) {
String ol... | 3.68 |
streampipes_EpProperties_listIntegerEp | /**
* Creates a new list-based event property of type integer and with the assigned domain property.
*
* @param label A human-readable label of the property
* @param runtimeName The field identifier of the event property at runtime.
* @param domainProperty The semantics of the list property as a String... | 3.68 |
hudi_MetadataMigrator_migrateToVersion | /**
* Migrate metadata to a specific version.
*
* @param metadata Hoodie Table Meta Client
* @param metadataVersion Metadata Version
* @param targetVersion Target Version
* @return Metadata conforming to the target version
*/
public T migrateToVersion(T metadata, int metadataVersion, int targetVersion) {
Vali... | 3.68 |
pulsar_AuthenticationDataSource_hasDataFromHttp | /**
* Check if data from HTTP are available.
*
* @return true if this authentication data contain data from HTTP
*/
default boolean hasDataFromHttp() {
return false;
} | 3.68 |
druid_MySqlStatementParser_parseProcedureStatementList | /**
* parse procedure statement block
*/
private void parseProcedureStatementList(List<SQLStatement> statementList, int max) {
for (; ; ) {
if (max != -1) {
if (statementList.size() >= max) {
return;
}
}
if (lexer.token() == Token.EOF) {
... | 3.68 |
hadoop_FlowRunRowKeyPrefix_getRowKeyPrefix | /*
* (non-Javadoc)
*
* @see
* org.apache.hadoop.yarn.server.timelineservice.storage.application.
* RowKeyPrefix#getRowKeyPrefix()
*/
public byte[] getRowKeyPrefix() {
// We know we're a FlowRunRowKey with null florRunId, so we can simply
// delegate
return super.getRowKey();
} | 3.68 |
hadoop_TimelineEntity_getEntityType | /**
* Get the entity type
*
* @return the entity type
*/
@XmlElement(name = "entitytype")
public String getEntityType() {
return entityType;
} | 3.68 |
morf_MergeStatement_deepCopy | /**
* @see org.alfasoftware.morf.util.DeepCopyableWithTransformation#deepCopy(org.alfasoftware.morf.util.DeepCopyTransformation)
*/
@Override
public MergeStatementBuilder deepCopy(DeepCopyTransformation transformer) {
return new MergeStatementBuilder(this, transformer);
} | 3.68 |
pulsar_ProducerImpl_maybeScheduleBatchFlushTask | // must acquire semaphore before calling
private void maybeScheduleBatchFlushTask() {
if (this.batchFlushTask != null || getState() != State.Ready) {
return;
}
scheduleBatchFlushTask(conf.getBatchingMaxPublishDelayMicros());
} | 3.68 |
hbase_BaseSourceImpl_setGauge | /**
* Set a single gauge to a value.
* @param gaugeName gauge name
* @param value the new value of the gauge.
*/
@Override
public void setGauge(String gaugeName, long value) {
MutableGaugeLong gaugeInt = metricsRegistry.getGauge(gaugeName, value);
gaugeInt.set(value);
} | 3.68 |
flink_TopNBuffer_lastEntry | /** Returns the last Entry in the buffer. Returns null if the TreeMap is empty. */
public Map.Entry<RowData, Collection<RowData>> lastEntry() {
return treeMap.lastEntry();
} | 3.68 |
morf_UpgradeStatusTableServiceImpl_tidyUp | /**
*
* @see org.alfasoftware.morf.upgrade.UpgradeStatusTableService#tidyUp(javax.sql.DataSource)
*/
@Override
public void tidyUp(DataSource dataSource) {
try {
new SqlScriptExecutorProvider(dataSource, sqlDialect).get().execute(sqlDialect.dropStatements(table(UpgradeStatusTableService.UPGRADE_STATUS)));
... | 3.68 |
flink_HsSpillingStrategyUtils_getBuffersByConsumptionPriorityInOrder | /**
* Calculate and get expected number of buffers with the highest consumption priority. For each
* buffer, The greater the difference between next buffer index to consume of subpartition it
* belongs to and buffer index, the higher the priority.
*
* @param nextBufferIndexToConsume downstream next buffer index to... | 3.68 |
framework_BrowserInfo_isIOS6 | /**
* Checks if the browser is run on iOS 6.
*
* @since 7.1.1
* @return true if the browser is run on iOS 6, false otherwise
*/
public boolean isIOS6() {
return isIOS() && getOperatingSystemMajorVersion() == 6;
} | 3.68 |
hbase_AccessControlClient_isAuthorizationEnabled | /**
* Return true if authorization is supported and enabled
* @param connection The connection to use
* @return true if authorization is supported and enabled, false otherwise
*/
public static boolean isAuthorizationEnabled(Connection connection) throws IOException {
return connection.getAdmin().getSecurityCapabi... | 3.68 |
hadoop_BalanceJob_execute | /**
* Run the state machine.
*/
public void execute() {
boolean quit = false;
try {
while (!jobDone && !quit && scheduler.isRunning()) {
if (curProcedure == null) { // Job done.
finish(null);
quit = true;
} else {
if (curProcedure == firstProcedure || lastProcedure != curPr... | 3.68 |
morf_AbstractSqlDialectTest_testSelectMultipleWhereScript | /**
* Tests a select with multiple where clauses
*/
@Test
public void testSelectMultipleWhereScript() {
SelectStatement stmt = new SelectStatement().from(new TableReference(TEST_TABLE))
.where(and(
eq(new FieldReference(STRING_FIELD), "A0001"),
greaterThan(new FieldReference(INT_FIELD), 200801... | 3.68 |
morf_H2Dialect_getSqlForLastDayOfMonth | /**
* @see org.alfasoftware.morf.jdbc.SqlDialect#getSqlForLastDayOfMonth
*/
@Override
protected String getSqlForLastDayOfMonth(AliasedField date) {
return "DATEADD(dd, -DAY(DATEADD(m,1," + getSqlFrom(date) + ")), DATEADD(m,1," + getSqlFrom(date) + "))";
} | 3.68 |
framework_Range_endsAfter | /**
* Checks whether this range ends after the end of another range.
*
* @param other
* the other range to compare against
* @return <code>true</code> if this range ends after the <code>other</code>
*/
public boolean endsAfter(final Range other) {
return getEnd() > other.getEnd();
} | 3.68 |
dubbo_ServiceConfigBase_setInterfaceClass | /**
* @param interfaceClass
* @see #setInterface(Class)
* @deprecated
*/
public void setInterfaceClass(Class<?> interfaceClass) {
setInterface(interfaceClass);
} | 3.68 |
hbase_ProcedurePrepareLatch_createLatch | /**
* Create a latch if the client does not have async proc support
* @param major major version with async proc support
* @param minor minor version with async proc support
* @return a CompatibilityLatch or a NoopLatch if the client has async proc support
*/
public static ProcedurePrepareLatch createLatch(int maj... | 3.68 |
morf_TableReference_as | /**
* Specifies the alias to use for the table.
*
* @param aliasName the name of the alias
* @return an updated {@link TableReference} (this will not be a new object)
*/
public TableReference as(String aliasName) {
if (AliasedField.immutableDslEnabled()) {
return new TableReference(this, aliasName);
} else... | 3.68 |
morf_AbstractSqlDialectTest_testJoinNoCriteria | /**
* Tests a join with no ON criteria.
*/
@Test
public void testJoinNoCriteria() {
SelectStatement testStatement = select().from(tableRef("TableOne")).crossJoin(tableRef("TableTwo"));
assertEquals(testDialect.convertStatementToSQL(testStatement), expectedJoinOnEverything());
} | 3.68 |
hadoop_WriteOperationHelper_writeFailed | /**
* Callback on a write failure.
* @param ex Any exception raised which triggered the failure.
*/
public void writeFailed(Exception ex) {
LOG.debug("Write to {} failed", this, ex);
} | 3.68 |
streampipes_Labels_withId | /**
* Creates a label with the string value of an enum.
* Static properties require a fully-specified label, see {@link #from(String, String, String)}
* @param internalId The internal identifier of the element, e.g., "LATITUDE-FIELD-MAPPING"
* @return
*/
public static Label withId(Enum<?> internalId) {
return ne... | 3.68 |
flink_ThriftObjectConversions_toString | /**
* Converts a {@link Throwable} object into a flattened list of texts including its stack trace
* and the stack traces of the nested causes.
*
* @param ex a {@link Throwable} object
* @return a flattened list of texts including the {@link Throwable} object's stack trace and
* the stack traces of the nested... | 3.68 |
framework_LogSection_setLimit | /**
* Sets the log row limit.
*
* @param limit
* The limit
*/
public void setLimit(int limit) {
this.limit = limit;
applyLimit();
// TODO should be persisted
} | 3.68 |
flink_EmbeddedRocksDBStateBackend_setWriteBatchSize | /**
* Sets the max batch size will be used in {@link RocksDBWriteBatchWrapper}, no positive value
* will disable memory size controller, just use item count controller.
*
* @param writeBatchSize The size will used to be used in {@link RocksDBWriteBatchWrapper}.
*/
public void setWriteBatchSize(long writeBatchSize)... | 3.68 |
pulsar_SecurityUtility_processConscryptTrustManager | // workaround https://github.com/google/conscrypt/issues/1015
private static void processConscryptTrustManager(TrustManager trustManager) {
if (trustManager.getClass().getName().equals("org.conscrypt.TrustManagerImpl")) {
try {
Class<?> conscryptClazz = Class.forName("org.conscrypt.Conscrypt");
... | 3.68 |
morf_SqlDialect_getDeleteLimitPreFromClause | /**
* Returns the SQL that specifies the deletion limit ahead of the FROM clause, if any, for the dialect.
*
* @param limit The delete limit.
* @return The SQL fragment.
*/
protected Optional<String> getDeleteLimitPreFromClause(@SuppressWarnings("unused") int limit) {
return Optional.empty();
} | 3.68 |
hibernate-validator_ConstraintMappingContextImplBase_addConstraint | /**
* Adds a constraint to the set of constraints managed by this creational context.
*
* @param constraint the constraint to add
*/
protected void addConstraint(ConfiguredConstraint<?> constraint) {
constraints.add( constraint );
} | 3.68 |
hadoop_RegistryDNSServer_launchDNSServer | /**
* Launch the server.
* @param conf configuration
* @param rdns registry dns instance
* @return
*/
static RegistryDNSServer launchDNSServer(Configuration conf,
RegistryDNS rdns) {
RegistryDNSServer dnsServer = null;
Thread.setDefaultUncaughtExceptionHandler(
new HadoopUncaughtExceptionHandler());... | 3.68 |
hadoop_ListResultEntrySchema_group | /**
* Get the group value.
*
* @return the group value
*/
public String group() {
return group;
} | 3.68 |
flink_FailureHandlingResultSnapshot_create | /**
* Creates a {@code FailureHandlingResultSnapshot} based on the passed {@link
* FailureHandlingResult} and {@link ExecutionVertex ExecutionVertices}.
*
* @param failureHandlingResult The {@code FailureHandlingResult} that is used for extracting
* the failure information.
* @param currentExecutionsLookup Th... | 3.68 |
hadoop_AdlFsInputStream_available | /**
* This method returns the remaining bytes in the stream, rather than the
* expected Java
* interpretation of {@link java.io.InputStream#available()}, which expects
* the
* number of remaining
* bytes in the local buffer. Moreover, it caps the value returned to a
* maximum of Integer.MAX_VALUE.
* These chang... | 3.68 |
flink_ZooKeeperLeaderElectionHaServices_cleanupZooKeeperPaths | /** Cleans up leftover ZooKeeper paths. */
private void cleanupZooKeeperPaths() throws Exception {
deleteOwnedZNode();
tryDeleteEmptyParentZNodes();
} | 3.68 |
flink_PartitionRequestListenerManager_removeExpiration | /**
* Remove the expire partition request listener and add it to the given timeoutListeners.
*
* @param now the timestamp
* @param timeout the timeout mills
* @param timeoutListeners the expire partition request listeners
*/
public void removeExpiration(
long now, long timeout, Collection<PartitionRequest... | 3.68 |
flink_BlobUtils_getStorageLocation | /**
* Returns the (designated) physical storage location of the BLOB with the given key.
*
* @param storageDir storage directory used be the BLOB service
* @param key the key identifying the BLOB
* @param jobId ID of the job for the incoming files (or <tt>null</tt> if job-unrelated)
* @return the (designated) phy... | 3.68 |
hbase_Filter_setReversed | /**
* alter the reversed scan flag
* @param reversed flag
*/
public void setReversed(boolean reversed) {
this.reversed = reversed;
} | 3.68 |
framework_RpcDataProviderExtension_addDataGenerator | /**
* Adds a {@link DataGenerator} for this {@code RpcDataProviderExtension}.
* DataGenerators are called when sending row data to client. If given
* DataGenerator is already added, this method does nothing.
*
* @since 7.6
* @param generator
* generator to add
*/
public void addDataGenerator(DataGene... | 3.68 |
flink_DefaultResourceCleaner_withRegularCleanup | /**
* Regular cleanups are resources for which the cleanup is triggered after all prioritized
* cleanups succeeded. All added regular cleanups will run concurrently to each other.
*
* @param label The label being used when logging errors in the given cleanup.
* @param regularCleanup The cleanup callback that is go... | 3.68 |
hadoop_GangliaConf_getTmax | /**
* @return the tmax
*/
int getTmax() {
return tmax;
} | 3.68 |
hbase_FanOutOneBlockAsyncDFSOutput_completed | // all lock-free to make it run faster
private void completed(Channel channel) {
for (Iterator<Callback> iter = waitingAckQueue.iterator(); iter.hasNext();) {
Callback c = iter.next();
// if the current unfinished replicas does not contain us then it means that we have already
// acked this one, let's ite... | 3.68 |
framework_DefaultConnectionStateHandler_doReconnect | /**
* Re-sends the payload to the server (if not null) or re-sends a heartbeat
* request immediately.
*
* @param payload
* the payload that did not reach the server, null if the problem
* was detected by a heartbeat
*/
protected void doReconnect(JsonObject payload) {
if (!connection.isA... | 3.68 |
hbase_MasterObserver_postMasterStoreFlush | /**
* Called after the master local region memstore is flushed to disk.
* @param ctx the environment to interact with the framework and master
*/
default void postMasterStoreFlush(final ObserverContext<MasterCoprocessorEnvironment> ctx)
throws IOException {
} | 3.68 |
flink_EntropyInjector_removeEntropyMarkerIfPresent | /**
* Removes the entropy marker string from the path, if the given file system is an
* entropy-injecting file system (implements {@link EntropyInjectingFileSystem}) and the entropy
* marker key is present. Otherwise, this returns the path as is.
*
* @param path The path to filter.
* @return The path without the ... | 3.68 |
flink_BinaryStringData_copy | /** Copy a new {@code BinaryStringData}. */
public BinaryStringData copy() {
ensureMaterialized();
byte[] copy =
BinarySegmentUtils.copyToBytes(
binarySection.segments, binarySection.offset, binarySection.sizeInBytes);
return new BinaryStringData(
new MemorySegmen... | 3.68 |
morf_Criterion_lessThanOrEqualTo | /**
* Helper method to create a new "LESS THAN OR EQUAL TO" expression.
*
* <blockquote><pre>
* Criterion.lessThanOrEqualTo(new Field("startdate"), 20091001);</pre></blockquote>
*
* @param field the field to evaluate in the expression (the left hand side of the expression)
* @param value the value to evaluate... | 3.68 |
hadoop_NMClientAsync_onCommitLastReInitialization | /**
* Callback for commit of last re-initialization.
*
* @param containerId the Id of the container to commit reInitialize.
*/
public void onCommitLastReInitialization(ContainerId containerId) {} | 3.68 |
framework_PureGWTTestApplication_addCommand | /**
* Adds a command item to the menu. When the entry is clicked, the
* command is executed.
*
* @param cmd
* a command object.
*/
public void addCommand(Command cmd) {
menubar.addItem(cmd.title, cmd.command);
items.add(cmd);
} | 3.68 |
AreaShop_RentingRegionEvent_getPlayer | /**
* Get the player that is trying to rent the region.
* @return The player that is trying to rent the region
*/
public OfflinePlayer getPlayer() {
return player;
} | 3.68 |
cron-utils_CronDefinitionBuilder_spring53 | /**
* Creates CronDefinition instance matching Spring (v5.2 onwards) specification.
* https://spring.io/blog/2020/11/10/new-in-spring-5-3-improved-cron-expressions
*
* <p>The cron expression is expected to be a string comprised of 6
* fields separated by white space. Fields can contain any of the allowed
* values... | 3.68 |
framework_EditorOpenEvent_getGrid | /**
* Gets the editors' grid.
*
* @return the editors' grid
*/
public Grid<T> getGrid() {
return getSource().getGrid();
} | 3.68 |
flink_FailureHandlingResult_getFailedExecution | /**
* Returns an {@code Optional} with the {@link Execution} causing this failure or an empty
* {@code Optional} if it's a global failure.
*
* @return The {@code Optional} with the failed {@code Execution} or an empty {@code Optional}
* if it's a global failure.
*/
public Optional<Execution> getFailedExecutio... | 3.68 |
hbase_MetricsSink_getFailedBatches | /**
* Get the count of the failed bathes
*/
protected long getFailedBatches() {
return mss.getFailedBatches();
} | 3.68 |
framework_VAbstractCalendarPanel_doSetDate | /**
* The actual implementation of the logic which sets the data of the Panel.
* The method {@link #setDate(Date)} just delegate a call to this method
* providing additional config parameters.
*
* @param currentDate
* currentDate The date to set
* @param needRerender
* if {@code true} then... | 3.68 |
dubbo_DataQueueCommand_getData | // for test
public byte[] getData() {
return data;
} | 3.68 |
flink_Over_orderBy | /**
* Specifies the time attribute on which rows are ordered.
*
* <p>For streaming tables, reference a rowtime or proctime time attribute here to specify the
* time mode.
*
* <p>For batch tables, refer to a timestamp or long attribute.
*
* @param orderBy field reference
* @return an over window with defined or... | 3.68 |
querydsl_GuavaGroupByBuilder_asSortedTable | /**
* Get the results as sorted table
*
* @param column column expression
* @param expression value expression
* @param rowComparator row comparator
* @param columnComparator column comparator
* @param <C> Column type
* @param <V> Value type
* @return new result transformer
*/
public <C, V> ResultTransformer<... | 3.68 |
hadoop_STSClientFactory_requestSessionCredentials | /**
* Request a set of session credentials.
*
* @param duration duration of the credentials
* @param timeUnit time unit of duration
* @return the role result
* @throws IOException on a failure of the request
*/
@Retries.RetryTranslated
public Credentials requestSessionCredentials(
final long duration,
fi... | 3.68 |
framework_VTree_getSubPartName | /*
* (non-Javadoc)
*
* @see com.vaadin.client.ui.SubPartAware#getSubPartName(com.google
* .gwt.user.client.Element)
*/
@Override
public String getSubPartName(
com.google.gwt.user.client.Element subElement) {
// Supported identifiers:
//
// n[index]/n[index]/n[index]{/expand}
//
// Ends ... | 3.68 |
framework_Button_readDesign | /*
* (non-Javadoc)
*
* @see com.vaadin.ui.AbstractComponent#readDesign(org.jsoup.nodes .Element,
* com.vaadin.ui.declarative.DesignContext)
*/
@Override
public void readDesign(Element design, DesignContext designContext) {
super.readDesign(design, designContext);
Attributes attr = design.attributes();
... | 3.68 |
hadoop_SubApplicationEntityReader_getTable | /**
* Uses the {@link SubApplicationTableRW}.
*/
protected BaseTableRW<?> getTable() {
return SUB_APPLICATION_TABLE;
} | 3.68 |
hbase_NamespaceStateManager_start | /**
* Starts the NamespaceStateManager. The boot strap of cache is done in the post master start hook
* of the NamespaceAuditor class.
* @throws IOException Signals that an I/O exception has occurred.
*/
public void start() throws IOException {
LOG.info("Namespace State Manager started.");
initialize();
} | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.