name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
streampipes_ChangedValueDetectionProcessor_declareModel | //TODO: Change Icon
@Override
public DataProcessorDescription declareModel() {
return ProcessingElementBuilder.create("org.apache.streampipes.processors.transformation.jvm.changed-value")
.category(DataProcessorType.VALUE_OBSERVER)
.withLocales(Locales.EN)
.withAssets(Assets.DOCUMENTATION)
.re... | 3.68 |
pulsar_BrokerService_unblockDispatchersOnUnAckMessages | /**
* Unblocks the dispatchers and removes it from the {@link #blockedDispatchers} list.
*
* @param dispatcherList
*/
public void unblockDispatchersOnUnAckMessages(List<PersistentDispatcherMultipleConsumers> dispatcherList) {
lock.writeLock().lock();
try {
dispatcherList.forEach(dispatcher -> {
... | 3.68 |
hadoop_RollingFileSystemSink_throwMetricsException | /**
* If the sink isn't set to ignore errors, throw a new
* {@link MetricsException}. The message parameter will be used as the
* new exception's message with the current file name
* ({@link #currentFilePath}) appended to it.
*
* @param message the exception message. The message will have a colon and
* the cur... | 3.68 |
flink_AbstractAutoCloseableRegistry_removeCloseableInternal | /** Removes a mapping from the registry map, respecting locking. */
protected final boolean removeCloseableInternal(R closeable) {
synchronized (getSynchronizationLock()) {
return closeableToRef.remove(closeable) != null;
}
} | 3.68 |
hadoop_LpSolver_generateOverAllocationConstraints | /**
* Generate over-allocation constraints.
*
* @param lpModel the LP model.
* @param cJobITimeK actual container allocation for job i in time
* interval k.
* @param oa container over-allocation.
* @param x predicted container allocati... | 3.68 |
hbase_ColumnFamilyDescriptorBuilder_setIndexBlockEncoding | /**
* Set index block encoding algorithm used in block cache.
* @param type What kind of index block encoding will be used.
* @return this (for chained invocation)
*/
public ModifyableColumnFamilyDescriptor setIndexBlockEncoding(IndexBlockEncoding type) {
return setValue(INDEX_BLOCK_ENCODING_BYTES,
type == nu... | 3.68 |
hibernate-validator_ConstraintHelper_findValidatorDescriptors | /**
* Returns those validator descriptors for the given constraint annotation
* matching the given target.
*
* @param annotationType The annotation of interest.
* @param validationTarget The target, either annotated element or parameters.
* @param <A> the type of the annotation
*
* @return A list with matching ... | 3.68 |
hbase_HRegionServer_abort | /**
* Cause the server to exit without closing the regions it is serving, the log it is using and
* without notifying the master. Used unit testing and on catastrophic events such as HDFS is
* yanked out from under hbase or we OOME. the reason we are aborting the exception that caused
* the abort, or null
*/
@Over... | 3.68 |
morf_TableOutputter_outputExampleData | /**
* Outputs the example data rows.
*
* @param numberOfExamples to output
* @param workSheet to add the data rows to
* @param table to get metadata from
* @param startRow to start adding the example rows at
* @param records to add as examples
* @return the new row to carry on outputting at
* @throws WriteExce... | 3.68 |
graphhopper_LandmarkStorage_setMaximumWeight | /**
* Specify the maximum possible value for your used area. With this maximum weight value you can influence the storage
* precision for your weights that help A* finding its way to the goal. The same value is used for all subnetworks.
* Note, if you pick this value too big then too similar weights are stored
* (s... | 3.68 |
framework_AbstractMedia_pause | /**
* Pauses the media.
*/
public void pause() {
getRpcProxy(MediaControl.class).pause();
} | 3.68 |
hmily_PropertyName_isIndexed | /**
* Whether the parameter of the index type list array.
*/
private boolean isIndexed(final int index) {
String element = getElement(index);
return isIndexed(element);
} | 3.68 |
framework_Table_setSortDisabled | /**
* Disables the sorting by the user altogether.
*
* @param sortDisabled
* True if sorting is disabled.
* @deprecated As of 7.0, use {@link #setSortEnabled(boolean)} instead
*/
@Deprecated
public void setSortDisabled(boolean sortDisabled) {
setSortEnabled(!sortDisabled);
} | 3.68 |
framework_ApplicationConfiguration_isQuietDebugMode | /**
* Checks whether debug logging should be quiet.
*
* @return <code>true</code> if debug logging should be quiet
*/
public static boolean isQuietDebugMode() {
String debugParameter = Window.Location.getParameter("debug");
return isDebugAvailable() && debugParameter != null
&& debugParameter.st... | 3.68 |
hadoop_VersionUtil_compareVersions | /**
* Compares two version name strings using maven's ComparableVersion class.
*
* @param version1
* the first version to compare
* @param version2
* the second version to compare
* @return a negative integer if version1 precedes version2, a positive
* integer if version2 precedes vers... | 3.68 |
hbase_Mutation_isReturnResults | /** Returns current value for returnResults */
// Used by Increment and Append only.
@InterfaceAudience.Private
protected boolean isReturnResults() {
byte[] v = getAttribute(RETURN_RESULTS);
return v == null ? true : Bytes.toBoolean(v);
} | 3.68 |
framework_EventHelper_updateFocusHandler | /**
* Adds or removes a focus handler depending on if the connector has focus
* listeners on the server side or not.
*
* @param connector
* The connector to update. Must implement focusHandler.
* @param handlerRegistration
* The old registration reference or null if no handler has been
* ... | 3.68 |
querydsl_Expressions_arrayPath | /**
* Create a new Path expression
*
* @param arrayType array type
* @param metadata path metadata
* @param <A> array type
* @param <E> element type
* @return path expression
*/
public static <A, E> ArrayPath<A, E> arrayPath(Class<A> arrayType, PathMetadata metadata) {
return new ArrayPath<A, E>(arrayType, ... | 3.68 |
hbase_BlockCache_notifyFileCachingCompleted | /**
* Notifies the cache implementation that the given file has been fully cached (all its blocks
* made into the cache).
* @param fileName the file that has been completely cached.
*/
default void notifyFileCachingCompleted(Path fileName, int totalBlockCount, int dataBlockCount,
long size) {
// noop
} | 3.68 |
hbase_MasterObserver_postCompletedTruncateTableAction | /**
* Called after {@link org.apache.hadoop.hbase.master.HMaster} truncates a table. Called as part
* of truncate table procedure and it is async to the truncate RPC call.
* @param ctx the environment to interact with the framework and master
* @param tableName the name of the table
*/
default void postCompl... | 3.68 |
flink_DataStreamStateTTLTestProgram_setBackendWithCustomTTLTimeProvider | /**
* Sets the state backend to a new {@link StubStateBackend} which has a {@link
* MonotonicTTLTimeProvider}.
*
* @param env The {@link StreamExecutionEnvironment} of the job.
*/
private static void setBackendWithCustomTTLTimeProvider(StreamExecutionEnvironment env) {
final MonotonicTTLTimeProvider ttlTimePro... | 3.68 |
framework_BindingValidationStatus_getMessage | /**
* Gets error validation message if status is {@link Status#ERROR}.
*
* @return an optional validation error status or an empty optional if
* status is not an error
*/
public Optional<String> getMessage() {
if (getStatus() == Status.OK || result == null) {
return Optional.empty();
}
... | 3.68 |
hudi_HashFunction_hash | /**
* Hashes a specified key into several integers.
*
* @param k The specified key.
* @return The array of hashed values.
*/
public int[] hash(Key k) {
byte[] b = k.getBytes();
if (b == null) {
throw new NullPointerException("buffer reference is null");
}
if (b.length == 0) {
throw new IllegalArgum... | 3.68 |
hbase_NamespaceAuditor_isInitialized | /**
* Checks if namespace auditor is initialized. Used only for testing.
* @return true, if is initialized
*/
public boolean isInitialized() {
return stateManager.isInitialized();
} | 3.68 |
framework_DateField_setRangeEnd | /**
* Sets the end range for this component. If the value is set after this
* date (taking the resolution into account), the component will not
* validate. If <code>endDate</code> is set to <code>null</code>, any value
* after <code>startDate</code> will be accepted by the range.
*
* @param endDate
* ... | 3.68 |
flink_HiveParserSemanticAnalyzer_processPTFSource | /*
* - a partitionTableFunctionSource can be a tableReference, a SubQuery or another
* PTF invocation.
* - For a TABLEREF: set the source to the alias returned by processTable
* - For a SubQuery: set the source to the alias returned by processSubQuery
* - For a PTF invocation: recursively call processPTFChain.
... | 3.68 |
hbase_Superusers_initialize | /**
* Should be called only once to pre-load list of super users and super groups from Configuration.
* This operation is idempotent.
* @param conf configuration to load users from
* @throws IOException if unable to initialize lists of superusers or super groups
* @throws IllegalStateException if current... | 3.68 |
hbase_Get_setCacheBlocks | /**
* Set whether blocks should be cached for this Get.
* <p>
* This is true by default. When true, default settings of the table and family are used (this
* will never override caching blocks if the block cache is disabled for that family or entirely).
* @param cacheBlocks if false, default settings are overridde... | 3.68 |
morf_AbstractSqlDialectTest_expectedLower | /**
* @return The expected SQL for the LOWER function.
*/
protected String expectedLower() {
return "SELECT LOWER(field1) FROM " + tableName("schedule");
} | 3.68 |
hudi_BaseConsistentHashingBucketClusteringPlanStrategy_checkPrecondition | /**
* TODO maybe add force config to schedule the clustering. It could allow clustering on partitions that are not doing write operation.
* Block clustering if there is any ongoing concurrent writers
*
* @return true if the schedule can proceed
*/
@Override
public boolean checkPrecondition() {
HoodieTimeline tim... | 3.68 |
framework_VaadinSession_getState | /**
* Returns the lifecycle state of this session.
*
* @since 7.2
* @return the current state
*/
public State getState() {
assert hasLock();
return state;
} | 3.68 |
hbase_ClusterStatusTracker_toByteArray | /** Returns Content of the clusterup znode as a serialized pb with the pb magic as prefix. */
static byte[] toByteArray() {
ZooKeeperProtos.ClusterUp.Builder builder = ZooKeeperProtos.ClusterUp.newBuilder();
builder.setStartDate(new java.util.Date().toString());
return ProtobufUtil.prependPBMagic(builder.build().... | 3.68 |
hadoop_CommitUtilsWithMR_getMagicTaskAttemptPath | /**
* Compute the path where the output of a task attempt is stored until
* that task is committed.
* This path is marked as a base path for relocations, so subdirectory
* information is preserved.
* @param context the context of the task attempt.
* @param jobUUID unique Job ID.
* @param dest The output path to ... | 3.68 |
hbase_HFileReaderImpl_getMetaBlock | /**
* @param cacheBlock Add block to cache, if found
* @return block wrapped in a ByteBuffer, with header skipped
*/
@Override
public HFileBlock getMetaBlock(String metaBlockName, boolean cacheBlock) throws IOException {
if (trailer.getMetaIndexCount() == 0) {
return null; // there are no meta blocks
}
if ... | 3.68 |
flink_EnrichedRowData_replaceMutableRow | /**
* Replaces the mutable {@link RowData} backing this {@link EnrichedRowData}.
*
* <p>This method replaces the mutable row data in place and does not return a new object. This
* is done for performance reasons.
*/
public EnrichedRowData replaceMutableRow(RowData mutableRow) {
this.mutableRow = mutableRow;
... | 3.68 |
hbase_DateTieredCompactionPolicy_needsCompaction | /**
* Heuristics for guessing whether we need minor compaction.
*/
@Override
@InterfaceAudience.Private
public boolean needsCompaction(Collection<HStoreFile> storeFiles,
List<HStoreFile> filesCompacting) {
ArrayList<HStoreFile> candidates = new ArrayList<>(storeFiles);
try {
return !selectMinorCompaction(ca... | 3.68 |
hadoop_RegistryPathUtils_getUsername | /**
* Return the username found in the ZK path.
*
* @param recPath the ZK recPath.
* @return the user name.
*/
public static String getUsername(String recPath) {
String user = "anonymous";
Matcher matcher = USER_NAME.matcher(recPath);
if (matcher.find()) {
user = matcher.group(1);
}
return user;
} | 3.68 |
hibernate-validator_TypeHelper_getErasedType | /**
* Gets the erased type of the specified type.
*
* @param type the type to perform erasure on
*
* @return the erased type, never a parameterized type nor a type variable
*
* @see <a href="http://docs.oracle.com/javase/specs/jls/se8/html/jls-4.html#jls-4.6">4.6 Type Erasure</a>
*/
public static Type getErased... | 3.68 |
pulsar_BaseResource_requestAsync | // do the authentication stage, and once authentication completed return a Builder
public CompletableFuture<Builder> requestAsync(final WebTarget target) {
CompletableFuture<Builder> builderFuture = new CompletableFuture<>();
CompletableFuture<Map<String, String>> authFuture = new CompletableFuture<>();
try... | 3.68 |
AreaShop_TeleportFeature_teleportPlayer | /**
* Teleport a player to the region when he has permissions for it.
* @param player Player that should be teleported
* @return true if the teleport succeeded, otherwise false
*/
public boolean teleportPlayer(Player player) {
return teleportPlayer(player, false, true);
} | 3.68 |
hbase_HMaster_getLoadedCoprocessors | /**
* The set of loaded coprocessors is stored in a static set. Since it's statically allocated, it
* does not require that HMaster's cpHost be initialized prior to accessing it.
* @return a String representation of the set of names of the loaded coprocessors.
*/
public static String getLoadedCoprocessors() {
ret... | 3.68 |
flink_DeltaIteration_registerAggregationConvergenceCriterion | /**
* Registers an {@link Aggregator} for the iteration together with a {@link
* ConvergenceCriterion}. For a general description of aggregators, see {@link
* #registerAggregator(String, Aggregator)} and {@link Aggregator}. At the end of each
* iteration, the convergence criterion takes the aggregator's global aggr... | 3.68 |
pulsar_AuthenticationProviderOpenID_validateAllowedAudiences | /**
* Validate the configured allow list of allowedAudiences. The allowedAudiences must be set because
* JWT must have an audience claim.
* See https://openid.net/specs/openid-connect-basic-1_0.html#IDTokenValidation.
* @param allowedAudiences
* @return the validated audiences
*/
String[] validateAllowedAudiences... | 3.68 |
flink_QuickSort_getMaxDepth | /** Deepest recursion before giving up and doing a heapsort. Returns 2 * ceil(log(n)). */
protected static int getMaxDepth(int x) {
if (x <= 0) {
throw new IllegalArgumentException("Undefined for " + x);
}
return (32 - Integer.numberOfLeadingZeros(x - 1)) << 2;
} | 3.68 |
morf_NamedParameterPreparedStatement_createFor | /**
* Create the prepared statement against the specified connection.
*
* @param connection the connection
* @return the prepared statement.
* @throws SQLException if the statement could not be created
*/
public NamedParameterPreparedStatement createFor(Connection connection) throws SQLException {
return new Na... | 3.68 |
hbase_MobUtils_createWriter | /**
* Creates a writer for the mob file in temp directory.
* @param conf The current configuration.
* @param fs The current file system.
* @param family The descriptor of the current column family.
* @param path The path for a temp directory.
* @... | 3.68 |
dubbo_NetUtils_isInvalidPort | /**
* Tells whether the port to test is an invalid port.
*
* @implNote Numeric comparison only.
* @param port port to test
* @return true if invalid
*/
public static boolean isInvalidPort(int port) {
return port < MIN_PORT || port > MAX_PORT;
} | 3.68 |
framework_DateField_getRangeStart | /**
* Returns the precise rangeStart used.
*
* @param startDate
*
*/
public Date getRangeStart() {
return getState(false).rangeStart;
} | 3.68 |
hadoop_MemoryPlacementConstraintManager_addConstraintToMap | /**
* Helper method that adds a constraint to a map for a given source tag.
* Assumes there is already a lock on the constraint map.
*
* @param constraintMap constraint map to which the constraint will be added
* @param sourceTags the source tags that will enable this constraint
* @param placementConstraint the n... | 3.68 |
dubbo_DubboBootstrap_await | /**
* Block current thread to be await.
*
* @return {@link DubboBootstrap}
*/
public DubboBootstrap await() {
// if has been waited, no need to wait again, return immediately
if (!awaited.get()) {
if (!isStopped()) {
executeMutually(() -> {
while (!awaited.get()) {
... | 3.68 |
Activiti_BaseEntityEventListener_isValidEvent | /**
* @return true, if the event is an {@link ActivitiEntityEvent} and (if needed) the entityClass set in this instance, is assignable from the entity class in the event.
*/
protected boolean isValidEvent(ActivitiEvent event) {
boolean valid = false;
if (event instanceof ActivitiEntityEvent) {
if (entityClass... | 3.68 |
hbase_MasterObserver_preRecommissionRegionServer | /**
* Called before recommission region server.
*/
default void preRecommissionRegionServer(ObserverContext<MasterCoprocessorEnvironment> ctx,
ServerName server, List<byte[]> encodedRegionNames) throws IOException {
} | 3.68 |
hbase_StorageClusterStatusModel_setLiveNodes | /**
* @param nodes the list of live node models
*/
public void setLiveNodes(List<Node> nodes) {
this.liveNodes = nodes;
} | 3.68 |
querydsl_DateExpression_month | /**
* Create a month expression (range 1-12 / JAN-DEC)
*
* @return month
*/
public NumberExpression<Integer> month() {
if (month == null) {
month = Expressions.numberOperation(Integer.class, Ops.DateTimeOps.MONTH, mixin);
}
return month;
} | 3.68 |
hudi_HoodieUnMergedLogRecordScanner_newBuilder | /**
* Returns the builder for {@code HoodieUnMergedLogRecordScanner}.
*/
public static HoodieUnMergedLogRecordScanner.Builder newBuilder() {
return new Builder();
} | 3.68 |
hbase_Bytes_toLong | /**
* Converts a byte array to a long value.
* @param bytes array of bytes
* @param offset offset into array
* @param length length of data (must be {@link #SIZEOF_LONG})
* @return the long value
* @throws IllegalArgumentException if length is not {@link #SIZEOF_LONG} or if there's not enough
* ... | 3.68 |
morf_AbstractSqlDialectTest_testOptimiseForRowCountOnSubquery | /**
* Check that we don't allow the use of the optimise for row count hint on a subquery.
*/
@Test(expected = IllegalArgumentException.class)
public void testOptimiseForRowCountOnSubquery() {
testDialect.convertStatementToSQL(
select().from(select().from("Foo").optimiseForRowCount(1))
);
} | 3.68 |
rocketmq-connect_RecordOffsetManagement_submitRecord | /**
* submit record
*
* @param position
* @return
*/
public SubmittedPosition submitRecord(RecordPosition position) {
SubmittedPosition submittedPosition = new SubmittedPosition(position);
records.computeIfAbsent(position.getPartition(), e -> new LinkedList<>()).add(submittedPosition);
// ensure thread... | 3.68 |
hbase_AuthManager_removeTable | /**
* Remove given table from AuthManager's table cache.
* @param table table name
*/
public void removeTable(TableName table) {
tableCache.remove(table);
} | 3.68 |
flink_StaticFileServerHandler_setDateAndCacheHeaders | /**
* Sets the "date" and "cache" headers for the HTTP Response.
*
* @param response The HTTP response object.
* @param fileToCache File to extract the modification timestamp from.
*/
public static void setDateAndCacheHeaders(HttpResponse response, File fileToCache) {
SimpleDateFormat dateFormatter = new Simpl... | 3.68 |
hudi_HoodieHiveUtils_getTimestampWriteable | /**
* Get timestamp writeable object from long value.
* Hive3 use TimestampWritableV2 to build timestamp objects and Hive2 use TimestampWritable.
* So that we need to initialize timestamp according to the version of Hive.
*/
public static Writable getTimestampWriteable(long value, boolean timestampMillis) {
retur... | 3.68 |
hbase_MetricsMaster_setNumTableInSpaceQuotaViolation | /**
* Sets the number of table in violation of a space quota.
* @see MetricsMasterQuotaSource#updateNumTablesInSpaceQuotaViolation(long)
*/
public void setNumTableInSpaceQuotaViolation(final long numTablesInViolation) {
masterQuotaSource.updateNumTablesInSpaceQuotaViolation(numTablesInViolation);
} | 3.68 |
flink_PlanReference_fromJsonString | /** Create a reference starting from a JSON string. */
public static PlanReference fromJsonString(String jsonString) {
Objects.requireNonNull(jsonString, "Json string cannot be null");
return new ContentPlanReference(jsonString);
} | 3.68 |
rocketmq-connect_AvroData_splitName | /**
* Split a full dotted-syntax name into a namespace and a single-component name.
*/
private static String[] splitName(String fullName) {
String[] result = new String[2];
int indexLastDot = fullName.lastIndexOf('.');
if (indexLastDot >= 0) {
result[0] = fullName.substring(0, indexLastDot);
... | 3.68 |
hadoop_ConverterUtils_getYarnUrlFromURI | /*
* This method is deprecated, use {@link URL#fromURI(URI)} instead.
*/
@Public
@Deprecated
public static URL getYarnUrlFromURI(URI uri) {
return URL.fromURI(uri);
} | 3.68 |
hbase_ColumnFamilyDescriptorBuilder_setEvictBlocksOnClose | /**
* Set the setEvictBlocksOnClose flag.
* @param value true if we should evict cached blocks from the blockcache on close
* @return this (for chained invocation)
*/
public ModifyableColumnFamilyDescriptor setEvictBlocksOnClose(boolean value) {
return setValue(EVICT_BLOCKS_ON_CLOSE_BYTES, Boolean.toString(value)... | 3.68 |
hbase_ReplicationSourceLogQueue_getQueue | /**
* Return queue for the given walGroupId Please don't add or remove elements from the returned
* queue. Use {@link #enqueueLog(Path, String)} and {@link #remove(String)} methods respectively.
* @param walGroupId walGroupId
*/
public PriorityBlockingQueue<Path> getQueue(String walGroupId) {
return queues.get(wa... | 3.68 |
flink_SharingPhysicalSlotRequestBulk_clearPendingRequests | /**
* Clear the pending requests.
*
* <p>The method can be used to make the bulk fulfilled and stop the fulfillability check in
* {@link PhysicalSlotRequestBulkChecker}.
*/
void clearPendingRequests() {
pendingRequests.clear();
} | 3.68 |
hadoop_OBSListing_accept | /**
* Accept all prefixes except the one for the base path, "self".
*
* @param keyPath qualified path to the entry
* @param prefix common prefix in listing.
* @return true if the entry is accepted (i.e. that a status entry should be
* generated.
*/
@Override
public boolean accept(final Path keyPath, final Strin... | 3.68 |
hbase_SnapshotQuotaObserverChore_getTimeUnit | /**
* Extracts the time unit for the chore period and initial delay from the configuration. The
* configuration value for {@link #SNAPSHOT_QUOTA_CHORE_TIMEUNIT_KEY} must correspond to a
* {@link TimeUnit} value.
* @param conf The configuration object.
* @return The configured time unit for the chore period and ini... | 3.68 |
hadoop_FederationCache_buildGetPoliciesConfigurationsCacheRequest | /**
* Build GetPoliciesConfigurations CacheRequest.
*
* @param cacheKey cacheKey.
* @return CacheRequest.
* @throws YarnException exceptions from yarn servers.
*/
protected CacheRequest<String, CacheResponse<SubClusterPolicyConfiguration>>
buildGetPoliciesConfigurationsCacheRequest(String cacheKey) throws Yar... | 3.68 |
hbase_MetricsSource_getWALReaderEditsBufferUsage | /**
* Returns the amount of memory in bytes used in this RegionServer by edits pending replication.
*/
public long getWALReaderEditsBufferUsage() {
return globalSourceSource.getWALReaderEditsBufferBytes();
} | 3.68 |
hbase_SnapshotManager_takeSnapshot | /**
* Take a snapshot based on the enabled/disabled state of the table.
* @throws HBaseSnapshotException when a snapshot specific exception occurs.
* @throws IOException when some sort of generic IO exception occurs.
*/
public void takeSnapshot(SnapshotDescription snapshot) throws IOException {
this.ta... | 3.68 |
pulsar_ManagedLedgerConfig_setLazyCursorRecovery | /**
* Whether to recover cursors lazily when trying to recover a
* managed ledger backing a persistent topic. It can improve write availability of topics.
* The caveat is now when recovered ledger is ready to write we're not sure if all old consumers last mark
* delete position can be recovered or not.
* @param la... | 3.68 |
hadoop_TimelineStateStore_serviceInit | /**
* Initialize the state storage
*
* @param conf the configuration
* @throws IOException
*/
@Override
public void serviceInit(Configuration conf) throws IOException {
initStorage(conf);
} | 3.68 |
hadoop_PathLocation_getDestinations | /**
* Get the list of locations found in the mount table.
* The first result is the highest priority path.
*
* @return List of remote locations.
*/
public List<RemoteLocation> getDestinations() {
return Collections.unmodifiableList(this.destinations);
} | 3.68 |
flink_ComponentMetricGroup_close | /**
* Closes the component group by removing and closing all metrics and subgroups (inherited from
* {@link AbstractMetricGroup}), plus closing and removing all dedicated component subgroups.
*/
@Override
public void close() {
synchronized (this) {
if (!isClosed()) {
// remove all metrics and... | 3.68 |
hadoop_RLESparseResourceAllocation_getCapacityAtTime | /**
* Returns the capacity, i.e. total resources allocated at the specified point
* of time.
*
* @param tick timeStap at which resource needs to be known
* @return the resources allocated at the specified time
*/
public Resource getCapacityAtTime(long tick) {
readLock.lock();
try {
Entry<Long, Resource> c... | 3.68 |
hudi_SimpleBloomFilter_write | // @Override
public void write(DataOutput out) throws IOException {
out.write(getUTF8Bytes(filter.toString()));
} | 3.68 |
open-banking-gateway_DateTimeFormatConfig_addFormatters | /**
* Swagger-codegen is not able to produce @DateTimeFormat annotation:
* https://github.com/swagger-api/swagger-codegen/issues/1235
* https://github.com/swagger-api/swagger-codegen/issues/4113
* To fix this - forcing formatters globally.
*/
@Override
public void addFormatters(FormatterRegistry registry) {
Da... | 3.68 |
flink_RocksDBStateBackend_setNumberOfTransferThreads | /**
* Sets the number of threads used to transfer files while snapshotting/restoring.
*
* @param numberOfTransferThreads The number of threads used to transfer files while
* snapshotting/restoring.
*/
public void setNumberOfTransferThreads(int numberOfTransferThreads) {
rocksDBStateBackend.setNumberOfTrans... | 3.68 |
rocketmq-connect_JsonConverterConfig_cacheSize | /**
* return cache size
*
* @return
*/
public int cacheSize() {
return cacheSize;
} | 3.68 |
streampipes_StreamRequirementsBuilder_requiredProperty | /**
* Sets a new property requirement, e.g., a property of a specific data type or with specific semantics
* a data stream that is connected to this pipeline element must provide.
*
* @param propertyRequirement The property requirement.
* Use {@link org.apache.streampipes.sdk.helpers.EpR... | 3.68 |
framework_CalendarDropHandler_getApplicationConnection | /*
* (non-Javadoc)
*
* @see com.vaadin.terminal.gwt.client.ui.dd.VDropHandler#
* getApplicationConnection ()
*/
@Override
public ApplicationConnection getApplicationConnection() {
return calendarConnector.getClient();
} | 3.68 |
hadoop_PathLocation_getDestinationOrder | /**
* Get the order for the destinations.
*
* @return Order for the destinations.
*/
public DestinationOrder getDestinationOrder() {
return this.destOrder;
} | 3.68 |
flink_FileChannelMemoryMappedBoundedData_close | /**
* Closes the file and unmaps all memory mapped regions. After calling this method, access to
* any ByteBuffer obtained from this instance will cause a segmentation fault.
*/
public void close() throws IOException {
IOUtils.closeQuietly(fileChannel);
for (ByteBuffer bb : memoryMappedRegions) {
Pl... | 3.68 |
framework_AbstractComponent_writeDesign | /*
* (non-Javadoc)
*
* @see com.vaadin.ui.Component#writeDesign(org.jsoup.nodes.Element,
* com.vaadin.ui.declarative.DesignContext)
*/
@Override
public void writeDesign(Element design, DesignContext designContext) {
AbstractComponent def = designContext.getDefaultInstance(this);
Attributes attr = design.at... | 3.68 |
flink_Tuple1_setFields | /**
* Sets new values to all fields of the tuple.
*
* @param f0 The value for field 0
*/
public void setFields(T0 f0) {
this.f0 = f0;
} | 3.68 |
framework_DataCommunicator_onRequestRows | /**
* Request the given rows to be available on the client side.
*
* @param firstRowIndex
* the index of the first requested row
* @param numberOfRows
* the number of requested rows
* @param firstCachedRowIndex
* the index of the first cached row
* @param cacheSize
* ... | 3.68 |
querydsl_CollectionExpressionBase_isEmpty | /**
* Create a {@code this.isEmpty()} expression
*
* <p>Evaluates to true, if this has no elements.</p>
*
* @return this.isEmpty()
*/
public final BooleanExpression isEmpty() {
if (empty == null) {
empty = Expressions.booleanOperation(Ops.COL_IS_EMPTY, mixin);
}
return empty;
} | 3.68 |
hadoop_TypedBytesInput_readRawMap | /**
* Reads the raw bytes following a <code>Type.MAP</code> code.
* @return the obtained bytes sequence
* @throws IOException
*/
public byte[] readRawMap() throws IOException {
Buffer buffer = new Buffer();
int length = readMapHeader();
buffer.append(new byte[] {
(byte) Type.MAP.code,
(byte) (0xff & (... | 3.68 |
flink_InputChannel_increaseBackoff | /**
* Increases the current backoff and returns whether the operation was successful.
*
* @return <code>true</code>, iff the operation was successful. Otherwise, <code>false</code>.
*/
protected boolean increaseBackoff() {
// Backoff is disabled
if (initialBackoff == 0) {
return false;
}
if... | 3.68 |
hudi_HoodieKeyLookupHandle_getLookupResult | /**
* Of all the keys, that were added, return a list of keys that were actually found in the file group.
*/
public HoodieKeyLookupResult getLookupResult() {
if (LOG.isDebugEnabled()) {
LOG.debug("#The candidate row keys for " + partitionPathFileIDPair + " => " + candidateRecordKeys);
}
HoodieBaseFile base... | 3.68 |
flink_StreamElement_asRecord | /**
* Casts this element into a StreamRecord.
*
* @return This element as a stream record.
* @throws java.lang.ClassCastException Thrown, if this element is actually not a stream record.
*/
@SuppressWarnings("unchecked")
public final <E> StreamRecord<E> asRecord() {
return (StreamRecord<E>) this;
} | 3.68 |
pulsar_MessageIdAdv_compareTo | /**
* The default implementation of {@link Comparable#compareTo(Object)}.
*/
default int compareTo(MessageId o) {
if (!(o instanceof MessageIdAdv)) {
throw new UnsupportedOperationException("Unknown MessageId type: "
+ ((o != null) ? o.getClass().getName() : "null"));
}
final Messa... | 3.68 |
flink_SystemProcessingTimeService_isAlive | /**
* @return {@code true} is the status of the service is {@link #STATUS_ALIVE}, {@code false}
* otherwise.
*/
@VisibleForTesting
boolean isAlive() {
return status.get() == STATUS_ALIVE;
} | 3.68 |
flink_CliFrontend_list | /**
* Executes the list action.
*
* @param args Command line arguments for the list action.
*/
protected void list(String[] args) throws Exception {
LOG.info("Running 'list' command.");
final Options commandOptions = CliFrontendParser.getListCommandOptions();
final CommandLine commandLine = getCommandL... | 3.68 |
hbase_ProcedureExecutor_isStarted | /**
* Return true if the procedure is started.
* @param procId the ID of the procedure to check
* @return true if the procedure execution is started, otherwise false.
*/
public boolean isStarted(long procId) {
Procedure<?> proc = procedures.get(procId);
if (proc == null) {
return completed.get(procId) != nu... | 3.68 |
flink_KubernetesUtils_getNamespacedServiceName | /** Generate namespaced name of the service. */
public static String getNamespacedServiceName(Service service) {
return service.getMetadata().getName() + "." + service.getMetadata().getNamespace();
} | 3.68 |
framework_VTree_isGrandParentOf | /**
* Travels up the hierarchy looking for this node.
*
* @param child
* The child which grandparent this is or is not
* @return True if this is a grandparent of the child node
*/
public boolean isGrandParentOf(TreeNode child) {
TreeNode currentNode = child;
boolean isGrandParent = false;
w... | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.