name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
morf_DatabaseMetaDataProvider_loadAllTableNames | /**
* Creates a map of all table names,
* indexed by their case-agnostic names.
*
* @return Map of real table names.
*/
protected Map<AName, RealName> loadAllTableNames() {
final ImmutableMap.Builder<AName, RealName> tableNameMappings = ImmutableMap.builder();
try {
final DatabaseMetaData databaseMetaData... | 3.68 |
hadoop_UnmanagedApplicationManager_getAMRMClientRelayer | /**
* Returns the rmProxy relayer of this UAM.
*
* @return rmProxy relayer of the UAM
*/
public AMRMClientRelayer getAMRMClientRelayer() {
return this.rmProxyRelayer;
} | 3.68 |
hbase_CacheConfig_shouldEvictOnClose | /**
* @return true if blocks should be evicted from the cache when an HFile reader is closed, false
* if not
*/
public boolean shouldEvictOnClose() {
return this.evictOnClose;
} | 3.68 |
hbase_Scan_addFamily | /**
* Get all columns from the specified family.
* <p>
* Overrides previous calls to addColumn for this family.
* @param family family name
*/
public Scan addFamily(byte[] family) {
familyMap.remove(family);
familyMap.put(family, null);
return this;
} | 3.68 |
flink_AdaptiveScheduler_computeReactiveModeVertexParallelismStore | /**
* Creates the parallelism store for a set of vertices, optionally with a flag to leave the
* vertex parallelism unchanged. If the flag is set, the parallelisms must be valid for
* execution.
*
* <p>We need to set parallelism to the max possible value when requesting resources, but when
* executing the graph w... | 3.68 |
hmily_HmilyTacLocalParticipantExecutor_cancel | /**
* Do cancel.
*
* @param participant hmily participant
*/
public static void cancel(final HmilyParticipant participant) {
List<HmilyParticipantUndo> undoList = HmilyParticipantUndoCacheManager.getInstance().get(participant.getParticipantId());
for (HmilyParticipantUndo undo : undoList) {
boolean ... | 3.68 |
flink_ExceptionUtils_tryEnrichOutOfMemoryError | /**
* Tries to enrich OutOfMemoryErrors being part of the passed root Throwable's cause tree.
*
* <p>This method improves error messages for direct and metaspace {@link OutOfMemoryError}. It
* adds description about the possible causes and ways of resolution.
*
* @param root The Throwable of which the cause tree ... | 3.68 |
hadoop_ResourceUsageMetrics_setVirtualMemoryUsage | /**
* Set the virtual memory usage.
*/
public void setVirtualMemoryUsage(long usage) {
virtualMemoryUsage = usage;
} | 3.68 |
hbase_ScheduledChore_initialChore | /**
* Override to run a task before we start looping.
* @return true if initial chore was successful
*/
protected boolean initialChore() {
// Default does nothing
return true;
} | 3.68 |
framework_Table_getUpdatedRowCount | /**
* Subclass and override this to enable partial row updates, bypassing the
* normal caching and lazy loading mechanism. This is useful for updating
* the state of certain rows, e.g. in the TreeTable the collapsed state of a
* single node is updated using this mechanism.
*
* @return the number of rows to update... | 3.68 |
rocketmq-connect_ConfigManagementService_configure | /**
* Configure class with the given key-value pairs
*
* @param config can be DistributedConfig or StandaloneConfig
*/
default void configure(WorkerConfig config) {
} | 3.68 |
framework_TouchScrollDelegate_isMoved | /**
* Has user moved the touch.
*
* @return
*/
public boolean isMoved() {
return moved;
} | 3.68 |
hbase_ClientTokenUtil_obtainAndCacheToken | /**
* Obtain an authentication token for the given user and add it to the user's credentials.
* @param conn The HBase cluster connection
* @param user The user for whom to obtain the token
* @throws IOException If making a remote call to the authentication service fails
* @throws InterruptedException If e... | 3.68 |
hbase_HFileReaderImpl_getCachedBlock | /**
* Retrieve block from cache. Validates the retrieved block's type vs {@code expectedBlockType}
* and its encoding vs. {@code expectedDataBlockEncoding}. Unpacks the block as necessary.
*/
private HFileBlock getCachedBlock(BlockCacheKey cacheKey, boolean cacheBlock, boolean useLock,
boolean updateCacheMetrics, ... | 3.68 |
hbase_SegmentScanner_backwardSeek | /**
* Seek the scanner at or before the row of specified Cell, it firstly tries to seek the scanner
* at or after the specified Cell, return if peek KeyValue of scanner has the same row with
* specified Cell, otherwise seek the scanner at the first Cell of the row which is the previous
* row of specified KeyValue
... | 3.68 |
hbase_VersionInfo_getUrl | /**
* Get the subversion URL for the root hbase directory.
* @return the url
*/
public static String getUrl() {
return Version.url;
} | 3.68 |
hadoop_Histogram_getCDF | /**
* Produces a discrete approximation of the CDF. The user provides the points
* on the {@code Y} axis he wants, and we give the corresponding points on the
* {@code X} axis, plus the minimum and maximum from the data.
*
* @param scale
* the denominator applied to every element of buckets. For example... | 3.68 |
flink_BinaryRowWriter_setNullAt | /** Default not null. */
@Override
public void setNullAt(int pos) {
setNullBit(pos);
segment.putLong(getFieldOffset(pos), 0L);
} | 3.68 |
hadoop_SinglePendingCommit_destinationPath | /**
* Build the destination path of the object.
* @return the path
* @throws IllegalStateException if the URI is invalid
*/
public Path destinationPath() {
Preconditions.checkState(StringUtils.isNotEmpty(uri), "Empty uri");
try {
return new Path(new URI(uri));
} catch (URISyntaxException e) {
throw ne... | 3.68 |
pulsar_PulsarAdminException_clone | /**
* This method is meant to be overriden by all subclasses.
* We cannot make it 'abstract' because it would be a breaking change in the public API.
* @return a new PulsarAdminException
*/
protected PulsarAdminException clone() {
return new PulsarAdminException(getMessage(), getCause(), httpError, statusCode);... | 3.68 |
hbase_AsyncScanSingleRegionRpcRetryingCaller_prepare | // return false if the scan has already been resumed. See the comment above for ScanResumerImpl
// for more details.
synchronized boolean prepare(ScanResponse resp, int numberOfCompleteRows) {
if (state == ScanResumerState.RESUMED) {
// user calls resume before we actually suspend the scan, just continue;
ret... | 3.68 |
framework_AbstractComponent_getWidthUnits | /*
* (non-Javadoc)
*
* @see com.vaadin.server.Sizeable#getWidthUnits()
*/
@Override
public Unit getWidthUnits() {
return widthUnit;
} | 3.68 |
hbase_ZKConfig_getClientZKQuorumServersString | /**
* Get the client ZK Quorum servers string
* @param conf the configuration to read
* @return Client quorum servers, or null if not specified
*/
public static String getClientZKQuorumServersString(Configuration conf) {
setZooKeeperClientSystemProperties(HConstants.ZK_CFG_PROPERTY_PREFIX, conf);
String clientQ... | 3.68 |
flink_TaskDeploymentDescriptor_getAttemptNumber | /** Returns the attempt number of the subtask. */
public int getAttemptNumber() {
return executionId.getAttemptNumber();
} | 3.68 |
querydsl_GeometryExpression_overlaps | /**
* Returns 1 (TRUE) if this geometric object “spatially overlaps” anotherGeometry.
*
* @param geometry other geometry
* @return true, if overlaps
*/
public BooleanExpression overlaps(Expression<? extends Geometry> geometry) {
return Expressions.booleanOperation(SpatialOps.OVERLAPS, mixin, geometry);
} | 3.68 |
flink_CommonTestUtils_assertThrows | /** Checks whether an exception with a message occurs when running a piece of code. */
public static void assertThrows(
String msg, Class<? extends Exception> expected, Callable<?> code) {
try {
Object result = code.call();
Assert.fail("Previous method call should have failed but it returned... | 3.68 |
hibernate-validator_PlatformResourceBundleLocator_run | /**
* Runs the given privileged action, using a privileged block if required.
* <p>
* <b>NOTE:</b> This must never be changed into a publicly available method to avoid execution of arbitrary
* privileged actions within HV's protection domain.
*/
@IgnoreForbiddenApisErrors(reason = "SecurityManager is deprecated in... | 3.68 |
shardingsphere-elasticjob_TaskContext_getIdForUnassignedSlave | /**
* Get unassigned task ID before job execute.
*
* @param id task ID
* @return unassigned task ID before job execute
*/
public static String getIdForUnassignedSlave(final String id) {
return id.replaceAll(TaskContext.from(id).getSlaveId(), UNASSIGNED_SLAVE_ID);
} | 3.68 |
morf_SchemaChangeSequence_applyToSchema | /**
* Applies the changes to the given schema.
*
* @param initialSchema The schema to apply changes to.
* @return the resulting schema after applying changes in this sequence
*/
public Schema applyToSchema(Schema initialSchema) {
Schema currentSchema = initialSchema;
for (UpgradeStepWithChanges changesForStep ... | 3.68 |
hadoop_SuccessData_serializer | /**
* Get a JSON serializer for this class.
* @return a serializer.
*/
public static JsonSerialization<SuccessData> serializer() {
return new JsonSerialization<>(SuccessData.class, false, false);
} | 3.68 |
flink_CachingLookupFunction_open | /**
* Open the {@link CachingLookupFunction}.
*
* <p>In order to reduce the memory usage of the cache, {@link LookupCacheManager} is used to
* provide a shared cache instance across subtasks of this function. Here we use {@link
* #functionIdentifier()} as the id of the cache, which is generated by MD5 of serialize... | 3.68 |
hbase_MetaTableAccessor_fullScanTables | /**
* Performs a full scan of <code>hbase:meta</code> for tables.
* @param connection connection we're using
* @param visitor Visitor invoked against each row in tables family.
*/
public static void fullScanTables(Connection connection,
final ClientMetaTableAccessor.Visitor visitor) throws IOException {
scan... | 3.68 |
hbase_TableMapReduceUtil_addHBaseDependencyJars | /**
* Add HBase and its dependencies (only) to the job configuration.
* <p>
* This is intended as a low-level API, facilitating code reuse between this class and its mapred
* counterpart. It also of use to external tools that need to build a MapReduce job that interacts
* with HBase but want fine-grained control o... | 3.68 |
hadoop_DynoInfraUtils_getNameNodeWebUri | /**
* Get the URI that can be used to access the launched NameNode's web UI, e.g.
* for JMX calls.
*
* @param nameNodeProperties The set of properties representing the
* information about the launched NameNode.
* @return The URI to the web UI.
*/
static URI getNameNodeWebUri(Properties ... | 3.68 |
shardingsphere-elasticjob_JobConfigurationPOJO_toJobConfiguration | /**
* Convert to job configuration.
*
* @return job configuration
*/
public JobConfiguration toJobConfiguration() {
JobConfiguration result = JobConfiguration.newBuilder(jobName, shardingTotalCount)
.cron(cron).timeZone(timeZone).shardingItemParameters(shardingItemParameters).jobParameter(jobParamet... | 3.68 |
pulsar_ManagedLedgerConfig_getEnsembleSize | /**
* @return the ensembleSize
*/
public int getEnsembleSize() {
return ensembleSize;
} | 3.68 |
hbase_IncrementalBackupManager_getLogFilesForNewBackup | /**
* For each region server: get all log files newer than the last timestamps but not newer than the
* newest timestamps.
* @param olderTimestamps the timestamp for each region server of the last backup.
* @param newestTimestamps the timestamp for each region server that the backup should lead to.
* @param conf ... | 3.68 |
graphhopper_Frequency_getId | /**
* Frequency entries have no ID in GTFS so we define one based on the fields in the frequency entry.
*
* It is possible to have two identical frequency entries in the GTFS, which under our understanding of the situation
* would mean that two sets of vehicles were randomly running the same trip at the same headwa... | 3.68 |
hadoop_AuxServiceRecord_launchTime | /**
* The time when the service was created, e.g. 2016-03-16T01:01:49.000Z.
**/
public AuxServiceRecord launchTime(Date time) {
this.launchTime = time == null ? null : (Date) time.clone();
return this;
} | 3.68 |
hibernate-validator_AbstractConstraintValidatorManagerImpl_resolveAssignableTypes | /**
* Tries to reduce all assignable classes down to a single class.
*
* @param assignableTypes The set of all classes which are assignable to the class of the value to be validated and
* which are handled by at least one of the validators for the specified constraint.
*/
private void resolveAssignableTypes(List<T... | 3.68 |
framework_VMenuBar_setSelected | /**
* Set the currently selected item of this menu.
*
* @param item
*/
public void setSelected(CustomMenuItem item) {
// If we had something selected, unselect
if (item != selected && selected != null) {
selected.setSelected(false);
}
// If we have a valid selection, select it
if (item !... | 3.68 |
hbase_SnapshotScannerHDFSAclHelper_addTableAcl | /**
* Add table user acls
* @param tableName the table
* @param users the table users with READ permission
* @return false if an error occurred, otherwise true
*/
public boolean addTableAcl(TableName tableName, Set<String> users, String operation) {
try {
long start = EnvironmentEdgeManager.currentTime()... | 3.68 |
shardingsphere-elasticjob_JobNodeStorage_getJobRootNodeData | /**
* Get job root node data.
*
* @return data of job node
*/
public String getJobRootNodeData() {
return regCenter.get("/" + jobName);
} | 3.68 |
hbase_HBaseZKTestingUtility_setupClusterTestDir | /**
* Creates a directory for the cluster, under the test data
*/
protected void setupClusterTestDir() {
if (clusterTestDir != null) {
return;
}
// Using randomUUID ensures that multiple clusters can be launched by
// a same test, if it stops & starts them
Path testDir = getDataTestDir("cluster_" + get... | 3.68 |
hbase_HBaseTestingUtility_createWal | /**
* Create an unmanaged WAL. Be sure to close it when you're through.
*/
public static WAL createWal(final Configuration conf, final Path rootDir, final RegionInfo hri)
throws IOException {
// The WAL subsystem will use the default rootDir rather than the passed in rootDir
// unless I pass along via the conf.... | 3.68 |
hadoop_MawoConfiguration_getZKAddress | /**
* Get ZooKeeper Address.
* @return value of ZooKeeper.address
*/
public String getZKAddress() {
return configsMap.get(ZK_ADDRESS);
} | 3.68 |
hadoop_OBSDataBlocks_flush | /**
* Flush operation will flush to disk.
*
* @throws IOException IOE raised on FileOutputStream
*/
@Override
void flush() throws IOException {
super.flush();
out.flush();
} | 3.68 |
hadoop_NativeTaskOutputFiles_getOutputFile | /**
* Return the path to local map output file created earlier
*/
public Path getOutputFile() throws IOException {
String path = String.format(OUTPUT_FILE_FORMAT_STRING, TASKTRACKER_OUTPUT, id);
return lDirAlloc.getLocalPathToRead(path, conf);
} | 3.68 |
hadoop_FlowActivityColumnPrefix_getColumnPrefix | /**
* @return the column name value
*/
public String getColumnPrefix() {
return columnPrefix;
} | 3.68 |
hadoop_AzureBlobFileSystem_getCanonicalServiceName | /**
* If Delegation tokens are enabled, the canonical service name of
* this filesystem is the filesystem URI.
* @return either the filesystem URI as a string, or null.
*/
@Override
public String getCanonicalServiceName() {
String name = null;
if (delegationTokenManager != null) {
name = delegationTokenMana... | 3.68 |
hibernate-validator_AbstractMethodOverrideCheck_getEnclosingTypeElementQualifiedName | /**
* Find a {@link String} representation of qualified name ({@link Name}) of corresponding {@link TypeElement} that
* contains a given {@link ExecutableElement}.
*
* @param currentMethod a method
* @return a class/interface qualified name represented by {@link String} to which a method belongs to
*/
protected S... | 3.68 |
shardingsphere-elasticjob_JobShutdownHookPlugin_start | /**
* <p>
* Called when the associated <code>Scheduler</code> is started, in order
* to let the plug-in know it can now make calls into the scheduler if it
* needs to.
* </p>
*/
@Override
public void start() {
} | 3.68 |
hbase_ThriftConnection_getAdmin | /**
* Get a ThriftAdmin, ThriftAdmin is NOT thread safe
* @return a ThriftAdmin
* @throws IOException IOException
*/
@Override
public Admin getAdmin() throws IOException {
Pair<THBaseService.Client, TTransport> client = clientBuilder.getClient();
return new ThriftAdmin(client.getFirst(), client.getSecond(), con... | 3.68 |
morf_InsertStatement_from | /**
* Specifies the table to source the data from
*
* @param sourceTable the table to source the data from
* @return a statement with the changes applied.
*
*/
public InsertStatement from(TableReference sourceTable) {
return copyOnWriteOrMutate(
b -> b.from(sourceTable),
() -> {
if (selectSt... | 3.68 |
druid_CharsetConvert_decode | /**
* 字符串解码
*
* @param s String
* @return String
* @throws UnsupportedEncodingException
*/
public String decode(String s) throws UnsupportedEncodingException {
if (enable && !isEmpty(s)) {
s = new String(s.getBytes(serverEncoding), clientEncoding);
}
return s;
} | 3.68 |
flink_AbstractBytesMultiMap_checkSkipReadForPointer | /** For pointer needing update, skip unaligned part (4 bytes) for convenient updating. */
private void checkSkipReadForPointer(AbstractPagedInputView source) throws IOException {
// skip if there is no enough size.
// Note: Use currentSegmentLimit instead of segmentSize.
int available = source.getCurrentSeg... | 3.68 |
hbase_HRegionServer_setupWALAndReplication | /**
* Setup WAL log and replication if enabled. Replication setup is done in here because it wants to
* be hooked up to WAL.
*/
private void setupWALAndReplication() throws IOException {
WALFactory factory = new WALFactory(conf, serverName, this);
// TODO Replication make assumptions here based on the default fi... | 3.68 |
framework_AbstractSplitPanelElement_getFirstComponent | /**
* Gets the first component of a split panel and wraps it in given class.
*
* @param clazz
* Components element class
* @return First component wrapped in given class
*/
public <T extends AbstractElement> T getFirstComponent(Class<T> clazz) {
return getContainedComponent(clazz, byFirstContainer)... | 3.68 |
hadoop_SuccessData_joinMap | /**
* Join any map of string to value into a string, sorting the keys first.
* @param map map to join
* @param prefix prefix before every entry
* @param middle string between key and value
* @param suffix suffix to each entry
* @return a string for reporting.
*/
protected static String joinMap(Map<String, ?> map... | 3.68 |
morf_TableOutputter_spreadsheetifyName | /**
* Converts camel capped names to something we can show in a spreadsheet.
*
* @param name Name to convert.
* @return A human readable version of the name wtih camel caps replaced by spaces.
*/
private String spreadsheetifyName(String name) {
return StringUtils.capitalize(name).replaceAll("([A-Z][a-z])", " $1"... | 3.68 |
hbase_ScannerModel_setFilter | /**
* @param filter the filter specification
*/
public void setFilter(String filter) {
this.filter = filter;
} | 3.68 |
hbase_ColumnFamilyDescriptorBuilder_setVersions | /**
* Set minimum and maximum versions to keep.
* @param minVersions minimal number of versions
* @param maxVersions maximum number of versions
* @return this (for chained invocation)
*/
public ModifyableColumnFamilyDescriptor setVersions(int minVersions, int maxVersions) {
if (minVersions <= 0) {
// TODO: A... | 3.68 |
flink_CompletedOperationCache_containsOperation | /** Returns whether this cache contains an operation under the given operation key. */
public boolean containsOperation(final K operationKey) {
return registeredOperationTriggers.containsKey(operationKey)
|| completedOperations.getIfPresent(operationKey) != null;
} | 3.68 |
framework_VaadinService_generateConnectorId | /**
* Generates a unique id to use for a newly attached connector.
*
* @see ConnectorIdGenerator
* @see #initConnectorIdGenerator(List)
*
* @since 8.1
*
* @param session
* the session to which the connector has been attached, not
* <code>null</code>
* @param connector
* the ... | 3.68 |
hbase_RegionState_toDescriptiveString | /**
* A slower (but more easy-to-read) stringification
*/
public String toDescriptiveString() {
long relTime = EnvironmentEdgeManager.currentTime() - stamp;
return hri.getRegionNameAsString() + " state=" + state + ", ts=" + new Date(stamp) + " ("
+ (relTime / 1000) + "s ago)" + ", server=" + serverName;
} | 3.68 |
flink_PendingCheckpointStats_toFailedCheckpoint | /**
* Reports a failed pending checkpoint.
*
* @param failureTimestamp Timestamp of the failure.
* @param cause Optional cause of the failure.
*/
FailedCheckpointStats toFailedCheckpoint(long failureTimestamp, @Nullable Throwable cause) {
return new FailedCheckpointStats(
checkpointId,
... | 3.68 |
hbase_ArrayBackedTag_getValueOffset | /** Returns Offset of actual tag bytes within the backed buffer */
@Override
public int getValueOffset() {
return this.offset + INFRASTRUCTURE_SIZE;
} | 3.68 |
framework_CompositeValidator_getErrorMessage | /**
* Gets the error message for the composite validator. If the error message
* is null, original error messages of the sub-validators are used instead.
*/
public String getErrorMessage() {
if (errorMessage != null) {
return errorMessage;
}
// TODO Return composite error message
return nul... | 3.68 |
hudi_AbstractTableFileSystemView_mergeCompactionPendingFileSlices | /**
* Helper to merge last 2 file-slices. These 2 file-slices do not have compaction done yet.
*
* @param lastSlice Latest File slice for a file-group
* @param penultimateSlice Penultimate file slice for a file-group in commit timeline order
*/
private static FileSlice mergeCompactionPendingFileSlices(FileSlice la... | 3.68 |
AreaShop_Utils_millisToTicks | /**
* Convert milliseconds to ticks.
* @param milliseconds Milliseconds to convert
* @return milliseconds divided by 50 (20 ticks per second)
*/
public static long millisToTicks(long milliseconds) {
return milliseconds / 50;
} | 3.68 |
hmily_XaResourceWrapped_rollback0 | /**
* 子类实现. Rollback 0.
*
* @param xid the xid
* @throws XAException the xa exception
*/
void rollback0(final Xid xid) throws XAException {
} | 3.68 |
flink_BinaryInputFormat_createStatistics | /**
* Fill in the statistics. The last modification time and the total input size are prefilled.
*
* @param files The files that are associated with this block input format.
* @param stats The pre-filled statistics.
*/
protected SequentialStatistics createStatistics(
List<FileStatus> files, FileBaseStatist... | 3.68 |
flink_LimitedConnectionsFileSystem_getMaxNumOpenInputStreams | /** Gets the maximum number of concurrently open input streams. */
public int getMaxNumOpenInputStreams() {
return maxNumOpenInputStreams;
} | 3.68 |
hbase_TablePermission_tableFieldsEqual | /**
* Check if fields of table in table permission equals.
* @param tp to be checked table permission
* @return true if equals, false otherwise
*/
public boolean tableFieldsEqual(TablePermission tp) {
if (tp == null) {
return false;
}
boolean tEq = (table == null && tp.table == null) || (table != null &&... | 3.68 |
pulsar_ModularLoadManagerStrategy_onActiveBrokersChange | /**
* Triggered when active brokers change.
*/
default void onActiveBrokersChange(Set<String> activeBrokers) {
} | 3.68 |
hbase_MapReduceBackupCopyJob_getSubTaskPercntgInWholeTask | /**
* Get the current copy task percentage within the whole task if multiple copies are needed.
* @return the current copy task percentage
*/
public float getSubTaskPercntgInWholeTask() {
return subTaskPercntgInWholeTask;
} | 3.68 |
hadoop_ConsumerRaisingIOE_andThen | /**
* after calling {@link #accept(Object)},
* invoke the next consumer in the chain.
* @param next next consumer
* @return the chain.
*/
default ConsumerRaisingIOE<T> andThen(
ConsumerRaisingIOE<? super T> next) {
return (T t) -> {
accept(t);
next.accept(t);
};
} | 3.68 |
graphhopper_GtfsStorage_postInit | // TODO: Refactor initialization
public void postInit() {
LocalDate latestStartDate = LocalDate.ofEpochDay(this.gtfsFeeds.values().stream().mapToLong(f -> f.getStartDate().toEpochDay()).max().getAsLong());
LocalDate earliestEndDate = LocalDate.ofEpochDay(this.gtfsFeeds.values().stream().mapToLong(f -> f.getEndDate().... | 3.68 |
framework_Tree_writeItems | /**
* Recursively writes the root items and their children to a design.
*
* @since 7.5.0
* @param design
* the element into which to insert the items
* @param context
* the DesignContext instance used in writing
*/
@Override
protected void writeItems(Element design, DesignContext context) ... | 3.68 |
flink_MutableRecordAndPosition_setPosition | /** Sets the position without setting a record. */
public void setPosition(long offset, long recordSkipCount) {
this.offset = offset;
this.recordSkipCount = recordSkipCount;
} | 3.68 |
hudi_BufferedRandomAccessFile_endOfBufferReached | /**
* @return whether currentPosition has reached the end of valid buffer.
*/
private boolean endOfBufferReached() {
return this.currentPosition >= this.validLastPosition;
} | 3.68 |
framework_AbstractComponentConnector_onDropTargetDetached | /**
* Invoked when a {@link DropTargetExtensionConnector} has been removed from
* this component.
* <p>
* By default, does nothing.
* <p>
* This is a framework internal method, and should not be invoked manually.
*
* @since 8.1
* @see #onDropTargetAttached()
*/
public void onDropTargetDetached() {
} | 3.68 |
pulsar_ConnectionPool_createConnection | /**
* Resolve DNS asynchronously and attempt to connect to any IP address returned by DNS server.
*/
private CompletableFuture<Channel> createConnection(InetSocketAddress logicalAddress,
InetSocketAddress unresolvedPhysicalAddress) {
CompletableFuture<List<InetS... | 3.68 |
flink_TableStats_merge | /**
* Merges two table stats. When the stats are unknown, whatever the other are, we need return
* unknown stats. See {@link #UNKNOWN}.
*
* @param other The other table stats to merge.
* @return The merged table stats.
*/
public TableStats merge(TableStats other, @Nullable Set<String> partitionKeys) {
if (thi... | 3.68 |
querydsl_MetaDataExporter_setGeneratedAnnotationClass | /**
* Set the fully qualified class name of the "generated" annotation added ot the generated sources
*
* @param generatedAnnotationClass the fully qualified class name of the <em>Single-Element Annotation</em> (with {@code String} element) to be used on
* the generated sources, or {... | 3.68 |
hbase_Cacheable_release | /**
* Decrease its reference count, and if no reference then free the memory of this object, its
* backend is usually a {@link org.apache.hadoop.hbase.nio.ByteBuff}, and we will put its NIO
* ByteBuffers back to {@link org.apache.hadoop.hbase.io.ByteBuffAllocator}
*/
default boolean release() {
return false;
} | 3.68 |
framework_AbsoluteLayout_getBottomValue | /**
* Gets the 'bottom' attributes value using current units.
*
* @return The value of the 'bottom' attribute, null if not set
* @see #getBottomUnits()
*/
public Float getBottomValue() {
return bottomValue;
} | 3.68 |
hadoop_TimelineStateStore_serviceStart | /**
* Start the state storage for use
*
* @throws IOException
*/
@Override
public void serviceStart() throws IOException {
startStorage();
} | 3.68 |
hbase_BitComparator_parseFrom | /**
* Parse a serialized representation of {@link BitComparator}
* @param pbBytes A pb serialized {@link BitComparator} instance
* @return An instance of {@link BitComparator} made from <code>bytes</code>
* @throws DeserializationException if an error occurred
* @see #toByteArray
*/
public static BitComparator pa... | 3.68 |
hbase_TableRecordReaderImpl_next | /**
* @param key HStoreKey as input key.
* @param value MapWritable as input value
* @return true if there was more data
*/
public boolean next(ImmutableBytesWritable key, Result value) throws IOException {
Result result;
try {
try {
result = this.scanner.next();
if (logScannerActivity) {
... | 3.68 |
flink_Plan_getDataSinks | /**
* Gets all the data sinks of this job.
*
* @return All sinks of the program.
*/
public Collection<? extends GenericDataSinkBase<?>> getDataSinks() {
return this.sinks;
} | 3.68 |
flink_MurmurHashUtils_hashUnsafeBytes | /**
* Hash unsafe bytes.
*
* @param base base unsafe object
* @param offset offset for unsafe object
* @param lengthInBytes length in bytes
* @return hash code
*/
public static int hashUnsafeBytes(Object base, long offset, int lengthInBytes) {
return hashUnsafeBytes(base, offset, lengthInBytes, DEFAULT_SEED)... | 3.68 |
hadoop_DynamicIOStatisticsBuilder_withAtomicLongMaximum | /**
* Add a maximum statistic to dynamically return the
* latest value of the source.
* @param key key of this statistic
* @param source atomic long maximum
* @return the builder.
*/
public DynamicIOStatisticsBuilder withAtomicLongMaximum(String key,
AtomicLong source) {
withLongFunctionMaximum(key, s -> so... | 3.68 |
hadoop_SelectBinding_opt | /**
* Resolve an option.
* @param builderOptions the options which came in from the openFile builder.
* @param fsConf configuration of the owning FS.
* @param base base option (no s3a: prefix)
* @param defVal default value. Must not be null.
* @param trim should the result be trimmed.
* @return the possibly trim... | 3.68 |
pulsar_Commands_readChecksum | /**
* Read the checksum and advance the reader index in the buffer.
*
* <p>Note: This method assume the checksum presence was already verified before.
*/
public static int readChecksum(ByteBuf buffer) {
buffer.skipBytes(2); //skip magic bytes
return buffer.readInt();
} | 3.68 |
hadoop_MutableStat_resetMinMax | /**
* Reset the all time min max of the metric
*/
public void resetMinMax() {
minMax.reset();
} | 3.68 |
hudi_CompactionUtils_getPendingCompactionOperations | /**
* Get pending compaction operations for both major and minor compaction.
*/
public static Stream<Pair<HoodieFileGroupId, Pair<String, HoodieCompactionOperation>>> getPendingCompactionOperations(
HoodieInstant instant, HoodieCompactionPlan compactionPlan) {
List<HoodieCompactionOperation> ops = compactionPla... | 3.68 |
flink_AdaptiveScheduler_computeVertexParallelismStore | /**
* Creates the parallelism store that should be used for determining scheduling requirements,
* which may choose different parallelisms than set in the {@link JobGraph} depending on the
* execution mode.
*
* @param jobGraph The job graph for execution.
* @param executionMode The mode of scheduler execution.
*... | 3.68 |
zxing_Detector_getBullsEyeCorners | /**
* Finds the corners of a bull-eye centered on the passed point.
* This returns the centers of the diagonal points just outside the bull's eye
* Returns [topRight, bottomRight, bottomLeft, topLeft]
*
* @param pCenter Center point
* @return The corners of the bull-eye
* @throws NotFoundException If no valid bu... | 3.68 |
dubbo_ConfigurationUtils_getDynamicGlobalConfiguration | /**
* For compact single instance
*
* @deprecated Replaced to {@link ConfigurationUtils#getDynamicGlobalConfiguration(ScopeModel)}
*/
@Deprecated
public static Configuration getDynamicGlobalConfiguration() {
return ApplicationModel.defaultModel()
.getDefaultModule()
.modelEnvironment()
... | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.