name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
morf_SchemaAdapter_tableNames | /**
* @see org.alfasoftware.morf.metadata.Schema#tableNames()
*/
@Override
public Collection<String> tableNames() {
return delegate.tableNames();
} | 3.68 |
hbase_AccessController_prePrepareBulkLoad | /**
* Authorization check for SecureBulkLoadProtocol.prepareBulkLoad()
* @param ctx the context
*/
@Override
public void prePrepareBulkLoad(ObserverContext<RegionCoprocessorEnvironment> ctx)
throws IOException {
requireAccess(ctx, "prePrepareBulkLoad",
ctx.getEnvironment().getRegion().getTableDescriptor().ge... | 3.68 |
flink_RoundRobinOperatorStateRepartitioner_groupByStateMode | /** Group by the different named states. */
@SuppressWarnings("unchecked, rawtype")
private GroupByStateNameResults groupByStateMode(
List<List<OperatorStateHandle>> previousParallelSubtaskStates) {
// Reorganize: group by (State Name -> StreamStateHandle + StateMetaInfo)
EnumMap<
O... | 3.68 |
dubbo_DubboAutoConfiguration_serviceAnnotationBeanProcessor | /**
* Creates {@link ServiceAnnotationPostProcessor} Bean
*
* @param packagesToScan the packages to scan
* @return {@link ServiceAnnotationPostProcessor}
*/
@ConditionalOnProperty(prefix = DUBBO_SCAN_PREFIX, name = BASE_PACKAGES_PROPERTY_NAME)
@ConditionalOnBean(name = BASE_PACKAGES_BEAN_NAME)
@Bean
public Service... | 3.68 |
flink_AbstractStreamOperatorV2_open | /**
* This method is called immediately before any elements are processed, it should contain the
* operator's initialization logic, e.g. state initialization.
*
* <p>The default implementation does nothing.
*
* @throws Exception An exception in this method causes the operator to fail.
*/
@Override
public void op... | 3.68 |
flink_HiveParserUtils_isNative | // TODO: we need a way to tell whether a function is built-in, for now just return false so that
// the unparser will quote them
public static boolean isNative(SqlOperator sqlOperator) {
return false;
} | 3.68 |
dubbo_DubboMergingDigest_recordAllData | /**
* Turns on internal data recording.
*/
@Override
public TDigest recordAllData() {
super.recordAllData();
data = new ArrayList<>();
tempData = new ArrayList<>();
return this;
} | 3.68 |
hbase_SnapshotReferenceUtil_visitRegionStoreFiles | /**
* Iterate over the snapshot store files in the specified region
* @param manifest snapshot manifest to inspect
* @param visitor callback object to get the store files
* @throws IOException if an error occurred while scanning the directory
*/
public static void visitRegionStoreFiles(final SnapshotRegionManifes... | 3.68 |
dubbo_MetricsSupport_increment | /**
* Incr method num
*/
public static void increment(
MetricsKey metricsKey,
MetricsPlaceValue placeType,
MethodMetricsCollector<TimeCounterEvent> collector,
MetricsEvent event) {
collector.increment(
event.getAttachmentValue(METHOD_METRICS),
new MetricsKey... | 3.68 |
morf_ResultSetComparer_valueCheck | /**
* Produces a mismatch if the specified column index mismatches.
*/
@SuppressWarnings("rawtypes")
private Optional<ResultSetMismatch> valueCheck(ResultSet left, ResultSet right, String[] keys, int i, int columnType, MismatchType checkForMismatchType) throws SQLException {
Comparable leftValue;
Comparable right... | 3.68 |
Activiti_DelegateInvocation_proceed | /**
* make the invocation proceed, performing the actual invocation of the user code.
*
* @throws Exception
* the exception thrown by the user code
*/
public void proceed() {
invoke();
} | 3.68 |
hadoop_ReadBufferManager_getBlock | /**
* {@link AbfsInputStream} calls this method read any bytes already available in a buffer (thereby saving a
* remote read). This returns the bytes if the data already exists in buffer. If there is a buffer that is reading
* the requested offset, then this method blocks until that read completes. If the data is qu... | 3.68 |
flink_FactoryUtil_discoverOptionalDecodingFormat | /**
* Discovers a {@link DecodingFormat} of the given type using the given option (if present)
* as factory identifier.
*/
public <I, F extends DecodingFormatFactory<I>>
Optional<DecodingFormat<I>> discoverOptionalDecodingFormat(
Class<F> formatFactoryClass, ConfigOption<String> formatOption)... | 3.68 |
morf_AbstractSelectStatementBuilder_getOrderBys | /**
* Gets the fields which the select is ordered by
*
* @return the order by fields
*/
List<AliasedField> getOrderBys() {
return orderBys;
} | 3.68 |
pulsar_SecurityUtil_loginKerberos | /**
* Initializes UserGroupInformation with the given Configuration and performs the login for the
* given principal and keytab. All logins should happen through this class to ensure other threads
* are not concurrently modifying UserGroupInformation.
* <p/>
* @param config the configuration instance
* @param ... | 3.68 |
morf_NamedParameterPreparedStatement_setQueryTimeout | /**
* Sets the timeout in <b>seconds</b> after which the query will time out on
* database side. In such case JDBC driver will throw an exception which is
* specific to database implementation but is likely to extend
* {@link SQLTimeoutException}.
*
* @param queryTimeout timeout in <b>seconds</b>
* @exception SQ... | 3.68 |
flink_HiveParserCalcitePlanner_genDistSortBy | // Generate plan for sort by, cluster by and distribute by. This is basically same as generating
// order by plan.
// Should refactor to combine them.
private Pair<RelNode, RelNode> genDistSortBy(
HiveParserQB qb, RelNode srcRel, boolean outermostOB) throws SemanticException {
RelNode res = null;
RelNod... | 3.68 |
hadoop_AzureBlobFileSystem_trailingPeriodCheck | /**
* Performs a check for (.) until root in the path to throw an exception.
* The purpose is to differentiate between dir/dir1 and dir/dir1.
* Without the exception the behavior seen is dir1. will appear
* to be present without it's actual creation as dir/dir1 and dir/dir1. are
* treated as identical.
* @param p... | 3.68 |
pulsar_LegacyHierarchicalLedgerRangeIterator_getLedgerRangeByLevel | /**
* Get a single node level1/level2.
*
* @param level1
* 1st level node name
* @param level2
* 2nd level node name
* @throws IOException
*/
LedgerManager.LedgerRange getLedgerRangeByLevel(final String level1, final String level2)
throws IOException {
StringBuilder nodeBuilder = t... | 3.68 |
hbase_EnableTableProcedure_getMaxReplicaId | /** Returns Maximum region replica id found in passed list of regions. */
private static int getMaxReplicaId(List<RegionInfo> regions) {
int max = 0;
for (RegionInfo regionInfo : regions) {
if (regionInfo.getReplicaId() > max) {
// Iterating through all the list to identify the highest replicaID region.
... | 3.68 |
hadoop_IOStatisticsContext_setThreadIOStatisticsContext | /**
* Set the IOStatisticsContext for the current thread.
* @param statisticsContext IOStatistics context instance for the
* current thread. If null, the context is reset.
*/
static void setThreadIOStatisticsContext(
IOStatisticsContext statisticsContext) {
IOStatisticsContextIntegration.setThreadIOStatistics... | 3.68 |
hibernate-validator_AnnotationMetaDataProvider_findConstraintAnnotations | /**
* Examines the given annotation to see whether it is a single- or multi-valued constraint annotation.
*
* @param constrainable The constrainable to check for constraints annotations
* @param annotation The annotation to examine
* @param type the element type on which the annotation/constraint is placed on
* @... | 3.68 |
framework_VAccordion_clearPaintables | /**
* {@inheritDoc}
*
* @deprecated This method is not called by the framework code anymore.
*/
@Deprecated
@Override
protected void clearPaintables() {
clear();
} | 3.68 |
hudi_HoodieTableConfig_getTableChecksum | /**
* Read the table checksum.
*/
private Long getTableChecksum() {
return getLong(TABLE_CHECKSUM);
} | 3.68 |
hbase_Procedure_isRunnable | /** Returns true if the procedure is in a RUNNABLE state. */
public synchronized boolean isRunnable() {
return state == ProcedureState.RUNNABLE;
} | 3.68 |
flink_ShuffleMaster_close | /**
* Closes this shuffle master service which should release all resources. A shuffle master will
* only be closed when the cluster is shut down.
*/
@Override
default void close() throws Exception {} | 3.68 |
hbase_HMaster_main | /**
* @see org.apache.hadoop.hbase.master.HMasterCommandLine
*/
public static void main(String[] args) {
LOG.info("STARTING service " + HMaster.class.getSimpleName());
VersionInfo.logVersion();
new HMasterCommandLine(HMaster.class).doMain(args);
} | 3.68 |
querydsl_CollectionUtils_unmodifiableSet | /**
* Return an unmodifiable copy of a set, or the same set if its already an unmodifiable type.
*
* @param set the set
* @param <T> element type
* @return unmodifiable copy of a set, or the same set if its already an unmodifiable type
*/
@SuppressWarnings("unchecked")
public static <T> Set<T> unmodifiableSet(Set... | 3.68 |
flink_CommittableMessageTypeInfo_noOutput | /**
* Returns the type information for a {@link CommittableMessage} with no committable.
*
* @return {@link TypeInformation} with {@link CommittableMessage}
*/
public static TypeInformation<CommittableMessage<Void>> noOutput() {
return new CommittableMessageTypeInfo<>(NoOutputSerializer::new);
} | 3.68 |
hibernate-validator_ExecutableHelper_isResolvedToSameMethodInHierarchy | /**
* Checks if a pair of given methods ({@code left} and {@code right}) are resolved to the same
* method based on the {@code mainSubType} type.
*
* @param mainSubType a type at the bottom of class hierarchy to be used to lookup the methods.
* @param left one of the methods to check
* @param right another of the... | 3.68 |
framework_Calendar_setStartDate | /**
* Sets start date for the calendar. This and {@link #setEndDate(Date)}
* control the range of dates visible on the component. The default range is
* one week.
*
* @param date
* First visible date to show.
*/
public void setStartDate(Date date) {
if (!date.equals(startDate)) {
startDate... | 3.68 |
hmily_OriginTrackedPropertiesLoader_isEndOfLine | /**
* Is end of line boolean.
*
* @return the boolean
*/
public boolean isEndOfLine() {
return this.character == -1 || (!this.escaped && this.character == '\n');
} | 3.68 |
hadoop_AbstractClientRequestInterceptor_getConf | /**
* Gets the {@link Configuration}.
*/
@Override
public Configuration getConf() {
return this.conf;
} | 3.68 |
framework_Window_addFocusListener | /*
* (non-Javadoc)
*
* @see
* com.vaadin.event.FieldEvents.FocusNotifier#addFocusListener(com.vaadin
* .event.FieldEvents.FocusListener)
*/
@Override
public Registration addFocusListener(FocusListener listener) {
return addListener(FocusEvent.EVENT_ID, FocusEvent.class, listener,
FocusListener.foc... | 3.68 |
hudi_RDDConsistentBucketBulkInsertPartitioner_initializeBucketIdentifier | /**
* Initialize hashing metadata of input records. The metadata of all related partitions will be loaded, and
* the mapping from partition to its bucket identifier is constructed.
*/
private Map<String, ConsistentBucketIdentifier> initializeBucketIdentifier(JavaRDD<HoodieRecord<T>> records) {
return records.map(H... | 3.68 |
framework_AbstractComponentTest_createActions | /**
* Create actions for the component. Remember to call super.createActions()
* when overriding.
*/
protected void createActions() {
createBooleanAction("Enabled", CATEGORY_STATE, true, enabledCommand);
createBooleanAction("Readonly", CATEGORY_STATE, false, readonlyCommand);
createBooleanAction("Visible... | 3.68 |
shardingsphere-elasticjob_ZookeeperRegistryCenter_waitForCacheClose | /*
* // TODO sleep 500ms, let cache client close first and then client, otherwise will throw exception reference:https://issues.apache.org/jira/browse/CURATOR-157
*/
private void waitForCacheClose() {
try {
Thread.sleep(500L);
} catch (final InterruptedException ex) {
Thread.currentThread().in... | 3.68 |
hudi_Hive3Shim_getDateWriteable | /**
* Get date writeable object from int value.
* Hive3 use DateWritableV2 to build date objects and Hive2 use DateWritable.
* So that we need to initialize date according to the version of Hive.
*/
public Writable getDateWriteable(int value) {
try {
return (Writable) DATE_WRITEABLE_V2_CONSTRUCTOR.newInstance... | 3.68 |
hadoop_MutableGaugeLong_decr | /**
* decrement by delta
* @param delta of the decrement
*/
public void decr(long delta) {
value.addAndGet(-delta);
setChanged();
} | 3.68 |
graphhopper_GraphHopper_setProfiles | /**
* Sets the routing profiles that shall be supported by this GraphHopper instance. The (and only the) given profiles
* can be used for routing without preparation and for CH/LM preparation.
* <p>
* Here is an example how to setup two CH profiles and one LM profile (via the Java API)
*
* <pre>
* {@code
* ho... | 3.68 |
flink_SavepointReader_readListState | /**
* Read operator {@code ListState} from a {@code Savepoint} when a custom serializer was used;
* e.g., a different serializer than the one returned by {@code
* TypeInformation#createSerializer}.
*
* @param identifier The identifier of the operator.
* @param name The (unique) name for the state.
* @param typeI... | 3.68 |
flink_HiveParserASTNodeOrigin_getUsageAlias | /**
* @return the alias of the object from which an HiveParserASTNode originated, e.g. "v1" (this
* can help with debugging context-dependent expansions)
*/
public String getUsageAlias() {
return usageAlias;
} | 3.68 |
hadoop_AbfsConfiguration_accountConf | /**
* Appends an account name to a configuration key yielding the
* account-specific form.
* @param key Account-agnostic configuration key
* @return Account-specific configuration key
*/
public String accountConf(String key) {
return key + "." + accountName;
} | 3.68 |
hadoop_LpSolver_generateUnderAllocationConstraints | /**
* Generate under-allocation constraints.
*
* @param lpModel the LP model.
* @param cJobITimeK actual container allocation for job i in time
* interval k.
* @param uaPredict absolute container under-allocation.
* @param ua recursive container u... | 3.68 |
AreaShop_RentRegion_setDuration | /**
* Set the duration of the rent.
* @param duration The duration of the rent (as specified on the documentation pages)
*/
public void setDuration(String duration) {
setSetting("rent.duration", duration);
} | 3.68 |
querydsl_PathMetadataFactory_forArrayAccess | /**
* Create a new PathMetadata instance for indexed array access
*
* @param parent parent path
* @param index index of element
* @return array access path
*/
public static PathMetadata forArrayAccess(Path<?> parent, @Range(from = 0, to = Integer.MAX_VALUE) int index) {
return new PathMetadata(parent, index, ... | 3.68 |
graphhopper_InstructionsOutgoingEdges_isLeavingCurrentStreet | /**
* If the name and prevName changes this method checks if either the current street is continued on a
* different edge or if the edge we are turning onto is continued on a different edge.
* If either of these properties is true, we can be quite certain that a turn instruction should be provided.
*/
public boolea... | 3.68 |
hadoop_SysInfoWindows_getNetworkBytesWritten | /** {@inheritDoc} */
@Override
public long getNetworkBytesWritten() {
refreshIfNeeded();
return netBytesWritten;
} | 3.68 |
hbase_ReplicationSourceLogQueue_getOldestWalTimestamp | /*
* Get the oldest wal timestamp from all the queues.
*/
private long getOldestWalTimestamp() {
long oldestWalTimestamp = Long.MAX_VALUE;
for (Map.Entry<String, PriorityBlockingQueue<Path>> entry : queues.entrySet()) {
PriorityBlockingQueue<Path> queue = entry.getValue();
Path path = queue.peek();
//... | 3.68 |
hbase_ReplicationSourceManager_logPositionAndCleanOldLogs | /**
* This method will log the current position to storage. And also clean old logs from the
* replication queue.
* @param source the replication source
* @param entryBatch the wal entry batch we just shipped
*/
public void logPositionAndCleanOldLogs(ReplicationSourceInterface source,
WALEntryBatch entryBatc... | 3.68 |
hbase_HFileWriterImpl_getPath | /** Returns Path or null if we were passed a stream rather than a Path. */
@Override
public Path getPath() {
return path;
} | 3.68 |
hadoop_AzureBlobFileSystemStore_isAppendBlobKey | /**
* Checks if the given key in Azure Storage should be stored as a page
* blob instead of block blob.
*/
public boolean isAppendBlobKey(String key) {
return isKeyForDirectorySet(key, appendBlobDirSet);
} | 3.68 |
graphhopper_ResponsePath_getPoints | /**
* This method returns all points on the path. Keep in mind that calculating the distance from
* these points might yield different results compared to getDistance as points could have been
* simplified on import or after querying.
*/
public PointList getPoints() {
check("getPoints");
return pointList;
} | 3.68 |
graphhopper_StringEncodedValue_indexOf | /**
* @param value the String to retrieve the index
* @return the non-zero index of the String or <i>0</i> if it couldn't be found
*/
public int indexOf(String value) {
return indexMap.getOrDefault(value, 0);
} | 3.68 |
pulsar_MessageRouter_choosePartition | /**
* Choose a partition based on msg and the topic metadata.
*
* @param msg message to route
* @param metadata topic metadata
* @return the partition to route the message.
* @since 1.22.0
*/
default int choosePartition(Message<?> msg, TopicMetadata metadata) {
return choosePartition(msg);
} | 3.68 |
hbase_ParseFilter_registerFilter | /**
* Register a new filter with the parser. If the filter is already registered, an
* IllegalArgumentException will be thrown.
* @param name a name for the filter
* @param filterClass fully qualified class name
*/
public static void registerFilter(String name, String filterClass) {
if (LOG.isInfoEnabled(... | 3.68 |
hbase_HbckChore_runChore | /**
* Request execution of this chore's action.
* @return {@code true} if the chore was executed, {@code false} if the chore is disabled or
* already running.
*/
public boolean runChore() {
// This function does the sanity checks of making sure the chore is not run when it is
// disabled or when it's al... | 3.68 |
hbase_KeyValue_matchingRowColumn | /**
* Compares the row and column of two keyvalues for equality
* @param left left cell to compare row and column
* @param right right cell to compare row and column
* @return True if same row and column.
*/
public boolean matchingRowColumn(final Cell left, final Cell right) {
short lrowlength = left.getRowLeng... | 3.68 |
hadoop_TaskInfo_getOutputBytes | /**
* @return Raw bytes written to the destination FileSystem. Note that this may
* not match output bytes.
*/
public long getOutputBytes() {
return bytesOut;
} | 3.68 |
hmily_MetricsTrackerFacade_start | /**
* Init for metrics tracker manager.
*
* @param metricsConfig metrics config
*/
public void start(final HmilyMetricsConfig metricsConfig) {
if (this.isStarted.compareAndSet(false, true)) {
metricsBootService = ExtensionLoaderFactory.load(MetricsBootService.class, metricsConfig.getMetricsName());
... | 3.68 |
hbase_ModifyRegionUtils_editRegions | /**
* Execute the task on the specified set of regions.
* @param exec Thread Pool Executor
* @param regions {@link RegionInfo} that describes the regions to edit
* @param task {@link RegionFillTask} custom code to edit the region
*/
public static void editRegions(final ThreadPoolExecutor exec,
final Collec... | 3.68 |
hbase_SnapshotInfo_getLogsSize | /** Returns the total log size */
public long getLogsSize() {
return logSize.get();
} | 3.68 |
flink_WrappingCollector_outputResult | /** Outputs the final result to the wrapped collector. */
public void outputResult(T result) {
this.collector.collect(result);
} | 3.68 |
framework_MethodPropertyDescriptor_writeObject | /* Special serialization to handle method references */
private void writeObject(ObjectOutputStream out)
throws IOException {
out.defaultWriteObject();
SerializerHelper.writeClass(out, propertyType);
if (writeMethod != null) {
out.writeObject(writeMethod.getName());
SerializerHelper... | 3.68 |
flink_Grouping_getCustomPartitioner | /**
* Gets the custom partitioner to be used for this grouping, or {@code null}, if none was
* defined.
*
* @return The custom partitioner to be used for this grouping.
*/
@Internal
public Partitioner<?> getCustomPartitioner() {
return this.customPartitioner;
} | 3.68 |
hbase_ScheduledChore_updateTimeTrackingBeforeRun | /**
* Update our time tracking members. Called at the start of an execution of this chore's run()
* method so that a correct decision can be made as to whether or not we missed the start time
*/
private synchronized void updateTimeTrackingBeforeRun() {
timeOfLastRun = timeOfThisRun;
timeOfThisRun = EnvironmentEd... | 3.68 |
hbase_KeyValueUtil_nextShallowCopy | /**
* Creates a new KeyValue object positioned in the supplied ByteBuffer and sets the ByteBuffer's
* position to the start of the next KeyValue. Does not allocate a new array or copy data.
*/
public static KeyValue nextShallowCopy(final ByteBuffer bb, final boolean includesMvccVersion,
boolean includesTags) {
i... | 3.68 |
flink_SourceCoordinatorContext_getAndRemoveUncheckpointedAssignment | /**
* Get the split to put back. This only happens when a source reader subtask has failed.
*
* @param subtaskId the failed subtask id.
* @param restoredCheckpointId the checkpoint that the task is recovered to.
* @return A list of splits that needs to be added back to the {@link SplitEnumerator}.
*/
List<SplitT>... | 3.68 |
flink_TypeInfoLogicalTypeConverter_fromLogicalTypeToTypeInfo | /** Use {@link BigDecimalTypeInfo} to retain precision and scale of decimal. */
public static TypeInformation fromLogicalTypeToTypeInfo(LogicalType type) {
DataType dataType =
fromLogicalTypeToDataType(type)
.nullable()
.bridgedTo(ClassLogicalTypeConverter.getDefa... | 3.68 |
flink_CheckpointStatsCounts_createSnapshot | /**
* Creates a snapshot of the current state.
*
* @return Snapshot of the current state.
*/
CheckpointStatsCounts createSnapshot() {
return new CheckpointStatsCounts(
numRestoredCheckpoints,
numTotalCheckpoints,
numInProgressCheckpoints,
numCompletedCheckpoints,
... | 3.68 |
hbase_GroupingTableMapper_extractKeyValues | /**
* Extract columns values from the current record. This method returns null if any of the columns
* are not found.
* <p>
* Override this method if you want to deal with nulls differently.
* @param r The current values.
* @return Array of byte values.
*/
protected byte[][] extractKeyValues(Result r) {
byte[]... | 3.68 |
hbase_AccessChecker_getUserGroups | /**
* Retrieve the groups of the given user.
* @param user User name
*/
public static List<String> getUserGroups(String user) {
try {
return groupService.getGroups(user);
} catch (IOException e) {
LOG.error("Error occurred while retrieving group for " + user, e);
return new ArrayList<>();
}
} | 3.68 |
framework_JsonPaintTarget_addData | /**
*
* @param s
* json string, object or array
*/
public void addData(String s) {
children.add(s);
} | 3.68 |
hibernate-validator_ConstraintAnnotationVisitor_checkConstraints | /**
* Retrieves the checks required for the given element and annotations,
* executes them and reports all occurred errors.
*
* @param annotatedElement The element to check.
* @param mirrors The annotations to check.
*/
private void checkConstraints(Element annotatedElement, List<AnnotationMirror> mirrors) {
for... | 3.68 |
querydsl_GroupByBuilder_collection | /**
* Get the results as a list
*
* @param expression projection
* @return new result transformer
*/
public <V, RES extends Collection<V>> ResultTransformer<RES> collection(Supplier<RES> resultFactory, FactoryExpression<V> expression) {
final FactoryExpression<V> transformation = FactoryExpressionUtils.wrap(ex... | 3.68 |
flink_EmbeddedRocksDBStateBackend_getWriteBatchSize | /** Gets the max batch size will be used in {@link RocksDBWriteBatchWrapper}. */
public long getWriteBatchSize() {
return writeBatchSize == UNDEFINED_WRITE_BATCH_SIZE
? WRITE_BATCH_SIZE.defaultValue().getBytes()
: writeBatchSize;
} | 3.68 |
flink_StandardDeCompressors_getDecompressorForFileName | /**
* Gets the decompressor for a file name. This checks the file against all known and supported
* file extensions. Returns null if there is no decompressor for this file name.
*/
@Nullable
public static InflaterInputStreamFactory<?> getDecompressorForFileName(String fileName) {
for (final Map.Entry<String, Inf... | 3.68 |
hudi_BootstrapExecutorUtils_syncHive | /**
* Sync to Hive.
*/
private void syncHive() {
if (cfg.enableHiveSync) {
TypedProperties metaProps = new TypedProperties();
metaProps.putAll(props);
metaProps.put(META_SYNC_DATABASE_NAME.key(), cfg.database);
metaProps.put(META_SYNC_TABLE_NAME.key(), cfg.tableName);
metaProps.put(META_SYNC_BAS... | 3.68 |
hbase_MasterObserver_postRegionOffline | /**
* Called after the region has been marked offline.
* @param ctx the environment to interact with the framework and master
*/
default void postRegionOffline(final ObserverContext<MasterCoprocessorEnvironment> ctx,
final RegionInfo regionInfo) throws IOException {
} | 3.68 |
rocketmq-connect_RocketMQKafkaSinkTaskContext_convertToTopicPartition | /**
* convert to kafka topic partition
*
* @param partitionMap
* @return
*/
public TopicPartition convertToTopicPartition(Map<String, ?> partitionMap) {
if (partitionMap.containsKey(TOPIC) && partitionMap.containsKey(QUEUE_ID)) {
return new TopicPartition(partitionMap.get(TOPIC).toString(), Integer.val... | 3.68 |
pulsar_MessageImpl_getPayload | /**
* used only for unit-test to validate payload's state and ref-cnt.
*
* @return
*/
@VisibleForTesting
ByteBuf getPayload() {
return payload;
} | 3.68 |
flink_TernaryBoolean_getOrDefault | /**
* Gets the boolean value corresponding to this value. If this is the 'undefined' value, the
* method returns the given default.
*
* @param defaultValue The value to be returned in case this ternary value is 'undefined'.
*/
public boolean getOrDefault(boolean defaultValue) {
return this == UNDEFINED ? defau... | 3.68 |
morf_DirectoryDataSet_openOutputStreamForTable | /**
* @see org.alfasoftware.morf.xml.XmlStreamProvider.XmlOutputStreamProvider#openOutputStreamForTable(java.lang.String)
*/
@Override
public OutputStream openOutputStreamForTable(String tableName) {
try {
return new FileOutputStream(new File(directory, tableName + ".xml"));
} catch (FileNotFoundException e) ... | 3.68 |
flink_MemorySegment_getShort | /**
* Reads a short integer value (16 bit, 2 bytes) from the given position, composing them into a
* short value according to the current byte order.
*
* @param index The position from which the memory will be read.
* @return The short value at the given position.
* @throws IndexOutOfBoundsException Thrown, if th... | 3.68 |
framework_DDEventHandleStrategy_handleMouseUp | /**
* Called to handle {@link Event#ONMOUSEUP} event.
*
* @param target
* target element over which DnD event has happened
* @param event
* ONMOUSEUP GWT event for active DnD operation
* @param mediator
* VDragAndDropManager data accessor
*/
protected void handleMouseUp(Element... | 3.68 |
zxing_UPCEANReader_findGuardPattern | /**
* @param row row of black/white values to search
* @param rowOffset position to start search
* @param whiteFirst if true, indicates that the pattern specifies white/black/white/...
* pixel counts, otherwise, it is interpreted as black/white/black/...
* @param pattern pattern of counts of number of black and wh... | 3.68 |
pulsar_ChannelFutures_toCompletableFuture | /**
* Convert a {@link ChannelFuture} into a {@link CompletableFuture}.
*
* @param channelFuture the {@link ChannelFuture}
* @return a {@link CompletableFuture} that completes successfully when the channelFuture completes successfully,
* and completes exceptionally if the channelFuture completes with a {@l... | 3.68 |
dubbo_MetricsApplicationListener_onFinishEventBuild | /**
* To end the monitoring normally, in addition to increasing the number of corresponding indicators,
* use the introspection method to calculate the relevant rt indicators
*
* @param metricsKey Monitor key
* @param collector Corresponding collector
*/
public static AbstractMetricsKeyListener onFinishEventBuil... | 3.68 |
morf_TestingDatabaseEquivalentStringComparator_configure | /**
* @see com.google.inject.AbstractModule#configure()
*/
@Override
protected void configure() {
binder().bind(DatabaseEquivalentStringComparator.class).to(TestingDatabaseEquivalentStringComparator.class);
} | 3.68 |
hadoop_EvaluatingStatisticsMap_values | /**
* Evaluate all the entries and provide a list of the results.
*
* This is not a snapshot, so if the evaluators actually return
* references to mutable objects (e.g. a MeanStatistic instance)
* then that value may still change.
* @return the current list of evaluated results.
*/
@Override
public Collection<E>... | 3.68 |
flink_StreamExecutionEnvironment_enableChangelogStateBackend | /**
* Enable the change log for current state backend. This change log allows operators to persist
* state changes in a very fine-grained manner. Currently, the change log only applies to keyed
* state, so non-keyed operator state and channel state are persisted as usual. The 'state' here
* refers to 'keyed state'.... | 3.68 |
framework_GridLayout_getCursorY | /**
* Gets the current y-position (row) of the cursor.
*
* <p>
* The cursor position points the position for the next component that is
* added without specifying its coordinates (grid cell). When the cursor
* position is occupied, the next component will be added to the first free
* position after the cursor.
... | 3.68 |
open-banking-gateway_BaseDatasafeDbStorageService_read | /**
* Open Datasafe object for reading.
* @param absoluteLocation Absolute path of the object to read. I.e. {@code db://storage/deadbeef}
* @return Stream to read data from.
*/
@Override
@SneakyThrows
@Transactional(noRollbackFor = BaseDatasafeDbStorageService.DbStorageEntityNotFoundException.class)
public InputStr... | 3.68 |
hadoop_FedBalanceContext_setBandwidthLimit | /**
* The bandwidth limit of the distcp job(MB).
* @param value the bandwidth.
* @return the builder.
*/
public Builder setBandwidthLimit(int value) {
this.bandwidthLimit = value;
return this;
} | 3.68 |
cron-utils_SecondsDescriptor_visit | /**
* Provide a human readable description for Every instance.
*
* @param every - Every
* @return human readable description - String
*/
@Override
public Every visit(final Every every) {
String description;
if (every.getPeriod().getValue() > 1) {
description = String.format("%s %s ", bundle.getStri... | 3.68 |
hmily_HmilyXaResource_getXaResource | /**
* Gets xa resource.
*
* @return the xa resource
*/
public XAResource getXaResource() {
return xaResource;
} | 3.68 |
framework_VDateField_getDefaultDate | /**
* Sets the default date when no date is selected.
*
* @return the default date
* @since 8.1.2
*/
public Date getDefaultDate() {
return defaultDate;
} | 3.68 |
flink_WindowedStream_process | /**
* Applies the given window function to each window. The window function is called for each
* evaluation of the window for each key individually. The output of the window function is
* interpreted as a regular non-windowed stream.
*
* <p>Note that this function requires that all data in the windows is buffered ... | 3.68 |
hbase_RegionServerObserver_preReplicateLogEntries | // TODO remove below 2 hooks when we implement AC as a core impl than a CP impl.
/**
* This will be called before executing replication request to shipping log entries.
* @param ctx the environment to interact with the framework and region server.
* @deprecated As of release 2.0.0 with out any replacement. This is m... | 3.68 |
hadoop_MountInterface_getValue | /** @return the int value representing the procedure. */
public int getValue() {
return ordinal();
} | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.