name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
MagicPlugin_Wand_checkSpellLevelsAndInventory | /**
* Covers the special case of a wand having spell levels and inventory slots that came from configs,
* but now we've modified the spells list and need to figure out if we also need to persist the levels and
* slots separately.
*
* <p>This should all be moved to CasterProperties at some point to handle the same ... | 3.68 |
hbase_MobFileCache_getMissCount | /**
* Gets the count of misses to the mob file cache.
* @return The count of misses to the mob file cache.
*/
public long getMissCount() {
return miss.sum();
} | 3.68 |
hadoop_PlacementConstraints_allocationTagWithNamespace | /**
* Constructs a target expression on a set of allocation tags under
* a certain namespace.
*
* @param namespace namespace of the allocation tags
* @param allocationTags allocation tags
* @return a target expression
*/
public static TargetExpression allocationTagWithNamespace(String namespace,
String... al... | 3.68 |
dubbo_ServiceAnnotationPostProcessor_scanServiceBeans | /**
* Scan and registers service beans whose classes was annotated {@link Service}
*
* @param packagesToScan The base packages to scan
* @param registry {@link BeanDefinitionRegistry}
*/
private void scanServiceBeans(Set<String> packagesToScan, BeanDefinitionRegistry registry) {
scanned = true;
if (... | 3.68 |
rocketmq-connect_Serdes_String | /**
* A serde for nullable {@code String} type.
*/
static public Serde<String> String() {
return new StringSerde();
} | 3.68 |
hbase_OnlineLogRecord_getScan | /**
* If {@value org.apache.hadoop.hbase.HConstants#SLOW_LOG_SCAN_PAYLOAD_ENABLED} is enabled then
* this value may be present and should represent the Scan that produced the given
* {@link OnlineLogRecord}
*/
public Optional<Scan> getScan() {
return scan;
} | 3.68 |
framework_AbstractDateField_setRangeEnd | /**
* Sets the end range for this component. If the value is set after this
* date (taking the resolution into account), the component will not
* validate. If {@code endDate} is set to {@code null}, any value after
* {@code startDate} will be accepted by the range.
* <p>
* Note: It's usually recommended to use on... | 3.68 |
flink_SplitDataProperties_splitsOrderedBy | /**
* Defines that the data within an input split is sorted on the fields defined by the field
* expressions in the specified orders. Multiple field expressions must be separated by the
* semicolon ';' character. All records of an input split must be emitted by the input format in
* the defined order.
*
* <p><b> ... | 3.68 |
graphhopper_Helper_intToDegree | /**
* Converts back the integer value.
*
* @return the degree value of the specified integer
*/
public static double intToDegree(int storedInt) {
if (storedInt == Integer.MAX_VALUE)
return Double.MAX_VALUE;
if (storedInt == -Integer.MAX_VALUE)
return -Double.MAX_VALUE;
return (double) st... | 3.68 |
querydsl_BooleanExpression_isTrue | /**
* Create a {@code this == true} expression
*
* @return this == true
*/
public BooleanExpression isTrue() {
return eq(true);
} | 3.68 |
framework_DragEndEvent_getDropEffect | /**
* Get drop effect of the dragend event. The value will be the desired
* action, that is the dropEffect value of the last dragenter or dragover
* event. The value depends on the effectAllowed parameter of the drag
* source, the dropEffect parameter of the drop target, and its drag over
* and drop criteria.
* <... | 3.68 |
zxing_BitMatrix_clear | /**
* Clears all bits (sets to false).
*/
public void clear() {
int max = bits.length;
for (int i = 0; i < max; i++) {
bits[i] = 0;
}
} | 3.68 |
framework_RowReference_getGrid | /**
* Gets the grid that contains the referenced row.
*
* @return the grid that contains referenced row
*/
public Grid<T> getGrid() {
return grid;
} | 3.68 |
flink_ExecutionConfig_getRegisteredTypesWithKryoSerializerClasses | /** Returns the registered types with their Kryo Serializer classes. */
public LinkedHashMap<Class<?>, Class<? extends Serializer<?>>>
getRegisteredTypesWithKryoSerializerClasses() {
return registeredTypesWithKryoSerializerClasses;
} | 3.68 |
hbase_RatioBasedCompactionPolicy_applyCompactionPolicy | /**
* -- Default minor compaction selection algorithm: choose CompactSelection from candidates --
* First exclude bulk-load files if indicated in configuration. Start at the oldest file and stop
* when you find the first file that meets compaction criteria: (1) a recently-flushed, small file
* (i.e. <= minCompactSi... | 3.68 |
hadoop_OBSDataBlocks_startUpload | /**
* Switch to the upload state and return a stream for uploading. Base class
* calls {@link #enterState(DestState, DestState)} to manage the state
* machine.
*
* @return the stream
* @throws IOException trouble
*/
Object startUpload() throws IOException {
LOG.debug("Start datablock[{}] upload", index);
ent... | 3.68 |
flink_ChannelReaderInputView_sendReadRequest | /**
* Sends a new read requests, if further requests remain. Otherwise, this method adds the
* segment directly to the readers return queue.
*
* @param seg The segment to use for the read request.
* @throws IOException Thrown, if the reader is in error.
*/
protected void sendReadRequest(MemorySegment seg) throws ... | 3.68 |
flink_InputChannel_setError | /**
* Atomically sets an error for this channel and notifies the input gate about available data to
* trigger querying this channel by the task thread.
*/
protected void setError(Throwable cause) {
if (this.cause.compareAndSet(null, checkNotNull(cause))) {
// Notify the input gate.
notifyChannelN... | 3.68 |
hbase_TableDescriptorChecker_sanityCheck | /**
* Checks whether the table conforms to some sane limits, and configured values (compression, etc)
* work. Throws an exception if something is wrong.
*/
public static void sanityCheck(final Configuration c, final TableDescriptor td)
throws IOException {
CompoundConfiguration conf = new CompoundConfiguration()... | 3.68 |
hbase_IpcClientSpanBuilder_getRpcPackageAndService | /**
* Retrieve the combined {@code $package.$service} value from {@code sd}.
*/
public static String getRpcPackageAndService(final Descriptors.ServiceDescriptor sd) {
// it happens that `getFullName` returns a string in the $package.$service format required by
// the otel RPC specification. Use it for now; might ... | 3.68 |
zxing_WifiConfigManager_changeNetworkUnEncrypted | // Adding an open, unsecured network
private static void changeNetworkUnEncrypted(WifiManager wifiManager, WifiParsedResult wifiResult) {
WifiConfiguration config = changeNetworkCommon(wifiResult);
config.allowedKeyManagement.set(WifiConfiguration.KeyMgmt.NONE);
updateNetwork(wifiManager, config);
} | 3.68 |
flink_DataSinkNode_getInputConnection | /**
* Gets the input of the sink.
*
* @return The input connection.
*/
public DagConnection getInputConnection() {
return this.input;
} | 3.68 |
hbase_ReplicationThrottler_resetStartTick | /**
* Reset the cycle start tick to NOW
*/
public void resetStartTick() {
if (this.enabled) {
this.cycleStartTick = EnvironmentEdgeManager.currentTime();
}
} | 3.68 |
shardingsphere-elasticjob_ElasticJobTracingConfiguration_tracingDataSource | /**
* Create a bean of tracing DataSource.
*
* @param tracingProperties tracing Properties
* @return tracing DataSource
*/
@Bean("tracingDataSource")
public DataSource tracingDataSource(final TracingProperties tracingProperties) {
DataSourceProperties dataSource = tracingProperties.getDataSource();
if (dat... | 3.68 |
flink_SqlLikeUtils_similar | /** SQL {@code SIMILAR} function with escape. */
public static boolean similar(String s, String pattern, String escape) {
final String regex = sqlToRegexSimilar(pattern, escape);
return Pattern.matches(regex, s);
} | 3.68 |
framework_AbstractClientConnector_addMethodInvocationToQueue | /**
* For internal use: adds a method invocation to the pending RPC call queue.
*
* @param interfaceName
* RPC interface name
* @param method
* RPC method
* @param parameters
* RPC all parameters
*
* @since 7.0
*/
protected void addMethodInvocationToQueue(String interfaceName... | 3.68 |
morf_SchemaChangeSequence_tableAdditions | /**
* @return The set of all table which are added by this sequence.
*/
public Set<String> tableAdditions() {
return tableAdditions;
} | 3.68 |
flink_DataStreamScanProvider_produceDataStream | /** Creates a scan Java {@link DataStream} from a {@link StreamExecutionEnvironment}. */
@Deprecated
default DataStream<RowData> produceDataStream(StreamExecutionEnvironment execEnv) {
throw new UnsupportedOperationException(
"This method is deprecated. "
+ "Use produceDataStream(Pro... | 3.68 |
framework_DataCommunicator_cleanUp | /**
* Executes the data destruction for dropped data that is not sent to
* the client. This method takes most recently sent data objects in a
* collection. Doing the clean up like this prevents the
* {@link ActiveDataHandler} from creating new keys for rows that were
* dropped but got re-requested by the client-si... | 3.68 |
flink_TaskLocalStateStoreImpl_discardLocalStateForCheckpoint | /**
* Helper method that discards state objects with an executor and reports exceptions to the log.
*/
private void discardLocalStateForCheckpoint(long checkpointID, Optional<TaskStateSnapshot> o) {
if (LOG.isTraceEnabled()) {
LOG.trace(
"Discarding local task state snapshot of checkpoint... | 3.68 |
pulsar_MessageIdAdv_getAckSet | /**
* Get the BitSet that indicates which messages in the batch.
*
* @implNote The message IDs of a batch should share a BitSet. For example, given 3 messages in the same batch whose
* size is 3, all message IDs of them should return "111" (i.e. a BitSet whose size is 3 and all bits are 1). If the
* 1st message ha... | 3.68 |
hadoop_StoragePolicySatisfyManager_getNextPathId | /**
* @return the next SPS path id, on which path users has invoked to satisfy
* storages.
*/
public Long getNextPathId() {
synchronized (pathsToBeTraversed) {
return pathsToBeTraversed.poll();
}
} | 3.68 |
framework_MarginInfo_hasRight | /**
* Checks if this MarginInfo object has the right edge margin enabled.
*
* @return true if right edge margin is enabled
*/
public boolean hasRight() {
return (bitMask & RIGHT) == RIGHT;
} | 3.68 |
querydsl_NumberExpression_shortValue | /**
* Create a {@code this.shortValue()} expression
*
* <p>Get the short expression of this numeric expression</p>
*
* @return this.shortValue()
* @see java.lang.Number#shortValue()
*/
public NumberExpression<Short> shortValue() {
return castToNum(Short.class);
} | 3.68 |
hbase_HRegionFileSystem_bulkLoadStoreFile | /**
* Bulk load: Add a specified store file to the specified family. If the source file is on the
* same different file-system is moved from the source location to the destination location,
* otherwise is copied over.
* @param familyName Family that will gain the file
* @param srcPath {@link Path} to the file t... | 3.68 |
flink_BinaryStringData_numChars | /** Returns the number of UTF-8 code points in the string. */
public int numChars() {
ensureMaterialized();
if (inFirstSegment()) {
int len = 0;
for (int i = 0;
i < binarySection.sizeInBytes;
i += numBytesForFirstByte(getByteOneSegment(i))) {
len++;
... | 3.68 |
hbase_SaslClientAuthenticationProvider_canRetry | /**
* Returns true if the implementation is capable of performing some action which may allow a
* failed authentication to become a successful authentication. Otherwise, returns false
*/
default boolean canRetry() {
return false;
} | 3.68 |
framework_Tree_removeExpandListener | /**
* Removes the expand listener.
*
* @param listener
* the Listener to be removed.
*/
public void removeExpandListener(ExpandListener listener) {
removeListener(ExpandEvent.class, listener,
ExpandListener.EXPAND_METHOD);
} | 3.68 |
flink_YarnClusterDescriptor_addShipFiles | /**
* Adds the given files to the list of files to ship.
*
* <p>Note that any file matching "<tt>flink-dist*.jar</tt>" will be excluded from the upload by
* {@link YarnApplicationFileUploader#registerMultipleLocalResources(Collection, String,
* LocalResourceType)} since we upload the Flink uber jar ourselves and d... | 3.68 |
flink_SourceCoordinatorSerdeUtils_writeCoordinatorSerdeVersion | /** Write the current serde version. */
static void writeCoordinatorSerdeVersion(DataOutputStream out) throws IOException {
out.writeInt(CURRENT_VERSION);
} | 3.68 |
flink_FileOutputFormat_initializeGlobal | /**
* Initialization of the distributed file system if it is used.
*
* @param parallelism The task parallelism.
*/
@Override
public void initializeGlobal(int parallelism) throws IOException {
final Path path = getOutputFilePath();
final FileSystem fs = path.getFileSystem();
// only distributed file sys... | 3.68 |
flink_ResultInfo_getColumnInfos | /** Get the column info of the data. */
public List<ColumnInfo> getColumnInfos() {
return Collections.unmodifiableList(columnInfos);
} | 3.68 |
morf_SqlServer_getXADataSource | /**
* Returns a SQL Server XA data source. Note that this method may fail at
* run-time if {@code SQLServerXADataSource} is not available on the classpath.
*
* @throws IllegalStateException If the data source cannot be created.
*
* @see org.alfasoftware.morf.jdbc.DatabaseType#getXADataSource(java.lang.String,
* ... | 3.68 |
MagicPlugin_MapController_forceReload | /**
* Force reload of the specific url and cropping.
*/
public void forceReload(String worldName, String url, int x, int y, int width, int height) {
get(worldName, url, x, y, width, height).reload();
} | 3.68 |
hbase_HRegionLocation_hashCode | /**
* @see java.lang.Object#hashCode()
*/
@Override
public int hashCode() {
return this.serverName.hashCode();
} | 3.68 |
framework_AbstractStringToNumberConverter_convertToNumber | /**
* Convert the value to a Number using the given locale and
* {@link #getFormat(Locale)}.
*
* @param value
* The value to convert
* @param locale
* The locale to use for conversion
* @return The converted value
* @throws ConversionException
* If there was a problem convert... | 3.68 |
hbase_Increment_isReturnResults | /** Returns current setting for returnResults */
// This method makes public the superclasses's protected method.
@Override
public boolean isReturnResults() {
return super.isReturnResults();
} | 3.68 |
hbase_ClientMetaTableAccessor_scanMeta | /**
* Performs a scan of META table for given table.
* @param metaTable scanner over meta table
* @param startRow Where to start the scan
* @param stopRow Where to stop the scan
* @param type scanned part of meta
* @param maxRows maximum rows to return
* @param visitor Visitor invoked against each ro... | 3.68 |
framework_CustomLayout_getTemplateName | /** Get the name of the template. */
public String getTemplateName() {
return getState(false).templateName;
} | 3.68 |
MagicPlugin_BaseSpell_onPlayerQuit | /**
* Listener method, called on player quit for registered spells.
*
* @param event The player who just quit
*/
public void onPlayerQuit(PlayerQuitEvent event)
{
} | 3.68 |
morf_PostgreSQL_getXADataSource | /**
* Returns a PostgreSQL XA data source.
*
* @throws IllegalStateException If the data source cannot be created.
*
* @see org.alfasoftware.morf.jdbc.DatabaseType#getXADataSource(String, String, String)
*/
@Override
public XADataSource getXADataSource(String jdbcUrl, String username, String password) {
try {
... | 3.68 |
flink_ConfigOptions_asList | /** Defines that the option's type should be a list of previously defined atomic type. */
public ListConfigOptionBuilder<T> asList() {
return new ListConfigOptionBuilder<>(key, clazz);
} | 3.68 |
flink_HiveParserSemanticAnalyzer_processPTFChain | /*
* - tree form is
* ^(TOK_PTBLFUNCTION name alias? partitionTableFunctionSource partitioningSpec? arguments*)
* - a partitionTableFunctionSource can be a tableReference, a SubQuery or another
* PTF invocation.
*/
private PartitionedTableFunctionSpec processPTFChain(HiveParserQB qb, HiveParserASTNode ptf)
... | 3.68 |
morf_AliasedField_getAlias | /**
* Gets the alias of the field.
*
* @return the alias
*/
public String getAlias() {
return alias;
} | 3.68 |
flink_DualInputOperator_clearFirstInput | /** Clears this operator's first input. */
public void clearFirstInput() {
this.input1 = null;
} | 3.68 |
dubbo_Pane_isTimeInWindow | /**
* Check whether given timestamp is in current pane.
*
* @param timeMillis timestamp in milliseconds.
* @return true if the given time is in current pane, otherwise false
*/
public boolean isTimeInWindow(long timeMillis) {
// [)
return startInMs <= timeMillis && timeMillis < endInMs;
} | 3.68 |
framework_AbstractGridRendererConnector_getColumnId | /**
* Gets the column id for a column.
* <p>
* In case this renderer wants be able to identify a column in such a way
* that the server also understands it, the column id is used for that.
* Columns are identified by unified ids between the client and the server.
*
* @param column
* the column object... | 3.68 |
hadoop_BlockBlobInputStream_getPos | /**
* Gets the read position of the stream.
* @return the zero-based byte offset of the read position.
* @throws IOException IO failure
*/
@Override
public synchronized long getPos() throws IOException {
checkState();
return (streamBuffer != null)
? streamPosition - streamBufferLength + streamBufferPositi... | 3.68 |
framework_AbstractSelect_isSelected | /**
* Tests if an item is selected.
*
* <p>
* In single select mode testing selection status of the item identified by
* {@link #getNullSelectionItemId()} returns true if the value of the
* property is null.
* </p>
*
* @param itemId
* the Id the of the item to be tested.
* @see #getNullSelectionIt... | 3.68 |
hbase_RegionLocator_getRegionLocations | /**
* Find all the replicas for the region on which the given row is being served.
* @param row Row to find.
* @return Locations for all the replicas of the row.
* @throws IOException if a remote or network exception occurs
*/
default List<HRegionLocation> getRegionLocations(byte[] row) throws IOException {
retu... | 3.68 |
flink_DefaultExecutionTopology_containsIntraRegionAllToAllEdge | /**
* Check if the {@link DefaultLogicalPipelinedRegion} contains intra-region all-to-all edges or
* not.
*/
private static boolean containsIntraRegionAllToAllEdge(
DefaultLogicalPipelinedRegion logicalPipelinedRegion) {
for (LogicalVertex vertex : logicalPipelinedRegion.getVertices()) {
for (Log... | 3.68 |
hbase_TableDescriptorBuilder_setNormalizerTargetRegionCount | /**
* Setting the target region count of table normalization .
* @param regionCount the target region count.
* @return the modifyable TD
*/
public ModifyableTableDescriptor setNormalizerTargetRegionCount(final int regionCount) {
return setValue(NORMALIZER_TARGET_REGION_COUNT_KEY, Integer.toString(regionCount));
} | 3.68 |
framework_NativeButtonIconAndText_buttonClick | /*
* (non-Javadoc)
*
* @see com.vaadin.ui.Button.ClickListener#buttonClick(com.vaadin.ui.Button.
* ClickEvent)
*/
@Override
public void buttonClick(ClickEvent event) {
Button b = event.getButton();
String was = b.getIconAlternateText();
if (was == null || was.isEmpty()) {
b.setIconAlternateText... | 3.68 |
hbase_CommonFSUtils_getWALRegionDir | /**
* Returns the WAL region directory based on the given table name and region name
* @param conf configuration to determine WALRootDir
* @param tableName Table that the region is under
* @param encodedRegionName Region name used for creating the final region directory
* @return the region di... | 3.68 |
hbase_RegionMover_rackManager | /**
* Set specific rackManager implementation. This setter method is for testing purpose only.
* @param rackManager rackManager impl
* @return RegionMoverBuilder object
*/
@InterfaceAudience.Private
public RegionMoverBuilder rackManager(RackManager rackManager) {
this.rackManager = rackManager;
return this;
} | 3.68 |
zilla_WsClientFactory_assembleHeader | // @return no bytes consumed to assemble websocket header
private int assembleHeader(
DirectBuffer buffer,
int offset,
int length)
{
int remaining = Math.min(length, MAXIMUM_HEADER_SIZE - headerLength);
// may copy more than actual header length (up to max header length), but will adjust at the end
... | 3.68 |
hudi_HoodieTable_getInvalidDataPaths | /**
* Returns the possible invalid data file name with given marker files.
*/
protected Set<String> getInvalidDataPaths(WriteMarkers markers) throws IOException {
return markers.createdAndMergedDataPaths(context, config.getFinalizeWriteParallelism());
} | 3.68 |
flink_MapDataUtil_convertToJavaMap | /**
* Converts a {@link MapData} into Java {@link Map}, the keys and values of the Java map still
* holds objects of internal data structures.
*/
public static Map<Object, Object> convertToJavaMap(
MapData map, LogicalType keyType, LogicalType valueType) {
ArrayData keyArray = map.keyArray();
ArrayDa... | 3.68 |
hbase_HRegionFileSystem_createTempName | /**
* Generate a unique temporary Path. Used in conjuction with commitStoreFile() to get a safer file
* creation. <code>
* Path file = fs.createTempName();
* ...StoreFile.Writer(file)...
* fs.commitStoreFile("family", file);
* </code>
* @param suffix extra information to append to the generated name
* @return U... | 3.68 |
framework_TabsheetBaseConnector_init | /*
* (non-Javadoc)
*
* @see com.vaadin.client.ui.AbstractConnector#init()
*/
@Override
protected void init() {
super.init();
getWidget().setClient(getConnection());
} | 3.68 |
framework_CompositeValidator_validate | /**
* Validates the given value.
* <p>
* The value is valid, if:
* <ul>
* <li><code>MODE_AND</code>: All of the sub-validators are valid
* <li><code>MODE_OR</code>: Any of the sub-validators are valid
* </ul>
*
* If the value is invalid, validation error is thrown. If the error message
* is set (non-null), it... | 3.68 |
flink_Tuple17_copy | /**
* Shallow tuple copy.
*
* @return A new Tuple with the same fields as this.
*/
@Override
@SuppressWarnings("unchecked")
public Tuple17<T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16>
copy() {
return new Tuple17<>(
this.f0, this.f1, this.f2, this.f3, this.f4, th... | 3.68 |
hibernate-validator_MethodValidationConfiguration_allowOverridingMethodAlterParameterConstraint | /**
* Define whether overriding methods that override constraints should throw a {@code ConstraintDefinitionException}.
* The default value is {@code false}, i.e. do not allow.
*
* See Section 5.6.5 of the Jakarta Bean Validation Specification, specifically
* <pre>
* "In sub types (be it sub classes/interfaces or... | 3.68 |
shardingsphere-elasticjob_RegExceptionHandler_handleException | /**
* Handle exception.
*
* @param cause exception to be handled
*/
public static void handleException(final Exception cause) {
if (null == cause) {
return;
}
if (isIgnoredException(cause) || null != cause.getCause() && isIgnoredException(cause.getCause())) {
log.debug("Elastic job: ign... | 3.68 |
hbase_CacheConfig_shouldCacheBloomsOnWrite | /**
* @return true if bloom blocks should be written to the cache when an HFile is written, false if
* not
*/
public boolean shouldCacheBloomsOnWrite() {
return this.cacheBloomsOnWrite;
} | 3.68 |
hbase_MasterDDLOperationHelper_deleteColumnFamilyFromFileSystem | /**
* Remove the column family from the file system
**/
public static void deleteColumnFamilyFromFileSystem(final MasterProcedureEnv env,
final TableName tableName, final List<RegionInfo> regionInfoList, final byte[] familyName,
final boolean hasMob) throws IOException {
final MasterFileSystem mfs = env.getMast... | 3.68 |
flink_CopyOnWriteStateMap_makeTable | /**
* Allocate a table of the given capacity and set the threshold accordingly.
*
* @param newCapacity must be a power of two
*/
private StateMapEntry<K, N, S>[] makeTable(int newCapacity) {
if (newCapacity < MAXIMUM_CAPACITY) {
threshold = (newCapacity >> 1) + (newCapacity >> 2); // 3/4 capacity
}... | 3.68 |
querydsl_Expressions_booleanTemplate | /**
* Create a new Template expression
*
* @param template template
* @param args template parameters
* @return template expression
*/
public static BooleanTemplate booleanTemplate(Template template, List<?> args) {
return new BooleanTemplate(template, args);
} | 3.68 |
hudi_RollbackUtils_getRollbackPlan | /**
* Get Latest version of Rollback plan corresponding to a clean instant.
*
* @param metaClient Hoodie Table Meta Client
* @param rollbackInstant Instant referring to rollback action
* @return Rollback plan corresponding to rollback instant
* @throws IOException
*/
public static HoodieRollbackPlan getRoll... | 3.68 |
AreaShop_GeneralRegion_getConfigurationSectionSetting | /**
* Get a configuration section setting for this region, defined as follows
* - If earlyResult is non-null, use that
* - Else if the region has the setting in its own file (/regions/regionName.yml), use that
* - Else if the region has groups, use the setting defined by the most important group, if any
* - Otherw... | 3.68 |
flink_Table_getSchema | /**
* Returns the schema of this table.
*
* @deprecated This method has been deprecated as part of FLIP-164. {@link TableSchema} has been
* replaced by two more dedicated classes {@link Schema} and {@link ResolvedSchema}. Use
* {@link Schema} for declaration in APIs. {@link ResolvedSchema} is offered by th... | 3.68 |
incubator-hugegraph-toolchain_HugeGraphLoader_executeParseTask | /**
* Execute parse task sync
*/
private void executeParseTask(InputStruct struct, ElementMapping mapping,
ParseTaskBuilder.ParseTask task) {
long start = System.currentTimeMillis();
// Sync parse
List<List<Record>> batches = task.get();
long end = System.currentTimeMilli... | 3.68 |
framework_VFilterSelect_filterOptions | /**
* Filters the options at certain page using the given filter
*
* @param page
* The page to filter
* @param filter
* The filter to apply to the options
* @param immediate
* Whether to send the options request immediately
*/
private void filterOptions(int page, String filter,... | 3.68 |
AreaShop_GeneralRegion_getFeature | /**
* Get a feature of this region.
* @param clazz The class of the feature to get
* @param <T> The feature to get
* @return The feature (either just instantiated or cached)
*/
public <T extends RegionFeature> T getFeature(Class<T> clazz) {
RegionFeature result = features.get(clazz);
if(result == null) {
resul... | 3.68 |
graphhopper_Measurement_start | // creates properties file in the format key=value
// Every value is one y-value in a separate diagram with an identical x-value for every Measurement.start call
void start(PMap args) throws IOException {
final String graphLocation = args.getString("graph.location", "");
final boolean useJson = args.getBool("me... | 3.68 |
framework_BeanValidator_getJavaxBeanValidatorFactory | /**
* Returns the underlying JSR-303 bean validator factory used. A factory is
* created using {@link Validation} if necessary.
*
* @return the validator factory to use
*/
protected static ValidatorFactory getJavaxBeanValidatorFactory() {
return LazyFactoryInitializer.FACTORY;
} | 3.68 |
flink_ExecutionEnvironment_registerType | /**
* Registers the given type with the serialization stack. If the type is eventually serialized
* as a POJO, then the type is registered with the POJO serializer. If the type ends up being
* serialized with Kryo, then it will be registered at Kryo to make sure that only tags are
* written.
*
* @param type The c... | 3.68 |
pulsar_LedgerMetadataUtils_buildMetadataForDelayedIndexBucket | /**
* Build additional metadata for a delayed message index bucket.
*
* @param bucketKey key of the delayed message bucket
* @param topicName name of the topic
* @param cursorName name of the cursor
* @return an immutable map which describes the schema
*/
public static Map<String, byte[]> buildMetadataForDelay... | 3.68 |
flink_Schema_fromFields | /** Adopts the given field names and field data types as physical columns of the schema. */
public Builder fromFields(
List<String> fieldNames, List<? extends AbstractDataType<?>> fieldDataTypes) {
Preconditions.checkNotNull(fieldNames, "Field names must not be null.");
Preconditions.checkNotNull(fieldD... | 3.68 |
flink_ProducerMergedPartitionFileWriter_flush | /** Called in single-threaded ioExecutor. Order is guaranteed. */
private void flush(
List<SubpartitionBufferContext> toWrite, CompletableFuture<Void> flushSuccessNotifier) {
try {
List<ProducerMergedPartitionFileIndex.FlushedBuffer> buffers = new ArrayList<>();
calculateSizeAndFlushBuffers(... | 3.68 |
hudi_ClientIds_builder | /**
* Returns the builder.
*/
public static Builder builder() {
return new Builder();
} | 3.68 |
hudi_SourceFormatAdapter_processErrorEvents | /**
* transform datasets with error events when error table is enabled
* @param eventsRow
* @return
*/
public Option<Dataset<Row>> processErrorEvents(Option<Dataset<Row>> eventsRow,
ErrorEvent.ErrorReason errorReason) {
return eventsRow.map(dataset -> {
... | 3.68 |
hbase_ZKListener_getWatcher | /** Returns The watcher associated with this listener */
public ZKWatcher getWatcher() {
return this.watcher;
} | 3.68 |
hbase_MemStore_stopReplayingFromWAL | /**
* This message intends to inform the MemStore that the replaying edits from WAL are done
*/
default void stopReplayingFromWAL() {
return;
} | 3.68 |
dubbo_MetadataInfo_getNoProtocolServiceInfo | /**
* Get service infos of an interface with specified group, version.
* There may have several service infos of different protocols, this method will simply pick the first one.
*
* @param serviceKeyWithoutProtocol key is of format '{group}/{interface name}:{version}'
* @return the first service info related to se... | 3.68 |
hbase_RegionNormalizerManager_getSplitPlanCount | /**
* Return the number of times a {@link SplitNormalizationPlan} has been submitted.
*/
public long getSplitPlanCount() {
return worker == null ? 0 : worker.getSplitPlanCount();
} | 3.68 |
querydsl_SQLExpressions_varSamp | /**
* returns the sample variance of a set of numbers after discarding the nulls in this set.
*
* @param expr argument
* @return var_samp(expr)
*/
public static <T extends Number> WindowOver<T> varSamp(Expression<T> expr) {
return new WindowOver<T>(expr.getType(), SQLOps.VARSAMP, expr);
} | 3.68 |
flink_PythonDependencyUtils_addPythonArchive | /**
* Adds a Python archive file (zip format). The file will be extracted and moved to a
* dedicated directory under the working directory of the Python UDF workers. The param
* `targetDir` is the name of the dedicated directory. The Python UDFs and the config option
* "python.executable" could access the extracted... | 3.68 |
hbase_Increment_numFamilies | /**
* Method for retrieving the number of families to increment from
* @return number of families
*/
@Override
public int numFamilies() {
return this.familyMap.size();
} | 3.68 |
flink_CheckpointProperties_getCheckpointType | /** Gets the type of the checkpoint (checkpoint / savepoint). */
public SnapshotType getCheckpointType() {
return checkpointType;
} | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.