name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_HiveTableUtil_validateConstraint | // returns a constraint trait that requires VALIDATE
public static byte validateConstraint(byte trait) {
return (byte) (trait | HIVE_CONSTRAINT_VALIDATE);
} | 3.68 |
framework_ThemeResource_equals | /**
* Tests if the given object equals this Resource.
*
* @param obj
* the object to be tested for equality.
* @return <code>true</code> if the given object equals this Icon,
* <code>false</code> if not.
* @see java.lang.Object#equals(Object)
*/
@Override
public boolean equals(Object obj) {
... | 3.68 |
hbase_SplitTableRegionProcedure_checkSplittable | /**
* Check whether the region is splittable
* @param env MasterProcedureEnv
* @param regionToSplit parent Region to be split
*/
private void checkSplittable(final MasterProcedureEnv env, final RegionInfo regionToSplit)
throws IOException {
// Ask the remote RS if this region is splittable.
// If we... | 3.68 |
graphhopper_BitUtil_toLong | /**
* See the counterpart {@link #fromLong(long)}
*/
public final long toLong(byte[] b) {
return toLong(b, 0);
} | 3.68 |
flink_EventsGenerator_nextInvalid | /**
* Creates an event for an illegal state transition of one of the internal state machines. If
* the generator has not yet started any state machines (for example, because no call to {@link
* #next(int, int)} was made, yet), this will return null.
*
* @return An event for a illegal state transition, or null, if ... | 3.68 |
hbase_MetricsTableRequests_getMetricRegistryInfo | // Visible for testing
public MetricRegistryInfo getMetricRegistryInfo() {
return registryInfo;
} | 3.68 |
shardingsphere-elasticjob_JobNodeStorage_addDataListener | /**
* Add data listener.
*
* @param listener data listener
*/
public void addDataListener(final DataChangedEventListener listener) {
Executor executor = ListenerNotifierManager.getInstance().getJobNotifyExecutor(jobName);
regCenter.watch("/" + jobName, listener, executor);
} | 3.68 |
framework_VTree_getNavigationStartKey | /**
* Get the key the moves the selection to the beginning of the table. By
* default this is the Home key but by overriding this you can change the
* key to whatever you want.
*
* @return
*/
protected int getNavigationStartKey() {
return KeyCodes.KEY_HOME;
} | 3.68 |
open-banking-gateway_PsuSecureStorage_registerPsu | /**
* Registers PSU in Datasafe
* @param psu PSU data
* @param password PSU KeyStore/Datasafe password.
*/
public void registerPsu(Psu psu, Supplier<char[]> password) {
this.userProfile()
.createDocumentKeystore(
psu.getUserIdAuth(password),
config.defaultPriv... | 3.68 |
hbase_RegionCoprocessorHost_preScannerNext | /**
* @param s the scanner
* @param results the result set returned by the region server
* @param limit the maximum number of results to return
* @return 'has next' indication to client if bypassing default behavior, or null otherwise
* @exception IOException Exception
*/
public Boolean preScannerNext(fin... | 3.68 |
hadoop_CurrentJHParser_canParse | /**
* Can this parser parse the input?
*
* @param input
* @return Whether this parser can parse the input.
* @throws IOException
*/
public static boolean canParse(InputStream input) throws IOException {
final DataInputStream in = new ForkedDataInputStream(input);
try {
final EventReader reader = new Eve... | 3.68 |
hbase_AbstractFSWAL_getNumRolledLogFiles | // public only until class moves to o.a.h.h.wal
/** Returns the number of rolled log files */
public int getNumRolledLogFiles() {
return walFile2Props.size();
} | 3.68 |
morf_SqlUtils_selectDistinct | /**
* Constructs a distinct Select Statement which optionally selects on a subset of fields.
* If no fields are specified then this is equivalent of selecting all
* fields (i.e. {@code SELECT DISTINCT * FROM x}).
*
* <p>Usage is discouraged; this method will be deprecated at some point. Use
* {@link SelectStateme... | 3.68 |
pulsar_Topics_delete | /**
* @see Topics#delete(String, boolean, boolean)
* IMPORTANT NOTICE: the application is not able to connect to the topic(delete then re-create with same name) again
* if the schema auto uploading is disabled. Besides, users should to use the truncate method to clean up
* data of the topic instead of delete method... | 3.68 |
dubbo_ExpiringMap_getTimeToLive | /**
* get time to live
*
* @return time to live
*/
public int getTimeToLive() {
return (int) timeToLiveMillis / 1000;
} | 3.68 |
framework_ExpandingContainer_addItemAt | /**
* @throws UnsupportedOperationException
* always
*/
@Override
public Item addItemAt(int index, Object newItemId) {
throw new UnsupportedOperationException();
} | 3.68 |
hadoop_Lz4Codec_getDecompressorType | /**
* Get the type of {@link Decompressor} needed by this {@link CompressionCodec}.
*
* @return the type of decompressor needed by this codec.
*/
@Override
public Class<? extends Decompressor> getDecompressorType() {
return Lz4Decompressor.class;
} | 3.68 |
framework_VAbstractCalendarPanel_isAcceptedByRangeStart | /**
* Accepts dates greater than or equal to rangeStart, depending on the
* resolution. If the resolution is set to DAY, the range will compare on a
* day-basis. If the resolution is set to YEAR, only years are compared. So
* even if the range is set to one millisecond in next year, also next year
* will be includ... | 3.68 |
hadoop_CoderUtil_getValidIndexes | /**
* Picking up indexes of valid inputs.
* @param inputs decoding input buffers
* @param <T>
*/
static <T> int[] getValidIndexes(T[] inputs) {
int[] validIndexes = new int[inputs.length];
int idx = 0;
for (int i = 0; i < inputs.length; i++) {
if (inputs[i] != null) {
validIndexes[idx++] = i;
}
... | 3.68 |
querydsl_MetaDataExporter_setNamePrefix | /**
* Override the name prefix for the classes (default: Q)
*
* @param namePrefix name prefix for querydsl-types (default: Q)
*/
public void setNamePrefix(String namePrefix) {
module.bind(CodegenModule.PREFIX, namePrefix);
} | 3.68 |
hadoop_ReadStatistics_getTotalEcDecodingTimeMillis | /**
* Return the total time in milliseconds used for erasure coding decoding.
*/
public synchronized long getTotalEcDecodingTimeMillis() {
return totalEcDecodingTimeMillis;
} | 3.68 |
hadoop_QuotaUsage_isTypeQuotaSet | /**
* Return true if any storage type quota has been set.
*
* @return if any storage type quota has been set true, not false.
* */
public boolean isTypeQuotaSet() {
if (typeQuota != null) {
for (StorageType t : StorageType.getTypesSupportingQuota()) {
if (typeQuota[t.ordinal()] > 0L) {
return tr... | 3.68 |
framework_AbstractSplitPanel_isUserOriginated | /**
* {@inheritDoc}
*
* @since 8.1
*/
@Override
public boolean isUserOriginated() {
return userOriginated;
} | 3.68 |
hudi_InstantRange_builder | /**
* Returns the builder.
*/
public static Builder builder() {
return new Builder();
} | 3.68 |
hbase_RootProcedureState_loadStack | /**
* Called on store load by the ProcedureExecutor to load part of the stack. Each procedure has its
* own stack-positions. Which means we have to write to the store only the Procedure we executed,
* and nothing else. on load we recreate the full stack by aggregating each procedure
* stack-positions.
*/
protected... | 3.68 |
morf_XmlDataSetConsumer_emptyElement | /**
* Output an empty (self-closing) element: <foo/>
*
* @param contentHandler The content handler
* @param name The element name
* @param attributes The attributes
* @throws SAXException When there's a writer error
*/
private void emptyElement(ContentHandler contentHandler, String name, Attributes attributes) ... | 3.68 |
framework_VComboBox_setAllowNewItems | /**
* Sets whether creation of new items when there is no match is allowed or
* not.
*
* @param allowNewItems
* true to allow creation of new items, false to only allow
* selection of existing items
*/
public void setAllowNewItems(boolean allowNewItems) {
this.allowNewItems = allowNewIt... | 3.68 |
rocketmq-connect_JdbcSinkTask_start | /**
* Start the component
* @param keyValue
*/
@Override
public void start(KeyValue keyValue) {
originalConfig = keyValue;
config = new JdbcSinkConfig(keyValue);
remainingRetries = config.getMaxRetries();
this.dialect = DatabaseDialectLoader.getDatabaseDialect(config);
log.info("Initializing writ... | 3.68 |
hbase_AvlUtil_prepend | /**
* Prepend a node to the tree before a specific node
* @param head the head of the linked list
* @param base the node which we want to add the {@code node} before it
* @param node the node which we want to add it before the {@code base} node
*/
public static <TNode extends AvlLinkedNode> TNode prepend(TNode hea... | 3.68 |
querydsl_NumberExpression_ceil | /**
* Create a {@code ceil(this)} expression
*
* <p>Returns the smallest (closest to negative infinity)
* {@code double} value that is greater than or equal to the
* argument and is equal to a mathematical integer</p>
*
* @return ceil(this)
* @see java.lang.Math#ceil(double)
*/
public NumberExpression<T> ceil(... | 3.68 |
flink_TieredStorageConfiguration_getMemoryTierNumBytesPerSegment | /**
* Get the segment size of memory tier.
*
* @return segment size.
*/
public int getMemoryTierNumBytesPerSegment() {
return memoryTierNumBytesPerSegment;
} | 3.68 |
hbase_LruBlockCache_containsBlock | /**
* Whether the cache contains block with specified cacheKey
* @return true if contains the block
*/
@Override
public boolean containsBlock(BlockCacheKey cacheKey) {
return map.containsKey(cacheKey);
} | 3.68 |
flink_PojoSerializerSnapshot_previousSerializerHasNonRegisteredSubclasses | /**
* Checks whether the previous serializer, represented by this snapshot, has non-registered
* subclasses.
*/
private static boolean previousSerializerHasNonRegisteredSubclasses(
LinkedOptionalMap<Class<?>, TypeSerializerSnapshot<?>>
nonRegisteredSubclassSerializerSnapshots) {
return no... | 3.68 |
hbase_AdaptiveLifoCoDelCallQueue_updateTunables | /**
* Update tunables.
* @param newCodelTargetDelay new CoDel target delay
* @param newCodelInterval new CoDel interval
* @param newLifoThreshold new Adaptive Lifo threshold
*/
public void updateTunables(int newCodelTargetDelay, int newCodelInterval,
double newLifoThreshold) {
this.codelTargetDelay = new... | 3.68 |
hbase_BucketAllocator_freeCount | /**
* How many more items can be allocated from the currently claimed blocks of this bucket size
*/
public long freeCount() {
return freeCount;
} | 3.68 |
hbase_MasterObserver_postEnableTable | /**
* Called after the enableTable operation has been requested. Called as part of enable table RPC
* call.
* @param ctx the environment to interact with the framework and master
* @param tableName the name of the table
*/
default void postEnableTable(final ObserverContext<MasterCoprocessorEnvironment> ctx,
... | 3.68 |
flink_Tuple11_toString | /**
* Creates a string representation of the tuple in the form (f0, f1, f2, f3, f4, f5, f6, f7, f8,
* f9, f10), where the individual fields are the value returned by calling {@link
* Object#toString} on that field.
*
* @return The string representation of the tuple.
*/
@Override
public String toString() {
ret... | 3.68 |
framework_Calendar_isClientChangeAllowed | /**
* Is the user allowed to trigger events which alters the events.
*
* @return true if the client is allowed to send changes to server
* @see #isEventClickAllowed()
*/
protected boolean isClientChangeAllowed() {
return !isReadOnly();
} | 3.68 |
flink_TestcontainersSettings_network | /**
* Sets the {@code network} and returns a reference to this Builder enabling method
* chaining.
*
* @param network The {@code network} to set.
* @return A reference to this Builder.
*/
public Builder network(Network network) {
this.network = network;
return this;
} | 3.68 |
hudi_TableSizeStats_readConfigFromFileSystem | /**
* Reads config from the file system.
*
* @param jsc {@link JavaSparkContext} instance.
* @param cfg {@link Config} instance.
* @return the {@link TypedProperties} instance.
*/
private TypedProperties readConfigFromFileSystem(JavaSparkContext jsc, Config cfg) {
return UtilHelpers.readConfig(jsc.hadoopConfigu... | 3.68 |
flink_SessionWithGap_on | /**
* Specifies the time attribute on which rows are grouped.
*
* <p>For streaming tables you can specify grouping by a event-time or processing-time
* attribute.
*
* <p>For batch tables you can specify grouping on a timestamp or long attribute.
*
* @param timeField time attribute for streaming and batch tables... | 3.68 |
morf_AbstractSelectStatementBuilder_getFromSelects | /**
* @return the fromSelects
*/
List<SelectStatement> getFromSelects() {
return fromSelects;
} | 3.68 |
hudi_SparkRDDReadClient_getPendingCompactions | /**
* Return all pending compactions with instant time for clients to decide what to compact next.
*
* @return
*/
public List<Pair<String, HoodieCompactionPlan>> getPendingCompactions() {
HoodieTableMetaClient metaClient =
HoodieTableMetaClient.builder().setConf(hadoopConf).setBasePath(hoodieTable.getMetaCl... | 3.68 |
flink_PropertiesUtil_flatten | /**
* Flatten a recursive {@link Properties} to a first level property map.
*
* <p>In some cases, {@code KafkaProducer#propsToMap} for example, Properties is used purely as
* a HashTable without considering its default properties.
*
* @param config Properties to be flattened
* @return Properties without defaults... | 3.68 |
hadoop_ServiceRecord_set | /**
* Handle unknown attributes by storing them in the
* {@link #attributes} map
* @param key attribute name
* @param value attribute value.
*/
@JsonAnySetter
public void set(String key, Object value) {
attributes.put(key, value.toString());
} | 3.68 |
graphhopper_VectorTile_clearUintValue | /**
* <code>optional uint64 uint_value = 5;</code>
*/
public Builder clearUintValue() {
bitField0_ = (bitField0_ & ~0x00000010);
uintValue_ = 0L;
onChanged();
return this;
} | 3.68 |
flink_SlotSharingGroup_setTaskOffHeapMemory | /** Set the task off-heap memory for this SlotSharingGroup. */
public Builder setTaskOffHeapMemory(MemorySize taskOffHeapMemory) {
this.taskOffHeapMemory = taskOffHeapMemory;
return this;
} | 3.68 |
pulsar_ProducerConfiguration_getEncryptionKeys | /**
*
* @return encryptionKeys
*
*/
public Set<String> getEncryptionKeys() {
return conf.getEncryptionKeys();
} | 3.68 |
hudi_FormatUtils_getBooleanWithAltKeys | /**
* Gets the boolean value for a {@link ConfigProperty} config from Flink configuration. The key and
* alternative keys are used to fetch the config. The default value of {@link ConfigProperty}
* config, if exists, is returned if the config is not found in the configuration.
*
* @param conf Configs in ... | 3.68 |
druid_SQLColumnDefinition_setIdentity | // for sqlserver
public void setIdentity(Identity identity) {
if (identity != null) {
identity.setParent(this);
}
this.identity = identity;
} | 3.68 |
framework_FlyweightRow_setSkipNext | /**
* Sets the number of cells to skip when {@link #next()} is called the
* next time. Cell hiding is also handled eagerly in this method.
*
* @param colspan
* the number of cells to skip on next invocation of
* {@link #next()}
*/
public void setSkipNext(final int colspan) {
assert cols... | 3.68 |
dubbo_InternalThread_setThreadLocalMap | /**
* Sets the internal data structure that keeps the threadLocal variables bound to this thread.
* Note that this method is for internal use only, and thus is subject to change at any time.
*/
public final void setThreadLocalMap(InternalThreadLocalMap threadLocalMap) {
this.threadLocalMap = threadLocalMap;
} | 3.68 |
shardingsphere-elasticjob_JobFacade_getShardingContexts | /**
* Get sharding contexts.
*
* @return sharding contexts
*/
public ShardingContexts getShardingContexts() {
boolean isFailover = configService.load(true).isFailover();
if (isFailover) {
List<Integer> failoverShardingItems = failoverService.getLocalFailoverItems();
if (!failoverShardingItem... | 3.68 |
hbase_RequestConverter_buildWarmupRegionRequest | /**
* Create a WarmupRegionRequest for a given region name
* @param regionInfo Region we are warming up
*/
public static WarmupRegionRequest buildWarmupRegionRequest(final RegionInfo regionInfo) {
WarmupRegionRequest.Builder builder = WarmupRegionRequest.newBuilder();
builder.setRegionInfo(ProtobufUtil.toRegionI... | 3.68 |
framework_TouchScrollDelegate_addElement | /**
* Registers the given element as scrollable.
*/
public void addElement(Element scrollable) {
scrollable.addClassName(SCROLLABLE_CLASSNAME);
if (requiresDelegate()) {
delegate.scrollableElements.add(scrollable);
}
} | 3.68 |
hadoop_AbstractConfigurableFederationPolicy_getIsDirty | /**
* Returns true whether the last reinitialization requires actual changes, or
* was "free" as the weights have not changed. This is used by subclasses
* overriding reinitialize and calling super.reinitialize() to know whether to
* quit early.
*
* @return whether more work is needed to initialize.
*/
public bo... | 3.68 |
morf_ValueConverters_booleanValue | /**
* Always returns true or false in line with the contract of {@link Boolean#valueOf(String)}.
*/
@Override
public Boolean booleanValue(T value) {
return Boolean.valueOf(value.toString());
} | 3.68 |
pulsar_ManagedLedgerConfig_getMinimumRolloverTimeMs | /**
* @return the minimum rollover time
*/
public int getMinimumRolloverTimeMs() {
return minimumRolloverTimeMs;
} | 3.68 |
hbase_StorageClusterVersionModel_getVersion | /** Returns the storage cluster version */
@XmlAttribute(name = "Version")
public String getVersion() {
return version;
} | 3.68 |
flink_UserDefinedFunctionHelper_prepareInstance | /** Prepares a {@link UserDefinedFunction} instance for usage in the API. */
public static void prepareInstance(ReadableConfig config, UserDefinedFunction function) {
validateClass(function.getClass(), false);
cleanFunction(config, function);
} | 3.68 |
hadoop_HsController_taskCounters | /*
* (non-Javadoc)
* @see org.apache.hadoop.mapreduce.v2.app.webapp.AppController#taskCounters()
*/
@Override
public void taskCounters() {
super.taskCounters();
} | 3.68 |
pulsar_TopicsBase_getSchemaData | // Build schemaData from passed in schema string.
private SchemaData getSchemaData(String keySchema, String valueSchema) {
try {
SchemaInfoImpl valueSchemaInfo = (valueSchema == null || valueSchema.isEmpty())
? (SchemaInfoImpl) StringSchema.utf8().getSchemaInfo() :
SCHEMA_INF... | 3.68 |
zilla_ManyToOneRingBuffer_tryClaim | /**
* {@inheritDoc}
*/
public int tryClaim(final int msgTypeId, final int length)
{
checkTypeId(msgTypeId);
checkMsgLength(length);
final AtomicBuffer buffer = this.buffer;
final int recordLength = length + HEADER_LENGTH;
final int recordIndex = claimCapacity(buffer, recordLength);
if (INSUF... | 3.68 |
hbase_CheckAndMutate_getQualifier | /** Returns the qualifier to check */
public byte[] getQualifier() {
return qualifier;
} | 3.68 |
hudi_HoodieAsyncService_shutdownCallback | /**
* Add shutdown callback for the completable future.
*
* @param callback The callback
*/
@SuppressWarnings("unchecked")
private void shutdownCallback(Function<Boolean, Boolean> callback) {
if (future == null) {
return;
}
future.whenComplete((resp, error) -> {
if (null != callback) {
callback... | 3.68 |
framework_PopupDateField_setAssistiveText | /**
* Set a description that explains the usage of the Widget for users of
* assistive devices.
*
* @param description
* String with the description
*/
public void setAssistiveText(String description) {
getState().descriptionForAssistiveDevices = description;
} | 3.68 |
hbase_WALPrettyPrinter_run | /**
* Pass one or more log file names and formatting options and it will dump out a text version of
* the contents on <code>stdout</code>. Command line arguments Thrown upon file system errors etc.
*/
public static void run(String[] args) throws IOException {
// create options
Options options = new Options();
... | 3.68 |
hmily_PropertyKeyParse_parse | /**
* Parse property key [ ].
*
* @param propertyName the property name
* @return the property key [ ]
*/
public PropertyKey[] parse(final String propertyName) {
// Use a local copy in case another thread changes things
LastKey<String> last = this.lastKeyStr;
if (last != null && last.isFrom(propertyNam... | 3.68 |
framework_VTree_handleKeyNavigation | /**
* Handles the keyboard navigation.
*
* @param keycode
* The keycode of the pressed key
* @param ctrl
* Was ctrl pressed
* @param shift
* Was shift pressed
* @return Returns true if the key was handled, else false
*/
protected boolean handleKeyNavigation(int keycode, boolea... | 3.68 |
flink_ServerConnection_createEstablishedConnection | /**
* Creates an established connection from the given channel.
*
* @param channel Channel to create an established connection from
*/
private InternalConnection<REQ, RESP> createEstablishedConnection(Channel channel) {
if (failureCause != null || !running) {
// Close the channel and we are done. Any qu... | 3.68 |
AreaShop_GeneralRegion_getIntegerSetting | /**
* Get a boolean setting for this region, defined as follows
* - If the region has the setting in its own file (/regions/regionName.yml), use that
* - If the region has groups, use the setting defined by the most important group, if any
* - Otherwise fallback to the default.yml file setting
* @param path The pa... | 3.68 |
flink_WindowedOperatorTransformation_apply | /**
* Applies the given window function to each window. The window function is called for each
* evaluation of the window for each key individually. The output of the window function is
* interpreted as a regular non-windowed stream.
*
* <p>Note that this function requires that all data in the windows is buffered ... | 3.68 |
framework_Form_isModified | /*
* Is the object modified but not committed? Don't add a JavaDoc comment
* here, we use the default one from the interface.
*/
@Override
public boolean isModified() {
for (final Object id : propertyIds) {
final Field<?> f = fields.get(id);
if (f != null && f.isModified()) {
return t... | 3.68 |
hadoop_XAttrCodec_decodeValue | /**
* Decode string representation of a value and check whether it's
* encoded. If the given string begins with 0x or 0X, it expresses
* a hexadecimal number. If the given string begins with 0s or 0S,
* base64 encoding is expected. If the given string is enclosed in
* double quotes, the inner string is treated a... | 3.68 |
flink_TimeWindow_getStart | /**
* Gets the starting timestamp of the window. This is the first timestamp that belongs to this
* window.
*
* @return The starting timestamp of this window.
*/
public long getStart() {
return start;
} | 3.68 |
flink_CreditBasedSequenceNumberingViewReader_getAvailabilityAndBacklog | /**
* Returns true only if the next buffer is an event or the reader has both available credits and
* buffers.
*
* @implSpec BEWARE: this must be in sync with {@link #getNextDataType(BufferAndBacklog)}, such
* that {@code getNextDataType(bufferAndBacklog) != NONE <=>
* AvailabilityWithBacklog#isAvailable(... | 3.68 |
flink_ExecutionConfig_disableGenericTypes | /**
* Disables the use of generic types (types that would be serialized via Kryo). If this option
* is used, Flink will throw an {@code UnsupportedOperationException} whenever it encounters a
* data type that would go through Kryo for serialization.
*
* <p>Disabling generic types can be helpful to eagerly find and... | 3.68 |
hibernate-validator_MethodInheritanceTree_hasParallelDefinitions | /**
* Checks if there are any parallel definitions of the method in the hierarchy.
*
* @return {@code true} if there are any parallel definitions of the method in the hierarchy, {@code false} otherwise
*/
public boolean hasParallelDefinitions() {
return topLevelMethods.size() > 1;
} | 3.68 |
open-banking-gateway_AccountInformationRequestCommon_fintech_calls_list_accounts_for_anton_brueckner | // Note that anton.brueckner is typically used for REDIRECT (real REDIRECT that is returned by bank, and not REDIRECT approach in table)
public SELF fintech_calls_list_accounts_for_anton_brueckner(String bankProfileId) {
return fintech_calls_list_accounts_for_anton_brueckner(bankProfileId, false);
} | 3.68 |
hbase_BalancerClusterState_getOrComputeRackLocalities | /**
* Retrieves and lazily initializes a field storing the locality of every region/server
* combination
*/
public float[][] getOrComputeRackLocalities() {
if (rackLocalities == null || regionsToMostLocalEntities == null) {
computeCachedLocalities();
}
return rackLocalities;
} | 3.68 |
flink_RocksDBOperationUtils_sanityCheckArenaBlockSize | /**
* Logs a warning if the arena block size is too high causing RocksDB to flush constantly.
* Essentially, the condition <a
* href="https://github.com/dataArtisans/frocksdb/blob/49bc897d5d768026f1eb816d960c1f2383396ef4/include/rocksdb/write_buffer_manager.h#L47">
* here</a> will always be true.
*
* @param write... | 3.68 |
morf_SchemaUtils_namesOfColumns | /**
* Turn a list of columns into a list of the columns' names.
*
* @param columns The columns
* @return The columns' names.
*/
public static List<String> namesOfColumns(List<Column> columns) {
return Lists.transform(columns, new Function<Column, String>() {
@Override
public String apply(Column column) {... | 3.68 |
flink_CheckpointProperties_discardOnJobFinished | /**
* Returns whether the checkpoint should be discarded when the owning job reaches the {@link
* JobStatus#FINISHED} state.
*
* @return <code>true</code> if the checkpoint should be discarded when the owning job reaches
* the {@link JobStatus#FINISHED} state; <code>false</code> otherwise.
* @see CompletedChe... | 3.68 |
rocketmq-connect_Worker_checkStoppedTasks | /**
* check stopped tasks
*/
private void checkStoppedTasks() {
for (Runnable runnable : stoppedTasks) {
WorkerTask workerTask = (WorkerTask) runnable;
Future future = taskToFutureMap.get(runnable);
try {
if (null != future) {
future.get(workerConfig.getMaxStart... | 3.68 |
hudi_AvroSchemaUtils_isAtomicSchemasCompatibleEvolution | /**
* Establishes whether {@code newReaderSchema} is compatible w/ {@code prevWriterSchema}, as
* defined by Avro's {@link AvroSchemaCompatibility}.
* {@code newReaderSchema} is considered compatible to {@code prevWriterSchema}, iff data written using {@code prevWriterSchema}
* could be read by {@code newReaderSche... | 3.68 |
dubbo_AbstractDynamicConfiguration_getTimeout | /**
* Get the timeout from {@link URL the specified connection URL}
*
* @param url {@link URL the specified connection URL}
* @return non-null
* @since 2.7.8
*/
protected static long getTimeout(URL url) {
return getParameter(url, TIMEOUT_PARAM_NAME, -1L);
} | 3.68 |
flink_ProgramOptionsUtils_isPythonEntryPoint | /**
* @return True if the commandline contains "-py" or "-pym" options or comes from PyFlink shell,
* false otherwise.
*/
public static boolean isPythonEntryPoint(CommandLine line) {
return line.hasOption(PY_OPTION.getOpt())
|| line.hasOption(PYMODULE_OPTION.getOpt())
|| "org.apache.f... | 3.68 |
framework_VScrollTable_setAlign | /**
* Set alignment of the text in the cell.
*
* @param c
* The alignment which can be ALIGN_CENTER, ALIGN_LEFT,
* ALIGN_RIGHT
*/
public void setAlign(char c) {
if (align != c) {
switch (c) {
case ALIGN_CENTER:
captionContainer.getStyle().setTextAlign(TextAlig... | 3.68 |
framework_TreeConnector_updateActionMap | /**
* Registers action for the root and also for individual nodes
*
* @param uidl
*/
private void updateActionMap(UIDL uidl) {
for (final Object child : uidl) {
final UIDL action = (UIDL) child;
final String key = action.getStringAttribute("key");
final String caption = action
... | 3.68 |
hudi_HoodieSparkQuickstart_pointInTimeQuery | /**
* Lets look at how to query data as of a specific time.
* The specific time can be represented by pointing endTime to a specific commit time
* and beginTime to “000” (denoting earliest possible commit time).
*/
public static void pointInTimeQuery(SparkSession spark, String tablePath, String tableName) {
List<... | 3.68 |
hbase_BucketCache_disableCache | /**
* Used to shut down the cache -or- turn it off in the case of something broken.
*/
private void disableCache() {
if (!cacheEnabled) return;
LOG.info("Disabling cache");
cacheEnabled = false;
ioEngine.shutdown();
this.scheduleThreadPool.shutdown();
for (int i = 0; i < writerThreads.length; ++i)
wri... | 3.68 |
hbase_RpcExecutor_getQueues | /** Returns the list of request queues */
protected List<BlockingQueue<CallRunner>> getQueues() {
return queues;
} | 3.68 |
morf_SchemaValidator_isSQLReservedWord | /**
* Method to establish if a given string is an SQL Reserved Word
*
* @param word the string to establish if its a SQL Reserved Word
* @return true if its a SQL Reserved Word otherwise false.
*/
boolean isSQLReservedWord(String word){
return sqlReservedWords.get().contains(word.toUpperCase());
} | 3.68 |
flink_SkipListUtils_getKeyDataOffset | /**
* Returns the offset of key data in the key space.
*
* @param level level of the key.
*/
public static int getKeyDataOffset(int level) {
return SkipListUtils.getKeyMetaLen(level);
} | 3.68 |
querydsl_MathExpressions_tan | /**
* Create a {@code tan(num)} expression
*
* <p>Returns the tangent of an angle of num radians.</p>
*
* @param num numeric expression
* @return tan(num)
*/
public static <A extends Number & Comparable<?>> NumberExpression<Double> tan(Expression<A> num) {
return Expressions.numberOperation(Double.class, Ops... | 3.68 |
framework_GridKeyPressEvent_getCharCode | /**
* Gets the char code for this event.
*
* @return the char code
*/
public char getCharCode() {
return (char) getUnicodeCharCode();
} | 3.68 |
hbase_BlockCache_getFullyCachedFiles | /**
* Returns an Optional containing the map of files that have been fully cached (all its blocks are
* present in the cache. This method may not be overridden by all implementing classes. In such
* cases, the returned Optional will be empty.
* @return empty optional if this method is not supported, otherwise the r... | 3.68 |
hbase_ClientMetaTableAccessor_getRegionLocation | /** Returns the HRegionLocation from meta for the given region */
public static CompletableFuture<Optional<HRegionLocation>>
getRegionLocation(AsyncTable<?> metaTable, byte[] regionName) {
CompletableFuture<Optional<HRegionLocation>> future = new CompletableFuture<>();
try {
RegionInfo parsedRegionInfo = Cata... | 3.68 |
AreaShop_AreaShop_getFileManager | /**
* Method to get the FileManager (loads/save regions and can be used to get regions).
* @return The fileManager
*/
public FileManager getFileManager() {
return fileManager;
} | 3.68 |
morf_SqlServerDialect_alterTableChangeColumnStatements | /**
* @see org.alfasoftware.morf.jdbc.SqlDialect#alterTableChangeColumnStatements(org.alfasoftware.morf.metadata.Table, org.alfasoftware.morf.metadata.Column, org.alfasoftware.morf.metadata.Column)
*/
@Override
public Collection<String> alterTableChangeColumnStatements(Table table, final Column oldColumn, Column newC... | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.