name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
hbase_HFileBlock_writeHeaderAndData | /**
* Similar to {@link #writeHeaderAndData(FSDataOutputStream)}, but records the offset of this
* block so that it can be referenced in the next block of the same type.
*/
void writeHeaderAndData(FSDataOutputStream out) throws IOException {
long offset = out.getPos();
if (startOffset != UNSET && offset != start... | 3.68 |
flink_RowDataLocalTimeZoneConverter_getSessionTimeZone | /** Get time zone from the given session config. */
private static ZoneId getSessionTimeZone(ReadableConfig sessionConfig) {
final String zone = sessionConfig.get(TableConfigOptions.LOCAL_TIME_ZONE);
return TableConfigOptions.LOCAL_TIME_ZONE.defaultValue().equals(zone)
? ZoneId.systemDefault()
... | 3.68 |
flink_BinaryStringData_byteAt | /**
* Returns the {@code byte} value at the specified index. An index ranges from {@code 0} to
* {@code binarySection.sizeInBytes - 1}.
*
* @param index the index of the {@code byte} value.
* @return the {@code byte} value at the specified index of this UTF-8 bytes.
* @exception IndexOutOfBoundsException if the {... | 3.68 |
hadoop_ShadedProtobufHelper_ipc | /**
* Evaluate a protobuf call, converting any ServiceException to an IOException.
* @param call invocation to make
* @return the result of the call
* @param <T> type of the result
* @throws IOException any translated protobuf exception
*/
public static <T> T ipc(IpcCall<T> call) throws IOException {
try {
... | 3.68 |
flink_AbstractRowTimeUnboundedPrecedingOver_processElement | /**
* Puts an element from the input stream into state if it is not late. Registers a timer for the
* next watermark.
*
* @param input The input value.
* @param ctx A {@link Context} that allows querying the timestamp of the element and getting
* TimerService for registering timers and querying the time. The ... | 3.68 |
flink_ClusterEntryPointExceptionUtils_tryEnrichClusterEntryPointError | /**
* Tries to enrich the passed exception or its causes with additional information.
*
* <p>This method improves error messages for direct and metaspace {@link OutOfMemoryError}. It
* adds descriptions about possible causes and ways of resolution.
*
* @param root The Throwable of which the cause tree shall be tr... | 3.68 |
hbase_HRegionServer_createRegionServerStatusStub | /**
* Get the current master from ZooKeeper and open the RPC connection to it. To get a fresh
* connection, the current rssStub must be null. Method will block until a master is available.
* You can break from this block by requesting the server stop.
* @param refresh If true then master address will be read from Z... | 3.68 |
flink_HashPartition_spillPartition | /**
* Spills this partition to disk and sets it up such that it continues spilling records that are
* added to it. The spilling process must free at least one buffer, either in the partition's
* record buffers, or in the memory segments for overflow buckets. The partition immediately
* takes back one buffer to use ... | 3.68 |
flink_SlideWithSize_every | /**
* Specifies the window's slide as time or row-count interval.
*
* <p>The slide determines the interval in which windows are started. Hence, sliding windows can
* overlap if the slide is smaller than the size of the window.
*
* <p>For example, you could have windows of size 15 minutes that slide by 3 minutes. ... | 3.68 |
hbase_OrderedFloat64_decodeDouble | /**
* Read a {@code double} value from the buffer {@code src}.
* @param src the {@link PositionedByteRange} to read the {@code double} from
* @return the {@code double} floating-point value with the same bit pattern
*/
public double decodeDouble(PositionedByteRange src) {
return OrderedBytes.decodeFloat64(src);
} | 3.68 |
hadoop_NameValuePair_getName | /**
* Get the name.
* @return The name.
*/
public String getName() {
return name;
} | 3.68 |
hbase_StoreFileWriter_getUniqueFile | /**
* @param dir Directory to create file in.
* @return random filename inside passed <code>dir</code>
*/
public static Path getUniqueFile(final FileSystem fs, final Path dir) throws IOException {
if (!fs.getFileStatus(dir).isDirectory()) {
throw new IOException("Expecting " + dir.toString() + " to be a direct... | 3.68 |
hbase_ProcedureStoreTracker_setDeletedIfModifiedInBoth | /**
* Similar with {@link #setDeletedIfModified(long...)}, but here the {@code procId} are given by
* the {@code tracker}. If a procedure is modified by us, and also by the given {@code tracker},
* then we mark it as deleted.
* @see #setDeletedIfModified(long...)
*/
public void setDeletedIfModifiedInBoth(Procedure... | 3.68 |
cron-utils_CronDefinitionBuilder_matchDayOfWeekAndDayOfMonth | /**
* Sets matchDayOfWeekAndDayOfMonth value to true.
*
* @return this CronDefinitionBuilder instance
*/
public CronDefinitionBuilder matchDayOfWeekAndDayOfMonth() {
matchDayOfWeekAndDayOfMonth = true;
return this;
} | 3.68 |
hadoop_DiskBalancerWorkStatus_parseJson | /**
* Returns a DiskBalancerWorkStatus object from the Json .
* @param json - json String
* @return DiskBalancerWorkStatus
* @throws IOException
*/
public static DiskBalancerWorkStatus parseJson(String json) throws
IOException {
return READER_WORKSTATUS.readValue(json);
} | 3.68 |
morf_AbstractSqlDialectTest_testSelectWithNestedConcatenations | /**
* Tests concatenation in a select with nested concatenations.
*/
@Test
public void testSelectWithNestedConcatenations() {
SelectStatement stmt = new SelectStatement(new ConcatenatedField(new FieldReference("field1"), new ConcatenatedField(
new FieldReference("field2"), new FieldLiteral("XYZ"))).as("test")).... | 3.68 |
graphhopper_LandmarkStorage__getInternalDA | /**
* For testing only
*/
DataAccess _getInternalDA() {
return landmarkWeightDA;
} | 3.68 |
flink_TypeInferenceUtil_createInvalidInputException | /** Returns an exception for invalid input arguments. */
public static ValidationException createInvalidInputException(
TypeInference typeInference, CallContext callContext, ValidationException cause) {
return new ValidationException(
String.format(
"Invalid input arguments. ... | 3.68 |
framework_TreeAction_execute | /**
* Sends message to server that this action has been fired. Messages are
* "standard" Vaadin messages whose value is comma separated pair of
* targetKey (row, treeNod ...) and actions id.
*
* Variablename is always "action".
*
* Actions are always sent immediatedly to server.
*/
@Override
public void execute... | 3.68 |
framework_LayoutManager_setNeedsVerticalLayout | /**
* Marks that a ManagedLayout should be layouted vertically in the next
* layout phase even if none of the elements managed by the layout have been
* resized vertically.
* <p>
* For SimpleManagedLayout which is always layouted in both directions, this
* has the same effect as {@link #setNeedsLayout(ManagedLayo... | 3.68 |
hbase_AvlUtil_readPrev | /**
* Return the predecessor of the current node
* @param node the current node
* @return the predecessor of the current node
*/
public static <TNode extends AvlLinkedNode> TNode readPrev(TNode node) {
return (TNode) node.iterPrev;
} | 3.68 |
open-banking-gateway_ProcessResultEventHandler_handleEvent | /**
* Spring event-bus listener to listen for BPMN process result.
*
* @param result BPMN process message to notify with the subscribers.
*/
@TransactionalEventListener
public void handleEvent(InternalProcessResult result) {
Consumer<InternalProcessResult> consumer;
synchronized (lock) {
InternalPr... | 3.68 |
framework_BasicEventResizeHandler_eventResize | /*
* (non-Javadoc)
*
* @see
* com.vaadin.addon.calendar.ui.CalendarComponentEvents.EventResizeHandler
* #eventResize
* (com.vaadin.addon.calendar.ui.CalendarComponentEvents.EventResize)
*/
@Override
public void eventResize(EventResize event) {
CalendarEvent calendarEvent = event.getCalendarEvent();
if (... | 3.68 |
flink_HiveCatalog_getOrderedFullPartitionValues | /**
* Get a list of ordered partition values by re-arranging them based on the given list of
* partition keys. If the partition value is null, it'll be converted into default partition
* name.
*
* @param partitionSpec a partition spec.
* @param partitionKeys a list of partition keys.
* @param tablePath path of t... | 3.68 |
framework_VaadinService_setClassLoader | /**
* Sets the class loader to use for loading classes loaded by name, e.g.
* custom UI classes. Invokers of this method should be careful to not break
* any existing class loader hierarchy, e.g. by ensuring that a class loader
* set for this service delegates to the previously set class loader if the
* class is n... | 3.68 |
hadoop_AbfsRestOperation_createNewTracingContext | /**
* Creates a new Tracing context before entering the retry loop of a rest operation.
* This will ensure all rest operations have unique
* tracing context that will be used for all the retries.
* @param tracingContext original tracingContext.
* @return tracingContext new tracingContext object created from origin... | 3.68 |
hmily_XidImpl_newBranchId | /**
* New branch id x id.
*
* @return the x id
*/
public XidImpl newBranchId() {
return new XidImpl(this);
} | 3.68 |
hudi_CompactionUtil_rollbackEarliestCompaction | /**
* Rolls back the earliest compaction if there exists.
*
* <p>Makes the strategy not that radical: firstly check whether there exists inflight compaction instants,
* rolls back the first inflight instant only if it has timed out. That means, if there are
* multiple timed out instants on the timeline, we only ro... | 3.68 |
framework_TooltipOnRequiredIndicator_setup | /*
* (non-Javadoc)
*
* @see com.vaadin.tests.components.AbstractTestUI#setup(com.vaadin.server.
* VaadinRequest)
*/
@Override
protected void setup(VaadinRequest request) {
VerticalLayout layout = new VerticalLayout();
TextField inVertical = new TextField();
inVertical.setRequired(true);
inVertical... | 3.68 |
pulsar_ResourceUnitRanking_removeLoadedServiceUnit | /**
* Remove a service unit from the loaded bundle list.
*/
public void removeLoadedServiceUnit(String suName, ResourceQuota quota) {
if (this.loadedBundles.remove(suName)) {
this.allocatedQuota.substract(quota);
estimateLoadPercentage();
}
} | 3.68 |
framework_VaadinSession_getBrowser | /**
* Get the web browser associated with this session.
*
* @return the web browser object
*
* @deprecated As of 7.0, use {@link Page#getWebBrowser()} instead.
*/
@Deprecated
public WebBrowser getBrowser() {
assert hasLock();
return browser;
} | 3.68 |
hadoop_CommitContext_getOuterSubmitter | /**
* Return a submitter.
* If created with 0 threads, this returns null so
* TaskPool knows to run it in the current thread.
* @return a submitter or null
*/
public synchronized TaskPool.Submitter getOuterSubmitter() {
return outerSubmitter;
} | 3.68 |
hadoop_ManifestSuccessData_getCommitter | /**
* @return committer name.
*/
public String getCommitter() {
return committer;
} | 3.68 |
hadoop_Validate_checkPositiveInteger | /**
* Validates that the given integer argument is not zero or negative.
* @param value the argument value to validate
* @param argName the name of the argument being validated.
*/
public static void checkPositiveInteger(long value, String argName) {
checkArgument(value > 0, "'%s' must be a positive integer.", ar... | 3.68 |
pulsar_MetadataStore_getMetadataCache | /**
* Create a metadata cache that uses a particular serde object.
*
* @param <T>
* @param serde
* the custom serialization/deserialization object
* @return the metadata cache object
*/
default <T> MetadataCache<T> getMetadataCache(MetadataSerde<T> serde) {
return getMetadataCache(serde, getDefaul... | 3.68 |
framework_AbstractField_setReadOnly | /**
* {@inheritDoc}
* <p>
* The server ignores (potentially forged) value change requests from the
* client to fields that are read-only. Programmatically changing the field
* value via {@link #setValue(T)} is still possible.
* <p>
* The read-only mode is distinct from the
* {@linkplain Component#setEnabled(boo... | 3.68 |
flink_Schema_fromResolvedSchema | /** Adopts all members from the given resolved schema. */
public Builder fromResolvedSchema(ResolvedSchema resolvedSchema) {
addResolvedColumns(resolvedSchema.getColumns());
addResolvedWatermarkSpec(resolvedSchema.getWatermarkSpecs());
resolvedSchema.getPrimaryKey().ifPresent(this::addResolvedConstraint);
... | 3.68 |
flink_EncodingUtils_repeat | /**
* Returns padding using the specified delimiter repeated to a given length.
*
* <pre>
* StringUtils.repeat('e', 0) = ""
* StringUtils.repeat('e', 3) = "eee"
* StringUtils.repeat('e', -2) = ""
* </pre>
*
* <p>Note: this method doesn't not support padding with <a
* href="http://www.unicode.org/glossary/#s... | 3.68 |
hadoop_Check_validIdentifier | /**
* Verifies a value is a valid identifier,
* <code>[a-zA-Z_][a-zA-Z0-9_\-]*</code>, up to a maximum length.
*
* @param value string to check if it is a valid identifier.
* @param maxLen maximun length.
* @param name the name to use in the exception message.
*
* @return the value.
*
* @throws IllegalArgumen... | 3.68 |
hbase_AsyncAdmin_balance | /**
* Invoke the balancer. Will run the balancer and if regions to move, it will go ahead and do the
* reassignments. If there is region in transition, force parameter of true would still run
* balancer. Can *not* run for other reasons. Check logs.
* @param forcible whether we should force balance even if there is ... | 3.68 |
shardingsphere-elasticjob_GuaranteeService_executeInLeaderForLastCompleted | /**
* Invoke doAfterJobExecutedAtLastCompleted method once after last completed.
*
* @param listener AbstractDistributeOnceElasticJobListener instance
* @param shardingContexts sharding contexts
*/
public void executeInLeaderForLastCompleted(final AbstractDistributeOnceElasticJobListener listener,
... | 3.68 |
flink_PythonEnvUtils_getLibFiles | /**
* Gets pyflink dependent libs in specified directory.
*
* @param libDir The lib directory
*/
private static List<java.nio.file.Path> getLibFiles(String libDir) {
final List<java.nio.file.Path> libFiles = new ArrayList<>();
SimpleFileVisitor<java.nio.file.Path> finder =
new SimpleFileVisitor<... | 3.68 |
querydsl_StringExpression_startsWithIgnoreCase | /**
* Create a {@code this.startsWithIgnoreCase(str)} expression
*
* @param str string
* @return this.startsWithIgnoreCase(str)
*/
public BooleanExpression startsWithIgnoreCase(String str) {
return startsWithIgnoreCase(ConstantImpl.create(str));
} | 3.68 |
framework_DataCommunicator_removeDataGenerator | /**
* Removes a data generator from this data communicator. If there is no such
* data generator, does nothing.
*
* @param generator
* the data generator to remove, not null
*/
public void removeDataGenerator(DataGenerator<T> generator) {
Objects.requireNonNull(generator, "generator cannot be null"... | 3.68 |
querydsl_JDOExpressions_selectFrom | /**
* Create a new detached {@link JDOQuery} instance with the given projection
*
* @param expr projection and source
* @param <T>
* @return select(expr).from(expr)
*/
public static <T> JDOQuery<T> selectFrom(EntityPath<T> expr) {
return select(expr).from(expr);
} | 3.68 |
hadoop_StoreContext_getOwner | /**
* Get the owner of the filesystem.
* @return the user who created this filesystem.
*/
public UserGroupInformation getOwner() {
return owner;
} | 3.68 |
framework_Form_setReadOnly | /**
* Sets the component's to read-only mode to the specified state.
*
* @see Component#setReadOnly(boolean)
*/
@Override
public void setReadOnly(boolean readOnly) {
super.setReadOnly(readOnly);
for (final Object id : propertyIds) {
fields.get(id).setReadOnly(readOnly);
}
} | 3.68 |
hbase_MasterObserver_preTruncateTable | /**
* Called before {@link org.apache.hadoop.hbase.master.HMaster} truncates a table. Called as part
* of truncate table RPC call.
* @param ctx the environment to interact with the framework and master
* @param tableName the name of the table
*/
default void preTruncateTable(final ObserverContext<MasterCopro... | 3.68 |
hadoop_InstrumentedReadLock_startLockTiming | /**
* Starts timing for the instrumented read lock.
* It records the time to ThreadLocal.
*/
@Override
protected void startLockTiming() {
if (readWriteLock.getReadHoldCount() == 1) {
readLockHeldTimeStamp.set(getTimer().monotonicNow());
}
} | 3.68 |
flink_TypeExtractionUtils_checkAndExtractLambda | /**
* Checks if the given function has been implemented using a Java 8 lambda. If yes, a
* LambdaExecutable is returned describing the method/constructor. Otherwise null.
*
* @throws TypeExtractionException lambda extraction is pretty hacky, it might fail for unknown
* JVM issues.
*/
public static LambdaExecu... | 3.68 |
flink_TemporalTableJoinUtil_isEventTimeTemporalJoin | /**
* Check if the given join condition is an initial temporal join condition or a rewrote join
* condition on event time.
*/
public static boolean isEventTimeTemporalJoin(@Nonnull RexNode joinCondition) {
RexVisitor<Void> temporalConditionFinder =
new RexVisitorImpl<Void>(true) {
@Ov... | 3.68 |
hudi_QuickstartUtils_generateUniqueUpdates | /**
* Generates new updates, one for each of the keys above
* list
*
* @param n Number of updates (must be no more than number of existing keys)
* @return list of hoodie record updates
*/
public List<HoodieRecord> generateUniqueUpdates(Integer n) {
if (numExistingKeys < n) {
throw new HoodieException("Data ... | 3.68 |
hadoop_LocatedFileStatus_setBlockLocations | /**
* Hook for subclasses to lazily set block locations. The {@link #locations}
* field should be null before this is called.
* @param locations Block locations for this instance.
*/
protected void setBlockLocations(BlockLocation[] locations) {
this.locations = locations;
} | 3.68 |
framework_VComboBox_onKeyUp | /**
* Triggered when a key was depressed.
*
* @param event
* The KeyUpEvent of the key depressed
*/
@Override
public void onKeyUp(KeyUpEvent event) {
if (enableDebug) {
debug("VComboBox: onKeyUp(" + event.getNativeKeyCode() + ")");
}
if (enabled && !readonly) {
switch (event.... | 3.68 |
hadoop_MappingRuleResult_getQueue | /**
* This method returns the result queue. Currently only makes sense when
* result == PLACE.
* @return the queue this result is about
*/
public String getQueue() {
return queue;
} | 3.68 |
hadoop_OBSFileSystem_getCannedACL | /**
* Get the bucket acl of user setting.
*
* @return bucket acl {@link AccessControlList}
*/
AccessControlList getCannedACL() {
return cannedACL;
} | 3.68 |
flink_SourceCoordinatorSerdeUtils_readAndVerifyCoordinatorSerdeVersion | /** Read and verify the serde version. */
static int readAndVerifyCoordinatorSerdeVersion(DataInputStream in) throws IOException {
int version = in.readInt();
if (version > CURRENT_VERSION) {
throw new IOException("Unsupported source coordinator serde version " + version);
}
return version;
} | 3.68 |
morf_BaseDataSetReader_availableStreamNames | /**
* @see org.alfasoftware.morf.xml.XmlStreamProvider.XmlInputStreamProvider#availableStreamNames()
*/
@Override
public Collection<String> availableStreamNames() {
return Lists.newArrayList(tableNames); // return a copy, as we are about to clear the list
} | 3.68 |
hadoop_DefaultOBSClientFactory_initConnectionSettings | /**
* Initializes all OBS SDK settings related to connection management.
*
* @param conf Hadoop configuration
* @param obsConf OBS SDK configuration
*/
@SuppressWarnings("deprecation")
private static void initConnectionSettings(final Configuration conf,
final ExtObsConfiguration obsConf) {
obsConf.setMax... | 3.68 |
hadoop_ManifestSuccessData_getTimestamp | /** @return timestamp of creation. */
public long getTimestamp() {
return timestamp;
} | 3.68 |
dubbo_HttpHeaderUtil_isRestAttachHeader | /**
* for judge rest header or rest attachment
*
* @param header
* @return
*/
public static boolean isRestAttachHeader(String header) {
if (StringUtils.isEmpty(header) || !header.startsWith(RestHeaderEnum.REST_HEADER_PREFIX.getHeader())) {
return false;
}
return true;
} | 3.68 |
hudi_HDFSParquetImporter_load | /**
* Imports records to Hoodie table.
*
* @param client Hoodie Client
* @param instantTime Instant Time
* @param hoodieRecords Hoodie Records
* @param <T> Type
*/
protected <T extends HoodieRecordPayload> JavaRDD<WriteStatus> load(SparkRDDWriteClient<T> client, String instantTime,
... | 3.68 |
hbase_FilterBase_hasFilterRow | /**
* Filters that never filter by modifying the returned List of Cells can inherit this
* implementation that does nothing. {@inheritDoc}
*/
@Override
public boolean hasFilterRow() {
return false;
} | 3.68 |
hbase_ZKClusterId_getUUIDForCluster | /**
* Get the UUID for the provided ZK watcher. Doesn't handle any ZK exceptions
* @param zkw watcher connected to an ensemble
* @return the UUID read from zookeeper
* @throws KeeperException if a ZooKeeper operation fails
*/
public static UUID getUUIDForCluster(ZKWatcher zkw) throws KeeperException {
String uui... | 3.68 |
flink_CollectSink_open | /**
* Initialize the connection with the Socket in the server.
*
* @param openContext the context.
*/
@Override
public void open(OpenContext openContext) throws Exception {
try {
client = new Socket(hostIp, port);
outputStream = client.getOutputStream();
streamWriter = new DataOutputView... | 3.68 |
hibernate-validator_ConstraintHelper_run | /**
* Runs the given privileged action, using a privileged block if required.
* <p>
* <b>NOTE:</b> This must never be changed into a publicly available method to avoid execution of arbitrary
* privileged actions within HV's protection domain.
*/
@IgnoreForbiddenApisErrors(reason = "SecurityManager is deprecated in... | 3.68 |
hbase_WALEdit_isMetaEditFamily | /**
* Replaying WALs can read Cell-at-a-time so need this method in those cases.
*/
public static boolean isMetaEditFamily(Cell cell) {
return CellUtil.matchingFamily(cell, METAFAMILY);
} | 3.68 |
aws-saas-boost_UpdateAction_addTarget | /**
* Adds a new target to this <code>UpdateAction</code>.
*
* e.g. if this is a <code>SERVICE</code>, the list of targets may be
* <code>[ "onboarding", "tenant" ]</code>.
*
* @param target the target to add
*/
public void addTarget(String target) {
targets.add(target);
} | 3.68 |
pulsar_ProxyConnection_handleLookup | /**
* handles discovery request from client ands sends next active broker address.
*/
@Override
protected void handleLookup(CommandLookupTopic lookup) {
checkArgument(state == State.ProxyLookupRequests);
lookupProxyHandler.handleLookup(lookup);
} | 3.68 |
hbase_AsyncAdmin_getCompactionState | /**
* Get the current compaction state of a table. It could be in a major compaction, a minor
* compaction, both, or none.
* @param tableName table to examine
* @return the current compaction state wrapped by a {@link CompletableFuture}
*/
default CompletableFuture<CompactionState> getCompactionState(TableName tab... | 3.68 |
flink_DataSource_withParameters | /**
* Pass a configuration to the InputFormat.
*
* @param parameters Configuration parameters
*/
public DataSource<OUT> withParameters(Configuration parameters) {
this.parameters = parameters;
return this;
} | 3.68 |
hadoop_OBSObjectBucketUtils_getObjectMetadata | /**
* Request object metadata; increments counters in the process.
*
* @param owner OBS File System instance
* @param key key
* @return the metadata
*/
protected static ObjectMetadata getObjectMetadata(final OBSFileSystem owner,
final String key) {
GetObjectMetadataRequest request = new GetObjectMetadataR... | 3.68 |
framework_SelectorPath_getLegacyLocatorQuery | /**
* Generates a legacy locator for SelectorPath.
*
* @return String containing Java code for element search and assignment
*/
private String getLegacyLocatorQuery() {
String name;
if (!path.isEmpty()) {
String[] frags = path.split("/");
name = getComponentName(frags[frags.length - 1]).subs... | 3.68 |
hudi_OptionsResolver_allOptions | /**
* Returns all the config options with the given class {@code clazz}.
*/
public static List<ConfigOption<?>> allOptions(Class<?> clazz) {
Field[] declaredFields = clazz.getDeclaredFields();
List<ConfigOption<?>> options = new ArrayList<>();
for (Field field : declaredFields) {
if (java.lang.reflect.Modif... | 3.68 |
hadoop_FedBalanceContext_setUseMountReadOnly | /**
* Use mount point readonly to disable write.
* @param value true if disabling write by setting mount point readonly.
* @return the builder.
*/
public Builder setUseMountReadOnly(boolean value) {
this.useMountReadOnly = value;
return this;
} | 3.68 |
framework_FlyweightRow_unattached | /**
* Creates a new iterator of unattached flyweight cells. A cell is
* unattached if it does not have a corresponding
* {@link FlyweightCell#getElement() DOM element} attached to the row
* element.
*
* @param cells
* the collection of cells to iterate
*/
public static CellIterator unattached(
... | 3.68 |
hudi_HoodieLogBlock_tryReadContent | /**
* Read or Skip block content of a log block in the log file. Depends on lazy reading enabled in
* {@link HoodieMergedLogRecordScanner}
*/
public static Option<byte[]> tryReadContent(FSDataInputStream inputStream, Integer contentLength, boolean readLazily)
throws IOException {
if (readLazily) {
// Seek ... | 3.68 |
hadoop_JobTokenIdentifier_readFields | /** {@inheritDoc} */
@Override
public void readFields(DataInput in) throws IOException {
jobid.readFields(in);
} | 3.68 |
hbase_KeyLocker_acquireLock | /**
* Return a lock for the given key. The lock is already locked.
*/
public ReentrantLock acquireLock(K key) {
if (key == null) throw new IllegalArgumentException("key must not be null");
lockPool.purge();
ReentrantLock lock = lockPool.get(key);
lock.lock();
return lock;
} | 3.68 |
pulsar_RocksdbMetadataStore_serialize | /**
* Note: we can only add new fields, but not change or remove existing fields.
*/
public byte[] serialize() {
byte[] result = new byte[HEADER_SIZE + data.length];
ByteBuffer buffer = ByteBuffer.wrap(result);
buffer.putInt(HEADER_SIZE);
buffer.putInt(FORMAT_VERSION_V1);
buffer.putLong(version);
... | 3.68 |
morf_Function_yyyymmddToDate | /**
* Helper method to create an instance of the "YYYYMMDDToDate" SQL function.
* {@code expression} must result in a string.
*
* @see Cast
* @param expression the expression to evaluate
* @return an instance of the YYYYMMDDToDate function
*/
public static Function yyyymmddToDate(AliasedField expression) {
ret... | 3.68 |
hudi_InternalSchemaChangeApplier_applyColumnCommentChange | /**
* Update col comment for hudi table.
*
* @param colName col name to be changed. if we want to change col from a nested filed, the fullName should be specify
* @param doc .
*/
public InternalSchema applyColumnCommentChange(String colName, String doc) {
TableChanges.ColumnUpdateChange updateChange = TableChang... | 3.68 |
flink_CrossOperator_projectTuple20 | /**
* Projects a pair of crossed elements to a {@link Tuple} with the previously selected
* fields.
*
* @return The projected data set.
* @see Tuple
* @see DataSet
*/
public <
T0,
T1,
T2,
T3,
T4,
T5,
T... | 3.68 |
hudi_HoodieSparkQuickstart_incrementalQuery | /**
* Hudi also provides capability to obtain a stream of records that changed since given commit timestamp.
* This can be achieved using Hudi’s incremental view and providing a begin time from which changes need to be streamed.
* We do not need to specify endTime, if we want all changes after the given commit (as i... | 3.68 |
dubbo_ApplicationModel_setServiceRepository | /**
* @deprecated only for ut
*/
@Deprecated
public void setServiceRepository(ServiceRepository serviceRepository) {
this.serviceRepository = serviceRepository;
} | 3.68 |
hadoop_TFile_hashCode | /**
* @see java.lang.Object#hashCode()
*/
@Override
public int hashCode() {
final int prime = 31;
int result = prime + blockIndex;
result = (int) (prime * result + recordIndex);
return result;
} | 3.68 |
hbase_KeyValue_shallowCopy | /**
* Creates a shallow copy of this KeyValue, reusing the data byte buffer.
* http://en.wikipedia.org/wiki/Object_copy
* @return Shallow copy of this KeyValue
*/
public KeyValue shallowCopy() {
KeyValue shallowCopy = new KeyValue(this.bytes, this.offset, this.length);
shallowCopy.setSequenceId(this.seqId);
r... | 3.68 |
hbase_WALKeyImpl_getClusterIds | /** Returns the set of cluster Ids that have consumed the change */
public List<UUID> getClusterIds() {
return clusterIds;
} | 3.68 |
hmily_HmilyHashLoadBalance_refresh | /**
* Refresh local invoker.
*
* @param invokers invokers
*/
@Override
public void refresh(final Collection<Invoker<T>> invokers) {
LOGGER.info("{} try to refresh RoundRobinLoadBalance's invoker cache, size= {} ", config.getSimpleObjectName(), CollectionUtils.isEmpty(invokers) ? 0 : invokers.size());
if (in... | 3.68 |
hbase_MasterObserver_preListDecommissionedRegionServers | /**
* Called before list decommissioned region servers.
*/
default void preListDecommissionedRegionServers(ObserverContext<MasterCoprocessorEnvironment> ctx)
throws IOException {
} | 3.68 |
hbase_RemoteProcedureDispatcher_addOperationToNode | /**
* Add a remote rpc.
* @param key the node identifier
*/
public void addOperationToNode(final TRemote key, RemoteProcedure rp)
throws NullTargetServerDispatchException, NoServerDispatchException, NoNodeDispatchException {
if (key == null) {
throw new NullTargetServerDispatchException(rp.toString());
}
... | 3.68 |
flink_TaskManagerServicesConfiguration_fromConfiguration | /**
* Utility method to extract TaskManager config parameters from the configuration and to sanity
* check them.
*
* @param configuration The configuration.
* @param resourceID resource ID of the task manager
* @param externalAddress identifying the IP address under which the TaskManager will be
* accessible... | 3.68 |
framework_VFilterSelect_updateStyleNames | /**
* Updates style names in suggestion popup to help theme building.
*
* @param uidl
* UIDL for the whole combo box
* @param componentState
* shared state of the combo box
*/
public void updateStyleNames(UIDL uidl,
AbstractComponentState componentState) {
debug("VFS.SP: updateS... | 3.68 |
hadoop_MultipartUploaderBuilderImpl_checksumOpt | /**
* Set checksum opt.
*/
@Override
public B checksumOpt(@Nonnull final ChecksumOpt chksumOpt) {
checkNotNull(chksumOpt);
checksumOpt = chksumOpt;
return getThisBuilder();
} | 3.68 |
hbase_QuotaTableUtil_createGetNamespaceSnapshotSize | /**
* Creates a {@code Get} to fetch the namespace's total snapshot size.
*/
static Get createGetNamespaceSnapshotSize(String namespace) {
Get g = new Get(getNamespaceRowKey(namespace));
g.addColumn(QUOTA_FAMILY_USAGE, QUOTA_SNAPSHOT_SIZE_QUALIFIER);
return g;
} | 3.68 |
framework_TreeTable_isCollapsed | /**
* Checks if Item with given identifier is collapsed in the UI.
*
* <p>
*
* @param itemId
* the identifier of the checked Item
* @return true if the Item with given id is collapsed
* @see Collapsible#isCollapsed(Object)
*/
public boolean isCollapsed(Object itemId) {
return !getContainerStrate... | 3.68 |
hibernate-validator_ConstraintDescriptorImpl_run | /**
* Runs the given privileged action, using a privileged block if required.
* <p>
* <b>NOTE:</b> This must never be changed into a publicly available method to avoid execution of arbitrary
* privileged actions within HV's protection domain.
*/
@IgnoreForbiddenApisErrors(reason = "SecurityManager is deprecated in... | 3.68 |
flink_StreamExchangeModeUtils_getGlobalStreamExchangeMode | /**
* The {@link GlobalStreamExchangeMode} should be determined by the {@link StreamGraphGenerator}
* in the future.
*/
@Deprecated
static Optional<GlobalStreamExchangeMode> getGlobalStreamExchangeMode(ReadableConfig config) {
return config.getOptional(ExecutionConfigOptions.TABLE_EXEC_SHUFFLE_MODE)
... | 3.68 |
morf_OracleDialect_createSequenceStartingFromExistingData | /**
* Returns a SQL statement to create a sequence for a table's autonumber column, where
* the sequence should start from the greater of either the autonumber column's start value
* or the maximum value for that column existing in the table.
*
* @param table Table for which the sequence should be created.
* @par... | 3.68 |
hbase_MiniHBaseCluster_startRegionServer | /**
* Starts a region server thread running
* @return New RegionServerThread
*/
public JVMClusterUtil.RegionServerThread startRegionServer() throws IOException {
final Configuration newConf = HBaseConfiguration.create(conf);
return startRegionServer(newConf);
} | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.