name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
druid_IPRange_getExtendedNetworkPrefix | /**
* Return the extended extended network prefix.
*
* @return Return the extended network prefix.
*/
public final int getExtendedNetworkPrefix() {
return extendedNetworkPrefix;
} | 3.68 |
framework_HierarchicalContainer_getChildren | /*
* Gets the IDs of the children of the specified Item. Don't add a JavaDoc
* comment here, we use the default documentation from implemented
* interface.
*/
@Override
public Collection<?> getChildren(Object itemId) {
LinkedList<Object> c;
if (filteredChildren != null) {
c = filteredChildren.get(i... | 3.68 |
pulsar_LeastResourceUsageWithWeight_selectBroker | /**
* Find a suitable broker to assign the given bundle to.
* This method is not thread safety.
*
* @param candidates The candidates for which the bundle may be assigned.
* @param bundleToAssign The data for the bundle to assign.
* @param loadData The load data from the leader broker.
* @param conf ... | 3.68 |
framework_Page_getPage | /**
* Gets the page in which the uri has changed.
*
* @return the page in which the uri has changed
*/
public Page getPage() {
return (Page) getSource();
} | 3.68 |
flink_CheckpointStorageWorkerView_toFileMergingStorage | /**
* Return {@link org.apache.flink.runtime.state.filesystem.FsMergingCheckpointStorageAccess} if
* file merging is enabled Otherwise, return itself. File merging is supported by subclasses of
* {@link org.apache.flink.runtime.state.filesystem.AbstractFsCheckpointStorageAccess}.
*/
default CheckpointStorageWorkerV... | 3.68 |
morf_CreateDeployedViews_execute | /**
* @see org.alfasoftware.morf.upgrade.UpgradeStep#execute(org.alfasoftware.morf.upgrade.SchemaEditor, org.alfasoftware.morf.upgrade.DataEditor)
*/
@Override
public void execute(SchemaEditor schema, DataEditor data) {
schema.addTable(table("DeployedViews").
columns(
colum... | 3.68 |
hadoop_LoggingAuditor_serviceInit | /**
* Service init, look for jobID and attach as an attribute in log entries.
* This is where the warning span is created, so the relevant attributes
* (and filtering options) are applied.
* @param conf configuration
* @throws Exception failure
*/
@Override
protected void serviceInit(final Configuration conf) thr... | 3.68 |
framework_CalendarTargetDetails_getDropTime | /**
* @return the date where the drop happened
*/
public Date getDropTime() {
if (hasDropTime) {
return (Date) getData("dropTime");
} else {
return (Date) getData("dropDay");
}
} | 3.68 |
hadoop_BlockStorageMovementNeeded_markScanCompleted | /**
* Mark directory scan is completed.
*/
public synchronized void markScanCompleted() {
this.fullyScanned = true;
} | 3.68 |
graphhopper_BikeCommonPriorityParser_handlePriority | /**
* In this method we prefer cycleways or roads with designated bike access and avoid big roads
* or roads with trams or pedestrian.
*
* @return new priority based on priorityFromRelation and on the tags in ReaderWay.
*/
int handlePriority(ReaderWay way, double wayTypeSpeed, Integer priorityFromRelation) {
T... | 3.68 |
flink_MiniCluster_terminateTaskManager | /**
* Terminates a TaskManager with the given index.
*
* <p>See {@link #startTaskManager()} to understand how TaskManagers are indexed. This method
* terminates a TaskManager with a given index but it does not clear the index. The index stays
* occupied for the lifetime of the MiniCluster and its TaskManager stays... | 3.68 |
hbase_SnapshotReferenceUtil_verifyStoreFile | /**
* Verify the validity of the snapshot store file
* @param conf The current {@link Configuration} instance.
* @param fs {@link FileSystem}
* @param snapshotDir {@link Path} to the Snapshot directory of the snapshot to verify
* @param snapshot the {@link SnapshotDescription} of the snapshot to... | 3.68 |
hbase_HFileReaderImpl_midKey | /**
* @return Midkey for this file. We work with block boundaries only so returned midkey is an
* approximation only.
*/
@Override
public Optional<Cell> midKey() throws IOException {
return Optional.ofNullable(dataBlockIndexReader.midkey(this));
} | 3.68 |
framework_GenericFontIcon_hashCode | /*
* (non-Javadoc)
*
* @see java.lang.Object#hashCode()
*/
@Override
public int hashCode() {
final int prime = 31;
int result = 1;
result = prime * result + codePoint;
result = prime * result
+ ((fontFamily == null) ? 0 : fontFamily.hashCode());
return result;
} | 3.68 |
hadoop_Utils_compareTo | /**
* Compare this version with another version.
*/
@Override
public int compareTo(Version that) {
if (major != that.major) {
return major - that.major;
}
return minor - that.minor;
} | 3.68 |
flink_TimestampsAndWatermarksTransformation_getInputType | /** Returns the {@code TypeInformation} for the elements of the input. */
public TypeInformation<IN> getInputType() {
return input.getOutputType();
} | 3.68 |
hadoop_QueuePriorityContainerCandidateSelector_isQueueSatisfied | /**
* Do we allow the demanding queue preempt resource from other queues?
* A satisfied queue is not allowed to preempt resource from other queues.
* @param demandingQueue
* @return allowed/not
*/
private boolean isQueueSatisfied(String demandingQueue,
String partition) {
TempQueuePerPartition tq = preemptio... | 3.68 |
hbase_ServerMetrics_getVersion | /** Returns the string type version of a regionserver. */
default String getVersion() {
return "0.0.0";
} | 3.68 |
hbase_FSTableDescriptors_getTableDir | /**
* Return the table directory in HDFS
*/
private Path getTableDir(TableName tableName) {
return CommonFSUtils.getTableDir(rootdir, tableName);
} | 3.68 |
framework_VLoadingIndicator_setSecondDelay | /**
* Sets the delay (in ms) which must pass before the loading indicator moves
* to its "second" state.
*
* @param secondDelay
* The delay (in ms) until the loading indicator moves into its
* "second" state. Counted from when {@link #trigger()} is
* called.
*/
public void setSe... | 3.68 |
hadoop_AMRMProxyService_stopApplication | /**
* Shuts down the request processing pipeline for the specified application
* attempt id.
*
* @param applicationId application id
*/
protected void stopApplication(ApplicationId applicationId) {
this.metrics.incrRequestCount();
Preconditions.checkArgument(applicationId != null, "applicationId is null");
R... | 3.68 |
hadoop_ActiveAuditManagerS3A_onExecutionFailure | /**
* Forward to the inner span.
* {@inheritDoc}
*/
@Override
public void onExecutionFailure(Context.FailedExecution context,
ExecutionAttributes executionAttributes) {
span.onExecutionFailure(context, executionAttributes);
} | 3.68 |
hbase_Get_setTimeRange | /**
* Get versions of columns only within the specified timestamp range, [minStamp, maxStamp).
* @param minStamp minimum timestamp value, inclusive
* @param maxStamp maximum timestamp value, exclusive
* @return this for invocation chaining
*/
public Get setTimeRange(long minStamp, long maxStamp) throws IOException... | 3.68 |
flink_DynamicSourceUtils_createRequiredMetadataColumns | /**
* Returns a list of required metadata columns. Ordered by the iteration order of {@link
* SupportsReadingMetadata#listReadableMetadata()}.
*
* <p>This method assumes that source and schema have been validated via {@link
* #prepareDynamicSource(String, ResolvedCatalogTable, DynamicTableSource, boolean,
* Reada... | 3.68 |
AreaShop_Utils_getDurationFromSecondsOrString | /**
* Get setting from config that could be only a number indicating seconds.
* or a string indicating a duration string
* @param path Path of the setting to read
* @return milliseconds that the setting indicates
*/
public static long getDurationFromSecondsOrString(String path) {
if(config.isLong(path) || config.... | 3.68 |
framework_VCalendarAction_getActionStartDate | /**
* Get the date and time when the action starts.
*
* @return
*/
public Date getActionStartDate() {
return actionStartDate;
} | 3.68 |
open-banking-gateway_QueryHeadersMapperTemplate_forExecution | /**
* Converts context object into object that can be used for ASPSP API call.
* @param context Context to convert
* @return Object that can be used with {@code Xs2aAdapter} to perform ASPSP API calls
*/
public ValidatedQueryHeaders<Q, H> forExecution(C context) {
return new ValidatedQueryHeaders<>(
... | 3.68 |
hadoop_FindOptions_getMinDepth | /**
* Returns the minimum depth for applying expressions.
*
* @return min depth
*/
public int getMinDepth() {
return this.minDepth;
} | 3.68 |
hbase_HFileBlockIndex_getNonRootIndexedKey | /**
* The indexed key at the ith position in the nonRootIndex. The position starts at 0.
* @param i the ith position
* @return The indexed key at the ith position in the nonRootIndex.
*/
static byte[] getNonRootIndexedKey(ByteBuff nonRootIndex, int i) {
int numEntries = nonRootIndex.getInt(0);
if (i < 0 || i >=... | 3.68 |
AreaShop_SetteleportCommand_canUse | /**
* Check if a person can set the teleport location of the region.
* @param person The person to check
* @param region The region to check for
* @return true if the person can set the teleport location, otherwise false
*/
public static boolean canUse(CommandSender person, GeneralRegion region) {
if(!(person ins... | 3.68 |
flink_CoGroupedStreams_equalTo | /**
* Specifies a {@link KeySelector} for elements from the second input with explicit type
* information for the key type.
*
* @param keySelector The KeySelector to be used for extracting the key for partitioning.
* @param keyType The type information describing the key type.
*/
public EqualTo equalTo(KeySelecto... | 3.68 |
hadoop_PeriodicService_getErrorCount | /**
* Get how many times we failed to run the periodic service.
*
* @return Times we failed to run the periodic service.
*/
protected long getErrorCount() {
return this.errorCount;
} | 3.68 |
hbase_StoreScanner_getEstimatedNumberOfKvsScanned | /** Returns The estimated number of KVs seen by this scanner (includes some skipped KVs). */
public long getEstimatedNumberOfKvsScanned() {
return this.kvsScanned;
} | 3.68 |
framework_TableScroll_initProperties | // set up the properties (columns)
private void initProperties(Table table) {
for (int i = 0; i < COLUMNS; i++) {
table.addContainerProperty("property" + i, String.class,
"some value");
}
} | 3.68 |
hbase_Scan_hasFilter | /** Returns true is a filter has been specified, false if not */
public boolean hasFilter() {
return filter != null;
} | 3.68 |
framework_AbstractRemoteDataSource_setRow | /**
* A method for the data source to update the row data.
*
* @param row
* the updated row object
*/
public void setRow(final T row) {
this.row = row;
assert getRowKey(row).equals(key) : "The old key does not "
+ "equal the new key for the given row (old: " + key
+ ", ne... | 3.68 |
flink_BlobServer_deletePermanent | /**
* Delete the uploaded data with the given {@link JobID} and {@link PermanentBlobKey}.
*
* @param jobId ID of the job this blob belongs to
* @param key the key of this blob
*/
@Override
public boolean deletePermanent(JobID jobId, PermanentBlobKey key) {
return deleteInternal(jobId, key);
} | 3.68 |
MagicPlugin_MagicController_initialize | /*
* Saving and loading
*/
public void initialize() {
warpController = new WarpController(this);
kitController = new KitController(this);
crafting = new CraftingController(this);
mobs = new MobController(this);
items = new ItemController(this);
enchanting = new EnchantingController(this);
... | 3.68 |
flink_StreamTaskActionExecutor_synchronizedExecutor | /** Returns an ExecutionDecorator that synchronizes each invocation on a given object. */
static SynchronizedStreamTaskActionExecutor synchronizedExecutor(Object mutex) {
return new SynchronizedStreamTaskActionExecutor(mutex);
} | 3.68 |
hbase_HRegionServer_getMasterAddressTracker | /** Returns Master address tracker instance. */
public MasterAddressTracker getMasterAddressTracker() {
return this.masterAddressTracker;
} | 3.68 |
Activiti_ReflectUtil_getSetter | /**
* Returns the setter-method for the given field name or null if no setter exists.
*/
public static Method getSetter(String fieldName, Class<?> clazz, Class<?> fieldType) {
String setterName = "set" + Character.toTitleCase(fieldName.charAt(0)) + fieldName.substring(1, fieldName.length());
try {
// Using ge... | 3.68 |
hbase_FamilyFilter_toByteArray | /** Returns The filter serialized using pb */
@Override
public byte[] toByteArray() {
FilterProtos.FamilyFilter.Builder builder = FilterProtos.FamilyFilter.newBuilder();
builder.setCompareFilter(super.convert());
return builder.build().toByteArray();
} | 3.68 |
hbase_BloomFilterMetrics_getRequestsCount | /** Returns Current value for bloom requests count */
public long getRequestsCount() {
return requests.sum();
} | 3.68 |
hbase_ClientModeStrategy_aggregateRecordsAndAddDistinct | /**
* Aggregate the records and count the unique values for the given distinctField
* @param records records to be processed
* @param groupBy Field on which group by needs to be done
* @param distinctField Field whose unique values needs to be counted
* @param uniqueCountAssigne... | 3.68 |
flink_JobGraph_getSavepointRestoreSettings | /**
* Returns the configured savepoint restore setting.
*
* @return The configured savepoint restore settings.
*/
public SavepointRestoreSettings getSavepointRestoreSettings() {
return savepointRestoreSettings;
} | 3.68 |
hadoop_PendingSet_getVersion | /** @return the version marker. */
public int getVersion() {
return version;
} | 3.68 |
framework_VAbstractPopupCalendar_setTextFieldTabIndex | /**
* Set correct tab index for disabled text field in IE as the value set in
* setTextFieldEnabled(...) gets overridden in
* TextualDateConnection.updateFromUIDL(...).
*
* @since 7.3.1
*/
public void setTextFieldTabIndex() {
if (BrowserInfo.get().isIE() && !textFieldEnabled) {
// index needs to be -2... | 3.68 |
hudi_HoodieTableConfig_getIndexDefinitionPath | /**
* @returns the index definition path.
*/
public Option<String> getIndexDefinitionPath() {
return Option.ofNullable(getString(INDEX_DEFINITION_PATH));
} | 3.68 |
hadoop_TrustedChannelResolver_isTrusted | /**
* Identify boolean value indicating whether a channel is trusted or not.
* @param peerAddress address of the peer
* @return true if the channel is trusted and false otherwise.
*/
public boolean isTrusted(InetAddress peerAddress) {
return false;
} | 3.68 |
flink_TableResultImpl_build | /** Returns a {@link TableResult} instance. */
public TableResultInternal build() {
if (printStyle == null) {
printStyle = PrintStyle.rawContent(resultProvider.getRowDataStringConverter());
}
return new TableResultImpl(
jobClient, resolvedSchema, resultKind, resultProvider, printStyle);
... | 3.68 |
streampipes_AdapterDescription_getCorrespondingServiceGroup | /**
* @deprecated check if the service group can be removed as a single pipeline element
* can correspond to different service groups
*/
@Deprecated
public String getCorrespondingServiceGroup() {
return correspondingServiceGroup;
} | 3.68 |
dubbo_ServiceConfig_isOnlyInJvm | /**
* Determine if it is injvm
*
* @return
*/
private boolean isOnlyInJvm() {
return getProtocols().size() == 1
&& LOCAL_PROTOCOL.equalsIgnoreCase(getProtocols().get(0).getName());
} | 3.68 |
flink_ApplicationDispatcherBootstrap_unwrapJobResultException | /**
* If the given {@link JobResult} indicates success, this passes through the {@link JobResult}.
* Otherwise, this returns a future that is finished exceptionally (potentially with an
* exception from the {@link JobResult}).
*/
private CompletableFuture<JobResult> unwrapJobResultException(
final Completab... | 3.68 |
hbase_MobUtils_hasMobReferenceTag | /**
* Whether the tag list has a mob reference tag.
* @param tags The tag list.
* @return True if the list has a mob reference tag, false if it doesn't.
*/
public static boolean hasMobReferenceTag(List<Tag> tags) {
if (!tags.isEmpty()) {
for (Tag tag : tags) {
if (tag.getType() == TagType.MOB_REFERENCE_... | 3.68 |
hadoop_AbstractManifestData_validateCollectionClass | /**
* Verify that all instances in a collection are of the given class.
* @param it iterator
* @param classname classname to require
* @throws IOException on a failure
*/
void validateCollectionClass(Iterable it, Class classname)
throws IOException {
for (Object o : it) {
verify(o.getClass().equals(class... | 3.68 |
flink_MultipleParameterTool_fromMultiMap | /**
* Returns {@link MultipleParameterTool} for the given multi map.
*
* @param multiMap A map of arguments. Key is String and value is a Collection.
* @return A {@link MultipleParameterTool}
*/
public static MultipleParameterTool fromMultiMap(Map<String, Collection<String>> multiMap) {
Preconditions.checkNotN... | 3.68 |
hadoop_S3ClientFactory_withTransferManagerExecutor | /**
* Set the executor that the transfer manager will use to execute background tasks.
* @param value new value
* @return the builder
*/
public S3ClientCreationParameters withTransferManagerExecutor(
final Executor value) {
transferManagerExecutor = value;
return this;
} | 3.68 |
hadoop_StoreContextBuilder_setAuditor | /**
* Set builder value.
* @param value new value
* @return the builder
*/
public StoreContextBuilder setAuditor(
final AuditSpanSource<AuditSpanS3A> value) {
auditor = value;
return this;
} | 3.68 |
framework_WebBrowser_isIE | /**
* Tests whether the user is using Internet Explorer.
*
* @return true if the user is using Internet Explorer, false if the user is
* not using Internet Explorer or if no information on the browser
* is present
*/
public boolean isIE() {
if (browserDetails == null) {
return false;
... | 3.68 |
hbase_LockManager_acquire | /**
* Acquire the lock, waiting indefinitely until the lock is released or the thread is
* interrupted.
* @throws InterruptedException If current thread is interrupted while waiting for the lock
*/
public boolean acquire() throws InterruptedException {
return tryAcquire(0);
} | 3.68 |
flink_LocalBufferPool_getEstimatedNumberOfRequestedMemorySegments | /**
* Estimates the number of requested buffers.
*
* @return the same value as {@link #getMaxNumberOfMemorySegments()} for bounded pools. For
* unbounded pools it returns an approximation based upon {@link
* #getNumberOfRequiredMemorySegments()}
*/
public int getEstimatedNumberOfRequestedMemorySegments() ... | 3.68 |
hbase_ClassSize_estimateBaseFromCoefficients | /**
* Estimate the static space taken up by a class instance given the coefficients returned by
* getSizeCoefficients.
* @param coeff the coefficients
* @param debug debug flag
* @return the size estimate, in bytes
*/
private static long estimateBaseFromCoefficients(int[] coeff, boolean debug) {
long prealign_s... | 3.68 |
hbase_PrivateCellUtil_estimatedSerializedSizeOf | /**
* Estimate based on keyvalue's serialization format in the RPC layer. Note that there is an extra
* SIZEOF_INT added to the size here that indicates the actual length of the cell for cases where
* cell's are serialized in a contiguous format (For eg in RPCs).
* @return Estimate of the <code>cell</code> size in ... | 3.68 |
hbase_AsyncAdmin_mergeSwitch | /**
* Turn the Merge switch on or off.
* @param enabled enabled or not
* @return Previous switch value wrapped by a {@link CompletableFuture}
*/
default CompletableFuture<Boolean> mergeSwitch(boolean enabled) {
return mergeSwitch(enabled, false);
} | 3.68 |
pulsar_ReaderConfiguration_getCryptoKeyReader | /**
* @return the CryptoKeyReader
*/
public CryptoKeyReader getCryptoKeyReader() {
return conf.getCryptoKeyReader();
} | 3.68 |
framework_TouchScrollDelegate_detectScrolledElement | /**
* Detects if a touch happens on a predefined element and the element has
* something to scroll.
*
* @param touch
* @return
*/
private boolean detectScrolledElement(Touch touch) {
Element target = touch.getTarget().cast();
for (Element el : scrollableElements) {
if (el.isOrHasChild(target)
... | 3.68 |
hadoop_FutureIOSupport_eval | /**
* Evaluate a CallableRaisingIOE in the current thread,
* converting IOEs to RTEs and propagating.
* See {@link FutureIO#eval(CallableRaisingIOE)}.
*
* @param callable callable to invoke
* @param <T> Return type.
* @return the evaluated result.
* @throws UnsupportedOperationException fail fast if unsupported... | 3.68 |
hbase_MetaFixer_fixHoles | /**
* If hole, it papers it over by adding a region in the filesystem and to hbase:meta. Does not
* assign.
*/
void fixHoles(CatalogJanitorReport report) {
final List<Pair<RegionInfo, RegionInfo>> holes = report.getHoles();
if (holes.isEmpty()) {
LOG.info("CatalogJanitor Report contains no holes to fix. Skip... | 3.68 |
hudi_KafkaConnectUtils_getDefaultHadoopConf | /**
* Returns the default Hadoop Configuration.
*
* @return
*/
public static Configuration getDefaultHadoopConf(KafkaConnectConfigs connectConfigs) {
Configuration hadoopConf = new Configuration();
// add hadoop config files
if (!StringUtils.isNullOrEmpty(connectConfigs.getHadoopConfDir())
|| !Stri... | 3.68 |
hadoop_ServiceLauncher_verifyConfigurationFilesExist | /**
* Verify that all the specified filenames exist.
* @param filenames a list of files
* @throws ServiceLaunchException if a file is not found
*/
protected void verifyConfigurationFilesExist(String[] filenames) {
if (filenames == null) {
return;
}
for (String filename : filenames) {
File file = new F... | 3.68 |
flink_DecimalData_isCompact | /** Returns whether the decimal value is small enough to be stored in a long. */
public static boolean isCompact(int precision) {
return precision <= MAX_COMPACT_PRECISION;
} | 3.68 |
hudi_ArrayColumnReader_setChildrenInfo | /**
* The lengths & offsets will be initialized as default size (1024), it should be set to the
* actual size according to the element number.
*/
private void setChildrenInfo(HeapArrayVector lcv, int itemNum, int elementNum) {
lcv.setSize(itemNum);
long[] lcvLength = new long[elementNum];
long[] lcvOffset = ne... | 3.68 |
flink_FunctionIdentifier_asSummaryString | /** Returns a string that summarizes this instance for printing to a console or log. */
public String asSummaryString() {
if (objectIdentifier != null) {
return String.join(
".",
objectIdentifier.getCatalogName(),
objectIdentifier.getDatabaseName(),
... | 3.68 |
hbase_AbstractFSWAL_atHeadOfRingBufferEventHandlerAppend | /**
* Exposed for testing only. Use to tricks like halt the ring buffer appending.
*/
protected void atHeadOfRingBufferEventHandlerAppend() {
// Noop
} | 3.68 |
hbase_RawCell_cloneTags | /**
* Allows cloning the tags in the cell to a new byte[]
* @return the byte[] having the tags
*/
default byte[] cloneTags() {
return PrivateCellUtil.cloneTags(this);
} | 3.68 |
pulsar_CustomCommandFactoryProvider_createCustomCommandFactories | /**
* create a Command Factory.
*/
public static List<CustomCommandFactory> createCustomCommandFactories(
Properties conf) throws IOException {
String names = conf.getProperty("customCommandFactories", "");
List<CustomCommandFactory> result = new ArrayList<>();
if (names.isEmpty()) {
// ea... | 3.68 |
flink_DynamicPartitionPruningUtils_isNewSource | /** Returns true if the source is FLIP-27 source, else false. */
private static boolean isNewSource(ScanTableSource scanTableSource) {
ScanTableSource.ScanRuntimeProvider provider =
scanTableSource.getScanRuntimeProvider(ScanRuntimeProviderContext.INSTANCE);
if (provider instanceof SourceProvider) {... | 3.68 |
framework_BeanItem_getBean | /**
* Gets the underlying JavaBean object.
*
* @return the bean object.
*/
public BT getBean() {
return bean;
} | 3.68 |
hadoop_AddMountAttributes_getMountTableForAddRequest | /**
* Create a new mount table object from the given mount point and update its attributes.
*
* @param mountSrc mount point src.
* @return MountTable object with updated attributes.
* @throws IOException If mount table instantiation fails.
*/
private MountTable getMountTableForAddRequest(String mountSrc) throws I... | 3.68 |
framework_VNotification_showError | /**
* Shows an error notification and redirects the user to the given URL when
* she clicks on the notification.
*
* If both message and caption are null, redirects the user to the url
* immediately
*
* @since 7.5.1
* @param connection
* A reference to the ApplicationConnection
* @param caption
* ... | 3.68 |
hadoop_AccessTokenTimer_setExpiresInMSSinceEpoch | /**
* Set when the access token will expire in milliseconds from epoch,
* as required by the WebHDFS configuration. This is a bit hacky and lame.
*
* @param expiresInMSSinceEpoch Access time expiration in ms since epoch.
*/
public void setExpiresInMSSinceEpoch(String expiresInMSSinceEpoch){
this.nextRefreshMSSi... | 3.68 |
open-banking-gateway_EncryptionWithInitVectorOper_decryption | /**
* Decryption cipher
* @param keyWithIv Symmetric key and initialization vector
* @return Symmetric decryption cipher
*/
@SneakyThrows
public Cipher decryption(SecretKeyWithIv keyWithIv) {
Cipher cipher = Cipher.getInstance(encSpec.getCipherAlgo());
cipher.init(
Cipher.DECRYPT_MODE, keyWithIv... | 3.68 |
hudi_StreamerUtil_isValidFile | /**
* Returns whether the give file is in valid hoodie format.
* For example, filtering out the empty or corrupt files.
*/
public static boolean isValidFile(FileStatus fileStatus) {
final String extension = FSUtils.getFileExtension(fileStatus.getPath().toString());
if (PARQUET.getFileExtension().equals(extension... | 3.68 |
flink_TaskStateSnapshot_putSubtaskStateByOperatorID | /**
* Maps the given operator id to the given subtask state. Returns the subtask state of a
* previous mapping, if such a mapping existed or null otherwise.
*/
public OperatorSubtaskState putSubtaskStateByOperatorID(
@Nonnull OperatorID operatorID, @Nonnull OperatorSubtaskState state) {
return subtaskSt... | 3.68 |
framework_GridSingleSelect_setUserSelectionAllowed | /**
* Sets whether the user is allowed to change the selection.
* <p>
* The check is done only for the client side actions. It doesn't affect
* selection requests sent from the server side.
*
* @param allowed
* <code>true</code> if the user is allowed to change the
* selection, <code>false... | 3.68 |
framework_ComponentDetail_setTooltipInfo | /**
* @param tooltipInfo
* the tooltipInfo to set
*/
public void setTooltipInfo(TooltipInfo tooltipInfo) {
this.tooltipInfo = tooltipInfo;
} | 3.68 |
pulsar_TopicsBase_publishMessagesToPartition | // Publish message to single partition of a partitioned topic.
protected void publishMessagesToPartition(AsyncResponse asyncResponse, ProducerMessages request,
boolean authoritative, int partition) {
if (topicName.isPartitioned()) {
asyncResponse.resume(new R... | 3.68 |
streampipes_JdbcClient_closeAll | /**
* Closes all open connections and statements of JDBC
*/
protected void closeAll() {
boolean error = false;
try {
if (this.statementHandler.statement != null) {
this.statementHandler.statement.close();
this.statementHandler.statement = null;
}
} catch (SQLException e) {
error = true;
... | 3.68 |
flink_StreamPhysicalPythonCorrelateRule_findTableFunction | // find only calc and table function
private boolean findTableFunction(FlinkLogicalCalc calc) {
RelNode child = ((RelSubset) calc.getInput()).getOriginal();
if (child instanceof FlinkLogicalTableFunctionScan) {
FlinkLogicalTableFunctionScan scan = (FlinkLogicalTableFunctionScan) child;
return Py... | 3.68 |
framework_ServerRpcQueue_showLoadingIndicator | /**
* Checks if a loading indicator should be shown when the RPCs have been
* sent to the server and we are waiting for a response.
*
* @return true if a loading indicator should be shown, false otherwise
*/
public boolean showLoadingIndicator() {
for (MethodInvocation invocation : getAll()) {
if (isLe... | 3.68 |
hbase_DigestSaslServerAuthenticationProvider_handle | /** {@inheritDoc} */
@Override
public void handle(Callback[] callbacks) throws InvalidToken, UnsupportedCallbackException {
NameCallback nc = null;
PasswordCallback pc = null;
AuthorizeCallback ac = null;
for (Callback callback : callbacks) {
if (callback instanceof AuthorizeCallback) {
ac = (Authoriz... | 3.68 |
hbase_UnsafeAccess_putLong | /**
* Put a long value out to the specified BB position in big-endian format.
* @param buf the byte buffer
* @param offset position in the buffer
* @param val long to write out
* @return incremented offset
*/
public static int putLong(ByteBuffer buf, int offset, long val) {
if (LITTLE_ENDIAN) {
val = ... | 3.68 |
hadoop_TimelineReaderClient_createTimelineReaderClient | /**
* Create a new instance of Timeline Reader Client.
*
* @return instance of Timeline Reader Client.
*/
@InterfaceAudience.Public
public static TimelineReaderClient createTimelineReaderClient() {
return new TimelineReaderClientImpl();
} | 3.68 |
flink_Channel_setReplicationFactor | /**
* Sets the replication factor of the connection.
*
* @param factor The replication factor of the connection.
*/
public void setReplicationFactor(int factor) {
this.replicationFactor = factor;
} | 3.68 |
flink_SocketStreamIterator_getPort | /**
* Returns the port on which the iterator is getting the data. (Used internally.)
*
* @return The port
*/
public int getPort() {
return socket.getLocalPort();
} | 3.68 |
querydsl_JTSGeometryExpressions_extent | /**
* Returns the bounding box that bounds rows of geometries.
*
* @param collection geometry collection
* @return geometry collection
*/
public static JTSGeometryExpression<?> extent(Expression<? extends GeometryCollection> collection) {
return geometryOperation(SpatialOps.EXTENT, collection);
} | 3.68 |
morf_SqlDialect_getSchemaName | /**
* Returns the database schema name. May be null.
* @return The schema name
*/
public String getSchemaName() {
return schemaName;
} | 3.68 |
framework_FreeformQuery_containsRowWithKey | /**
* This implementation of the containsRowWithKey method rewrites existing
* WHERE clauses in the query string. The logic is, however, not very
* complex and some times can do the Wrong Thing<sup>TM</sup>. For the
* situations where this logic is not enough, you can implement the
* getContainsRowQueryString meth... | 3.68 |
flink_MemorySegment_getInt | /**
* Reads an int value (32bit, 4 bytes) from the given position, in the system's native byte
* order. This method offers the best speed for integer reading and should be used unless a
* specific byte order is required. In most cases, it suffices to know that the byte order in
* which the value is written is the s... | 3.68 |
Activiti_TreeBuilderException_getExpression | /**
* @return the expression string
*/
public String getExpression() {
return expression;
} | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.