name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
druid_IPRange_getExtendedNetworkPrefix
/** * Return the extended extended network prefix. * * @return Return the extended network prefix. */ public final int getExtendedNetworkPrefix() { return extendedNetworkPrefix; }
3.68
framework_HierarchicalContainer_getChildren
/* * Gets the IDs of the children of the specified Item. Don't add a JavaDoc * comment here, we use the default documentation from implemented * interface. */ @Override public Collection<?> getChildren(Object itemId) { LinkedList<Object> c; if (filteredChildren != null) { c = filteredChildren.get(i...
3.68
pulsar_LeastResourceUsageWithWeight_selectBroker
/** * Find a suitable broker to assign the given bundle to. * This method is not thread safety. * * @param candidates The candidates for which the bundle may be assigned. * @param bundleToAssign The data for the bundle to assign. * @param loadData The load data from the leader broker. * @param conf ...
3.68
framework_Page_getPage
/** * Gets the page in which the uri has changed. * * @return the page in which the uri has changed */ public Page getPage() { return (Page) getSource(); }
3.68
flink_CheckpointStorageWorkerView_toFileMergingStorage
/** * Return {@link org.apache.flink.runtime.state.filesystem.FsMergingCheckpointStorageAccess} if * file merging is enabled Otherwise, return itself. File merging is supported by subclasses of * {@link org.apache.flink.runtime.state.filesystem.AbstractFsCheckpointStorageAccess}. */ default CheckpointStorageWorkerV...
3.68
morf_CreateDeployedViews_execute
/** * @see org.alfasoftware.morf.upgrade.UpgradeStep#execute(org.alfasoftware.morf.upgrade.SchemaEditor, org.alfasoftware.morf.upgrade.DataEditor) */ @Override public void execute(SchemaEditor schema, DataEditor data) { schema.addTable(table("DeployedViews"). columns( colum...
3.68
hadoop_LoggingAuditor_serviceInit
/** * Service init, look for jobID and attach as an attribute in log entries. * This is where the warning span is created, so the relevant attributes * (and filtering options) are applied. * @param conf configuration * @throws Exception failure */ @Override protected void serviceInit(final Configuration conf) thr...
3.68
framework_CalendarTargetDetails_getDropTime
/** * @return the date where the drop happened */ public Date getDropTime() { if (hasDropTime) { return (Date) getData("dropTime"); } else { return (Date) getData("dropDay"); } }
3.68
hadoop_BlockStorageMovementNeeded_markScanCompleted
/** * Mark directory scan is completed. */ public synchronized void markScanCompleted() { this.fullyScanned = true; }
3.68
graphhopper_BikeCommonPriorityParser_handlePriority
/** * In this method we prefer cycleways or roads with designated bike access and avoid big roads * or roads with trams or pedestrian. * * @return new priority based on priorityFromRelation and on the tags in ReaderWay. */ int handlePriority(ReaderWay way, double wayTypeSpeed, Integer priorityFromRelation) { T...
3.68
flink_MiniCluster_terminateTaskManager
/** * Terminates a TaskManager with the given index. * * <p>See {@link #startTaskManager()} to understand how TaskManagers are indexed. This method * terminates a TaskManager with a given index but it does not clear the index. The index stays * occupied for the lifetime of the MiniCluster and its TaskManager stays...
3.68
hbase_SnapshotReferenceUtil_verifyStoreFile
/** * Verify the validity of the snapshot store file * @param conf The current {@link Configuration} instance. * @param fs {@link FileSystem} * @param snapshotDir {@link Path} to the Snapshot directory of the snapshot to verify * @param snapshot the {@link SnapshotDescription} of the snapshot to...
3.68
hbase_HFileReaderImpl_midKey
/** * @return Midkey for this file. We work with block boundaries only so returned midkey is an * approximation only. */ @Override public Optional<Cell> midKey() throws IOException { return Optional.ofNullable(dataBlockIndexReader.midkey(this)); }
3.68
framework_GenericFontIcon_hashCode
/* * (non-Javadoc) * * @see java.lang.Object#hashCode() */ @Override public int hashCode() { final int prime = 31; int result = 1; result = prime * result + codePoint; result = prime * result + ((fontFamily == null) ? 0 : fontFamily.hashCode()); return result; }
3.68
hadoop_Utils_compareTo
/** * Compare this version with another version. */ @Override public int compareTo(Version that) { if (major != that.major) { return major - that.major; } return minor - that.minor; }
3.68
flink_TimestampsAndWatermarksTransformation_getInputType
/** Returns the {@code TypeInformation} for the elements of the input. */ public TypeInformation<IN> getInputType() { return input.getOutputType(); }
3.68
hadoop_QueuePriorityContainerCandidateSelector_isQueueSatisfied
/** * Do we allow the demanding queue preempt resource from other queues? * A satisfied queue is not allowed to preempt resource from other queues. * @param demandingQueue * @return allowed/not */ private boolean isQueueSatisfied(String demandingQueue, String partition) { TempQueuePerPartition tq = preemptio...
3.68
hbase_ServerMetrics_getVersion
/** Returns the string type version of a regionserver. */ default String getVersion() { return "0.0.0"; }
3.68
hbase_FSTableDescriptors_getTableDir
/** * Return the table directory in HDFS */ private Path getTableDir(TableName tableName) { return CommonFSUtils.getTableDir(rootdir, tableName); }
3.68
framework_VLoadingIndicator_setSecondDelay
/** * Sets the delay (in ms) which must pass before the loading indicator moves * to its "second" state. * * @param secondDelay * The delay (in ms) until the loading indicator moves into its * "second" state. Counted from when {@link #trigger()} is * called. */ public void setSe...
3.68
hadoop_AMRMProxyService_stopApplication
/** * Shuts down the request processing pipeline for the specified application * attempt id. * * @param applicationId application id */ protected void stopApplication(ApplicationId applicationId) { this.metrics.incrRequestCount(); Preconditions.checkArgument(applicationId != null, "applicationId is null"); R...
3.68
hadoop_ActiveAuditManagerS3A_onExecutionFailure
/** * Forward to the inner span. * {@inheritDoc} */ @Override public void onExecutionFailure(Context.FailedExecution context, ExecutionAttributes executionAttributes) { span.onExecutionFailure(context, executionAttributes); }
3.68
hbase_Get_setTimeRange
/** * Get versions of columns only within the specified timestamp range, [minStamp, maxStamp). * @param minStamp minimum timestamp value, inclusive * @param maxStamp maximum timestamp value, exclusive * @return this for invocation chaining */ public Get setTimeRange(long minStamp, long maxStamp) throws IOException...
3.68
flink_DynamicSourceUtils_createRequiredMetadataColumns
/** * Returns a list of required metadata columns. Ordered by the iteration order of {@link * SupportsReadingMetadata#listReadableMetadata()}. * * <p>This method assumes that source and schema have been validated via {@link * #prepareDynamicSource(String, ResolvedCatalogTable, DynamicTableSource, boolean, * Reada...
3.68
AreaShop_Utils_getDurationFromSecondsOrString
/** * Get setting from config that could be only a number indicating seconds. * or a string indicating a duration string * @param path Path of the setting to read * @return milliseconds that the setting indicates */ public static long getDurationFromSecondsOrString(String path) { if(config.isLong(path) || config....
3.68
framework_VCalendarAction_getActionStartDate
/** * Get the date and time when the action starts. * * @return */ public Date getActionStartDate() { return actionStartDate; }
3.68
open-banking-gateway_QueryHeadersMapperTemplate_forExecution
/** * Converts context object into object that can be used for ASPSP API call. * @param context Context to convert * @return Object that can be used with {@code Xs2aAdapter} to perform ASPSP API calls */ public ValidatedQueryHeaders<Q, H> forExecution(C context) { return new ValidatedQueryHeaders<>( ...
3.68
hadoop_FindOptions_getMinDepth
/** * Returns the minimum depth for applying expressions. * * @return min depth */ public int getMinDepth() { return this.minDepth; }
3.68
hbase_HFileBlockIndex_getNonRootIndexedKey
/** * The indexed key at the ith position in the nonRootIndex. The position starts at 0. * @param i the ith position * @return The indexed key at the ith position in the nonRootIndex. */ static byte[] getNonRootIndexedKey(ByteBuff nonRootIndex, int i) { int numEntries = nonRootIndex.getInt(0); if (i < 0 || i >=...
3.68
AreaShop_SetteleportCommand_canUse
/** * Check if a person can set the teleport location of the region. * @param person The person to check * @param region The region to check for * @return true if the person can set the teleport location, otherwise false */ public static boolean canUse(CommandSender person, GeneralRegion region) { if(!(person ins...
3.68
flink_CoGroupedStreams_equalTo
/** * Specifies a {@link KeySelector} for elements from the second input with explicit type * information for the key type. * * @param keySelector The KeySelector to be used for extracting the key for partitioning. * @param keyType The type information describing the key type. */ public EqualTo equalTo(KeySelecto...
3.68
hadoop_PeriodicService_getErrorCount
/** * Get how many times we failed to run the periodic service. * * @return Times we failed to run the periodic service. */ protected long getErrorCount() { return this.errorCount; }
3.68
hbase_StoreScanner_getEstimatedNumberOfKvsScanned
/** Returns The estimated number of KVs seen by this scanner (includes some skipped KVs). */ public long getEstimatedNumberOfKvsScanned() { return this.kvsScanned; }
3.68
framework_TableScroll_initProperties
// set up the properties (columns) private void initProperties(Table table) { for (int i = 0; i < COLUMNS; i++) { table.addContainerProperty("property" + i, String.class, "some value"); } }
3.68
hbase_Scan_hasFilter
/** Returns true is a filter has been specified, false if not */ public boolean hasFilter() { return filter != null; }
3.68
framework_AbstractRemoteDataSource_setRow
/** * A method for the data source to update the row data. * * @param row * the updated row object */ public void setRow(final T row) { this.row = row; assert getRowKey(row).equals(key) : "The old key does not " + "equal the new key for the given row (old: " + key + ", ne...
3.68
flink_BlobServer_deletePermanent
/** * Delete the uploaded data with the given {@link JobID} and {@link PermanentBlobKey}. * * @param jobId ID of the job this blob belongs to * @param key the key of this blob */ @Override public boolean deletePermanent(JobID jobId, PermanentBlobKey key) { return deleteInternal(jobId, key); }
3.68
MagicPlugin_MagicController_initialize
/* * Saving and loading */ public void initialize() { warpController = new WarpController(this); kitController = new KitController(this); crafting = new CraftingController(this); mobs = new MobController(this); items = new ItemController(this); enchanting = new EnchantingController(this); ...
3.68
flink_StreamTaskActionExecutor_synchronizedExecutor
/** Returns an ExecutionDecorator that synchronizes each invocation on a given object. */ static SynchronizedStreamTaskActionExecutor synchronizedExecutor(Object mutex) { return new SynchronizedStreamTaskActionExecutor(mutex); }
3.68
hbase_HRegionServer_getMasterAddressTracker
/** Returns Master address tracker instance. */ public MasterAddressTracker getMasterAddressTracker() { return this.masterAddressTracker; }
3.68
Activiti_ReflectUtil_getSetter
/** * Returns the setter-method for the given field name or null if no setter exists. */ public static Method getSetter(String fieldName, Class<?> clazz, Class<?> fieldType) { String setterName = "set" + Character.toTitleCase(fieldName.charAt(0)) + fieldName.substring(1, fieldName.length()); try { // Using ge...
3.68
hbase_FamilyFilter_toByteArray
/** Returns The filter serialized using pb */ @Override public byte[] toByteArray() { FilterProtos.FamilyFilter.Builder builder = FilterProtos.FamilyFilter.newBuilder(); builder.setCompareFilter(super.convert()); return builder.build().toByteArray(); }
3.68
hbase_BloomFilterMetrics_getRequestsCount
/** Returns Current value for bloom requests count */ public long getRequestsCount() { return requests.sum(); }
3.68
hbase_ClientModeStrategy_aggregateRecordsAndAddDistinct
/** * Aggregate the records and count the unique values for the given distinctField * @param records records to be processed * @param groupBy Field on which group by needs to be done * @param distinctField Field whose unique values needs to be counted * @param uniqueCountAssigne...
3.68
flink_JobGraph_getSavepointRestoreSettings
/** * Returns the configured savepoint restore setting. * * @return The configured savepoint restore settings. */ public SavepointRestoreSettings getSavepointRestoreSettings() { return savepointRestoreSettings; }
3.68
hadoop_PendingSet_getVersion
/** @return the version marker. */ public int getVersion() { return version; }
3.68
framework_VAbstractPopupCalendar_setTextFieldTabIndex
/** * Set correct tab index for disabled text field in IE as the value set in * setTextFieldEnabled(...) gets overridden in * TextualDateConnection.updateFromUIDL(...). * * @since 7.3.1 */ public void setTextFieldTabIndex() { if (BrowserInfo.get().isIE() && !textFieldEnabled) { // index needs to be -2...
3.68
hudi_HoodieTableConfig_getIndexDefinitionPath
/** * @returns the index definition path. */ public Option<String> getIndexDefinitionPath() { return Option.ofNullable(getString(INDEX_DEFINITION_PATH)); }
3.68
hadoop_TrustedChannelResolver_isTrusted
/** * Identify boolean value indicating whether a channel is trusted or not. * @param peerAddress address of the peer * @return true if the channel is trusted and false otherwise. */ public boolean isTrusted(InetAddress peerAddress) { return false; }
3.68
flink_TableResultImpl_build
/** Returns a {@link TableResult} instance. */ public TableResultInternal build() { if (printStyle == null) { printStyle = PrintStyle.rawContent(resultProvider.getRowDataStringConverter()); } return new TableResultImpl( jobClient, resolvedSchema, resultKind, resultProvider, printStyle); ...
3.68
streampipes_AdapterDescription_getCorrespondingServiceGroup
/** * @deprecated check if the service group can be removed as a single pipeline element * can correspond to different service groups */ @Deprecated public String getCorrespondingServiceGroup() { return correspondingServiceGroup; }
3.68
dubbo_ServiceConfig_isOnlyInJvm
/** * Determine if it is injvm * * @return */ private boolean isOnlyInJvm() { return getProtocols().size() == 1 && LOCAL_PROTOCOL.equalsIgnoreCase(getProtocols().get(0).getName()); }
3.68
flink_ApplicationDispatcherBootstrap_unwrapJobResultException
/** * If the given {@link JobResult} indicates success, this passes through the {@link JobResult}. * Otherwise, this returns a future that is finished exceptionally (potentially with an * exception from the {@link JobResult}). */ private CompletableFuture<JobResult> unwrapJobResultException( final Completab...
3.68
hbase_MobUtils_hasMobReferenceTag
/** * Whether the tag list has a mob reference tag. * @param tags The tag list. * @return True if the list has a mob reference tag, false if it doesn't. */ public static boolean hasMobReferenceTag(List<Tag> tags) { if (!tags.isEmpty()) { for (Tag tag : tags) { if (tag.getType() == TagType.MOB_REFERENCE_...
3.68
hadoop_AbstractManifestData_validateCollectionClass
/** * Verify that all instances in a collection are of the given class. * @param it iterator * @param classname classname to require * @throws IOException on a failure */ void validateCollectionClass(Iterable it, Class classname) throws IOException { for (Object o : it) { verify(o.getClass().equals(class...
3.68
flink_MultipleParameterTool_fromMultiMap
/** * Returns {@link MultipleParameterTool} for the given multi map. * * @param multiMap A map of arguments. Key is String and value is a Collection. * @return A {@link MultipleParameterTool} */ public static MultipleParameterTool fromMultiMap(Map<String, Collection<String>> multiMap) { Preconditions.checkNotN...
3.68
hadoop_S3ClientFactory_withTransferManagerExecutor
/** * Set the executor that the transfer manager will use to execute background tasks. * @param value new value * @return the builder */ public S3ClientCreationParameters withTransferManagerExecutor( final Executor value) { transferManagerExecutor = value; return this; }
3.68
hadoop_StoreContextBuilder_setAuditor
/** * Set builder value. * @param value new value * @return the builder */ public StoreContextBuilder setAuditor( final AuditSpanSource<AuditSpanS3A> value) { auditor = value; return this; }
3.68
framework_WebBrowser_isIE
/** * Tests whether the user is using Internet Explorer. * * @return true if the user is using Internet Explorer, false if the user is * not using Internet Explorer or if no information on the browser * is present */ public boolean isIE() { if (browserDetails == null) { return false; ...
3.68
hbase_LockManager_acquire
/** * Acquire the lock, waiting indefinitely until the lock is released or the thread is * interrupted. * @throws InterruptedException If current thread is interrupted while waiting for the lock */ public boolean acquire() throws InterruptedException { return tryAcquire(0); }
3.68
flink_LocalBufferPool_getEstimatedNumberOfRequestedMemorySegments
/** * Estimates the number of requested buffers. * * @return the same value as {@link #getMaxNumberOfMemorySegments()} for bounded pools. For * unbounded pools it returns an approximation based upon {@link * #getNumberOfRequiredMemorySegments()} */ public int getEstimatedNumberOfRequestedMemorySegments() ...
3.68
hbase_ClassSize_estimateBaseFromCoefficients
/** * Estimate the static space taken up by a class instance given the coefficients returned by * getSizeCoefficients. * @param coeff the coefficients * @param debug debug flag * @return the size estimate, in bytes */ private static long estimateBaseFromCoefficients(int[] coeff, boolean debug) { long prealign_s...
3.68
hbase_PrivateCellUtil_estimatedSerializedSizeOf
/** * Estimate based on keyvalue's serialization format in the RPC layer. Note that there is an extra * SIZEOF_INT added to the size here that indicates the actual length of the cell for cases where * cell's are serialized in a contiguous format (For eg in RPCs). * @return Estimate of the <code>cell</code> size in ...
3.68
hbase_AsyncAdmin_mergeSwitch
/** * Turn the Merge switch on or off. * @param enabled enabled or not * @return Previous switch value wrapped by a {@link CompletableFuture} */ default CompletableFuture<Boolean> mergeSwitch(boolean enabled) { return mergeSwitch(enabled, false); }
3.68
pulsar_ReaderConfiguration_getCryptoKeyReader
/** * @return the CryptoKeyReader */ public CryptoKeyReader getCryptoKeyReader() { return conf.getCryptoKeyReader(); }
3.68
framework_TouchScrollDelegate_detectScrolledElement
/** * Detects if a touch happens on a predefined element and the element has * something to scroll. * * @param touch * @return */ private boolean detectScrolledElement(Touch touch) { Element target = touch.getTarget().cast(); for (Element el : scrollableElements) { if (el.isOrHasChild(target) ...
3.68
hadoop_FutureIOSupport_eval
/** * Evaluate a CallableRaisingIOE in the current thread, * converting IOEs to RTEs and propagating. * See {@link FutureIO#eval(CallableRaisingIOE)}. * * @param callable callable to invoke * @param <T> Return type. * @return the evaluated result. * @throws UnsupportedOperationException fail fast if unsupported...
3.68
hbase_MetaFixer_fixHoles
/** * If hole, it papers it over by adding a region in the filesystem and to hbase:meta. Does not * assign. */ void fixHoles(CatalogJanitorReport report) { final List<Pair<RegionInfo, RegionInfo>> holes = report.getHoles(); if (holes.isEmpty()) { LOG.info("CatalogJanitor Report contains no holes to fix. Skip...
3.68
hudi_KafkaConnectUtils_getDefaultHadoopConf
/** * Returns the default Hadoop Configuration. * * @return */ public static Configuration getDefaultHadoopConf(KafkaConnectConfigs connectConfigs) { Configuration hadoopConf = new Configuration(); // add hadoop config files if (!StringUtils.isNullOrEmpty(connectConfigs.getHadoopConfDir()) || !Stri...
3.68
hadoop_ServiceLauncher_verifyConfigurationFilesExist
/** * Verify that all the specified filenames exist. * @param filenames a list of files * @throws ServiceLaunchException if a file is not found */ protected void verifyConfigurationFilesExist(String[] filenames) { if (filenames == null) { return; } for (String filename : filenames) { File file = new F...
3.68
flink_DecimalData_isCompact
/** Returns whether the decimal value is small enough to be stored in a long. */ public static boolean isCompact(int precision) { return precision <= MAX_COMPACT_PRECISION; }
3.68
hudi_ArrayColumnReader_setChildrenInfo
/** * The lengths & offsets will be initialized as default size (1024), it should be set to the * actual size according to the element number. */ private void setChildrenInfo(HeapArrayVector lcv, int itemNum, int elementNum) { lcv.setSize(itemNum); long[] lcvLength = new long[elementNum]; long[] lcvOffset = ne...
3.68
flink_FunctionIdentifier_asSummaryString
/** Returns a string that summarizes this instance for printing to a console or log. */ public String asSummaryString() { if (objectIdentifier != null) { return String.join( ".", objectIdentifier.getCatalogName(), objectIdentifier.getDatabaseName(), ...
3.68
hbase_AbstractFSWAL_atHeadOfRingBufferEventHandlerAppend
/** * Exposed for testing only. Use to tricks like halt the ring buffer appending. */ protected void atHeadOfRingBufferEventHandlerAppend() { // Noop }
3.68
hbase_RawCell_cloneTags
/** * Allows cloning the tags in the cell to a new byte[] * @return the byte[] having the tags */ default byte[] cloneTags() { return PrivateCellUtil.cloneTags(this); }
3.68
pulsar_CustomCommandFactoryProvider_createCustomCommandFactories
/** * create a Command Factory. */ public static List<CustomCommandFactory> createCustomCommandFactories( Properties conf) throws IOException { String names = conf.getProperty("customCommandFactories", ""); List<CustomCommandFactory> result = new ArrayList<>(); if (names.isEmpty()) { // ea...
3.68
flink_DynamicPartitionPruningUtils_isNewSource
/** Returns true if the source is FLIP-27 source, else false. */ private static boolean isNewSource(ScanTableSource scanTableSource) { ScanTableSource.ScanRuntimeProvider provider = scanTableSource.getScanRuntimeProvider(ScanRuntimeProviderContext.INSTANCE); if (provider instanceof SourceProvider) {...
3.68
framework_BeanItem_getBean
/** * Gets the underlying JavaBean object. * * @return the bean object. */ public BT getBean() { return bean; }
3.68
hadoop_AddMountAttributes_getMountTableForAddRequest
/** * Create a new mount table object from the given mount point and update its attributes. * * @param mountSrc mount point src. * @return MountTable object with updated attributes. * @throws IOException If mount table instantiation fails. */ private MountTable getMountTableForAddRequest(String mountSrc) throws I...
3.68
framework_VNotification_showError
/** * Shows an error notification and redirects the user to the given URL when * she clicks on the notification. * * If both message and caption are null, redirects the user to the url * immediately * * @since 7.5.1 * @param connection * A reference to the ApplicationConnection * @param caption * ...
3.68
hadoop_AccessTokenTimer_setExpiresInMSSinceEpoch
/** * Set when the access token will expire in milliseconds from epoch, * as required by the WebHDFS configuration. This is a bit hacky and lame. * * @param expiresInMSSinceEpoch Access time expiration in ms since epoch. */ public void setExpiresInMSSinceEpoch(String expiresInMSSinceEpoch){ this.nextRefreshMSSi...
3.68
open-banking-gateway_EncryptionWithInitVectorOper_decryption
/** * Decryption cipher * @param keyWithIv Symmetric key and initialization vector * @return Symmetric decryption cipher */ @SneakyThrows public Cipher decryption(SecretKeyWithIv keyWithIv) { Cipher cipher = Cipher.getInstance(encSpec.getCipherAlgo()); cipher.init( Cipher.DECRYPT_MODE, keyWithIv...
3.68
hudi_StreamerUtil_isValidFile
/** * Returns whether the give file is in valid hoodie format. * For example, filtering out the empty or corrupt files. */ public static boolean isValidFile(FileStatus fileStatus) { final String extension = FSUtils.getFileExtension(fileStatus.getPath().toString()); if (PARQUET.getFileExtension().equals(extension...
3.68
flink_TaskStateSnapshot_putSubtaskStateByOperatorID
/** * Maps the given operator id to the given subtask state. Returns the subtask state of a * previous mapping, if such a mapping existed or null otherwise. */ public OperatorSubtaskState putSubtaskStateByOperatorID( @Nonnull OperatorID operatorID, @Nonnull OperatorSubtaskState state) { return subtaskSt...
3.68
framework_GridSingleSelect_setUserSelectionAllowed
/** * Sets whether the user is allowed to change the selection. * <p> * The check is done only for the client side actions. It doesn't affect * selection requests sent from the server side. * * @param allowed * <code>true</code> if the user is allowed to change the * selection, <code>false...
3.68
framework_ComponentDetail_setTooltipInfo
/** * @param tooltipInfo * the tooltipInfo to set */ public void setTooltipInfo(TooltipInfo tooltipInfo) { this.tooltipInfo = tooltipInfo; }
3.68
pulsar_TopicsBase_publishMessagesToPartition
// Publish message to single partition of a partitioned topic. protected void publishMessagesToPartition(AsyncResponse asyncResponse, ProducerMessages request, boolean authoritative, int partition) { if (topicName.isPartitioned()) { asyncResponse.resume(new R...
3.68
streampipes_JdbcClient_closeAll
/** * Closes all open connections and statements of JDBC */ protected void closeAll() { boolean error = false; try { if (this.statementHandler.statement != null) { this.statementHandler.statement.close(); this.statementHandler.statement = null; } } catch (SQLException e) { error = true; ...
3.68
flink_StreamPhysicalPythonCorrelateRule_findTableFunction
// find only calc and table function private boolean findTableFunction(FlinkLogicalCalc calc) { RelNode child = ((RelSubset) calc.getInput()).getOriginal(); if (child instanceof FlinkLogicalTableFunctionScan) { FlinkLogicalTableFunctionScan scan = (FlinkLogicalTableFunctionScan) child; return Py...
3.68
framework_ServerRpcQueue_showLoadingIndicator
/** * Checks if a loading indicator should be shown when the RPCs have been * sent to the server and we are waiting for a response. * * @return true if a loading indicator should be shown, false otherwise */ public boolean showLoadingIndicator() { for (MethodInvocation invocation : getAll()) { if (isLe...
3.68
hbase_DigestSaslServerAuthenticationProvider_handle
/** {@inheritDoc} */ @Override public void handle(Callback[] callbacks) throws InvalidToken, UnsupportedCallbackException { NameCallback nc = null; PasswordCallback pc = null; AuthorizeCallback ac = null; for (Callback callback : callbacks) { if (callback instanceof AuthorizeCallback) { ac = (Authoriz...
3.68
hbase_UnsafeAccess_putLong
/** * Put a long value out to the specified BB position in big-endian format. * @param buf the byte buffer * @param offset position in the buffer * @param val long to write out * @return incremented offset */ public static int putLong(ByteBuffer buf, int offset, long val) { if (LITTLE_ENDIAN) { val = ...
3.68
hadoop_TimelineReaderClient_createTimelineReaderClient
/** * Create a new instance of Timeline Reader Client. * * @return instance of Timeline Reader Client. */ @InterfaceAudience.Public public static TimelineReaderClient createTimelineReaderClient() { return new TimelineReaderClientImpl(); }
3.68
flink_Channel_setReplicationFactor
/** * Sets the replication factor of the connection. * * @param factor The replication factor of the connection. */ public void setReplicationFactor(int factor) { this.replicationFactor = factor; }
3.68
flink_SocketStreamIterator_getPort
/** * Returns the port on which the iterator is getting the data. (Used internally.) * * @return The port */ public int getPort() { return socket.getLocalPort(); }
3.68
querydsl_JTSGeometryExpressions_extent
/** * Returns the bounding box that bounds rows of geometries. * * @param collection geometry collection * @return geometry collection */ public static JTSGeometryExpression<?> extent(Expression<? extends GeometryCollection> collection) { return geometryOperation(SpatialOps.EXTENT, collection); }
3.68
morf_SqlDialect_getSchemaName
/** * Returns the database schema name. May be null. * @return The schema name */ public String getSchemaName() { return schemaName; }
3.68
framework_FreeformQuery_containsRowWithKey
/** * This implementation of the containsRowWithKey method rewrites existing * WHERE clauses in the query string. The logic is, however, not very * complex and some times can do the Wrong Thing<sup>TM</sup>. For the * situations where this logic is not enough, you can implement the * getContainsRowQueryString meth...
3.68
flink_MemorySegment_getInt
/** * Reads an int value (32bit, 4 bytes) from the given position, in the system's native byte * order. This method offers the best speed for integer reading and should be used unless a * specific byte order is required. In most cases, it suffices to know that the byte order in * which the value is written is the s...
3.68
Activiti_TreeBuilderException_getExpression
/** * @return the expression string */ public String getExpression() { return expression; }
3.68