name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
morf_HumanReadableStatementHelper_generatePortableSqlStatementString
/** * Generates a human-readable description of a raw SQL operation. * * @param statement the data upgrade statement to describe. * @param preferredSQLDialect the SQL dialect to use, if available. If this is {@code null} or the preferred * dialect is not available in the statement bundle then an arbitrary...
3.68
hadoop_BondedS3AStatisticsContext_newInputStreamStatistics
/** * Create a stream input statistics instance. * The FileSystem.Statistics instance of the {@link #statisticsSource} * is used as the reference to FileSystem statistics to update * @return the new instance */ @Override public S3AInputStreamStatistics newInputStreamStatistics() { return getInstrumentation().new...
3.68
framework_Design_read
/** * Loads a design from the given input stream. * * @param design * The stream to read the design from * @return The root component of the design */ public static Component read(InputStream design) { DesignContext context = read(design, null); return context.getRootComponent(); }
3.68
framework_EventRouter_removeListener
/* * Removes the event listener method matching the given given parameters. * Don't add a JavaDoc comment here, we use the default documentation from * implemented interface. */ @Override @Deprecated public void removeListener(Class<?> eventType, Object target, String methodName) { // Find the correct ...
3.68
flink_GettingStartedExample_eval
// the 'eval()' method defines input and output types (reflectively extracted) // and contains the runtime logic public String eval(String street, String zipCode, String city) { return normalize(street) + ", " + normalize(zipCode) + ", " + normalize(city); }
3.68
hudi_TableServicePipeline_execute
/** * Run all table services task sequentially. */ public void execute() { tableServiceTasks.forEach(TableServiceTask::run); }
3.68
hbase_ServerSideScanMetrics_createCounter
/** * Create a new counter with the specified name * @return {@link AtomicLong} instance for the counter with counterName */ protected AtomicLong createCounter(String counterName) { AtomicLong c = new AtomicLong(0); counters.put(counterName, c); return c; }
3.68
dubbo_BaseServiceMetadata_getDisplayServiceKey
/** * Format : interface:version * * @return */ public String getDisplayServiceKey() { StringBuilder serviceNameBuilder = new StringBuilder(); serviceNameBuilder.append(serviceInterfaceName); serviceNameBuilder.append(COLON_SEPARATOR).append(version); return serviceNameBuilder.toString(); }
3.68
flink_ProjectOperator_types
/** @deprecated Deprecated method only kept for compatibility. */ @SuppressWarnings("unchecked") @Deprecated @PublicEvolving public <R extends Tuple> ProjectOperator<IN, R> types(Class<?>... types) { TupleTypeInfo<R> typeInfo = (TupleTypeInfo<R>) this.getResultType(); if (types.length != typeInfo.getArity()) {...
3.68
hbase_InclusiveStopFilter_toByteArray
/** Returns The filter serialized using pb */ @Override public byte[] toByteArray() { FilterProtos.InclusiveStopFilter.Builder builder = FilterProtos.InclusiveStopFilter.newBuilder(); if (this.stopRowKey != null) builder.setStopRowKey(UnsafeByteOperations.unsafeWrap(this.stopRowKey)); return builder.build...
3.68
framework_AriaHelper_ensureHasId
/** * Makes sure that the provided element has an id attribute. Adds a new * unique id if not. * * @param element * Element to check * @return String with the id of the element */ public static String ensureHasId(Element element) { assert element != null : "Valid Element required"; String id ...
3.68
hbase_HMaster_getSnapshotManager
/** Returns the underlying snapshot manager */ @Override public SnapshotManager getSnapshotManager() { return this.snapshotManager; }
3.68
shardingsphere-elasticjob_TransactionOperation_opCheckExists
/** * Operation check exists. * * @param key key * @return TransactionOperation */ public static TransactionOperation opCheckExists(final String key) { return new TransactionOperation(Type.CHECK_EXISTS, key, null); }
3.68
framework_AbstractMedia_setPreload
/** * Sets the preload attribute that is intended to provide a hint to the * browser how the media should be preloaded. Valid values are 'none', * 'metadata', 'preload', see the <a href= * "https://developer.mozilla.org/en/docs/Web/HTML/Element/video#attr-preload"> * Mozilla Developer Network</a> for details. * ...
3.68
framework_PointerDownEvent_getType
/** * Gets the event type associated with PointerDownEvent events. * * @return the handler type */ public static Type<PointerDownHandler> getType() { return TYPE; }
3.68
framework_ComboBox_setPopupWidth
/** * Sets the suggestion pop-up's width as a CSS string. By using relative * units (e.g. "50%") it's possible to set the popup's width relative to the * ComboBox itself. * * @see #getPopupWidth() * @since 7.7 * @param width * the width */ public void setPopupWidth(String width) { suggestionPopu...
3.68
hbase_MasterObserver_preDeleteTable
/** * Called before {@link org.apache.hadoop.hbase.master.HMaster} deletes a table. Called as part of * delete table RPC call. * @param ctx the environment to interact with the framework and master * @param tableName the name of the table */ default void preDeleteTable(final ObserverContext<MasterCoprocessor...
3.68
flink_SlotProfile_getPreferredLocations
/** Returns the preferred locations for the slot. */ public Collection<TaskManagerLocation> getPreferredLocations() { return preferredLocations; }
3.68
hbase_RegionCoprocessorHost_preBulkLoadHFile
/** * @param familyPaths pairs of { CF, file path } submitted for bulk load */ public void preBulkLoadHFile(final List<Pair<byte[], String>> familyPaths) throws IOException { execOperation(coprocEnvironments.isEmpty() ? null : new RegionObserverOperationWithoutResult() { @Override public void call(RegionObs...
3.68
dubbo_CtClassBuilder_getQualifiedClassName
/** * get full qualified class name * * @param className super class name, maybe qualified or not */ protected String getQualifiedClassName(String className) { if (className.contains(".")) { return className; } if (fullNames.containsKey(className)) { return fullNames.get(className); ...
3.68
hadoop_ExecutorServiceFuturePool_shutdown
/** * Utility to shutdown the {@link ExecutorService} used by this class. Will wait up to a * certain timeout for the ExecutorService to gracefully shutdown. * * @param logger Logger * @param timeout the maximum time to wait * @param unit the time unit of the timeout argument */ public void shutdown(Logger logge...
3.68
flink_DataTypeTemplate_isForceAnyPattern
/** Returns whether the given class must be treated as RAW type. */ boolean isForceAnyPattern(@Nullable Class<?> clazz) { if (forceRawPattern == null || clazz == null) { return false; } final String className = clazz.getName(); for (String pattern : forceRawPattern) { if (className.start...
3.68
hudi_HoodieFlinkWriteClient_waitForCleaningFinish
/** * Blocks and wait for the async cleaning service to finish. * * <p>The Flink write client is designed to write data set as buckets * but cleaning action should trigger after all the write actions within a * checkpoint finish. */ public void waitForCleaningFinish() { if (tableServiceClient.asyncCleanerServic...
3.68
hbase_StorageClusterStatusModel_getRootIndexSizeKB
/** Returns The current total size of root-level indexes for the region, in KB. */ @XmlAttribute public int getRootIndexSizeKB() { return rootIndexSizeKB; }
3.68
hudi_HoodieTableMetadataUtil_deleteMetadataTablePartition
/** * Delete a partition within the metadata table. * <p> * This can be used to delete a partition so that it can be re-bootstrapped. * * @param dataMetaClient {@code HoodieTableMetaClient} of the dataset for which metadata table is to be deleted * @param context instance of {@code HoodieEngineContext}. *...
3.68
hadoop_ShortWritable_equals
/** Returns true iff <code>o</code> is a ShortWritable with the same value. */ @Override public boolean equals(Object o) { if (!(o instanceof ShortWritable)) return false; ShortWritable other = (ShortWritable) o; return this.value == other.value; }
3.68
flink_HighAvailabilityServices_getWebMonitorLeaderRetriever
/** * This retriever should no longer be used on the cluster side. The web monitor retriever is * only required on the client-side and we have a dedicated high-availability services for the * client, named {@link ClientHighAvailabilityServices}. See also FLINK-13750. * * @return the leader retriever for web monito...
3.68
hbase_SaslAuthMethod_getAuthMethod
/** * Returns the Hadoop {@link AuthenticationMethod} for this method. */ public AuthenticationMethod getAuthMethod() { return method; }
3.68
pulsar_TxnMetaImpl_checkTxnStatus
/** * Check if the transaction is in an expected status. * * @param expectedStatus */ private synchronized void checkTxnStatus(TxnStatus expectedStatus) throws InvalidTxnStatusException { if (this.txnStatus != expectedStatus) { throw new InvalidTxnStatusException( txnID, expectedStatus, txnS...
3.68
AreaShop_Utils_getImportantWorldEditRegions
/** * Get a list of regions around a location. * - Returns highest priority, child instead of parent regions * @param location The location to check for regions * @return empty list if no regions found, 1 member if 1 region is a priority, more if regions with the same priority */ public static List<ProtectedRegion...
3.68
hbase_ClientMetaTableAccessor_getResults
/** Returns Collected results; wait till visits complete to collect all possible results */ List<T> getResults() { return this.results; }
3.68
hudi_CleanPlanner_getSavepointedDataFiles
/** * Get the list of data file names savepointed. */ public Stream<String> getSavepointedDataFiles(String savepointTime) { if (!hoodieTable.getSavepointTimestamps().contains(savepointTime)) { throw new HoodieSavepointException( "Could not get data files for savepoint " + savepointTime + ". No such save...
3.68
AreaShop_AreaShop_onEnable
/** * Called on start or reload of the server. */ @Override public void onEnable() { AreaShop.instance = this; Do.init(this); managers = new HashSet<>(); boolean error = false; // Find WorldEdit integration version to load String weVersion = null; String rawWeVersion = null; String weBeta = null; Plugin plu...
3.68
flink_JoinedStreams_apply
/** * Completes the join operation with the user function that is executed for each combination * of elements with the same key in a window. * * <p>Note: This method's return type does not support setting an operator-specific * parallelism. Due to binary backwards compatibility, this cannot be altered. Use the * ...
3.68
framework_VComboBox_updateSelectedIconPosition
/** * Positions the icon vertically in the middle. Should be called after the * icon has loaded */ private void updateSelectedIconPosition() { // Position icon vertically to middle int availableHeight = 0; availableHeight = getOffsetHeight(); int iconHeight = WidgetUtil.getRequiredHeight(selectedIte...
3.68
rocketmq-connect_Worker_stop
/** * We can choose to persist in-memory task status * so we can view history tasks */ public void stop() { // stop and await connectors if (!connectors.isEmpty()) { log.warn("Shutting down connectors {} uncleanly; herder should have shut down connectors before the Worker is stopped", connectors.key...
3.68
flink_EnrichedRowData_from
/** * Creates a new {@link EnrichedRowData} with the provided {@code fixedRow} as the immutable * static row, and uses the {@code producedRowFields}, {@code fixedRowFields} and {@code * mutableRowFields} arguments to compute the indexes mapping. * * <p>The {@code producedRowFields} should include the name of field...
3.68
flink_S3TestCredentials_getS3SecretKey
/** * Gets the S3 Secret Key. * * <p>This method throws an exception if the key is not available. Tests should use {@link * #assumeCredentialsAvailable()} to skip tests when credentials are not available. */ public static String getS3SecretKey() { if (S3_TEST_SECRET_KEY != null) { return S3_TEST_SECRET...
3.68
morf_Cast_as
/** * @see org.alfasoftware.morf.sql.element.AliasedField#as(java.lang.String) */ @Override public Cast as(String aliasName) { return (Cast) super.as(aliasName); }
3.68
flink_RoundRobinOperatorStateRepartitioner_repartitionBroadcastState
/** Repartition BROADCAST state. */ private void repartitionBroadcastState( Map<String, List<Tuple2<StreamStateHandle, OperatorStateHandle.StateMetaInfo>>> broadcastState, List<Map<StreamStateHandle, OperatorStateHandle>> mergeMapList) { int newParallelism = mergeMapList.size(); ...
3.68
flink_CrossOperator_projectTuple6
/** * Projects a pair of crossed elements to a {@link Tuple} with the previously selected * fields. * * @return The projected data set. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5> ProjectCross<I1, I2, Tuple6<T0, T1, T2, T3, T4, T5>> projectTuple6() { TypeInformation<?>[] fTypes = e...
3.68
morf_SqlUtils_substring
/** * Returns a SQL DSL expression to return the substring of <strong>field</strong> * of <strong>length</strong> characters, starting from <strong>from</strong>. * * @param field The source expression. * @param from The start character offset. * @param length The length of the substring. * @return The SQL DSL e...
3.68
hadoop_ManifestStoreOperationsThroughFileSystem_storePreservesEtagsThroughRenames
/** * Probe filesystem capabilities. * @param path path to probe. * @return true if the FS declares its renames work. */ @Override public boolean storePreservesEtagsThroughRenames(Path path) { try { return fileSystem.hasPathCapability(path, CommonPathCapabilities.ETAGS_PRESERVED_IN_RENAME); } catch ...
3.68
framework_Tree_setItemDescriptionGenerator
/** * Set the item description generator which generates tooltips for the tree * items. * * @param generator * The generator to use or null to disable */ public void setItemDescriptionGenerator( ItemDescriptionGenerator generator) { if (generator != itemDescriptionGenerator) { itemD...
3.68
MagicPlugin_MagicController_getNPCSuppliers
/** * @return The supplier set that is used. */ public NPCSupplierSet getNPCSuppliers() { return npcSuppliers; }
3.68
morf_AliasedField_getImpliedName
/** * Returns the name of the field either implied by its source or by its alias. * @return The implied name of the field */ public String getImpliedName() { return getAlias(); }
3.68
hudi_HoodieTableFactory_inferAvroSchema
/** * Inferences the deserialization Avro schema from the table schema (e.g. the DDL) * if both options {@link FlinkOptions#SOURCE_AVRO_SCHEMA_PATH} and * {@link FlinkOptions#SOURCE_AVRO_SCHEMA} are not specified. * * @param conf The configuration * @param rowType The specified table row type */ private stati...
3.68
hudi_DFSHoodieDatasetInputReader_iteratorSize
/** * Returns the number of elements remaining in {@code iterator}. The iterator will be left exhausted: its {@code hasNext()} method will return {@code false}. */ private static int iteratorSize(Iterator<?> iterator) { int count = 0; while (iterator.hasNext()) { iterator.next(); count++; } return cou...
3.68
querydsl_BeanMap_reinitialise
/** * Reinitializes this bean. Called during {@link #setBean(Object)}. * Does introspection to find properties. */ protected void reinitialise() { readMethods.clear(); writeMethods.clear(); types.clear(); initialise(); }
3.68
hadoop_ProxyCombiner_combine
/** * Combine two or more proxies which together comprise a single proxy * interface. This can be used for a protocol interface which {@code extends} * multiple other protocol interfaces. The returned proxy will implement * all of the methods of the combined proxy interface, delegating calls * to which proxy imple...
3.68
hadoop_AbfsClient_checkUserError
/** * Returns true if the status code lies in the range of user error. * @param responseStatusCode http response status code. * @return True or False. */ private boolean checkUserError(int responseStatusCode) { return (responseStatusCode >= HttpURLConnection.HTTP_BAD_REQUEST && responseStatusCode < HttpURLC...
3.68
morf_SqlDialect_getSqlFrom
/** * Convert a {@link WindowFunction} into standards compliant SQL. * @param windowFunctionField The field to convert * @return The resulting SQL **/ protected String getSqlFrom(WindowFunction windowFunctionField) { if (requiresOrderByForWindowFunction(windowFunctionField.getFunction()) && windowFunctionField.g...
3.68
flink_AbstractParameterTool_getShort
/** * Returns the Short value for the given key. If the key does not exists it will return the * default value given. The method fails if the value is not a Short. */ public short getShort(String key, short defaultValue) { addToDefaults(key, Short.toString(defaultValue)); String value = get(key); if (val...
3.68
morf_ConnectionResourcesBean_setHostName
/** * @see org.alfasoftware.morf.jdbc.AbstractConnectionResources#setHostName(java.lang.String) */ @Override public void setHostName(String hostName) { this.hostName = hostName; }
3.68
morf_AbstractSelectStatementBuilder_getJoins
/** * Gets the list of joined tables in the order they are joined * * @return the joined tables */ List<Join> getJoins() { return joins; }
3.68
AreaShop_BuyRegion_getPlayerName
/** * Get the name of the player that owns this region. * @return The name of the player that owns this region, if unavailable by UUID it will return the old cached name, if that is unavailable it will return &lt;UNKNOWN&gt; */ public String getPlayerName() { String result = Utils.toName(getBuyer()); if(result == ...
3.68
hadoop_StoragePolicySatisfyManager_removeAllPathIds
/** * Clean up all sps path ids. */ public void removeAllPathIds() { synchronized (pathsToBeTraversed) { pathsToBeTraversed.clear(); } }
3.68
flink_ResourceManager_deregisterApplication
/** * Cleanup application and shut down cluster. * * @param finalStatus of the Flink application * @param diagnostics diagnostics message for the Flink application or {@code null} */ @Override public CompletableFuture<Acknowledge> deregisterApplication( final ApplicationStatus finalStatus, @Nullable final ...
3.68
hadoop_OuterJoinRecordReader_combine
/** * Emit everything from the collector. */ protected boolean combine(Object[] srcs, TupleWritable dst) { assert srcs.length == dst.size(); return true; }
3.68
hadoop_CSQueueStore_add
/** * Method for adding a queue to the store. * @param queue Queue to be added */ public void add(CSQueue queue) { String fullName = queue.getQueuePath(); String shortName = queue.getQueueShortName(); try { modificationLock.writeLock().lock(); fullNameQueues.put(fullName, queue); getMap.put(fullN...
3.68
dubbo_ReferenceBean_getVersion
/** * The version of the service */ public String getVersion() { // Compatible with seata-1.4.0: io.seata.rm.tcc.remoting.parser.DubboRemotingParser#getServiceDesc() return referenceConfig.getVersion(); }
3.68
hadoop_BoundedResourcePool_release
/** * Releases a previously acquired resource. * * @throws IllegalArgumentException if item is null. */ @Override public void release(T item) { checkNotNull(item, "item"); synchronized (createdItems) { if (!createdItems.contains(item)) { throw new IllegalArgumentException("This item is not a part of ...
3.68
framework_DesignContext_getComponentLocalId
/** * Returns the local id for a component. * * @since 7.5.0 * * @param component * The component whose local id to get. * @return the local id of the component, or null if the component has no * local id assigned */ public String getComponentLocalId(Component component) { return compone...
3.68
hadoop_SnappyCompressor_getBytesRead
/** * Return number of bytes given to this compressor since last reset. */ @Override public long getBytesRead() { return bytesRead; }
3.68
flink_SqlAlterTable_getPartitionKVs
/** Get partition spec as key-value strings. */ public LinkedHashMap<String, String> getPartitionKVs() { return SqlPartitionUtils.getPartitionKVs(getPartitionSpec()); }
3.68
querydsl_SQLExpressions_lead
/** * expr evaluated at the row that is one row after the current row within the partition; * * @param expr expression * @return lead(expr) */ public static <T> WindowOver<T> lead(Expression<T> expr) { return new WindowOver<T>(expr.getType(), SQLOps.LEAD, expr); }
3.68
hbase_LockAndQueue_releaseExclusiveLock
/** Returns whether we should wake the procedures waiting on the lock here. */ public boolean releaseExclusiveLock(Procedure<?> proc) { if ( exclusiveLockOwnerProcedure == null || exclusiveLockOwnerProcedure.getProcId() != proc.getProcId() ) { // We are not the lock owner, it is probably inherited fro...
3.68
hbase_SnapshotScannerHDFSAclHelper_revokeAcl
/** * Remove acl when grant or revoke user permission * @param userPermission the user and permission * @param skipNamespaces the namespace set to skip remove acl * @param skipTables the table set to skip remove acl * @return false if an error occurred, otherwise true */ public boolean revokeAcl(UserPermissio...
3.68
morf_OracleDialect_selectStatementPreFieldDirectives
/** * @see org.alfasoftware.morf.jdbc.SqlDialect#selectStatementPreFieldDirectives(org.alfasoftware.morf.sql.SelectStatement) */ @Override protected String selectStatementPreFieldDirectives(SelectStatement selectStatement) { StringBuilder builder = new StringBuilder(); for (Hint hint : selectStatement.getHints()...
3.68
hbase_BlockIOUtils_readFullyWithHeapBuffer
/** * Copying bytes from InputStream to {@link ByteBuff} by using an temporary heap byte[] (default * size is 1024 now). * @param in the InputStream to read * @param out the destination {@link ByteBuff} * @param length to read * @throws IOException if any io error encountered. */ public static void readFu...
3.68
framework_View_getViewComponent
/** * Gets the component to show when navigating to the view. * * By default casts this View to a {@link Component} if possible, otherwise * throws an IllegalStateException. * * @since 8.1 * @return the component to show, by default the view instance itself */ public default Component getViewComponent() { i...
3.68
hadoop_NamenodeStatusReport_getNumDecommissioningDatanodes
/** * Get the number of decommissionining nodes. * * @return The number of decommissionining nodes. */ public int getNumDecommissioningDatanodes() { return this.decomDatanodes; }
3.68
hmily_MetricsReporter_counterIncrement
/** * Counter increment by count. * * @param name name * @param labelValues label values * @param count count */ public static void counterIncrement(final String name, final String[] labelValues, final long count) { Optional.ofNullable(metricsRegister).ifPresent(register -> register.counterIncrement(name, lab...
3.68
flink_ExceptionUtils_assertThrowableWithMessage
/** * The same as {@link #findThrowableWithMessage(Throwable, String)}, but rethrows original * exception if the expected exception was not found. */ public static <T extends Throwable> void assertThrowableWithMessage( Throwable throwable, String searchMessage) throws T { if (!findThrowableWithMessage(th...
3.68
hadoop_S3AReadOpContext_withAuditSpan
/** * Set builder value. * @param value new value * @return the builder */ public S3AReadOpContext withAuditSpan(final AuditSpan value) { auditSpan = value; return this; }
3.68
flink_SlotSharingGroup_setManagedMemoryMB
/** Set the task managed memory for this SlotSharingGroup in MB. */ public Builder setManagedMemoryMB(int managedMemoryMB) { this.managedMemory = MemorySize.ofMebiBytes(managedMemoryMB); return this; }
3.68
pulsar_TestRetrySupport_incrementSetupNumber
/** * This method should be called in the setup method of the concrete class. * * This increases an internal counter and resets the failure state which are used to determine * whether cleanup is needed before a test method is called. * */ protected final void incrementSetupNumber() { currentSetupNumber++; ...
3.68
flink_HybridSource_build
/** Build the source. */ public HybridSource<T> build() { return new HybridSource(sources); }
3.68
hbase_RegionReplicaUtil_addReplicas
/** * Create any replicas for the regions (the default replicas that was already created is passed to * the method) * @param regions existing regions * @param oldReplicaCount existing replica count * @param newReplicaCount updated replica count due to modify table * @return the combined list of default an...
3.68
framework_VCustomLayout_clear
/** Clear all widgets from the layout. */ @Override public void clear() { super.clear(); locationToWidget.clear(); childWidgetToCaptionWrapper.clear(); }
3.68
hbase_HMaster_isOnline
/** * Report whether this master is started This method is used for testing. * @return true if master is ready to go, false if not. */ public boolean isOnline() { return serviceStarted; }
3.68
hudi_SparkInternalSchemaConverter_collectColNamesFromSparkStruct
/** * Collect all the leaf nodes names. * * @param sparkSchema a spark schema * @return leaf nodes full names. */ public static List<String> collectColNamesFromSparkStruct(StructType sparkSchema) { List<String> result = new ArrayList<>(); collectColNamesFromStructType(sparkSchema, new LinkedList<>(), result);...
3.68
hudi_AvroSchemaUtils_isNullable
/** * Returns true in case provided {@link Schema} is nullable (ie accepting null values), * returns false otherwise */ public static boolean isNullable(Schema schema) { if (schema.getType() != Schema.Type.UNION) { return false; } List<Schema> innerTypes = schema.getTypes(); return innerTypes.size() > 1...
3.68
hbase_UnsafeAccess_toShort
/** * Reads a short value at the given Object's offset considering it was written in big-endian * format. * @return short value at offset */ public static short toShort(Object ref, long offset) { if (LITTLE_ENDIAN) { return Short.reverseBytes(HBasePlatformDependent.getShort(ref, offset)); } return HBasePl...
3.68
framework_StateChangeEvent_getChangedPropertiesFastSet
/** * Gets the properties that have changed. * * @return a set of names of the changed properties * * @deprecated As of 7.0.1, use {@link #hasPropertyChanged(String)} instead * for improved performance. */ @Deprecated public FastStringSet getChangedPropertiesFastSet() { if (changedProperties == n...
3.68
flink_AsynchronousFileIOChannel_registerAllRequestsProcessedListener
/** * Registers a listener to be notified when all outstanding requests have been processed. * * <p>New requests can arrive right after the listener got notified. Therefore, it is not safe * to assume that the number of outstanding requests is still zero after a notification unless * there was a close right before...
3.68
hbase_SingleColumnValueFilter_parseFrom
/** * Parse a serialized representation of {@link SingleColumnValueFilter} * @param pbBytes A pb serialized {@link SingleColumnValueFilter} instance * @return An instance of {@link SingleColumnValueFilter} made from <code>bytes</code> * @throws DeserializationException if an error occurred * @see #toByteArray */ ...
3.68
hbase_MetricsTableRequests_updateCheckAndDelete
/** * Update the CheckAndDelete time histogram. * @param time time it took */ public void updateCheckAndDelete(long time) { if (isEnableTableLatenciesMetrics()) { checkAndDeleteTimeHistogram.update(time); } }
3.68
flink_AbstractCheckpointStats_getTaskStateStats
/** * Returns the task state stats for the given job vertex ID or <code>null</code> if no task with * such an ID is available. * * @param jobVertexId Job vertex ID of the task stats to look up. * @return The task state stats instance for the given ID or <code>null</code>. */ public TaskStateStats getTaskStateStat...
3.68
hadoop_RpcProgramPortmap_set
/** * When a program first becomes available on a machine, it registers itself * with the port mapper program on the same machine. The program passes its * program number "prog", version number "vers", transport protocol number * "prot", and the port "port" on which it awaits service request. The * procedure retur...
3.68
hbase_CacheConfig_setEvictOnClose
/** * Only used for testing. * @param evictOnClose whether blocks should be evicted from the cache when an HFile reader is * closed */ public void setEvictOnClose(boolean evictOnClose) { this.evictOnClose = evictOnClose; }
3.68
hudi_CleanActionExecutor_runPendingClean
/** * Executes the Cleaner plan stored in the instant metadata. */ HoodieCleanMetadata runPendingClean(HoodieTable<T, I, K, O> table, HoodieInstant cleanInstant) { try { HoodieCleanerPlan cleanerPlan = CleanerUtils.getCleanerPlan(table.getMetaClient(), cleanInstant); return runClean(table, cleanInstant, cle...
3.68
druid_IPRange_computeMaskFromNetworkPrefix
/** * Convert a extended network prefix integer into an IP number. * * @param prefix The network prefix number. * @return Return the IP number corresponding to the extended network prefix. */ private IPAddress computeMaskFromNetworkPrefix(int prefix) { /* * int subnet = 0; for (int i=0; i<prefix; i++) { s...
3.68
hadoop_ApplicationMaster_main
/** * @param args Command line args */ public static void main(String[] args) { boolean result = false; try { ApplicationMaster appMaster = new ApplicationMaster(); LOG.info("Initializing ApplicationMaster"); boolean doRun = appMaster.init(args); if (!doRun) { System.exit(0); } resul...
3.68
hadoop_ManifestCommitter_abortTask
/** * Abort a task. * @param context task context * @throws IOException failure during the delete */ @Override public void abortTask(final TaskAttemptContext context) throws IOException { ManifestCommitterConfig committerConfig = enterCommitter(true, context); try { new AbortTaskStage( com...
3.68
framework_ServerRpcManager_applyInvocation
/** * Invoke a method in a server side RPC target class. This method is to be * used by the RPC framework and unit testing tools only. * * @param invocation * method invocation to perform */ public void applyInvocation(ServerRpcMethodInvocation invocation) throws RpcInvocationException { Me...
3.68
morf_OracleDialect_getSqlForAddDays
/** * @see org.alfasoftware.morf.jdbc.SqlDialect#getSqlForAddDays(org.alfasoftware.morf.sql.element.Function) */ @Override protected String getSqlForAddDays(Function function) { return String.format( "(%s) + (%s)", getSqlFrom(function.getArguments().get(0)), getSqlFrom(function.getArguments().get(1)) ...
3.68
framework_AbstractLegacyComponent_isImmediate
/** * Returns the immediate mode of the component. * <p> * Since Vaadin 8, the default mode is immediate. * * @return true if the component is in immediate mode (explicitly or * implicitly set), false if the component if not in immediate mode */ public boolean isImmediate() { if (explicitImmediateVal...
3.68
hadoop_ResourceRequest_executionType
/** * Set the <code>executionTypeRequest</code> of the request with 'ensure * execution type' flag set to true. * @see ResourceRequest#setExecutionTypeRequest( * ExecutionTypeRequest) * @param executionType <code>executionType</code> of the request. * @return {@link ResourceRequestBuilder} */ @Public @Evolving p...
3.68
querydsl_Expressions_stringOperation
/** * Create a new Operation expression * * @param operator operator * @param args operation arguments * @return operation expression */ public static StringOperation stringOperation(Operator operator, Expression<?>... args) { return new StringOperation(operator, args); }
3.68