name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
flink_StringUtils_writeString
/** * Writes a String to the given output. The written string can be read with {@link * #readString(DataInputView)}. * * @param str The string to write * @param out The output to write to * @throws IOException Thrown, if the writing or the serialization fails. */ public static void writeString(@Nonnull String st...
3.68
hbase_HBaseConfiguration_merge
/** * Merge two configurations. * @param destConf the configuration that will be overwritten with items from the srcConf * @param srcConf the source configuration **/ public static void merge(Configuration destConf, Configuration srcConf) { for (Map.Entry<String, String> e : srcConf) { destConf.set(e.getKey(...
3.68
framework_VFilterSelect_updatePopupPositionOnScroll
/** * Make the popup follow the position of the ComboBox when the page is * scrolled. */ private void updatePopupPositionOnScroll() { if (!scrollPending) { AnimationScheduler.get() .requestAnimationFrame(new AnimationCallback() { public void execute(double timestamp) {...
3.68
flink_StreamExecutionEnvironment_registerJobListener
/** * Register a {@link JobListener} in this environment. The {@link JobListener} will be notified * on specific job status changed. */ @PublicEvolving public void registerJobListener(JobListener jobListener) { checkNotNull(jobListener, "JobListener cannot be null"); jobListeners.add(jobListener); }
3.68
flink_FutureUtils_composeAfterwards
/** * Run the given asynchronous action after the completion of the given future. The given future * can be completed normally or exceptionally. In case of an exceptional completion, the * asynchronous action's exception will be added to the initial exception. * * @param future to wait for its completion * @param...
3.68
framework_HierarchicalDataProvider_fetch
/** * Fetches data from this HierarchicalDataProvider using given * {@code query}. Only the immediate children of * {@link HierarchicalQuery#getParent()} will be returned. * * @param query * given query to request data with * @return a stream of data objects resulting from the query * * @throws Ille...
3.68
hbase_RegionSplitter_split2
/** * Divide 2 numbers in half (for split algorithm) * @param a number #1 * @param b number #2 * @return the midpoint of the 2 numbers */ public BigInteger split2(BigInteger a, BigInteger b) { return a.add(b).divide(BigInteger.valueOf(2)).abs(); }
3.68
hudi_ClusteringCommand_runClustering
/** * Run clustering table service. * <p> * Example: * > connect --path {path to hudi table} * > clustering scheduleAndExecute --sparkMaster local --sparkMemory 2g */ @ShellMethod(key = "clustering scheduleAndExecute", value = "Run Clustering. Make a cluster plan first and execute that plan immediately") public S...
3.68
querydsl_NumberExpression_divide
/** * Create a {@code this / right} expression * * <p>Get the result of the operation this / right</p> * * @param right * @return this / right */ public <N extends Number & Comparable<?>> NumberExpression<T> divide(N right) { @SuppressWarnings("unchecked") Class<T> type = (Class< T>) getDivisionType(getT...
3.68
hbase_QuotaUtil_disableTableIfNotDisabled
/** * Method to disable a table, if not already disabled. This method suppresses * {@link TableNotEnabledException}, if thrown while disabling the table. * @param conn connection to re-use * @param tableName table name which has moved into space quota violation */ public static void disableTableIfNotDisabled(...
3.68
hibernate-validator_ValidationXmlParser_run
/** * Runs the given privileged action, using a privileged block if required. * <p> * <b>NOTE:</b> This must never be changed into a publicly available method to avoid execution of arbitrary * privileged actions within HV's protection domain. */ @IgnoreForbiddenApisErrors(reason = "SecurityManager is deprecated in...
3.68
hbase_RegionStateNode_offline
/** * Put region into OFFLINE mode (set state and clear location). * @return Last recorded server deploy */ public ServerName offline() { setState(State.OFFLINE); return setRegionLocation(null); }
3.68
flink_TableConfig_setPlannerConfig
/** * Sets the configuration of Planner for Table API and SQL queries. Changing the configuration * has no effect after the first query has been defined. */ public void setPlannerConfig(PlannerConfig plannerConfig) { this.plannerConfig = Preconditions.checkNotNull(plannerConfig); }
3.68
framework_StringToBooleanConverter_getTrueString
/** * Gets the locale-depended string representation for true. Default is * locale-independent value provided by {@link #getTrueString()} * * @since 7.5.4 * @param locale * to be used * @return the string representation for true */ protected String getTrueString(Locale locale) { return getTrueStr...
3.68
hbase_ProcedureExecutor_submitProcedure
// ========================================================================== // Submit/Abort Procedure // ========================================================================== /** * Add a new root-procedure to the executor. * @param proc the new procedure to execute. * @return the procedure id, that can be use...
3.68
flink_StreamOperatorWrapper_close
/** Close the operator. */ public void close() throws Exception { closed = true; wrapped.close(); }
3.68
querydsl_StringExpression_trim
/** * Create a {@code this.trim()} expression * * <p>Create a copy of the string, with leading and trailing whitespace * omitted.</p> * * @return this.trim() * @see java.lang.String#trim() */ public StringExpression trim() { if (trim == null) { trim = Expressions.stringOperation(Ops.TRIM, mixin); ...
3.68
hadoop_FederationMembershipStateStoreInputValidator_checkTimestamp
/** * Validate if the timestamp is positive or not. * * @param timestamp the timestamp to be verified * @throws FederationStateStoreInvalidInputException if the timestamp is * invalid */ private static void checkTimestamp(long timestamp) throws FederationStateStoreInvalidInputException { if (times...
3.68
rocketmq-connect_WorkerSinkTask_pauseAll
//pause all consumer topic queue private void pauseAll() { consumer.pause(messageQueues); }
3.68
hadoop_NamenodeStatusReport_statsValid
/** * If the statistics are valid. * * @return If the statistics are valid. */ public boolean statsValid() { return this.statsValid; }
3.68
framework_AbsoluteLayoutRelativeSizeContent_createFullOnFixed
/** * Creates an {@link AbsoluteLayout} of fixed size that contains a * fixed-sized {@link AbsoluteLayout}. * * @return the created layout */ private Component createFullOnFixed() { AbsoluteLayout absoluteLayout = new AbsoluteLayout(); absoluteLayout.setWidth(200, Unit.PIXELS); absoluteLayout.setHeight...
3.68
hadoop_BaseService_getServiceDependencies
/** * Returns the service dependencies of this service. The service will be * instantiated only if all the service dependencies are already initialized. * <p> * This method returns an empty array (size 0) * * @return an empty array (size 0). */ @Override public Class[] getServiceDependencies() { return new Cla...
3.68
framework_CellReference_set
/** * Sets the identifying information for this cell. * <p> * The difference between {@link #columnIndexDOM} and {@link #columnIndex} * comes from hidden columns. * * @param columnIndexDOM * the index of the column in the DOM * @param columnIndex * the index of the column * @param column...
3.68
flink_CsvBulkWriter_forPojo
/** * Builds a writer based on a POJO class definition. * * @param pojoClass The class of the POJO. * @param stream The output stream. * @param <T> The type of the elements accepted by this writer. */ static <T> CsvBulkWriter<T, T, Void> forPojo(Class<T> pojoClass, FSDataOutputStream stream) { final Converter...
3.68
zxing_PDF417Writer_bitMatrixFromEncoder
/** * Takes encoder, accounts for width/height, and retrieves bit matrix */ private static BitMatrix bitMatrixFromEncoder(PDF417 encoder, String contents, int errorCorrectionLevel, ...
3.68
flink_WindowSavepointReader_process
/** * Reads window state generated without any preaggregation such as {@code WindowedStream#apply} * and {@code WindowedStream#process}. * * @param uid The uid of the operator. * @param readerFunction The window reader function. * @param keyType The key type of the window. * @param stateType The type of records ...
3.68
framework_VTree_getNextSibling
/** * Gets the next sibling in the tree * * @param node * The node to get the sibling for * @return The sibling node or null if the node is the last sibling */ private TreeNode getNextSibling(TreeNode node) { TreeNode parent = node.getParentNode(); List<TreeNode> children; if (parent == nul...
3.68
flink_RemoteInputChannel_notifyBufferAvailable
/** * The unannounced credit is increased by the given amount and might notify increased credit to * the producer. */ @Override public void notifyBufferAvailable(int numAvailableBuffers) throws IOException { if (numAvailableBuffers > 0 && unannouncedCredit.getAndAdd(numAvailableBuffers) == 0) { notifyCre...
3.68
flink_OrcLegacyTimestampColumnVector_createFromConstant
// creates a Hive ColumnVector of constant timestamp value public static ColumnVector createFromConstant(int batchSize, Object value) { LongColumnVector res = new LongColumnVector(batchSize); if (value == null) { res.noNulls = false; res.isNull[0] = true; res.isRepeating = true; } el...
3.68
framework_AbstractColorPicker_getSwatchesVisibility
/** * Gets the visibility of the Swatches Tab. * * @since 7.5.0 * @return visibility of the swatches tab */ public boolean getSwatchesVisibility() { return swatchesVisible; }
3.68
flink_AbstractKeyedStateBackend_getOrCreateKeyedState
/** @see KeyedStateBackend */ @Override @SuppressWarnings("unchecked") public <N, S extends State, V> S getOrCreateKeyedState( final TypeSerializer<N> namespaceSerializer, StateDescriptor<S, V> stateDescriptor) throws Exception { checkNotNull(namespaceSerializer, "Namespace serializer"); checkNo...
3.68
graphhopper_MiniPerfTest_getMin
/** * @return minimum time of every call, in ms */ public double getMin() { return min / NS_PER_MS; }
3.68
hadoop_BufferData_setDone
/** * Indicates that this block is no longer of use and can be reclaimed. */ public synchronized void setDone() { if (this.checksum != 0) { if (getChecksum(this.buffer) != this.checksum) { throw new IllegalStateException("checksum changed after setReady()"); } } this.state = State.DONE; this.act...
3.68
framework_VTabsheet_setFocusedTab
/** * Sets the tab that has the focus currently. * * @param focusedTab * the focused tab or {@code null} if no tab should be * focused anymore */ private void setFocusedTab(Tab focusedTab) { this.focusedTab = focusedTab; }
3.68
flink_BinarySegmentUtils_readBinary
/** * Get binary, if len less than 8, will be include in variablePartOffsetAndLen. * * <p>Note: Need to consider the ByteOrder. * * @param baseOffset base offset of composite binary format. * @param fieldOffset absolute start offset of 'variablePartOffsetAndLen'. * @param variablePartOffsetAndLen a long value, r...
3.68
framework_FocusableFlowPanel_addKeyPressHandler
/* * (non-Javadoc) * * @see * com.google.gwt.event.dom.client.HasKeyPressHandlers#addKeyPressHandler * (com.google.gwt.event.dom.client.KeyPressHandler) */ @Override public HandlerRegistration addKeyPressHandler(KeyPressHandler handler) { return addDomHandler(handler, KeyPressEvent.getType()); }
3.68
hadoop_TextView_echoWithoutEscapeHtml
/** * Print strings as is (no newline, a la php echo). * @param args the strings to print */ public void echoWithoutEscapeHtml(Object... args) { PrintWriter out = writer(); for (Object s : args) { out.print(s); } }
3.68
hbase_SegmentFactory_createImmutableSegmentByCompaction
// create new flat immutable segment from compacting old immutable segments // for compaction public ImmutableSegment createImmutableSegmentByCompaction(final Configuration conf, final CellComparator comparator, MemStoreSegmentsIterator iterator, int numOfCells, CompactingMemStore.IndexType idxType, MemStoreCompact...
3.68
flink_ProjectOperator_projectTuple8
/** * Projects a {@link Tuple} {@link DataSet} to the previously selected fields. * * @return The projected DataSet. * @see Tuple * @see DataSet */ public <T0, T1, T2, T3, T4, T5, T6, T7> ProjectOperator<T, Tuple8<T0, T1, T2, T3, T4, T5, T6, T7>> projectTuple8() { TypeInformation<?>[] fTypes = extract...
3.68
hadoop_ExitUtil_getFirstExitException
/** * @return the first {@code ExitException} thrown, null if none thrown yet. */ public static ExitException getFirstExitException() { return FIRST_EXIT_EXCEPTION.get(); }
3.68
flink_SplitFetcher_pauseOrResumeSplits
/** * Called when some splits of this source instance progressed too much beyond the global * watermark of all subtasks. If the split reader implements {@link SplitReader}, it will relay * the information asynchronously through the split fetcher thread. * * @param splitsToPause the splits to pause * @param splits...
3.68
framework_UIEvents_getUI
/** * Get the {@link UI} instance that received the poll request. * * @return the {@link UI} that received the poll request. Never * <code>null</code>. */ public UI getUI() { /* * This cast is safe to make, since this class' constructor * constrains the source to be a UI instance. */ ...
3.68
flink_FlinkContainersSettings_jobManagerHostname
/** * Sets the job manager hostname and returns a reference to this Builder enabling method * chaining. * * @param jobManagerHostname The job manager hostname to set. * @return A reference to this Builder. */ public Builder jobManagerHostname(String jobManagerHostname) { return setConfigOption(JobManagerOptio...
3.68
framework_ComponentRendererConnector_createConnectorHierarchyChangeHandler
/** * Adds a listener for grid hierarchy changes to find detached connectors * previously handled by this renderer in order to detach from DOM their * widgets before {@link AbstractComponentConnector#onUnregister()} is * invoked otherwise an error message is logged. */ private void createConnectorHierarchyChangeHa...
3.68
hadoop_StartupProgress_beginPhase
/** * Begins execution of the specified phase. * * @param phase Phase to begin */ public void beginPhase(Phase phase) { if (!isComplete()) { phases.get(phase).beginTime = monotonicNow(); } LOG.debug("Beginning of the phase: {}", phase); }
3.68
dubbo_SerializingExecutor_execute
/** * Runs the given runnable strictly after all Runnables that were submitted * before it, and using the {@code executor} passed to the constructor. . */ @Override public void execute(Runnable r) { runQueue.add(r); schedule(r); }
3.68
hbase_TaskMonitor_getTasks
/** * Produces a list containing copies of the current state of all non-expired MonitoredTasks * handled by this TaskMonitor. * @param filter type of wanted tasks * @return A filtered list of MonitoredTasks. */ public synchronized List<MonitoredTask> getTasks(String filter) { purgeExpiredTasks(); TaskFilter ta...
3.68
flink_TieredStorageConfiguration_getMinReserveDiskSpaceFraction
/** * Minimum reserved disk space fraction in disk tier. * * @return the fraction. */ public float getMinReserveDiskSpaceFraction() { return minReserveDiskSpaceFraction; }
3.68
hadoop_RemoteSASKeyGeneratorImpl_makeRemoteRequest
/** * Helper method to make a remote request. * * @param urls - Urls to use for the remote request * @param path - hadoop.auth token for the remote request * @param queryParams - queryParams to be used. * @return RemoteSASKeyGenerationResponse */ private RemoteSASKeyGenerationResponse makeRemoteReq...
3.68
flink_AbstractBlobCache_setBlobServerAddress
/** * Sets the address of the {@link BlobServer}. * * @param blobServerAddress address of the {@link BlobServer}. */ public void setBlobServerAddress(InetSocketAddress blobServerAddress) { serverAddress = checkNotNull(blobServerAddress); }
3.68
morf_SelectStatement_useImplicitJoinOrder
/** * If supported by the dialect, hints to the database that joins should be applied in the order * they are written in the SQL statement. * * <p>This is supported to greater or lesser extends on different SQL dialects. For instance, * MySQL has no means to force ordering on anything except inner joins, but we d...
3.68
morf_Criterion_like
/** * Helper method to create a new "LIKE" expression. * * <blockquote><pre> * Criterion.like(new Field("agreementnumber"), "A%");</pre></blockquote> * * <p>Note the escape character is set to '\' (backslash) by the underlying system.</p> * * @param field the field to evaluate in the expression (the left han...
3.68
pulsar_SubscriptionStatsImpl_add
// if the stats are added for the 1st time, we will need to make a copy of these stats and add it to the current // stats public SubscriptionStatsImpl add(SubscriptionStatsImpl stats) { Objects.requireNonNull(stats); this.msgRateOut += stats.msgRateOut; this.msgThroughputOut += stats.msgThroughputOut; t...
3.68
hadoop_DomainColumn_getColumnQualifier
/** * @return the column name value */ private String getColumnQualifier() { return columnQualifier; }
3.68
hbase_MasterRpcServices_checkMasterProcedureExecutor
/** * @throws ServiceException If no MasterProcedureExecutor */ private void checkMasterProcedureExecutor() throws ServiceException { if (this.server.getMasterProcedureExecutor() == null) { throw new ServiceException("Master's ProcedureExecutor not initialized; retry later"); } }
3.68
flink_GatewayRetriever_getNow
/** * Returns the currently retrieved gateway if there is such an object. Otherwise it returns an * empty optional. * * @return Optional object to retrieve */ default Optional<T> getNow() { CompletableFuture<T> leaderFuture = getFuture(); if (leaderFuture != null) { if (leaderFuture.isCompletedExce...
3.68
flink_SharedBufferAccessor_releaseNode
/** * Decreases the reference counter for the given entry so that it can be removed once the * reference counter reaches 0. * * @param node id of the entry * @param version dewey number of the (potential) edge that locked the given node * @throws Exception Thrown if the system cannot access the state. */ public ...
3.68
querydsl_AbstractCollQuery_bind
/** * Bind the given collection to an already existing query source * * @param <A> type of expression * @param entity Path for the source * @param col content of the source * @return current object */ public <A> Q bind(Path<A> entity, Iterable<? extends A> col) { iterables.put(entity, col); return queryM...
3.68
hbase_HRegionServer_getRegions
/** * Gets the online regions of the specified table. This method looks at the in-memory * onlineRegions. It does not go to <code>hbase:meta</code>. Only returns <em>online</em> regions. * If a region on this table has been closed during a disable, etc., it will not be included in * the returned list. So, the retur...
3.68
hbase_MetricsHeapMemoryManager_setCurBlockCacheSizeGauge
/** * Set the current blockcache size used gauge * @param blockCacheSize the current memory usage in blockcache, in bytes. */ public void setCurBlockCacheSizeGauge(final long blockCacheSize) { source.setCurBlockCacheSizeGauge(blockCacheSize); }
3.68
flink_Costs_addHeuristicCpuCost
/** * Adds the given heuristic CPU cost to the current heuristic CPU cost for this Costs object. * * @param cost The heuristic CPU cost to add. */ public void addHeuristicCpuCost(double cost) { if (cost <= 0) { throw new IllegalArgumentException("Heuristic costs must be positive."); } this.heuri...
3.68
framework_VDragAndDropManager_executeWhenReady
/** * Method to execute commands when all existing dd related tasks are * completed (some may require server visit). * <p> * Using this method may be handy if criterion that uses lazy initialization * are used. Check * <p> * TODO Optimization: consider if we actually only need to keep the last * command in queu...
3.68
morf_OracleDialect_getSqlForNow
/** * @see org.alfasoftware.morf.jdbc.SqlDialect#getSqlForNow(org.alfasoftware.morf.sql.element.Function) */ @Override protected String getSqlForNow(Function function) { return "SYSTIMESTAMP AT TIME ZONE 'UTC'"; }
3.68
flink_FlinkDatabaseMetaData_storesMixedCaseIdentifiers
/** Flink sql is mixed case as sensitive. */ @Override public boolean storesMixedCaseIdentifiers() throws SQLException { return true; }
3.68
flink_KubernetesEntrypointUtils_loadConfiguration
/** * For non-HA cluster, {@link JobManagerOptions#ADDRESS} has be set to Kubernetes service name * on client side. See {@link KubernetesClusterDescriptor#deployClusterInternal}. So the * TaskManager will use service address to contact with JobManager. For HA cluster, {@link * JobManagerOptions#ADDRESS} will be set...
3.68
querydsl_PathBuilder_getBoolean
/** * Create a new Boolean typed path * * @param propertyName property name * @return property path */ public BooleanPath getBoolean(String propertyName) { validate(propertyName, Boolean.class); return super.createBoolean(propertyName); }
3.68
pulsar_Commands_peekAndCopyMessageMetadata
/** * Peek the message metadata from the buffer and return a deep copy of the metadata. * * If you want to hold multiple {@link MessageMetadata} instances from multiple buffers, you must call this method * rather than {@link Commands#peekMessageMetadata(ByteBuf, String, long)}, which returns a thread local referenc...
3.68
flink_UnorderedStreamElementQueue_emitCompleted
/** * Pops one completed elements into the given output. Because an input element may produce * an arbitrary number of output elements, there is no correlation between the size of the * collection and the popped elements. * * @return the number of popped input elements. */ int emitCompleted(TimestampedCollector<O...
3.68
open-banking-gateway_PathHeadersBodyMapperTemplate_forExecution
/** * Converts context object into object that can be used for ASPSP API call. * @param context Context to convert * @return Object that can be used with {@code Xs2aAdapter} to perform ASPSP API calls */ public ValidatedPathHeadersBody<P, H, B> forExecution(C context) { return new ValidatedPathHeadersBody<>( ...
3.68
flink_RestServerEndpoint_shutDownInternal
/** * Stops this REST server endpoint. * * @return Future which is completed once the shut down has been finished. */ protected CompletableFuture<Void> shutDownInternal() { synchronized (lock) { CompletableFuture<?> channelFuture = new CompletableFuture<>(); if (serverChannel != null) { ...
3.68
hbase_RequestConverter_buildGetServerInfoRequest
/** * Create a new GetServerInfoRequest * @return a GetServerInfoRequest */ public static GetServerInfoRequest buildGetServerInfoRequest() { return GetServerInfoRequest.getDefaultInstance(); }
3.68
hbase_HBaseTestingUtility_getOtherRegionServer
/** * Find any other region server which is different from the one identified by parameter * @return another region server */ public HRegionServer getOtherRegionServer(HRegionServer rs) { for (JVMClusterUtil.RegionServerThread rst : getMiniHBaseCluster().getRegionServerThreads()) { if (!(rst.getRegionServer() ...
3.68
hbase_CellCodec_write
/** * Write int length followed by array bytes. */ private void write(final byte[] bytes, final int offset, final int length) throws IOException { // TODO add BB backed os check and do for write. Pass Cell this.out.write(Bytes.toBytes(length)); this.out.write(bytes, offset, length); }
3.68
hadoop_WriteOperationHelper_select
/** * Execute an S3 Select operation. * On a failure, the request is only logged at debug to avoid the * select exception being printed. * * @param source source for selection * @param request Select request to issue. * @param action the action for use in exception creation * @return response * @throws IOExc...
3.68
pulsar_Schema_generic
/** * Returns a generic schema of existing schema info. * * <p>Only supports AVRO and JSON. * * @param schemaInfo schema info * @return a generic schema instance */ static GenericSchema<GenericRecord> generic(SchemaInfo schemaInfo) { return DefaultImplementation.getDefaultImplementation().getGenericSchema(sc...
3.68
hbase_RegionLocations_mergeLocations
/** * Merges this RegionLocations list with the given list assuming same range, and keeping the most * up to date version of the HRegionLocation entries from either list according to seqNum. If * seqNums are equal, the location from the argument (other) is taken. * @param other the locations to merge with * @retur...
3.68
hibernate-validator_AnnotationTypeMemberCheck_checkMessageAttribute
/** * Checks that the given type element * <p/> * <ul> * <li>has a method with name "message",</li> * <li>the return type of this method is {@link String}.</li> * </ul> * * @param element The element of interest. * * @return A possibly non-empty set of constraint check errors, never null. */ private Set<Cons...
3.68
dubbo_MetadataResolver_resolveConsumerServiceMetadata
/** * for consumer * * @param targetClass target service class * @param url consumer url * @return rest metadata * @throws CodeStyleNotSupportException not support type */ public static ServiceRestMetadata resolveConsumerServiceMetadata( Class<?> targetClass, URL url, String contextPathFromUrl) {...
3.68
framework_UIDL_getIntAttribute
/** * Gets the named attribute as an int. * * @param name * the name of the attribute to get * @return the attribute value */ public int getIntAttribute(String name) { return attr().getInt(name); }
3.68
hudi_S3EventsMetaSelector_createSourceSelector
/** * Factory method for creating custom CloudObjectsMetaSelector. Default selector to use is {@link * S3EventsMetaSelector} */ public static S3EventsMetaSelector createSourceSelector(TypedProperties props) { String sourceSelectorClass = getStringWithAltKeys( props, DFSPathSelectorConfig.SOURCE_INPUT_SELECTO...
3.68
framework_UIDL_getStringArrayAttribute
/** * Gets the named attribute as an array of Strings. * * @param name * the name of the attribute to get * @return the attribute value */ public String[] getStringArrayAttribute(String name) { return attr().getStringArray(name); }
3.68
flink_CheckpointProperties_discardOnJobFailed
/** * Returns whether the checkpoint should be discarded when the owning job reaches the {@link * JobStatus#FAILED} state. * * @return <code>true</code> if the checkpoint should be discarded when the owning job reaches * the {@link JobStatus#FAILED} state; <code>false</code> otherwise. * @see CompletedCheckpo...
3.68
flink_IOUtils_skipFully
/** * Similar to readFully(). Skips bytes in a loop. * * @param in The InputStream to skip bytes from * @param len number of bytes to skip * @throws IOException if it could not skip requested number of bytes for any reason (including * EOF) */ public static void skipFully(final InputStream in, long len) thro...
3.68
graphhopper_VectorTile_addValues
/** * <pre> * Dictionary encoding for values * </pre> * * <code>repeated .vector_tile.Tile.Value values = 4;</code> */ public Builder addValues( int index, vector_tile.VectorTile.Tile.Value.Builder builderForValue) { if (valuesBuilder_ == null) { ensureValuesIsMutable(); values_.add(index, builderFo...
3.68
querydsl_Expressions_list
/** * Combine the given expressions into a list expression * * @param exprs list elements * @return list expression */ public static Expression<Tuple> list(Expression<?>... exprs) { return list(Tuple.class, exprs); }
3.68
hbase_PrivateCellUtil_writeRow
/** * Writes the row from the given cell to the output stream * @param out The outputstream to which the data has to be written * @param cell The cell whose contents has to be written * @param rlength the row length */ public static void writeRow(OutputStream out, Cell cell, short rlength) throws IOExceptio...
3.68
hadoop_HdfsFileStatus_isEmptyLocalName
/** * Check if the local name is empty. * @return true if the name is empty */ default boolean isEmptyLocalName() { return getLocalNameInBytes().length == 0; }
3.68
hudi_HoodieAvroUtils_generateProjectionSchema
/** * Generate a reader schema off the provided writeSchema, to just project out the provided columns. */ public static Schema generateProjectionSchema(Schema originalSchema, List<String> fieldNames) { Map<String, Field> schemaFieldsMap = originalSchema.getFields().stream() .map(r -> Pair.of(r.name().toLowerC...
3.68
hbase_Query_setReplicaId
/** * Specify region replica id where Query will fetch data from. Use this together with * {@link #setConsistency(Consistency)} passing {@link Consistency#TIMELINE} to read data from a * specific replicaId. <br> * <b> Expert: </b>This is an advanced API exposed. Only use it if you know what you are doing */ public...
3.68
hbase_StorageClusterStatusModel_setMemStoreSizeMB
/** * @param memstoreSizeMB memstore size, in MB */ public void setMemStoreSizeMB(int memstoreSizeMB) { this.memstoreSizeMB = memstoreSizeMB; }
3.68
morf_WindowFunction_getPartitionBys
/** * @return the fields to partition by. */ public ImmutableList<AliasedField> getPartitionBys() { return partitionBys; }
3.68
flink_JaasModule_generateDefaultConfigFile
/** Generate the default JAAS config file. */ private static File generateDefaultConfigFile(String workingDir) { checkArgument(workingDir != null, "working directory should not be null."); final File jaasConfFile; try { Path path = Paths.get(workingDir); if (Files.notExists(path)) { ...
3.68
hbase_SimpleRpcServer_getNumOpenConnections
/** * The number of open RPC conections * @return the number of open rpc connections */ @Override public int getNumOpenConnections() { return connectionManager.size(); }
3.68
pulsar_ReaderInterceptor_onPartitionsChange
/** * This method is called when partitions of the topic (partitioned-topic) changes. * * @param topicName topic name * @param partitions new updated number of partitions */ default void onPartitionsChange(String topicName, int partitions) { }
3.68
hbase_CompactingMemStore_snapshot
/** * Push the current active memstore segment into the pipeline and create a snapshot of the tail of * current compaction pipeline Snapshot must be cleared by call to {@link #clearSnapshot}. * {@link #clearSnapshot(long)}. * @return {@link MemStoreSnapshot} */ @Override public MemStoreSnapshot snapshot() { // I...
3.68
flink_QueryableStateUtils_createKvStateServer
/** * Initializes the {@link KvStateServer server} responsible for sending the requested internal * state to the {@link KvStateClientProxy client proxy}. * * @param address the address to bind to. * @param ports the range of ports the state server will attempt to listen to (see {@link * org.apache.flink.confi...
3.68
hbase_ReplicationSourceManager_releaseBufferQuota
/** * To release the buffer quota which acquired by * {@link ReplicationSourceManager#acquireBufferQuota}. */ void releaseBufferQuota(long size) { if (size < 0) { throw new IllegalArgumentException("size should not less than 0"); } addTotalBufferUsed(-size); }
3.68
hadoop_RouterObserverReadProxyProvider_isRead
/** * Check if a method is read-only. * * @return whether the 'method' is a read-only operation. */ private static boolean isRead(Method method) { if (!method.isAnnotationPresent(ReadOnly.class)) { return false; } return !method.getAnnotationsByType(ReadOnly.class)[0].activeOnly(); }
3.68
hadoop_EntityTypeReader_getNextRowKey
/** * Gets the possibly next row key prefix given current prefix and type. * * @param currRowKeyPrefix The current prefix that contains user, cluster, * flow, run, and application id. * @param entityType Current entity type. * @return A new prefix for the possibly immediately next row key....
3.68
framework_AbstractOrderedLayoutConnector_needsExpand
/** * Does the layout need to expand? */ private boolean needsExpand() { return needsExpand; }
3.68