name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
hadoop_SelectTool_run
/** * Execute the select operation. * @param args argument list * @param out output stream * @return an exit code * @throws IOException IO failure * @throws ExitUtil.ExitException managed failure */ public int run(String[] args, PrintStream out) throws IOException, ExitUtil.ExitException { final List<Strin...
3.68
querydsl_AbstractGroupExpression_as
/** * Create an alias for the expression * * @return alias expression */ public DslExpression<R> as(String alias) { return as(ExpressionUtils.path(getType(), alias)); }
3.68
framework_HasStyleNames_removeStyleNames
/** * Removes one or more style names from component. Multiple styles can be * specified by using multiple parameters. * * <p> * The parameter must be a valid CSS style name. Only user-defined style * names added with {@link #addStyleName(String) addStyleName()} or * {@link #setStyleName(String) setStyleName()} ...
3.68
flink_ExceptionHistoryEntry_createGlobal
/** Creates an {@code ExceptionHistoryEntry} that is not based on an {@code Execution}. */ public static ExceptionHistoryEntry createGlobal( Throwable cause, CompletableFuture<Map<String, String>> failureLabels) { return new ExceptionHistoryEntry( cause, System.currentTimeMillis(), ...
3.68
hbase_StoreScanner_parallelSeek
/** * Seek storefiles in parallel to optimize IO latency as much as possible * @param scanners the list {@link KeyValueScanner}s to be read from * @param kv the KeyValue on which the operation is being requested */ private void parallelSeek(final List<? extends KeyValueScanner> scanners, final Cell kv) thro...
3.68
framework_VCaption_updateCaption
/** * Updates the caption from UIDL. * * This method may only be called when the caption has an owner - otherwise, * use {@link #updateCaptionWithoutOwner(UIDL, String, boolean, boolean)}. * * @return true if the position where the caption should be placed has * changed */ public boolean updateCaption()...
3.68
hbase_TableDescriptorBuilder_isNormalizationEnabled
/** * Check if normalization enable flag of the table is true. If flag is false then no region * normalizer won't attempt to normalize this table. * @return true if region normalization is enabled for this table **/ @Override public boolean isNormalizationEnabled() { return getOrDefault(NORMALIZATION_ENABLED_KEY,...
3.68
hbase_TableDescriptorBuilder_isReadOnly
/** * Check if the readOnly flag of the table is set. If the readOnly flag is set then the contents * of the table can only be read from but not modified. * @return true if all columns in the table should be read only */ @Override public boolean isReadOnly() { return getOrDefault(READONLY_KEY, Boolean::valueOf, D...
3.68
hadoop_PlacementConstraints_nodePartition
/** * Constructs a target expression on a node partition. It is satisfied if * the specified node partition has one of the specified nodePartitions. * * @param nodePartitions the set of values that the attribute should take * values from * @return the resulting expression on the node attribute */ public...
3.68
pulsar_AuthenticationDataHttps_hasDataFromTls
/* * TLS */ @Override public boolean hasDataFromTls() { return (certificates != null); }
3.68
hadoop_DiskBalancerDataNode_getDataNodeIP
/** * Returns the IP address of this Node. * * @return IP Address string */ public String getDataNodeIP() { return dataNodeIP; }
3.68
hadoop_AbfsHttpOperation_getConnUrl
/** * Gets the connection url. * @return url. */ URL getConnUrl() { return connection.getURL(); }
3.68
hudi_IncrSourceHelper_generateQueryInfo
/** * Find begin and end instants to be set for the next fetch. * * @param jssc Java Spark Context * @param srcBasePath Base path of Hudi source table * @param numInstantsPerFetch Max Instants per fetch * @param beginInstant Last Checkpoint String * @param mi...
3.68
hbase_SnapshotInfo_getSnapshotDescription
/** Returns the snapshot descriptor */ public SnapshotDescription getSnapshotDescription() { return ProtobufUtil.createSnapshotDesc(this.snapshot); }
3.68
hadoop_StartupProgress_getCounter
/** * Returns a counter associated with the specified phase and step. Typical * usage is to increment a counter within a tight loop. Callers may use this * method to obtain a counter once and then increment that instance repeatedly * within a loop. This prevents redundant lookup operations and object * creation...
3.68
dubbo_Invoker_invoke
// This method will never be called for a legacy invoker. @Override default org.apache.dubbo.rpc.Result invoke(org.apache.dubbo.rpc.Invocation invocation) throws org.apache.dubbo.rpc.RpcException { return null; }
3.68
dubbo_Server_start
/** * start server, bind port */ public void start() throws Throwable { if (!started.compareAndSet(false, true)) { return; } boss = new NioEventLoopGroup(1, new DefaultThreadFactory("qos-boss", true)); worker = new NioEventLoopGroup(0, new DefaultThreadFactory("qos-worker", true)); ServerB...
3.68
framework_TableQuery_writeObject
/** * Custom writeObject to call rollback() if object is serialized. */ private void writeObject(java.io.ObjectOutputStream out) throws IOException { try { rollback(); } catch (SQLException ignored) { } out.defaultWriteObject(); }
3.68
flink_EmbeddedRocksDBStateBackend_getDbStoragePaths
/** * Gets the configured local DB storage paths, or null, if none were configured. * * <p>Under these directories on the TaskManager, RocksDB stores its SST files and metadata * files. These directories do not need to be persistent, they can be ephermeral, meaning that * they are lost on a machine failure, becaus...
3.68
flink_SolutionSetNode_getOperator
/** * Gets the contract object for this data source node. * * @return The contract. */ @Override public SolutionSetPlaceHolder<?> getOperator() { return (SolutionSetPlaceHolder<?>) super.getOperator(); }
3.68
flink_ResultPartitionType_isHybridResultPartition
/** * {@link #isHybridResultPartition()} is used to judge whether it is the specified {@link * #HYBRID_FULL} or {@link #HYBRID_SELECTIVE} resultPartitionType. * * <p>this method suitable for judgment conditions related to the specific implementation of * {@link ResultPartitionType}. * * <p>this method not relate...
3.68
flink_HiveParserTypeCheckProcFactory_getIntervalExprProcessor
/** Factory method to get IntervalExprProcessor. */ public HiveParserTypeCheckProcFactory.IntervalExprProcessor getIntervalExprProcessor() { return new HiveParserTypeCheckProcFactory.IntervalExprProcessor(); }
3.68
morf_SqlServerDialect_dropPrimaryKey
/** * Drops the primary key from a {@link Table}. * * @param table The table to drop the primary key from */ private String dropPrimaryKey(Table table) { StringBuilder dropPkStatement = new StringBuilder(); dropPkStatement.append("ALTER TABLE ").append(schemaNamePrefix()).append(table.getName()).append(" DROP "...
3.68
flink_SourceTestSuiteBase_testTaskManagerFailure
/** * Test connector source with task manager failover. * * <p>This test will create 1 split in the external system, write test record set A into the * split, restart task manager to trigger job failover, write test record set B into the split, * and terminate the Flink job finally. * * <p>The number and order o...
3.68
hadoop_TypedBytesOutput_writeFloat
/** * Writes a float as a typed bytes sequence. * * @param f the float to be written * @throws IOException */ public void writeFloat(float f) throws IOException { out.write(Type.FLOAT.code); out.writeFloat(f); }
3.68
pulsar_AuthenticationProviderToken_getTokenAudienceClaim
// get Token Audience Claim from configuration, if not configured return null. private String getTokenAudienceClaim(ServiceConfiguration conf) throws IllegalArgumentException { String tokenAudienceClaim = (String) conf.getProperty(confTokenAudienceClaimSettingName); if (StringUtils.isNotBlank(tokenAudienceClaim...
3.68
hbase_KeyValueHeap_requestSeek
/** * {@inheritDoc} */ @Override public boolean requestSeek(Cell key, boolean forward, boolean useBloom) throws IOException { return generalizedSeek(true, key, forward, useBloom); }
3.68
hudi_TableHeader_addTableHeaderField
/** * Add a field (column) to table. * * @param fieldName field Name */ public TableHeader addTableHeaderField(String fieldName) { fieldNames.add(fieldName); return this; }
3.68
flink_PluginLoader_load
/** * Returns in iterator over all available implementations of the given service interface (SPI) * for the plugin. * * @param service the service interface (SPI) for which implementations are requested. * @param <P> Type of the requested plugin service. * @return An iterator of all implementations of the given s...
3.68
flink_SegmentsUtil_setFloat
/** * set float from segments. * * @param segments target segments. * @param offset value offset. */ public static void setFloat(MemorySegment[] segments, int offset, float value) { if (inFirstSegment(segments, offset, 4)) { segments[0].putFloat(offset, value); } else { setFloatMultiSegment...
3.68
framework_Escalator_scrollToRow
/** * Scrolls the body vertically so that the row at the given index is visible * and there is at least {@literal padding} pixels to the given scroll * destination. * * @param rowIndex * the index of the logical row to scroll to * @param destination * where the row should be aligned visual...
3.68
morf_ResultSetMismatch_getKey
/** * @return key identifying the mismatch. */ public String[] getKey() { return Arrays.copyOf(key, key.length); }
3.68
hadoop_RegistryTypeUtils_urlEndpoint
/** * Create a URL endpoint from a list of URIs * @param api implemented API * @param protocolType protocol type * @param uris URIs * @return a new endpoint */ public static Endpoint urlEndpoint(String api, String protocolType, URI... uris) { return new Endpoint(api, protocolType, uris); }
3.68
hadoop_IOStatisticsLogging_ioStatisticsToPrettyString
/** * Convert IOStatistics to a string form, with all the metrics sorted * and empty value stripped. * This is more expensive than the simple conversion, so should only * be used for logging/output where it's known/highly likely that the * caller wants to see the values. Not for debug logging. * @param statistics...
3.68
zxing_BitMatrix_getRow
/** * A fast method to retrieve one row of data from the matrix as a BitArray. * * @param y The row to retrieve * @param row An optional caller-allocated BitArray, will be allocated if null or too small * @return The resulting BitArray - this reference should always be used even when passing * your own ro...
3.68
dubbo_ScriptStateRouter_getEngine
/** * create ScriptEngine instance by type from url parameters, then cache it */ private ScriptEngine getEngine(URL url) { String type = url.getParameter(TYPE_KEY, DEFAULT_SCRIPT_TYPE_KEY); return ConcurrentHashMapUtils.computeIfAbsent(ENGINES, type, t -> { ScriptEngine scriptEngine = new ScriptEngin...
3.68
flink_SharedObjectsExtension_create
/** * Creates a new instance. Usually that should be done inside a JUnit test class as an * instance-field annotated with {@link org.junit.Rule}. */ public static SharedObjectsExtension create() { return new SharedObjectsExtension(LAST_ID.getAndIncrement()); }
3.68
hbase_ReflectionUtils_invokeMethod
/** * Get and invoke the target method from the given object with given parameters * @param obj the object to get and invoke method from * @param methodName the name of the method to invoke * @param params the parameters for the method to invoke * @return the return value of the method invocation */ @N...
3.68
framework_VTooltip_connectHandlersToWidget
/** * Connects DOM handlers to widget that are needed for tooltip presentation. * * @param widget * Widget which DOM handlers are connected */ public void connectHandlersToWidget(Widget widget) { Profiler.enter("VTooltip.connectHandlersToWidget"); widget.addDomHandler(tooltipEventHandler, MouseO...
3.68
hadoop_JavaCommandLineBuilder_addConfOptionToCLI
/** * Ass a configuration option to the command line of the application * @param conf configuration * @param key key * @param defVal default value * @return the resolved configuration option * @throws IllegalArgumentException if key is null or the looked up value * is null (that is: the argument is missing and ...
3.68
open-banking-gateway_AuthorizationPossibleErrorHandler_handlePossibleAuthorizationError
/** * Swallows retryable (like wrong password) authorization exceptions. * @param tryAuthorize Authorization function to call * @param onFail Fallback function to call if retryable exception occurred. */ public void handlePossibleAuthorizationError(Runnable tryAuthorize, Consumer<ErrorResponseException> onFail) { ...
3.68
hbase_PermissionStorage_isAclTable
/** * Returns {@code true} if the given table is {@code _acl_} metadata table. */ static boolean isAclTable(TableDescriptor desc) { return ACL_TABLE_NAME.equals(desc.getTableName()); }
3.68
morf_DataSourceAdapter_unwrap
/** * @see java.sql.Wrapper#unwrap(java.lang.Class) */ @Override public <T> T unwrap(Class<T> iface) throws SQLException { throw new UnsupportedOperationException("Wrappers not supported"); }
3.68
hudi_BaseHoodieTableServiceClient_scheduleCompactionAtInstant
/** * Schedules a new compaction instant with passed-in instant time. * * @param instantTime Compaction Instant Time * @param extraMetadata Extra Metadata to be stored */ public boolean scheduleCompactionAtInstant(String instantTime, Option<Map<String, String>> extraMetadata) throws HoodieIOException { return ...
3.68
hibernate-validator_SizeValidatorForArraysOfDouble_isValid
/** * Checks the number of entries in an array. * * @param array The array to validate. * @param constraintValidatorContext context in which the constraint is evaluated. * * @return Returns {@code true} if the array is {@code null} or the number of entries in * {@code array} is between the specified {@co...
3.68
flink_TableConfig_getPlannerConfig
/** Returns the current configuration of Planner for Table API and SQL queries. */ public PlannerConfig getPlannerConfig() { return plannerConfig; }
3.68
hbase_MasterObserver_postTableFlush
/** * Called after the table memstore is flushed to disk. * @param ctx the environment to interact with the framework and master * @param tableName the name of the table */ default void postTableFlush(final ObserverContext<MasterCoprocessorEnvironment> ctx, final TableName tableName) throws IOException { }
3.68
hbase_CompactingMemStore_setCompositeSnapshot
// the following three methods allow to manipulate the settings of composite snapshot public void setCompositeSnapshot(boolean useCompositeSnapshot) { this.compositeSnapshot = useCompositeSnapshot; }
3.68
flink_OperatingSystem_readOSFromSystemProperties
/** * Parses the operating system that the JVM runs on from the java system properties. If the * operating system was not successfully determined, this method returns {@code UNKNOWN}. * * @return The enum constant for the operating system, or {@code UNKNOWN}, if it was not * possible to determine. */ private ...
3.68
framework_VScrollTable_resizeCaptionContainer
/** * Makes room for the sorting indicator in case the column that the * header cell belongs to is sorted. This is done by resizing the width * of the caption container element by the correct amount */ public void resizeCaptionContainer(int rightSpacing) { int captionContainerWidth = width - colResizeWidget.get...
3.68
flink_VertexInputInfoComputationUtils_computeVertexInputInfoForAllToAll
/** * Compute the {@link JobVertexInputInfo} for a {@link DistributionPattern#ALL_TO_ALL} edge. * This computation algorithm will evenly distribute subpartitions to downstream subtasks * according to the number of subpartitions. Different downstream subtasks consume roughly the * same number of subpartitions. * *...
3.68
hadoop_TFile_getValueLength
/** * Get the length of the value. isValueLengthKnown() must be tested * true. * * @return the length of the value. */ public int getValueLength() { if (vlen >= 0) { return vlen; } throw new RuntimeException("Value length unknown."); }
3.68
framework_GridLayout_isHideEmptyRowsAndColumns
/** * Checks whether whether empty rows and columns should be considered as * non-existent when rendering or not. * * @see #setHideEmptyRowsAndColumns(boolean) * @since 7.3 * @return true if empty rows and columns are hidden, false otherwise */ public boolean isHideEmptyRowsAndColumns() { return getState(fal...
3.68
hadoop_SolverPreprocessor_getDiscreteSkyline
/** * Discretize job's lifespan into intervals, and return the number of * containers used by the job within each interval. * <p> Note that here we assume all containers allocated to the job have the * same {@link Resource}. This is due to the limit of * {@link RLESparseResourceAllocation}. * * @param skyList ...
3.68
framework_AbstractExtension_extend
/** * Add this extension to the target connector. This method is protected to * allow subclasses to require a more specific type of target. * * @param target * the connector to attach this extension to */ protected void extend(AbstractClientConnector target) { target.addExtension(this); }
3.68
hudi_BootstrapExecutorUtils_execute
/** * Executes Bootstrap. */ public void execute() throws IOException { initializeTable(); try (SparkRDDWriteClient bootstrapClient = new SparkRDDWriteClient(new HoodieSparkEngineContext(jssc), bootstrapConfig)) { HashMap<String, String> checkpointCommitMetadata = new HashMap<>(); checkpointCommitMetadat...
3.68
hbase_SnapshotManifest_open
/** * Return a SnapshotManifest instance with the information already loaded in-memory. * SnapshotManifest manifest = SnapshotManifest.open(...) TableDescriptor htd = * manifest.getDescriptor() for (SnapshotRegionManifest regionManifest: * manifest.getRegionManifests()) hri = regionManifest.getRegionInfo() for * (...
3.68
framework_Escalator_isScrollLocked
/** * Checks whether or not an direction is locked for scrolling. * * @param direction * the direction of the scroll of which to check the lock status * @return <code>true</code> if the direction is locked */ public boolean isScrollLocked(ScrollbarBundle.Direction direction) { switch (direction) { ...
3.68
hadoop_MappingRuleActionBase_setFallbackSkip
/** * Sets the fallback method to skip, if the action cannot be executed * We move onto the next rule, ignoring this one. * @return MappingRuleAction The same object for method chaining. */ public MappingRuleAction setFallbackSkip() { fallback = MappingRuleResult.createSkipResult(); return this; }
3.68
flink_SkipListUtils_getKeyLen
/** * Returns the length of the key. * * @param memorySegment memory segment for key space. * @param offset offset of key space in the memory segment. */ public static int getKeyLen(MemorySegment memorySegment, int offset) { return memorySegment.getInt(offset + KEY_LEN_OFFSET); }
3.68
zxing_BinaryBitmap_crop
/** * Returns a new object with cropped image data. Implementations may keep a reference to the * original data rather than a copy. Only callable if isCropSupported() is true. * * @param left The left coordinate, which must be in [0,getWidth()) * @param top The top coordinate, which must be in [0,getHeight()) * @...
3.68
framework_ComboBox_setTextInputAllowed
/** * Sets whether it is possible to input text into the field or whether the * field area of the component is just used to show what is selected. By * disabling text input, the comboBox will work in the same way as a * {@link NativeSelect} * * @see #isTextInputAllowed() * * @param textInputAllowed * ...
3.68
flink_FlinkContainersSettings_getHaStoragePath
/** * Gets HA storage path. * * @return The ha storage path. */ public String getHaStoragePath() { return haStoragePath; }
3.68
flink_StreamExecutionEnvironment_fromParallelCollection
// private helper for passing different names private <OUT> DataStreamSource<OUT> fromParallelCollection( SplittableIterator<OUT> iterator, TypeInformation<OUT> typeInfo, String operatorName) { return addSource( new FromSplittableIteratorFunction<>(iterator), operatorName, ...
3.68
graphhopper_Entity_checkRangeInclusive
/** @return whether the number actual is in the range [min, max] */ protected boolean checkRangeInclusive(double min, double max, double actual) { if (actual < min || actual > max) { feed.errors.add(new RangeError(tableName, row, null, min, max, actual)); // TODO set column name in loader so it's available ...
3.68
flink_KeyMap_getCurrentTableCapacity
/** * Gets the current table capacity, i.e., the number of slots in the hash table, without and * overflow chaining. * * @return The number of slots in the hash table. */ public int getCurrentTableCapacity() { return table.length; }
3.68
hudi_HoodieFlinkTableServiceClient_initMetadataWriter
/** * Initialize the table metadata writer, for e.g, bootstrap the metadata table * from the filesystem if it does not exist. */ private HoodieBackedTableMetadataWriter initMetadataWriter(Option<String> latestPendingInstant) { return (HoodieBackedTableMetadataWriter) FlinkHoodieBackedTableMetadataWriter.create( ...
3.68
flink_Channel_setSerializer
/** * Sets the serializer for this Channel. * * @param serializer The serializer to set. */ public void setSerializer(TypeSerializerFactory<?> serializer) { this.serializer = serializer; }
3.68
pulsar_ClientCnx_addPendingLookupRequests
// caller of this method needs to be protected under pendingLookupRequestSemaphore private void addPendingLookupRequests(long requestId, TimedCompletableFuture<LookupDataResult> future) { pendingRequests.put(requestId, future); requestTimeoutQueue.add(new RequestTime(requestId, RequestType.Lookup)); }
3.68
dubbo_MetadataService_getServiceDefinition
/** * Interface definition. * * @return */ default String getServiceDefinition(String interfaceName, String version, String group) { return getServiceDefinition(buildKey(interfaceName, group, version)); }
3.68
framework_VTabsheet_sendTabClosedEvent
/** * Informs the server that closing of a tab has been requested. * * @param tabIndex * the index of the closed to close */ void sendTabClosedEvent(int tabIndex) { getRpcProxy().closeTab(tabKeys.get(tabIndex)); }
3.68
hbase_SimpleRpcServer_bind
/** * A convenience method to bind to a given address and report better exceptions if the address is * not a valid host. * @param socket the socket to bind * @param address the address to bind to * @param backlog the number of connections allowed in the queue * @throws BindException if the address can't b...
3.68
hbase_StructBuilder_toStruct
/** * Retrieve the {@link Struct} represented by {@code this}. */ public Struct toStruct() { return new Struct(fields.toArray(new DataType<?>[fields.size()])); }
3.68
hbase_MasterProcedureUtil_submitProcedure
/** * Helper used to deal with submitting procs with nonce. Internally the * NonceProcedureRunnable.run() will be called only if no one else registered the nonce. any * Exception thrown by the run() method will be collected/handled and rethrown. <code> * long procId = MasterProcedureUtil.submitProcedure( * ne...
3.68
hadoop_S3ClientFactory_withRequesterPays
/** * Set requester pays option. * @param value new value * @return the builder */ public S3ClientCreationParameters withRequesterPays( final boolean value) { requesterPays = value; return this; }
3.68
hbase_MemStoreLABImpl_getOrMakeChunk
/** * Get the current chunk, or, if there is no current chunk, allocate a new one from the JVM. */ private Chunk getOrMakeChunk() { // Try to get the chunk Chunk c = currChunk.get(); if (c != null) { return c; } // No current chunk, so we want to allocate one. We race // against other allocators to CA...
3.68
hadoop_AbfsClientThrottlingIntercept_getReadThrottler
/** * Returns the analyzer for read operations. * @return AbfsClientThrottlingAnalyzer for read. */ AbfsClientThrottlingAnalyzer getReadThrottler() { return readThrottler; }
3.68
hbase_MultithreadedTableMapper_run
/** * Run the application's maps using a thread pool. */ @Override public void run(Context context) throws IOException, InterruptedException { outer = context; int numberOfThreads = getNumberOfThreads(context); mapClass = getMapperClass(context); if (LOG.isDebugEnabled()) { LOG.debug("Configuring multithr...
3.68
hadoop_FilterFileSystem_getDefaultBlockSize
// path variants delegate to underlying filesystem @Override public long getDefaultBlockSize(Path f) { return fs.getDefaultBlockSize(f); }
3.68
hbase_TableMapReduceUtil_getConfiguredInputFormat
/** * @return {@link TableInputFormat} .class unless Configuration has something else at * {@link #TABLE_INPUT_CLASS_KEY}. */ private static Class<? extends InputFormat> getConfiguredInputFormat(Job job) { return (Class<? extends InputFormat>) job.getConfiguration().getClass(TABLE_INPUT_CLASS_KEY, Tabl...
3.68
flink_ExecutionEnvironment_resetContextEnvironment
/** * Un-sets the context environment factory. After this method is called, the call to {@link * #getExecutionEnvironment()} will again return a default local execution environment, and it * is possible to explicitly instantiate the LocalEnvironment and the RemoteEnvironment. */ protected static void resetContextEn...
3.68
flink_ChangelogTruncateHelper_checkpoint
/** * Set the highest {@link SequenceNumber} of changelog used by the given checkpoint. * * @param lastUploadedTo exclusive */ public void checkpoint(long checkpointId, SequenceNumber lastUploadedTo) { checkpointedUpTo.put(checkpointId, lastUploadedTo); }
3.68
morf_OracleDialect_getSqlForAddMonths
/** * @see org.alfasoftware.morf.jdbc.SqlDialect#getSqlForAddMonths(org.alfasoftware.morf.sql.element.Function) */ @Override protected String getSqlForAddMonths(Function function) { return "ADD_MONTHS(" + getSqlFrom(function.getArguments().get(0)) + ", " + getSqlFrom(function.getArguments().get(1)) + ")"; }
3.68
morf_RemoveIndex_getIndexToBeRemoved
/** * @return The index to be removed. */ public Index getIndexToBeRemoved() { return indexToBeRemoved; }
3.68
hadoop_BalanceJob_removeAfterDone
/** * Automatically remove this job from the scheduler cache when the job is * done. */ public Builder removeAfterDone(boolean remove) { removeAfterDone = remove; return this; }
3.68
hbase_HRegionServer_getFavoredNodesForRegion
/** * Return the favored nodes for a region given its encoded name. Look at the comment around * {@link #regionFavoredNodesMap} on why we convert to InetSocketAddress[] here. * @param encodedRegionName the encoded region name. * @return array of favored locations */ @Override public InetSocketAddress[] getFavoredN...
3.68
graphhopper_Path_getWeight
/** * This weight will be updated during the algorithm. The initial value is maximum double. */ public double getWeight() { return weight; }
3.68
hadoop_JavaCommandLineBuilder_defineIfSet
/** * Add a <code>-D key=val</code> command to the CLI if <code>val</code> * is not null * @param key key * @param val value */ public boolean defineIfSet(String key, String val) { Preconditions.checkArgument(key != null, "null key"); if (val != null) { define(key, val); return true; } else { ret...
3.68
hadoop_Nfs3Constant_getValue
/** @return the int value representing the procedure. */ public int getValue() { return ordinal(); }
3.68
hadoop_CachingBlockManager_numReadErrors
/** * Number of errors encountered when reading. * * @return the number of errors encountered when reading. */ public int numReadErrors() { return numReadErrors.get(); }
3.68
hbase_ConfigurationUtil_setKeyValues
/** * Store a collection of Map.Entry's in conf, with each entry separated by ',' and key values * delimited by delimiter. * @param conf configuration to store the collection in * @param key overall key to store keyValues under * @param keyValues kvps to be stored under key in conf * @param delimiter c...
3.68
flink_OutputFormatProvider_of
/** Helper method for creating a static provider with a provided sink parallelism. */ static OutputFormatProvider of(OutputFormat<RowData> outputFormat, Integer sinkParallelism) { return new OutputFormatProvider() { @Override public OutputFormat<RowData> createOutputFormat() { return out...
3.68
framework_SerializerHelper_readClass
/** * Deserializes a class reference serialized by * {@link #writeClass(ObjectOutputStream, Class)}. Supports null class * references. * * @param in * {@code ObjectInputStream} to read from. * @return Class reference to the resolved class * @throws ClassNotFoundException * If the class c...
3.68
flink_DamBehavior_isMaterializing
/** * Checks whether this enumeration represents some form of materialization, either with a full * dam or without. * * @return True, if this enumeration constant represents a materializing behavior, false * otherwise. */ public boolean isMaterializing() { return this != PIPELINED; }
3.68
hbase_Response_setCode
/** * @param code the HTTP response code */ public void setCode(int code) { this.code = code; }
3.68
flink_HybridShuffleConfiguration_getBufferPoolSizeCheckIntervalMs
/** Check interval of buffer pool's size. */ public long getBufferPoolSizeCheckIntervalMs() { return bufferPoolSizeCheckIntervalMs; }
3.68
dubbo_StringUtils_join
/** * join string like javascript. * * @param array String array. * @param split split * @return String. */ public static String join(String[] array, String split) { if (ArrayUtils.isEmpty(array)) { return EMPTY_STRING; } StringBuilder sb = new StringBuilder(); for (int i = 0; i < array.le...
3.68
flink_MutableHashTable_getNewInMemoryPartition
/** * Returns a new inMemoryPartition object. This is required as a plug for * ReOpenableMutableHashTable. */ protected HashPartition<BT, PT> getNewInMemoryPartition(int number, int recursionLevel) { return new HashPartition<BT, PT>( this.buildSideSerializer, this.probeSideSerializer, ...
3.68
morf_SqlUtils_clobLiteral
/** * Constructs a new ClobFieldLiteral from a given string. * @param value - the literal value to use. * @return ClobFieldLiteral */ public static ClobFieldLiteral clobLiteral(String value) { return new ClobFieldLiteral(value); }
3.68
flink_ResultPartitionType_mustBePipelinedConsumed
/** return if this partition's upstream and downstream must be scheduled in the same time. */ public boolean mustBePipelinedConsumed() { return consumingConstraint == ConsumingConstraint.MUST_BE_PIPELINED; }
3.68