name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
hadoop_AbfsClient_appendSASTokenToQuery
/** * If configured for SAS AuthType, appends SAS token to queryBuilder. * @param path * @param operation * @param queryBuilder * @param cachedSasToken - previously acquired SAS token to be reused. * @return sasToken - returned for optional re-use. * @throws SASTokenProviderException */ private String appendSAS...
3.68
flink_StreamGraphHasherV2_generateNodeLocalHash
/** * Applies the {@link Hasher} to the {@link StreamNode} . The hasher encapsulates the current * state of the hash. * * <p>The specified ID is local to this node. We cannot use the {@link StreamNode#id}, because * it is incremented in a static counter. Therefore, the IDs for identical jobs will otherwise * be d...
3.68
hadoop_BinaryRecordOutput_get
/** * Get a thread-local record output for the supplied DataOutput. * @param out data output stream * @return binary record output corresponding to the supplied DataOutput. */ public static BinaryRecordOutput get(DataOutput out) { BinaryRecordOutput bout = B_OUT.get(); bout.setDataOutput(out); return bout; }
3.68
flink_VertexFlameGraph_waiting
// Indicates that it is waiting for the first samples to creating the flame graph public static VertexFlameGraph waiting() { return new VertexFlameGraph(-3, null); }
3.68
MagicPlugin_Mage_sendMessage
/** * Send a message to this Mage. * <p/> * Use this to send messages to the player that are important. * * @param message The message to send */ @Override public void sendMessage(String message) { sendMessage(controller.getMessagePrefix(), message); }
3.68
flink_LeaderInformationRegister_of
/** Creates a single-entry instance containing only the passed information. */ public static LeaderInformationRegister of( String componentId, LeaderInformation leaderInformation) { return new LeaderInformationRegister( Collections.singletonMap(componentId, leaderInformation)); }
3.68
hbase_Result_setStatistics
/** * Set load information about the region to the information about the result * @param loadStats statistics about the current region from which this was returned */ @InterfaceAudience.Private public void setStatistics(RegionLoadStats loadStats) { this.stats = loadStats; }
3.68
flink_RichFunction_open
/** * Initialization method for the function. It is called before the actual working methods (like * <i>map</i> or <i>join</i>) and thus suitable for one time setup work. For functions that are * part of an iteration, this method will be invoked at the beginning of each iteration * superstep. * * <p>The openConte...
3.68
hbase_TableDescriptorBuilder_setRegionMemStoreReplication
/** * Enable or Disable the memstore replication from the primary region to the replicas. The * replication will be used only for meta operations (e.g. flush, compaction, ...) * @param memstoreReplication true if the new data written to the primary region should be * replicated. false if ...
3.68
morf_HumanReadableStatementHelper_generateChangePrimaryKeyColumnsString
/** * @param tableName - the table name which needs its primary key columns changed * @param newPrimaryKeyColumns - the list of table names to primary key columns will become * @return a string containing the human-readable version of the action */ public static String generateChangePrimaryKeyColumnsString(String t...
3.68
flink_JobMasterPartitionTracker_getAllTrackedNonClusterPartitions
/** Gets all the non-cluster partitions under tracking. */ default Collection<ResultPartitionDeploymentDescriptor> getAllTrackedNonClusterPartitions() { return getAllTrackedPartitions().stream() .filter(descriptor -> !descriptor.getPartitionType().isPersistent()) .collect(Collectors.toList()...
3.68
hadoop_RouterQuotaUsage_verifyQuotaByStorageType
/** * Verify space quota by storage type is violated once quota is set. Relevant * method {@link DirectoryWithQuotaFeature#verifyQuotaByStorageType}. * @throws DSQuotaExceededException If the quota is exceeded. */ public void verifyQuotaByStorageType() throws DSQuotaExceededException { for (StorageType t: Storage...
3.68
streampipes_AbstractConfigurablePipelineElementBuilder_requiredSecret
/** * Assigns a new secret text-based configuration parameter (e.g., a password) which is required * by the processing element. * * @param label The {@link org.apache.streampipes.sdk.helpers.Label} that describes why this parameter is needed in a * user-friendly manner. * @return */ public K require...
3.68
rocketmq-connect_WorkerSinkTask_preCommit
/** * reset offset by custom */ private void preCommit() { Map<MessageQueue, Long> offsets = sinkTaskContext.queuesOffsets(); if (offsets.isEmpty()) { return; } for (Map.Entry<MessageQueue, Long> entry : offsets.entrySet()) { MessageQueue queue = entry.getKey(); Long offset = e...
3.68
hmily_EtcdClient_setClient
/** * set client. * * @param client client */ public void setClient(final Client client) { this.client = client; }
3.68
hadoop_DatanodeAdminProperties_getMaintenanceExpireTimeInMS
/** * Get the maintenance expiration time in milliseconds. * @return the maintenance expiration time in milliseconds. */ public long getMaintenanceExpireTimeInMS() { return this.maintenanceExpireTimeInMS; }
3.68
hbase_ZKWatcher_getQuorum
/** * Get the quorum address of this instance. * @return quorum string of this zookeeper connection instance */ public String getQuorum() { return quorum; }
3.68
flink_LogicalTypeChecks_hasScale
/** Checks the scale of all types that define a scale implicitly or explicitly. */ public static boolean hasScale(LogicalType logicalType, int scale) { return getScale(logicalType) == scale; }
3.68
hbase_ByteBufferKeyOnlyKeyValue_setKey
/** * A setter that helps to avoid object creation every time and whenever there is a need to create * new OffheapKeyOnlyKeyValue. * @param key - the key part of the cell * @param offset - offset of the cell * @param length - length of the cell * @param rowLen - the rowlen part of the cell */ public void setK...
3.68
morf_UpgradeGraph_isPackageNameValid
//Determine if the package name version is valid private boolean isPackageNameValid(Class<? extends UpgradeStep> stepClass) { if (stepClass.getPackage() == null) { return false; } return stepClass.getPackage().getName().matches(".*\\.upgrade\\.v[0-9]+_[0-9]+(_[0-9]+[a-z]?)*$"); }
3.68
flink_SavepointWriter_removeOperator
/** * Drop an existing operator from the savepoint. * * @param identifier The identifier of the operator. * @return A modified savepoint. */ public SavepointWriter removeOperator(OperatorIdentifier identifier) { metadata.removeOperator(identifier); return this; }
3.68
hadoop_BlockData_getStartOffset
/** * Gets the start offset of the given block. * @param blockNumber the id of the given block. * @return the start offset of the given block. * @throws IllegalArgumentException if blockNumber is invalid. */ public long getStartOffset(int blockNumber) { throwIfInvalidBlockNumber(blockNumber); return blockNumb...
3.68
pulsar_OffloadIndexBlock_getStreamSize
/** * @return the number of bytes in the stream. */ public long getStreamSize() { return streamSize; }
3.68
hadoop_ConnectionPool_getNumIdleConnections
/** * Number of usable i.e. no active thread connections. * * @return Number of idle connections */ protected int getNumIdleConnections() { int ret = 0; List<ConnectionContext> tmpConnections = this.connections; for (ConnectionContext conn : tmpConnections) { if (conn.isIdle()) { ret++; } } ...
3.68
streampipes_DataStreamBuilder_property
/** * Assigns a new event property to the stream's schema. * * @param property The event property that should be added. * Use {@link org.apache.streampipes.sdk.helpers.EpProperties} * for defining simple property definitions or * {@link org.apache.streampipes.sdk.bu...
3.68
hbase_ProcedureExecutor_init
/** * Initialize the procedure executor, but do not start workers. We will start them later. * <p/> * It calls ProcedureStore.recoverLease() and ProcedureStore.load() to recover the lease, and * ensure a single executor, and start the procedure replay to resume and recover the previous * pending and in-progress pr...
3.68
framework_AbstractComponent_getCaption
/* * Get's the component's caption. Don't add a JavaDoc comment here, we use * the default documentation from implemented interface. */ @Override public String getCaption() { return getState(false).caption; }
3.68
flink_DataSetUtils_checksumHashCode
/** * Convenience method to get the count (number of elements) of a DataSet as well as the checksum * (sum over element hashes). * * @return A ChecksumHashCode that represents the count and checksum of elements in the data * set. * @deprecated This method will be removed at some point. */ @Deprecated public ...
3.68
flink_FlinkDatabaseMetaData_nullsAreSortedLow
/** In flink null value will be used as low value for sort. */ @Override public boolean nullsAreSortedLow() throws SQLException { return true; }
3.68
flink_AbstractCollectResultBuffer_reset
/** Clear the whole buffer and discard all results. */ protected void reset() { buffer.clear(); userVisibleHead = 0; userVisibleTail = 0; offset = 0; }
3.68
streampipes_DataLakeResourceV4_isIgnoreMissingValues
// Checks if the parameter for missing value behaviour is set private boolean isIgnoreMissingValues(String missingValueBehaviour) { boolean ignoreMissingValues; if ("ignore".equals(missingValueBehaviour)) { ignoreMissingValues = true; } else { ignoreMissingValues = false; } return ignoreMissingValues;...
3.68
hadoop_Endpoint_newAddresses
/** * Create a new address structure of the requested size * @param size size to create * @return the new list */ private List<Map<String, String>> newAddresses(int size) { return new ArrayList<Map<String, String>>(size); }
3.68
hbase_BufferedDataBlockEncoder_copyFromNext
/** * Copy the state from the next one into this instance (the previous state placeholder). Used to * save the previous state when we are advancing the seeker to the next key/value. */ protected void copyFromNext(SeekerState nextState) { if (keyBuffer.length != nextState.keyBuffer.length) { keyBuffer = nextSta...
3.68
hbase_RefCnt_hasRecycler
/** * Returns true if this refCnt has a recycler. */ public boolean hasRecycler() { return recycler != ByteBuffAllocator.NONE; }
3.68
hbase_FsDelegationToken_releaseDelegationToken
/** * Releases a previously acquired delegation token. */ public void releaseDelegationToken() { if (userProvider.isHadoopSecurityEnabled()) { if (userToken != null && !hasForwardedToken) { try { userToken.cancel(this.fs.getConf()); } catch (Exception e) { LOG.warn("Failed to cancel ...
3.68
flink_DataStream_getPreferredResources
/** * Gets the preferred resources for this operator. * * @return The preferred resources set for this operator. */ @PublicEvolving public ResourceSpec getPreferredResources() { return transformation.getPreferredResources(); }
3.68
morf_MySqlDialect_alterAutoincrementStatement
/** * Returns a statement which will update the statistics for a specific table. */ private String alterAutoincrementStatement(Table table,Column autoIncrementColumn) { return "ALTER TABLE " + table.getName() + " AUTO_INCREMENT = " + autoIncrementColumn.getAutoNumberStart(); }
3.68
flink_AbstractPythonFunctionOperator_getConfiguration
/** Returns the {@link Configuration}. */ public Configuration getConfiguration() { return config; }
3.68
framework_HasHierarchicalDataProvider_setTreeData
/** * Sets a new {@link TreeDataProvider} wrapping the given {@link TreeData}. * * @param treeData * the tree data to set */ public default void setTreeData(TreeData<T> treeData) { setDataProvider(new TreeDataProvider<>(treeData)); }
3.68
hbase_ChoreService_getNumberOfChoresMissingStartTime
/** * Return number of chores that this service currently has scheduled that are missing their * scheduled start time */ int getNumberOfChoresMissingStartTime() { return choresMissingStartTime.size(); }
3.68
hadoop_ComponentContainers_addContainer
/** * Add a container. * @param container container */ public void addContainer(Container container) { containers.add(container); }
3.68
framework_VTooltip_getFinalY
/** * Return the final Y-coordinate of the tooltip based on cursor * position, size of the tooltip, size of the page and necessary * margins. * * @param offsetHeight * @return The final y-coordinate * */ private int getFinalY(int offsetHeight) { int y = 0; int heightNeeded = 10 + offsetHeight; int r...
3.68
framework_Buffered_getSource
/** * Gets a source of the exception. * * @return the Buffered object which generated this exception. */ public Buffered getSource() { return source; }
3.68
flink_WindowedOperatorTransformation_evictor
/** * Sets the {@code Evictor} that should be used to evict elements from a window before emission. * * <p>Note: When using an evictor window performance will degrade significantly, since * incremental aggregation of window results cannot be used. */ @PublicEvolving public WindowedOperatorTransformation<T, K, W> e...
3.68
zilla_HpackContext_staticIndex19
// Index in static table for the given name of length 19 private static int staticIndex19(DirectBuffer name) { switch (name.getByte(18)) { case 'e': if (STATIC_TABLE[43].name.equals(name)) // if-unmodified-since { return 43; } break; case 'n': if (S...
3.68
hadoop_SnappyCompressor_needsInput
/** * Returns true if the input data buffer is empty and * #setInput() should be called to provide more input. * * @return <code>true</code> if the input data buffer is empty and * #setInput() should be called in order to provide more input. */ @Override public boolean needsInput() { return !(compressed...
3.68
framework_ApplicationConfiguration_getUIId
/** * Gets the UI id of the server-side UI associated with this client-side * instance. The UI id should be included in every request originating from * this instance in order to associate the request with the right UI * instance on the server. * * @return the UI id */ public int getUIId() { return uiId; }
3.68
hadoop_StorageReceivedDeletedBlocks_getStorageID
/** * @deprecated Use {@link #getStorage()} instead */ @Deprecated public String getStorageID() { return storage.getStorageID(); }
3.68
framework_ContainerOrderedWrapper_addToOrderWrapper
/** * Registers the specified Item after the specified itemId in the wrapper's * internal ordering. The underlying container is not modified. Given item * id must be in the container, or must be null. * * @param id * the ID of the Item to be added to the ordering. * @param previousItemId * ...
3.68
flink_ColumnOperationUtils_renameColumns
/** * Creates a projection list that renames existing columns to new names. * * <p><b>NOTE:</b> Resulting expression are still unresolved. * * @param inputFields names of current columns * @param newAliases new aliases for current columns * @return projection expressions */ static List<Expression> renameColumns...
3.68
flink_CopyOnWriteSkipListStateMap_helpGetValueLen
/** Returns the length of the value. */ int helpGetValueLen(long valuePointer) { return SkipListUtils.helpGetValueLen(valuePointer, spaceAllocator); }
3.68
hbase_QuotaTableUtil_getNamespaceSnapshotSize
/** * Fetches the computed size of all snapshots against tables in a namespace for space quotas. */ static long getNamespaceSnapshotSize(Connection conn, String namespace) throws IOException { try (Table quotaTable = conn.getTable(QuotaTableUtil.QUOTA_TABLE_NAME)) { Result r = quotaTable.get(createGetNamespaceS...
3.68
hadoop_DistributedCache_addLocalArchives
/** * Add a archive that has been localized to the conf. Used * by internal DistributedCache code. * @param conf The conf to modify to contain the localized caches * @param str a comma separated list of local archives */ @Deprecated public static void addLocalArchives(Configuration conf, String str) { String ar...
3.68
pulsar_AbstractTopic_updateBrokerSubscriptionTypesEnabled
// subscriptionTypesEnabled is dynamic and can be updated online. public void updateBrokerSubscriptionTypesEnabled() { topicPolicies.getSubscriptionTypesEnabled().updateBrokerValue( subTypeStringsToEnumSet(brokerService.pulsar().getConfiguration().getSubscriptionTypesEnabled())); }
3.68
morf_SqlServerDialect_getSqlforLength
/** * @see org.alfasoftware.morf.jdbc.SqlDialect#getSqlforLength(Function) */ @Override protected String getSqlforLength(Function function){ return String.format("LEN(%s)", getSqlFrom(function.getArguments().get(0))); }
3.68
flink_VertexThreadInfoTrackerBuilder_setNumSamples
/** * Sets {@code numSamples}. * * @param numSamples Number of thread info samples to collect for each subtask. * @return Builder. */ public VertexThreadInfoTrackerBuilder setNumSamples(int numSamples) { this.numSamples = numSamples; return this; }
3.68
framework_FocusableHTML_focus
/** * Focus the panel. */ @Override public void focus() { setFocus(true); }
3.68
hadoop_OBSInputStream_toString
/** * String value includes statistics as well as stream state. <b>Important: * there are no guarantees as to the stability of this value.</b> * * @return a string value for printing in logs/diagnostics */ @Override @InterfaceStability.Unstable public String toString() { synchronized (this) { return "OBSInpu...
3.68
framework_VTabsheet_setTabIndex
/** * For internal use only. May be renamed or removed in a future release. * <p> * Sets the tabulator index for the active tab of the tab sheet. The active * tab represents the entire tab sheet in the browser's focus cycle * (excluding any focusable elements within the content panel). * <p> * This value is dele...
3.68
morf_XmlDataSetProducer_getView
/** * @see org.alfasoftware.morf.metadata.Schema#getView(java.lang.String) */ @Override public View getView(String name) { throw new IllegalArgumentException("No view named [" + name + "]. Views not supported in XML datasets"); }
3.68
flink_FlinkAggregateExpandDistinctAggregatesRule_convertSingletonDistinct
/** * Converts an aggregate with one distinct aggregate and one or more non-distinct aggregates to * multi-phase aggregates (see reference example below). * * @param relBuilder Contains the input relational expression * @param aggregate Original aggregate * @param argLists Arguments and filters to the distinct ag...
3.68
rocketmq-connect_DatabaseDialect_buildDropTableStatement
// drop table default String buildDropTableStatement(TableId table, boolean ifExists, boolean cascade) { ExpressionBuilder builder = expressionBuilder(); builder.append("DROP TABLE "); builder.append(table); if (ifExists) { builder.append(" IF EXISTS"); } if (cascade) { builder.a...
3.68
flink_IterativeDataSet_getAggregators
/** * Gets the registry for aggregators. On the registry, one can add {@link Aggregator}s and an * aggregator-based {@link ConvergenceCriterion}. This method offers an alternative way to * registering the aggregators via {@link #registerAggregator(String, Aggregator)} and {@link * #registerAggregationConvergenceCri...
3.68
flink_DeclarativeSlotPoolService_onStart
/** * This method is called when the slot pool service is started. It can be overridden by * subclasses. * * @param componentMainThreadExecutor componentMainThreadExecutor used by this slot pool service */ protected void onStart(ComponentMainThreadExecutor componentMainThreadExecutor) {}
3.68
framework_ConnectorTracker_notifyMarkedAsDirtyListeners
/** * Notify all registered MarkedAsDirtyListeners the given client connector * has been marked as dirty. * * @param connector * client connector marked as dirty * @since 8.4 */ public void notifyMarkedAsDirtyListeners(ClientConnector connector) { MarkedAsDirtyConnectorEvent event = new MarkedAsDi...
3.68
flink_CallExpression_permanent
/** * Creates a {@link CallExpression} to a resolved built-in function. It assumes that the {@link * BuiltInFunctionDefinition} instance is provided by the framework (usually the core module). */ @Internal public static CallExpression permanent( BuiltInFunctionDefinition builtInFunctionDefinition, Li...
3.68
framework_Calendar_getMonthNamesShort
/** * Localized display names for months starting from January. Returned * array's length is always 12. * * @return Array of localized month names. */ protected String[] getMonthNamesShort() { DateFormatSymbols s = new DateFormatSymbols(getLocale()); return Arrays.copyOf(s.getShortMonths(), 12); }
3.68
hbase_MetricsConnection_getHedgedReadWin
/** hedgedReadWin metric */ public Counter getHedgedReadWin() { return hedgedReadWin; }
3.68
flink_StructuredOptionsSplitter_escapeWithSingleQuote
/** * Escapes the given string with single quotes, if the input string contains a double quote or * any of the given {@code charsToEscape}. Any single quotes in the input string will be escaped * by doubling. * * <p>Given that the escapeChar is (;) * * <p>Examples: * * <ul> * <li>A,B,C,D => A,B,C,D * <li...
3.68
flink_JoinInputSideSpec_getUniqueKeySelector
/** * Returns the {@link KeySelector} to extract unique key from the input row. Returns null if the * input hasn't unique key. */ @Nullable public KeySelector<RowData, RowData> getUniqueKeySelector() { return uniqueKeySelector; }
3.68
hadoop_SingleFilePerBlockCache_size
/** * Gets the number of blocks in this cache. */ @Override public int size() { return blocks.size(); }
3.68
flink_TimestampUtil_isHiveTimestampColumnVector
// whether a ColumnVector is the new TimestampColumnVector public static boolean isHiveTimestampColumnVector(ColumnVector vector) { return hiveTSColVectorClz != null && hiveTSColVectorClz.isAssignableFrom(vector.getClass()); }
3.68
dubbo_DefaultApplicationDeployer_supportsExtension
/** * Supports the extension with the specified class and name * * @param extensionClass the {@link Class} of extension * @param name the name of extension * @return if supports, return <code>true</code>, or <code>false</code> * @since 2.7.8 */ private boolean supportsExtension(Class<?> extensionClass,...
3.68
querydsl_MultiCurveExpression_isClosed
/** * Returns 1 (TRUE) if this MultiCurve is closed [StartPoint ( ) = EndPoint ( ) for each * Curve in this MultiCurve]. * * @return closed */ public BooleanExpression isClosed() { if (closed == null) { closed = Expressions.booleanOperation(SpatialOps.IS_CLOSED, mixin); } return closed; }
3.68
flink_OneInputTransformation_setStateKeySelector
/** * Sets the {@link KeySelector} that must be used for partitioning keyed state of this * operation. * * @param stateKeySelector The {@code KeySelector} to set */ public void setStateKeySelector(KeySelector<IN, ?> stateKeySelector) { this.stateKeySelector = stateKeySelector; updateManagedMemoryStateBacke...
3.68
flink_MemorySegment_getChar
/** * Reads a char value from the given position, in the system's native byte order. * * @param index The position from which the memory will be read. * @return The char value at the given position. * @throws IndexOutOfBoundsException Thrown, if the index is negative, or larger than the * segment size minus 2...
3.68
hbase_LruBlockCache_clearCache
/** Clears the cache. Used in tests. */ public void clearCache() { this.map.clear(); this.elements.set(0); }
3.68
hbase_RegionServerSnapshotManager_waitForOutstandingTasks
/** * Wait for all of the currently outstanding tasks submitted via {@link #submitTask(Callable)}. * This *must* be called after all tasks are submitted via submitTask. * @return <tt>true</tt> on success, <tt>false</tt> otherwise * @throws SnapshotCreationException if the snapshot failed while we were waiting */ b...
3.68
framework_DefaultDeploymentConfiguration_getHeartbeatInterval
/** * {@inheritDoc} * <p> * The default interval is 300 seconds (5 minutes). */ @Override public int getHeartbeatInterval() { return heartbeatInterval; }
3.68
framework_ColumnProperty_getOldValue
/** * Returns the original non-modified value of this property if it has been * modified. * * @return The original value if <code>isModified()</code> is true, * <code>getValue()</code> otherwise. */ public Object getOldValue() { return value; }
3.68
flink_HighAvailabilityServices_getWebMonitorLeaderElection
/** * Gets the {@link LeaderElection} for the cluster's rest endpoint. * * @deprecated Use {@link #getClusterRestEndpointLeaderElection()} instead. */ @Deprecated default LeaderElection getWebMonitorLeaderElection() { throw new UnsupportedOperationException( "getWebMonitorLeaderElectionService shoul...
3.68
hbase_MetaTableAccessor_put
/** * @param t Table to use * @param p put to make */ private static void put(Table t, Put p) throws IOException { debugLogMutation(p); t.put(p); }
3.68
hadoop_Check_notNullElements
/** * Verifies a list does not have any NULL elements. * * @param list the list to check. * @param name the name to use in the exception message. * * @return the list. * * @throws IllegalArgumentException if the list has NULL elements. */ public static <T> List<T> notNullElements(List<T> list, String name) { ...
3.68
flink_StatusWatermarkValve_inputWatermarkStatus
/** * Feed a {@link WatermarkStatus} into the valve. This may trigger the valve to output either a * new Watermark Status, for which {@link DataOutput#emitWatermarkStatus(WatermarkStatus)} will * be called, or a new Watermark, for which {@link DataOutput#emitWatermark(Watermark)} will be * called. * * @param wate...
3.68
hadoop_Lz4Codec_createOutputStream
/** * Create a {@link CompressionOutputStream} that will write to the given * {@link OutputStream} with the given {@link Compressor}. * * @param out the location for the final output stream * @param compressor compressor to use * @return a stream the user can write uncompressed data to have it compressed ...
3.68
hadoop_NativeRuntime_createNativeObject
/** * create native object We use it to create native handlers */ public synchronized static long createNativeObject(String clazz) { assertNativeLibraryLoaded(); final long ret = JNICreateNativeObject(clazz.getBytes(StandardCharsets.UTF_8)); if (ret == 0) { LOG.warn("Can't create NativeObject for class " + ...
3.68
flink_FileLock_init
/** * Check whether the locking file exists in the file system. Create it if it does not exist. * Then create a FileOutputStream for it. * * @throws IOException If the file path is invalid or the parent dir does not exist */ private void init() throws IOException { if (!this.file.exists()) { this.file....
3.68
hadoop_RMAppKillByClientEvent_getCallerUGI
/** * returns the {@link UserGroupInformation} information. * @return UserGroupInformation */ public final UserGroupInformation getCallerUGI() { return callerUGI; }
3.68
hbase_ProcedureMember_defaultPool
/** * Default thread pool for the procedure * @param procThreads the maximum number of threads to allow in the pool * @param keepAliveMillis the maximum time (ms) that excess idle threads will wait for new tasks */ public static ThreadPoolExecutor defaultPool(String memberName, int procThreads, long keepAlive...
3.68
flink_Types_LIST
/** * Returns type information for a Java {@link java.util.List}. A list must not be null. Null * values in elements are not supported. * * <p>By default, lists are untyped and treated as a generic type in Flink; therefore, it is * useful to pass type information whenever a list is used. * * <p><strong>Note:</st...
3.68
hadoop_AppPriorityACLConfigurationParser_getUserOrGroupACLStringFromConfig
/* * This method will help to append user/group acl string against given * priority. For example "user1,user2 group1,group2" */ private StringBuilder getUserOrGroupACLStringFromConfig(String value) { // ACL strings could be generate for USER or GRUOP. // aclList in map contains two entries. 1. USER, 2. GROUP. ...
3.68
morf_UpgradePath_upgradeInProgress
/** * Returns whether or not this upgrade knew that an upgrade was in progress * at the point it was created. * * @return true if there was an upgrade in progress. */ public boolean upgradeInProgress() { return upgradeStatus != null && upgradeStatus != UpgradeStatus.NONE; }
3.68
hbase_RequestConverter_buildSetCleanerChoreRunningRequest
/** * Creates a request for enabling/disabling the cleaner chore * @return A {@link SetCleanerChoreRunningRequest} */ public static SetCleanerChoreRunningRequest buildSetCleanerChoreRunningRequest(boolean on) { return SetCleanerChoreRunningRequest.newBuilder().setOn(on).build(); }
3.68
hadoop_DatanodeLocalInfo_getDatanodeLocalReport
/** A formatted string for printing the status of the DataNode. */ public String getDatanodeLocalReport() { return ("Uptime: " + getUptime()) + ", Software version: " + getSoftwareVersion() + ", Config version: " + getConfigVersion(); }
3.68
flink_DeltaTrigger_of
/** * Creates a delta trigger from the given threshold and {@code DeltaFunction}. * * @param threshold The threshold at which to trigger. * @param deltaFunction The delta function to use * @param stateSerializer TypeSerializer for the data elements. * @param <T> The type of elements on which this trigger can oper...
3.68
hudi_HadoopConfigurations_getHadoopConf
/** * Creates a new hadoop configuration that is initialized with the given flink configuration. */ public static org.apache.hadoop.conf.Configuration getHadoopConf(Configuration conf) { org.apache.hadoop.conf.Configuration hadoopConf = FlinkClientUtil.getHadoopConf(); Map<String, String> options = FlinkOptions.g...
3.68
morf_GraphBasedUpgradeNode_isRoot
/** * @return true if this node is a no-op root node of the graph */ public boolean isRoot() { return sequence == 0; }
3.68
hbase_VisibilityController_buildException
/** Returns NameValuePair of the exception name to stringified version os exception. */ // Copied from ResponseConverter and made private. Only used in here. private static NameBytesPair buildException(final Throwable t) { NameBytesPair.Builder parameterBuilder = NameBytesPair.newBuilder(); parameterBuilder.setName...
3.68
flink_HiveDDLUtils_defaultTrait
// a constraint is by default ENABLE NOVALIDATE RELY public static byte defaultTrait() { byte res = enableConstraint((byte) 0); res = relyConstraint(res); return res; }
3.68
pulsar_SchemaStorage_put
/** * Put the schema to the schema storage. * * @param key The schema ID * @param fn The function to calculate the value and hash that need to put to the schema storage * The input of the function is all the existing schemas that used to do the schemas compatibility check * @return The schema version of...
3.68