name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
hbase_MasterObserver_preGrant | /**
* Called before granting user permissions.
* @param ctx the coprocessor instance's environment
* @param userPermission the user and permissions
* @param mergeExistingPermissions True if merge with previous granted permissions
*/
default void preGrant(ObserverContext<MasterCoproce... | 3.68 |
morf_ArchiveDataSetWriter_clearDestination | /**
* @see org.alfasoftware.morf.xml.XmlStreamProvider.XmlOutputStreamProvider#clearDestination()
*/
@Override
public void clearDestination() {
// No-op. Done by the open method.
} | 3.68 |
hadoop_ResourceTypeInfo_newInstance | /**
* Create a new instance of ResourceTypeInfo from name.
*
* @param name name of resource type
* @return the new ResourceTypeInfo object
*/
@InterfaceAudience.Public
@InterfaceStability.Unstable
public static ResourceTypeInfo newInstance(String name) {
return ResourceTypeInfo.newInstance(name, "");
} | 3.68 |
hbase_LogLevel_main | /**
* A command line implementation
*/
public static void main(String[] args) throws Exception {
CLI cli = new CLI(new Configuration());
System.exit(cli.run(args));
} | 3.68 |
hudi_HoodieRepairTool_readConfigFromFileSystem | /**
* Reads config from the file system.
*
* @param jsc {@link JavaSparkContext} instance.
* @param cfg {@link Config} instance.
* @return the {@link TypedProperties} instance.
*/
private TypedProperties readConfigFromFileSystem(JavaSparkContext jsc, Config cfg) {
return UtilHelpers.readConfig(jsc.hadoopConfigu... | 3.68 |
dubbo_ReflectUtils_name2class | /**
* name to class.
* "boolean" => boolean.class
* "java.util.Map[][]" => java.util.Map[][].class
*
* @param cl ClassLoader instance.
* @param name name.
* @return Class instance.
*/
private static Class<?> name2class(ClassLoader cl, String name) throws ClassNotFoundException {
int c = 0, index = name.in... | 3.68 |
flink_HiveParserQB_containsQueryWithoutSourceTable | /**
* returns true, if the query block contains any query, or subquery without a source table. Like
* select current_user(), select current_database()
*
* @return true, if the query block contains any query without a source table
*/
public boolean containsQueryWithoutSourceTable() {
for (HiveParserQBExpr qbexp... | 3.68 |
pulsar_LinuxInfoUtils_getTotalNicLimit | /**
* Get all physical nic limit.
* @param nics All nic path
* @param bitRateUnit Bit rate unit
* @return Total nic limit
*/
public static double getTotalNicLimit(List<String> nics, BitRateUnit bitRateUnit) {
return bitRateUnit.convert(nics.stream().mapToDouble(nicPath -> {
try {
return rea... | 3.68 |
flink_TypeInformation_of | /**
* Creates a TypeInformation for a generic type via a utility "type hint". This method can be
* used as follows:
*
* <pre>{@code
* TypeInformation<Tuple2<String, Long>> info = TypeInformation.of(new TypeHint<Tuple2<String, Long>>(){});
* }</pre>
*
* @param typeHint The hint for the generic type.
* @param <T... | 3.68 |
framework_DefaultEditorEventHandler_isOpenEvent | /**
* Returns whether the given event should open the editor. The default
* implementation returns true if and only if the event is a doubleclick or
* if it is a keydown event and the keycode is {@link #KEYCODE_OPEN}.
*
* @param event
* the received event
* @return true if the event is an open event, ... | 3.68 |
rocketmq-connect_LocalStateManagementServiceImpl_prePersist | /**
* pre persist
*/
private void prePersist() {
Map<String, ConnAndTaskStatus.CacheEntry<ConnectorStatus>> connectors = connAndTaskStatus.getConnectors();
if (connectors.isEmpty()) {
return;
}
connectors.forEach((connectName, connectorStatus) -> {
connectorStatusStore.put(connectName,... | 3.68 |
hadoop_Cluster_getSystemDir | /**
* Grab the jobtracker system directory path where
* job-specific files will be placed.
*
* @return the system directory where job-specific files are to be placed.
*/
public Path getSystemDir() throws IOException, InterruptedException {
if (sysDir == null) {
sysDir = new Path(client.getSystemDir());
... | 3.68 |
pulsar_NamespacesBase_internalRemoveBacklogQuota | /**
* Base method for removeBacklogQuota v1 and v2.
* Notion: don't re-use this logic.
*/
protected void internalRemoveBacklogQuota(AsyncResponse asyncResponse, BacklogQuotaType backlogQuotaType) {
validateNamespacePolicyOperationAsync(namespaceName, PolicyName.BACKLOG, PolicyOperation.WRITE)
.thenCo... | 3.68 |
framework_VAbstractCalendarPanel_onCancel | /**
* Notifies submit-listeners of a cancel event
*/
private void onCancel() {
if (getSubmitListener() != null) {
getSubmitListener().onCancel();
}
} | 3.68 |
rocketmq-connect_ConnectMetrics_registry | /**
* get metric registry
*
* @return
*/
public MetricRegistry registry() {
return metricRegistry;
} | 3.68 |
flink_ExpandColumnFunctionsRule_indexOfName | /** Find the index of targetName in the list. Return -1 if not found. */
private static int indexOfName(
List<UnresolvedReferenceExpression> inputFieldReferences, String targetName) {
int i;
for (i = 0; i < inputFieldReferences.size(); ++i) {
if (inputFieldReferences.get(i).getName().equals(targ... | 3.68 |
hbase_ReplicationPeer_isPeerEnabled | /**
* Test whether the peer is enabled.
* @return {@code true} if enabled, otherwise {@code false}.
*/
default boolean isPeerEnabled() {
return getPeerState() == PeerState.ENABLED;
} | 3.68 |
hbase_HFileArchiveUtil_getTableArchivePath | /**
* Get the path to the table archive directory based on the configured archive directory.
* <p>
* Assumed that the table should already be archived.
* @param conf {@link Configuration} to read the archive directory property. Can be null
* @param tableName Name of the table to be archived. Cannot be null.
... | 3.68 |
hadoop_AuxServiceFile_srcFile | /**
* This provides the source location of the configuration file, the content
* of which is dumped to dest_file post property substitutions, in the format
* as specified in type. Typically the src_file would point to a source
* controlled network accessible file maintained by tools like puppet, chef,
* or hdfs et... | 3.68 |
framework_FieldGroup_addCommitHandler | /**
* Adds a commit handler.
* <p>
* The commit handler is called before the field values are committed to the
* item ( {@link CommitHandler#preCommit(CommitEvent)}) and after the item
* has been updated ({@link CommitHandler#postCommit(CommitEvent)}). If a
* {@link CommitHandler} throws a CommitException the who... | 3.68 |
graphhopper_PathMerger_updateInstructionsWithContext | /**
* This method iterates over all instructions and uses the available context to improve the instructions.
* If the requests contains a heading, this method can transform the first continue to a u-turn if the heading
* points into the opposite direction of the route.
* At a waypoint it can transform the continue ... | 3.68 |
framework_GeneratedPropertyContainer_getContainerPropertyIds | /**
* Returns a list of propety ids available in this container. This
* collection will contain properties for generated properties. Removed
* properties will not show unless there is a generated property overriding
* those.
*/
@Override
public Collection<?> getContainerPropertyIds() {
Set<Object> wrappedPrope... | 3.68 |
hbase_ReplicationPeers_getPeer | /**
* Returns the ReplicationPeerImpl for the specified cached peer. This ReplicationPeer will
* continue to track changes to the Peer's state and config. This method returns null if no peer
* has been cached with the given peerId.
* @param peerId id for the peer
* @return ReplicationPeer object
*/
public Replica... | 3.68 |
morf_SqlDialect_getSqlForInsertInto | /**
* Returns the INSERT INTO statement.
*
* @param insertStatement he {@linkplain InsertStatement} object which can be used by the overriding methods to customize the INSERT statement.
* @return the INSERT INTO statement.
*/
protected String getSqlForInsertInto(@SuppressWarnings("unused") InsertStatement insertSt... | 3.68 |
hbase_HeterogeneousRegionCountCostFunction_readFile | /**
* used to read the rule files from either HDFS or local FS
*/
private List<String> readFile(final String filename) {
if (null == filename) {
return null;
}
try {
if (filename.startsWith("file:")) {
return readFileFromLocalFS(filename);
}
return readFileFromHDFS(filename);
} catch (IO... | 3.68 |
AreaShop_BuyRegion_setPrice | /**
* Change the price of the region.
* @param price The price to set this region to
*/
public void setPrice(Double price) {
setSetting("buy.price", price);
} | 3.68 |
hadoop_S3APrefetchingInputStream_available | /**
* Returns the number of bytes available for reading without blocking.
*
* @return the number of bytes available for reading without blocking.
* @throws IOException if there is an IO error during this operation.
*/
@Override
public synchronized int available() throws IOException {
throwIfClosed();
return in... | 3.68 |
hmily_HmilySQLUtil_getExactlyNumber | /**
* Get exactly number value and type.
*
* @param value string to be converted
* @param radix radix
* @return exactly number value and type
*/
public static Number getExactlyNumber(final String value, final int radix) {
try {
return getBigInteger(value, radix);
} catch (final NumberFormatExcepti... | 3.68 |
morf_Version2to4TransformingReader_readVersion | /**
* Tests whether a given input stream contains XML format 2, and therefore
* should have the transform applied.
* <p>
* This is designed to match the known output format of
* {@link XmlDataSetConsumer} which previously produced invalid XML. It is
* deliberately brittle. There is no need for a more intelligent ... | 3.68 |
querydsl_QueryBase_having | /**
* Add filters for aggregation
*
* @param o having conditions
* @return the current object
*/
public Q having(Predicate... o) {
return queryMixin.having(o);
} | 3.68 |
flink_HsMemoryDataManager_spillBuffers | /**
* Spill buffers for each subpartition in a decision.
*
* <p>Note that: The method should not be locked, it is the responsibility of each subpartition
* to maintain thread safety itself.
*
* @param toSpill All buffers that need to be spilled in a decision.
*/
private void spillBuffers(Map<Integer, List<Buffer... | 3.68 |
hbase_SegmentFactory_createImmutableSegmentByMerge | // create new flat immutable segment from merging old immutable segments
// for merge
public ImmutableSegment createImmutableSegmentByMerge(final Configuration conf,
final CellComparator comparator, MemStoreSegmentsIterator iterator, int numOfCells,
List<ImmutableSegment> segments, CompactingMemStore.IndexType idxT... | 3.68 |
hbase_IndividualBytesFieldCell_getRowArray | /**
* Implement Cell interface
*/
// 1) Row
@Override
public byte[] getRowArray() {
// If row is null, the constructor will reject it, by {@link KeyValue#checkParameters()},
// so it is safe to return row without checking.
return row;
} | 3.68 |
framework_VaadinFinderLocatorStrategy_findConnectorsByPath | /**
* Recursively finds connectors for the elements identified by the provided
* path by traversing the connector hierarchy starting from {@code parents}
* connectors.
*
* @param path
* The path identifying elements.
* @param parents
* The list of connectors to start traversing from.
* @r... | 3.68 |
hbase_MetricsSink_getAppliedOps | /**
* Gets the total number of OPs delivered to this sink.
*/
public long getAppliedOps() {
return this.mss.getSinkAppliedOps();
} | 3.68 |
pulsar_SubscribeRateLimiter_getAvailableSubscribeRateLimit | /**
* returns available subscribes if subscribe-throttling is enabled else it returns -1.
*
* @return
*/
public long getAvailableSubscribeRateLimit(ConsumerIdentifier consumerIdentifier) {
return subscribeRateLimiter.get(consumerIdentifier)
== null ? -1 : subscribeRateLimiter.get(consumerIdentifier)... | 3.68 |
dubbo_NetUtils_getIpByHost | /**
* @param hostName
* @return ip address or hostName if UnknownHostException
*/
public static String getIpByHost(String hostName) {
try {
return InetAddress.getByName(hostName).getHostAddress();
} catch (UnknownHostException e) {
return hostName;
}
} | 3.68 |
flink_MetricListener_getMeter | /**
* Get registered {@link Meter} with identifier relative to the root metric group.
*
* @param identifier identifier relative to the root metric group
* @return Optional registered meter
*/
public Optional<Meter> getMeter(String... identifier) {
return getMetric(Meter.class, identifier);
} | 3.68 |
flink_ProjectedRowData_from | /**
* Create an empty {@link ProjectedRowData} starting from a {@link Projection}.
*
* <p>Throws {@link IllegalStateException} if the provided {@code projection} array contains
* nested projections, which are not supported by {@link ProjectedRowData}.
*
* @see Projection
* @see ProjectedRowData
*/
public static... | 3.68 |
AreaShop_Utils_getDurationFromMinutesOrString | /**
* Get setting from config that could be only a number indicating minutes.
* or a string indicating a duration string.
* @param path Path of the setting to read
* @return milliseconds that the setting indicates
*/
public static long getDurationFromMinutesOrString(String path) {
if(config.isLong(path) || config... | 3.68 |
morf_MySqlDialect_connectionTestStatement | /**
* @see org.alfasoftware.morf.jdbc.SqlDialect#connectionTestStatement()
*/
@Override
public String connectionTestStatement() {
return "select 1";
} | 3.68 |
flink_QueryableStateConfiguration_numProxyQueryThreads | /** Returns the number of query threads for the queryable state client proxy. */
public int numProxyQueryThreads() {
return numPQueryThreads;
} | 3.68 |
hudi_SparkBasedReader_readAvro | // Spark anyways globs the path and gets all the paths in memory so take the List<filePaths> as an argument.
// https://github.com/apache/spark/.../org/apache/spark/sql/execution/datasources/DataSource.scala#L251
public static JavaRDD<GenericRecord> readAvro(SparkSession sparkSession, String schemaStr, List<String> lis... | 3.68 |
querydsl_FunctionalHelpers_wrap | /**
* Wrap a Querydsl expression into a Guava function
*
* @param projection projection to wrap
* @return Guava function
*/
public static <F,T> Function<F,T> wrap(Expression<T> projection) {
Path<?> path = projection.accept(PathExtractor.DEFAULT, null);
if (path != null) {
final Evaluator<T> ev = c... | 3.68 |
flink_CastRule_create | /** Create a casting context. */
static Context create(
boolean isPrinting,
boolean legacyBehaviour,
ZoneId zoneId,
ClassLoader classLoader) {
return new Context() {
@Override
public boolean isPrinting() {
return isPrinting;
}
@Override
... | 3.68 |
querydsl_GroupBy_map | /**
* Create a new aggregating map expression using a backing LinkedHashMap
*
* @param key key for the map entries
* @param value value for the map entries
* @return wrapper expression
*/
public static <K, V, T, U> AbstractGroupExpression<Pair<K, V>, Map<T, U>> map(GroupExpression<K, T> key,
... | 3.68 |
flink_OSSTestCredentials_getOSSAccessKey | /**
* Get OSS access key.
*
* @return OSS access key
*/
public static String getOSSAccessKey() {
if (ACCESS_KEY != null) {
return ACCESS_KEY;
} else {
throw new IllegalStateException("OSS access key is not available");
}
} | 3.68 |
hudi_AvroInternalSchemaConverter_nullableSchema | /** Returns schema with nullable true. */
public static Schema nullableSchema(Schema schema) {
if (schema.getType() == UNION) {
if (!isOptional(schema)) {
throw new HoodieSchemaException(String.format("Union schemas are not supported: %s", schema));
}
return schema;
} else {
return Schema.crea... | 3.68 |
hudi_ParquetSchemaConverter_toParquetType | /**
* Converts Flink Internal Type to Parquet schema.
*
* @param typeInformation Flink type information
* @param legacyMode is standard LIST and MAP schema or back-compatible schema
* @return Parquet schema
*/
public static MessageType toParquetType(
TypeInformation<?> typeInformation, boolean legacyMode... | 3.68 |
hadoop_FederationUtil_newSecretManager | /**
* Creates an instance of DelegationTokenSecretManager from the
* configuration.
*
* @param conf Configuration that defines the token manager class.
* @return New delegation token secret manager.
*/
public static AbstractDelegationTokenSecretManager<DelegationTokenIdentifier>
newSecretManager(Configuration... | 3.68 |
hadoop_ServiceShutdownHook_shutdown | /**
* Shutdown operation.
* <p>
* Subclasses may extend it, but it is primarily
* made available for testing.
* @return true if the service was stopped and no exception was raised.
*/
protected boolean shutdown() {
Service service;
boolean result = false;
synchronized (this) {
service = serviceRef.get()... | 3.68 |
flink_GlobalProperties_filterBySemanticProperties | /**
* Filters these GlobalProperties by the fields that are forwarded to the output as described by
* the SemanticProperties.
*
* @param props The semantic properties holding information about forwarded fields.
* @param input The index of the input.
* @return The filtered GlobalProperties
*/
public GlobalPropert... | 3.68 |
framework_AbstractSingleComponentContainerConnector_getContent | /**
* Returns the content (only/first child) of the container.
*
* @return child connector or null if none (e.g. invisible or not set on
* server)
*/
protected ComponentConnector getContent() {
List<ComponentConnector> children = getChildComponents();
if (children.isEmpty()) {
return null;
... | 3.68 |
hbase_QuotaObserverChore_addTableQuotaTable | /**
* Adds a table with a table quota.
*/
public void addTableQuotaTable(TableName tn) {
tablesWithTableQuotas.add(tn);
} | 3.68 |
flink_FlinkRexBuilder_makeIn | /**
* Convert the conditions into the {@code IN} and fix [CALCITE-4888]: Unexpected {@link RexNode}
* when call {@link RelBuilder#in} to create an {@code IN} predicate with a list of varchar
* literals which have different length in {@link RexBuilder#makeIn}.
*
* <p>The bug is because the origin implementation doe... | 3.68 |
hadoop_LoadManifestsStage_toString | /**
* To String includes all summary info except statistics.
* @return string value
*/
@Override
public String toString() {
final StringBuilder sb = new StringBuilder(
"SummaryInfo{");
sb.append("manifestCount=").append(getManifestCount());
sb.append(", fileCount=").append(getFileCount());
sb.append(",... | 3.68 |
hbase_IncrementalBackupManager_getIncrBackupLogFileMap | /**
* Obtain the list of logs that need to be copied out for this incremental backup. The list is set
* in BackupInfo.
* @return The new HashMap of RS log time stamps after the log roll for this incremental backup.
* @throws IOException exception
*/
public Map<String, Long> getIncrBackupLogFileMap() throws IOExcep... | 3.68 |
hbase_HFileReaderImpl_getFirstRowKey | /**
* TODO left from {@link HFile} version 1: move this to StoreFile after Ryan's patch goes in to
* eliminate {@link KeyValue} here.
* @return the first row key, or null if the file is empty.
*/
@Override
public Optional<byte[]> getFirstRowKey() {
// We have to copy the row part to form the row key alone
retur... | 3.68 |
flink_FlinkAssertions_chainOfCauses | /**
* You can use this method in combination with {@link
* AbstractThrowableAssert#extracting(Function, AssertFactory)} to perform assertions on a chain
* of causes. For example:
*
* <pre>{@code
* assertThat(throwable)
* .extracting(FlinkAssertions::chainOfCauses, FlinkAssertions.STREAM_THROWABLE)
* }</pre>... | 3.68 |
hadoop_IOStatisticsLogging_demandStringifyIOStatisticsSource | /**
* On demand stringifier of an IOStatisticsSource instance.
* <p>
* Whenever this object's toString() method is called, it evaluates the
* statistics.
* <p>
* This is designed to affordable to use in log statements.
* @param source source of statistics -may be null.
* @return an object whose toString() opera... | 3.68 |
rocketmq-connect_DebeziumOracleConnector_taskClass | /**
* Return the current connector class
*
* @return task implement class
*/
@Override
public Class<? extends Task> taskClass() {
return DebeziumOracleSource.class;
} | 3.68 |
hadoop_ActiveAuditManagerS3A_createExecutionInterceptors | /**
* Return a list of execution interceptors for the AWS SDK which
* relays to this class.
* @return a list of execution interceptors.
*/
@Override
public List<ExecutionInterceptor> createExecutionInterceptors()
throws IOException {
// wire up the AWS SDK To call back into this class when
// preparing to ... | 3.68 |
hadoop_ProducerConsumer_addWorker | /**
* Add another worker that will consume WorkRequest{@literal <T>} items
* from input queue, process each item using supplied processor, and for
* every processed item output WorkReport{@literal <R>} to output queue.
*
* @param processor Processor implementing WorkRequestProcessor interface.
*
*/
public v... | 3.68 |
framework_UIDL_getDoubleVariable | /**
* Gets the value of the named variable.
*
* @param name
* the name of the variable
* @return the value of the variable
*/
public double getDoubleVariable(String name) {
return var().getRawNumber(name);
} | 3.68 |
dubbo_ServiceBean_getService | /**
* Gets associated {@link Service}
*
* @return associated {@link Service}
*/
public Service getService() {
return service;
} | 3.68 |
hbase_DynamicMetricsRegistry_getGauge | /**
* Get a MetricMutableGaugeLong from the storage. If it is not there atomically put it.
* @param gaugeName name of the gauge to create or get.
* @param potentialStartingValue value of the new gauge if we have to create it.
*/
public MutableGaugeLong getGauge(String gaugeName, long potentialStartingV... | 3.68 |
flink_LocatableInputSplitAssigner_getNextUnassignedMinLocalCountSplit | /**
* Retrieves a LocatableInputSplit with minimum local count. InputSplits which have already
* been assigned (i.e., which are not contained in the provided set) are filtered out. The
* returned input split is NOT removed from the provided set.
*
* @param unassignedSplits Set of unassigned input splits.
* @retur... | 3.68 |
framework_RendererCellReference_set | /**
* Sets the identifying information for this cell.
*
* @param cell
* the flyweight cell to reference
* @param columnIndex
* the index of the column in the grid, including hidden cells
* @param column
* the column to reference
*/
public void set(FlyweightCell cell, int column... | 3.68 |
hbase_CompactingMemStore_stopCompaction | /**
* The request to cancel the compaction asynchronous task (caused by in-memory flush) The
* compaction may still happen if the request was sent too late Non-blocking request
*/
private void stopCompaction() {
if (inMemoryCompactionInProgress.get()) {
compactor.stop();
}
} | 3.68 |
framework_FormLayout_getExpandRatio | /**
* @deprecated This method currently has no effect as expand ratios are not
* implemented in FormLayout
*/
@Override
@Deprecated
public float getExpandRatio(Component component) {
return super.getExpandRatio(component);
} | 3.68 |
hbase_RegionCoprocessorHost_postFlush | /**
* Invoked after a memstore flush
*/
public void postFlush(HStore store, HStoreFile storeFile, FlushLifeCycleTracker tracker)
throws IOException {
if (coprocEnvironments.isEmpty()) {
return;
}
execOperation(new RegionObserverOperationWithoutResult() {
@Override
public void call(RegionObserver o... | 3.68 |
hadoop_StageConfig_getName | /**
* Get name of task/job.
* @return name for logging.
*/
public String getName() {
return name;
} | 3.68 |
flink_PendingCheckpoint_setCancellerHandle | /**
* Sets the handle for the canceller to this pending checkpoint. This method fails with an
* exception if a handle has already been set.
*
* @return true, if the handle was set, false, if the checkpoint is already disposed;
*/
public boolean setCancellerHandle(ScheduledFuture<?> cancellerHandle) {
synchroni... | 3.68 |
hadoop_FileIoProvider_listDirectory | /**
* Get a listing of the given directory using
* {@link IOUtils#listDirectory(File, FilenameFilter)}.
*
* @param volume target volume. null if unavailable.
* @param dir Directory to list.
* @param filter {@link FilenameFilter} to filter the directory entries.
* @throws IOException
*/
public List<String> listD... | 3.68 |
flink_FileChannelMemoryMappedBoundedData_finishWrite | /**
* Finishes the current region and prevents further writes. After calling this method, further
* calls to {@link #writeBuffer(Buffer)} will fail.
*/
@Override
public void finishWrite() throws IOException {
mapRegionAndStartNext();
fileChannel.close();
} | 3.68 |
hadoop_AbfsHttpOperation_parseListFilesResponse | /**
* Parse the list file response
*
* @param stream InputStream contains the list results.
* @throws IOException
*/
private void parseListFilesResponse(final InputStream stream) throws IOException {
if (stream == null) {
return;
}
if (listResultSchema != null) {
// already parse the response
re... | 3.68 |
framework_VFilterSelect_getStyle | /**
* Gets the style set for this suggestion item. Styles are typically set
* by a server-side {@link com.vaadin.ui.ComboBox.ItemStyleGenerator}.
* The returned style is prefixed by <code>v-filterselect-item-</code>.
*
* @since 7.5.6
* @return the style name to use, or <code>null</code> to not apply any
* ... | 3.68 |
hbase_Query_getConsistency | /**
* Returns the consistency level for this operation
* @return the consistency level
*/
public Consistency getConsistency() {
return consistency;
} | 3.68 |
hbase_HFileBlockIndex_getCacheOnWrite | /**
* @return true if we are using cache-on-write. This is configured by the caller of the
* constructor by either passing a valid block cache or null.
*/
@Override
public boolean getCacheOnWrite() {
return cacheConf != null && cacheConf.shouldCacheIndexesOnWrite();
} | 3.68 |
hadoop_BlockPoolTokenSecretManager_checkAccess | /**
* See {@link BlockTokenSecretManager#checkAccess(Token, String,
* ExtendedBlock, BlockTokenIdentifier.AccessMode,
* StorageType[], String[])}
*/
public void checkAccess(Token<BlockTokenIdentifier> token,
String userId, ExtendedBlock block, AccessMode mode,
StorageType[] sto... | 3.68 |
hbase_FileLink_getBackReferencesDir | /**
* Get the directory to store the link back references
* <p>
* To simplify the reference count process, during the FileLink creation a back-reference is added
* to the back-reference directory of the specified file.
* @param storeDir Root directory for the link reference folder
* @param fileName File Name with... | 3.68 |
querydsl_StringExpression_equalsIgnoreCase | /**
* Create a {@code this.equalsIgnoreCase(str)} expression
*
* <p>Compares this {@code StringExpression} to another {@code StringExpression}, ignoring case
* considerations.</p>
*
* @param str string
* @return this.equalsIgnoreCase(str)
* @see java.lang.String#equalsIgnoreCase(String)
*/
public BooleanExpres... | 3.68 |
framework_SQLContainer_setPageLengthInternal | /**
* Sets the page length internally, without refreshing the container.
*
* @param pageLength
* the new page length
*/
private void setPageLengthInternal(int pageLength) {
this.pageLength = pageLength > 0 ? pageLength : DEFAULT_PAGE_LENGTH;
cacheOverlap = getPageLength();
cachedItems.setCac... | 3.68 |
framework_ContainerOrderedWrapper_updateOrderWrapper | /**
* Updates the wrapper's internal ordering information to include all Items
* in the underlying container.
* <p>
* Note : If the contents of the wrapped container change without the
* wrapper's knowledge, this method needs to be called to update the
* ordering information of the Items.
* </p>
*/
public void ... | 3.68 |
hadoop_SessionTokenIdentifier_getMarshalledCredentials | /**
* Get the marshalled credentials.
* @return marshalled AWS credentials.
*/
public MarshalledCredentials getMarshalledCredentials() {
return marshalledCredentials;
} | 3.68 |
hadoop_ResourceRequestSetKey_extractMatchingKey | /**
* Extract the corresponding ResourceRequestSetKey for an allocated container
* from a given set. Return null if not found.
*
* @param container the allocated container
* @param keys the set of keys to look from
* @return ResourceRequestSetKey
*/
public static ResourceRequestSetKey extractMatchingKey(Containe... | 3.68 |
hbase_MasterCoprocessorHost_preSplitRegionAction | /**
* Invoked just before a split
* @param tableName the table where the region belongs to
* @param splitRow the split point
* @param user the user
*/
public void preSplitRegionAction(final TableName tableName, final byte[] splitRow,
final User user) throws IOException {
execOperation(coprocEnvironments.... | 3.68 |
shardingsphere-elasticjob_JobRegistry_getInstance | /**
* Get instance of job registry.
*
* @return instance of job registry
*/
public static JobRegistry getInstance() {
if (null == instance) {
synchronized (JobRegistry.class) {
if (null == instance) {
instance = new JobRegistry();
}
}
}
return ins... | 3.68 |
shardingsphere-elasticjob_QueryParameterMap_toSingleValueMap | /**
* Convert to a single value map, abandon values except the first of each parameter.
*
* @return single value map
*/
public Map<String, String> toSingleValueMap() {
return queryMap.entrySet().stream().collect(Collectors.toMap(Map.Entry::getKey, entry -> entry.getValue().get(0)));
} | 3.68 |
hbase_WALKeyImpl_getEncodedRegionName | /** Returns encoded region name */
@Override
public byte[] getEncodedRegionName() {
return encodedRegionName;
} | 3.68 |
hadoop_RouterRMAdminService_initializePipeline | /**
* Initializes the request interceptor pipeline for the specified user.
*
* @param user
*/
private RequestInterceptorChainWrapper initializePipeline(String user) {
synchronized (this.userPipelineMap) {
if (this.userPipelineMap.containsKey(user)) {
LOG.info("Request to start an already existing user: ... | 3.68 |
flink_CheckpointConfig_getMinPauseBetweenCheckpoints | /**
* Gets the minimal pause between checkpointing attempts. This setting defines how soon the
* checkpoint coordinator may trigger another checkpoint after it becomes possible to trigger
* another checkpoint with respect to the maximum number of concurrent checkpoints (see {@link
* #getMaxConcurrentCheckpoints()})... | 3.68 |
flink_TSetClientInfoResp_findByThriftIdOrThrow | /**
* Find the _Fields constant that matches fieldId, throwing an exception if it is not found.
*/
public static _Fields findByThriftIdOrThrow(int fieldId) {
_Fields fields = findByThriftId(fieldId);
if (fields == null)
throw new java.lang.IllegalArgumentException(
"Field " + fieldId +... | 3.68 |
graphhopper_Service_checkOverlap | /**
* Checks for overlapping days of week between two service calendars
* @param s1
* @param s2
* @return true if both calendars simultaneously operate on at least one day of the week
*/
public static boolean checkOverlap (Service s1, Service s2) {
if (s1.calendar == null || s2.calendar == null) {
retu... | 3.68 |
dubbo_BaseServiceMetadata_revertDisplayServiceKey | /**
* revert of org.apache.dubbo.common.ServiceDescriptor#getDisplayServiceKey()
*
* @param displayKey
* @return
*/
public static BaseServiceMetadata revertDisplayServiceKey(String displayKey) {
String[] eles = StringUtils.split(displayKey, COLON_SEPARATOR);
if (eles == null || eles.length < 1 || eles.leng... | 3.68 |
dubbo_FrameworkModel_getAllApplicationModels | /**
* Get all application models including the internal application model.
*/
public List<ApplicationModel> getAllApplicationModels() {
synchronized (globalLock) {
return Collections.unmodifiableList(applicationModels);
}
} | 3.68 |
flink_SqlLikeChainChecker_checkEnd | /** Matches the ending of each string to its pattern. */
private static boolean checkEnd(
BinaryStringData pattern, MemorySegment[] segments, int start, int len) {
int lenSub = pattern.getSizeInBytes();
return len >= lenSub
&& SegmentsUtil.equals(
pattern.getSegments(), 0... | 3.68 |
flink_ZooKeeperUtils_useNamespaceAndEnsurePath | /**
* Returns a facade of the client that uses the specified namespace, and ensures that all nodes
* in the path exist.
*
* @param client ZK client
* @param path the new namespace
* @return ZK Client that uses the new namespace
* @throws Exception ZK errors
*/
public static CuratorFramework useNamespaceAndEnsur... | 3.68 |
flink_ProjectOperator_projectTuple18 | /**
* Projects a {@link Tuple} {@link DataSet} to the previously selected fields.
*
* @return The projected DataSet.
* @see Tuple
* @see DataSet
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17>
ProjectOperator<
T,
Tu... | 3.68 |
morf_SqlUtils_field | /**
* Constructs a new field with a given name
*
* <p>Consider using {@link TableReference#field(String)} instead.</p>
*
* @param fieldName the name of the field
* @return {@link FieldReference}
* @see TableReference#field(String)
*/
public static FieldReference field(String fieldName) {
return new FieldRefer... | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.