name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_TableEnvironmentInternal_explainInternal | /**
* Returns the AST of this table and the execution plan to compute the result of this table.
*
* @param operations The operations to be explained.
* @param extraDetails The extra explain details which the explain result should include, e.g.
* estimated cost, changelog mode for streaming
* @return AST and t... | 3.68 |
querydsl_SurfaceExpression_pointOnSurface | /**
* A Point guaranteed to be on this Surface.
*
* @return point on surface
*/
public PointExpression<Point> pointOnSurface() {
if (pointOnSurface == null) {
pointOnSurface = GeometryExpressions.pointOperation(SpatialOps.POINT_ON_SURFACE, mixin);
}
return pointOnSurface;
} | 3.68 |
flink_RocksDBNativeMetricOptions_fromConfig | /** Creates a {@link RocksDBNativeMetricOptions} based on an external configuration. */
public static RocksDBNativeMetricOptions fromConfig(ReadableConfig config) {
RocksDBNativeMetricOptions options = new RocksDBNativeMetricOptions();
configurePropertyMetrics(options, config);
configureStatisticsMetrics(op... | 3.68 |
hbase_HBaseServerBase_getStartcode | /** Returns time stamp in millis of when this server was started */
public long getStartcode() {
return this.startcode;
} | 3.68 |
hadoop_MRProtoUtils_convertToProtoFormat | /*
* TaskType
*/
public static TaskTypeProto convertToProtoFormat(TaskType e) {
return TaskTypeProto.valueOf(e.name());
} | 3.68 |
hadoop_InMemoryConfigurationStore_getLogs | /**
* Configuration mutations not logged (i.e. not persisted) but directly
* confirmed. As such, a list of persisted configuration mutations does not
* exist.
* @return null Configuration mutation list not applicable for this store.
*/
@Override
protected LinkedList<LogMutation> getLogs() {
// Unimplemented.
r... | 3.68 |
hbase_CachedClusterId_getClusterId | /**
* Returns a cached copy of the cluster ID. null if the cache is not populated.
*/
private String getClusterId() {
if (!isClusterIdSet.get()) {
return null;
}
// It is ok to read without a lock since clusterId is immutable once set.
return clusterId.toString();
} | 3.68 |
rocketmq-connect_BufferedRecords_add | /**
* add record
*
* @param record
* @return
* @throws SQLException
*/
public List<ConnectRecord> add(ConnectRecord record) throws SQLException {
recordValidator.validate(record);
final List<ConnectRecord> flushed = new ArrayList<>();
boolean schemaChanged = false;
if (!Objects.equals(keySchema, r... | 3.68 |
flink_RestClientConfiguration_getIdlenessTimeout | /** {@see RestOptions#IDLENESS_TIMEOUT}. */
public long getIdlenessTimeout() {
return idlenessTimeout;
} | 3.68 |
flink_FileIOChannel_getPath | /** Returns the path to the underlying temporary file. */
public String getPath() {
return path.getAbsolutePath();
} | 3.68 |
rocketmq-connect_ExpressionBuilder_append | /**
* Append to this builder's expression the specified object surrounded by quotes. If the object
* is {@link Expressable}, then this builder delegates to the object's
* {@link Expressable#appendTo(ExpressionBuilder, boolean)} method. Otherwise, the string
* representation of the object is appended to the expressi... | 3.68 |
flink_RequestJobsOverview_readResolve | /** Preserve the singleton property by returning the singleton instance */
private Object readResolve() {
return INSTANCE;
} | 3.68 |
hbase_BaseSourceImpl_incGauge | /**
* Add some amount to a gauge.
* @param gaugeName The name of the gauge to increment.
* @param delta The amount to increment the gauge by.
*/
@Override
public void incGauge(String gaugeName, long delta) {
MutableGaugeLong gaugeInt = metricsRegistry.getGauge(gaugeName, 0L);
gaugeInt.incr(delta);
} | 3.68 |
querydsl_TypeCategory_isSubCategoryOf | /**
* transitive and reflexive subCategoryOf check
*
* @param ancestor
* @return
*/
public boolean isSubCategoryOf(TypeCategory ancestor) {
if (this == ancestor) {
return true;
} else if (superType == null) {
return false;
} else {
return superType == ancestor || superType.isSu... | 3.68 |
pulsar_TokenClient_buildClientCredentialsBody | /**
* Constructing http request parameters.
* @param req object with relevant request parameters
* @return Generate the final request body from a map.
*/
String buildClientCredentialsBody(ClientCredentialsExchangeRequest req) {
Map<String, String> bodyMap = new TreeMap<>();
bodyMap.put("grant_type", "client... | 3.68 |
hadoop_RouterClientMetrics_incInvokedConcurrent | /**
* Increase the concurrent metrics based on the method being invoked.
* @param method concurrently invoked method
*/
public void incInvokedConcurrent(Method method){
switch (method.getName()) {
case "setReplication":
concurrentSetReplicationOps.incr();
break;
case "setPermission":
concurrentSetP... | 3.68 |
hudi_HiveSchemaUtil_createHiveMap | /**
* Create a 'Map' schema from Parquet map field.
*/
private static String createHiveMap(String keyType, String valueType, boolean doFormat) {
return (doFormat ? "MAP< " : "MAP<") + keyType + (doFormat ? ", " : ",") + valueType + ">";
} | 3.68 |
hbase_ServerManager_unregisterListener | /**
* Remove the listener from the notification list.
* @param listener The ServerListener to unregister
*/
public boolean unregisterListener(final ServerListener listener) {
return this.listeners.remove(listener);
} | 3.68 |
querydsl_SimpleExpression_neAny | /**
* Create a {@code this != any right} expression
*
* @param right
* @return this != any right
*/
public BooleanExpression neAny(CollectionExpression<?, ? super T> right) {
return ne(ExpressionUtils.any(right));
} | 3.68 |
hbase_BucketAllocator_totalBytes | /**
* Combined {@link #totalCount()} * {@link #itemSize()}
*/
public long totalBytes() {
return totalCount * itemSize;
} | 3.68 |
hadoop_PeriodicRLESparseResourceAllocation_getTimePeriod | /**
* Get time period of PeriodicRLESparseResourceAllocation.
*
* @return timePeriod time period represented in ms.
*/
public long getTimePeriod() {
return this.timePeriod;
} | 3.68 |
pulsar_TxnBatchedPositionImpl_equals | /**
* It's exactly the same as {@link PositionImpl},make sure that when {@link TxnBatchedPositionImpl} used as the key
* of map same as {@link PositionImpl}. {@link #batchSize} and {@link #batchIndex} should not be involved in
* calculate, just like {@link PositionImpl#ackSet} is not involved in calculate.
* Note: ... | 3.68 |
hadoop_TwoColumnLayout_footer | /**
* @return the class that will render the footer.
*/
protected Class<? extends SubView> footer() {
return FooterBlock.class;
} | 3.68 |
framework_PureGWTTestApplication_createMenuPath | /**
* Create a menu path, if one doesn't already exist, and return the last
* menu in the series.
*
* @param path
* a varargs list or array of strings describing a menu path,
* e.g. "File", "Recent", "User Files", which would result in the
* File menu having a submenu called "Rec... | 3.68 |
querydsl_AntMetaDataExporter_addTypeMapping | /**
* Adds TypeMapping instance, called by Ant
*/
public void addTypeMapping(TypeMapping mapping) {
typeMappings.add(mapping);
} | 3.68 |
hbase_BoundedRecoveredHFilesOutputSink_createRecoveredHFileWriter | /**
* @return Returns a base HFile without compressions or encodings; good enough for recovery given
* hfile has metadata on how it was written.
*/
private StoreFileWriter createRecoveredHFileWriter(TableName tableName, String regionName,
long seqId, String familyName, boolean isMetaTable) throws IOExcepti... | 3.68 |
hudi_AvroSchemaConverter_recordTypesOfSameNumFields | /**
* Returns true if all the types are RECORD type with same number of fields.
*/
private static boolean recordTypesOfSameNumFields(List<Schema> types) {
if (types == null || types.size() == 0) {
return false;
}
if (types.stream().anyMatch(s -> s.getType() != Schema.Type.RECORD)) {
return false;
}
... | 3.68 |
pulsar_AutoConsumeSchema_unwrapInternalSchema | /**
* Get a specific schema version, fetching from the Registry if it is not loaded yet.
* This method is not intended to be used by applications.
* @param schemaVersion the version
* @return the Schema at the specific version
* @see #atSchemaVersion(byte[])
*/
public Schema<?> unwrapInternalSchema(byte[] schemaV... | 3.68 |
shardingsphere-elasticjob_IpUtils_getIp | /**
* Get IP address for localhost.
*
* @return IP address for localhost
*/
public static String getIp() {
if (null != cachedIpAddress) {
return cachedIpAddress;
}
NetworkInterface networkInterface = findNetworkInterface();
if (null != networkInterface) {
Enumeration<InetAddress> ip... | 3.68 |
morf_SqlServerMetaDataProvider_setAdditionalColumnMetadata | /**
* SQL Server does not return information on auto-increment columns in any way which JDBC can pick up,
* so we need a customised method for fetching this information.
*
* @see org.alfasoftware.morf.jdbc.DatabaseMetaDataProvider#setAdditionalColumnMetadata(RealName, ColumnBuilder, ResultSet)
*/
@Override
protect... | 3.68 |
hadoop_OBSFileSystem_getScheme | /**
* Return the protocol scheme for the FileSystem.
*
* @return "obs"
*/
@Override
public String getScheme() {
return "obs";
} | 3.68 |
hbase_AvlUtil_appendList | /**
* Append a list of nodes to the tree
* @param head the head of the current linked list
* @param otherHead the head of the list to append to the current list
* @return the new head of the current list
*/
public static <TNode extends AvlLinkedNode> TNode appendList(TNode head, TNode otherHead) {
if (head ... | 3.68 |
Activiti_Builder_main | /**
* Dump out abstract syntax tree for a given expression
*
* @param args array with one element, containing the expression string
*/
public static void main(String[] args) {
if (args.length != 1) {
System.err.println(
"usage: java " +
Builder.class.getName() +
" <ex... | 3.68 |
morf_Function_leftTrim | /**
* Helper method to create an instance of the "leftTrim" SQL function,
* which will result in argument having leading spaces removed.
*
* @param expression the field to evaluate.
* @return an instance of the leftTrim function.
*/
public static Function leftTrim(AliasedField expression) {
return new Function(... | 3.68 |
flink_CliFrontend_handleParametrizationException | /**
* Displays an optional exception message for incorrect program parametrization.
*
* @param e The exception to display.
* @return The return code for the process.
*/
private static int handleParametrizationException(ProgramParametrizationException e) {
LOG.error("Program has not been parametrized properly."... | 3.68 |
hudi_MarkerUtils_getAllMarkerDir | /**
* Gets all marker directories.
*
* @param tempPath Temporary folder under .hoodie.
* @param fs File system to use.
* @return All marker directories.
* @throws IOException upon error.
*/
public static List<Path> getAllMarkerDir(Path tempPath, FileSystem fs) throws IOException {
return Arrays.stream(fs... | 3.68 |
hadoop_InterruptEscalator_run | /**
* Shutdown callback: stop the service and set an atomic boolean
* if it stopped within the shutdown time.
*/
@Override
public void run() {
if (service != null) {
service.stop();
serviceWasShutdown.set(
service.waitForServiceToStop(shutdownTimeMillis));
} else {
serviceWasShutdown.set(true... | 3.68 |
framework_Notification_getDescription | /**
* Gets the description part of the notification message.
*
* @return The message description
*/
public String getDescription() {
return getState(false).description;
} | 3.68 |
flink_LookupCacheManager_getInstance | /** Get the shared instance of {@link LookupCacheManager}. */
public static synchronized LookupCacheManager getInstance() {
if (instance == null) {
instance = new LookupCacheManager();
}
return instance;
} | 3.68 |
hbase_CellChunkImmutableSegment_useIndexChunks | // Assuming we are going to use regular data chunks as index chunks,
// we check here how much free space will remain in the last allocated chunk
// (the least occupied one).
// If the percentage of its remaining free space is above the INDEX_CHUNK_UNUSED_SPACE
// threshold, then we will use index chunks (which are sma... | 3.68 |
hadoop_MapTaskAttemptInfo_getMapRuntime | /**
* Get the runtime for the <b>map</b> phase of the map-task attempt.
*
* @return the runtime for the <b>map</b> phase of the map-task attempt
*/
public long getMapRuntime() {
return runtime;
} | 3.68 |
framework_LayoutManager_getMarginWidth | /**
* Gets the combined left & right margin of the given element, provided that
* they have been measured. These elements are guaranteed to be measured:
* <ul>
* <li>ManagedLayouts and their child Connectors
* <li>Elements for which there is at least one ElementResizeListener
* <li>Elements for which at least one... | 3.68 |
flink_SecurityOptions_isRestSSLAuthenticationEnabled | /** Checks whether mutual SSL authentication for the external REST endpoint is enabled. */
public static boolean isRestSSLAuthenticationEnabled(Configuration sslConfig) {
checkNotNull(sslConfig, "sslConfig");
return isRestSSLEnabled(sslConfig) && sslConfig.getBoolean(SSL_REST_AUTHENTICATION_ENABLED);
} | 3.68 |
flink_AbstractFileSource_processStaticFileSet | /**
* Sets this source to bounded (batch) mode.
*
* <p>In this mode, the source processes the files that are under the given paths when the
* application is started. Once all files are processed, the source will finish.
*
* <p>This setting is also the default behavior. This method is mainly here to "switch back"
... | 3.68 |
streampipes_SimpleEstimator_isLowQuality | /**
* Given the statistics of the document before and after applying the {@link BoilerpipeExtractor},
* can we regard the extraction quality (too) low?
*
* Works well with {@link DefaultExtractor}, {@link ArticleExtractor} and others.
*
* @param dsBefore
* @param dsAfter
* @return true if low quality is to be e... | 3.68 |
querydsl_DateTimeExpression_second | /**
* Create a seconds expression (range 0-59)
*
* @return second
*/
public NumberExpression<Integer> second() {
if (seconds == null) {
seconds = Expressions.numberOperation(Integer.class, Ops.DateTimeOps.SECOND, mixin);
}
return seconds;
} | 3.68 |
flink_MapValue_keySet | /*
* (non-Javadoc)
* @see java.util.Map#keySet()
*/
@Override
public Set<K> keySet() {
return this.map.keySet();
} | 3.68 |
framework_VAbstractDropHandler_updateAcceptRules | /**
* Implementor/user of {@link VAbstractDropHandler} must pass the UIDL
* painted by {@link AcceptCriterion} to this method. Practically the
* details about {@link AcceptCriterion} are saved.
*
* @param uidl
* the accept criterion UIDL
*/
public void updateAcceptRules(UIDL uidl) {
criterioUIDL =... | 3.68 |
framework_Criterion_getValue | /**
* Gets the value of the payload to be compared.
*
* @return value of the payload to be compared
*/
public String getValue() {
return value;
} | 3.68 |
flink_MurmurHashUtil_hashBytesByWords | /**
* Hash bytes in MemorySegment, length must be aligned to 4 bytes.
*
* @param segment segment.
* @param offset offset for MemorySegment
* @param lengthInBytes length in MemorySegment
* @return hash code
*/
public static int hashBytesByWords(MemorySegment segment, int offset, int lengthInBytes) {
return ha... | 3.68 |
dubbo_ConfigUtils_getSystemProperty | /**
* System environment -> System properties
*
* @param key key
* @return value
*/
public static String getSystemProperty(String key) {
String value = System.getenv(key);
if (StringUtils.isEmpty(value)) {
value = System.getProperty(key);
}
return value;
} | 3.68 |
querydsl_HibernateInsertClause_setLockMode | /**
* Set the lock mode for the given path.
* @return the current object
*/
@SuppressWarnings("unchecked")
public HibernateInsertClause setLockMode(Path<?> path, LockMode lockMode) {
lockModes.put(path, lockMode);
return this;
} | 3.68 |
flink_HiveTableUtil_extractRowType | /** Create the Hive table's row type. */
public static DataType extractRowType(
HiveConf hiveConf,
Table hiveTable,
HiveMetastoreClientWrapper client,
HiveShim hiveShim) {
Tuple4<List<FieldSchema>, List<FieldSchema>, Set<String>, Optional<UniqueConstraint>>
hiveTableInfo... | 3.68 |
querydsl_OrderSpecifier_isAscending | /**
* Get whether the order is ascending or not
*
* @return ascending order
*/
public boolean isAscending() {
return order == Order.ASC;
} | 3.68 |
hbase_RegionServerFlushTableProcedureManager_stop | /**
* Gracefully shutdown the thread pool. An ongoing HRegion.flush() should not be interrupted
* (see HBASE-13877)
*/
void stop() {
if (this.stopped) return;
this.stopped = true;
this.executor.shutdown();
} | 3.68 |
hbase_Constraints_getKeyValueForClass | /**
* Get the kv {@link Entry} in the descriptor builder for the specified class
* @param builder {@link TableDescriptorBuilder} to read
* @param clazz To search for
* @return The {@link Pair} of {@literal <key, value>} in the table, if that class is present.
* {@code null} otherwise.
*/
private static ... | 3.68 |
rocketmq-connect_PatternFilter_filter | /**
* filter map
*
* @param record
* @param map
* @return
*/
R filter(R record, Map map) {
for (Object field : map.keySet()) {
if (!this.fields.contains(field)) {
continue;
}
Object value = map.get(field);
if (value instanceof String) {
String input = (S... | 3.68 |
hbase_FavoredNodeAssignmentHelper_generateMissingFavoredNodeSingleRack | /*
* Generate FN for a single rack scenario, don't generate from one of the excluded nodes. Helps
* when we would like to find a replacement node.
*/
private ServerName generateMissingFavoredNodeSingleRack(List<ServerName> favoredNodes,
List<ServerName> excludeNodes) throws IOException {
ServerName newServer = n... | 3.68 |
hadoop_FileIoProvider_dirSync | /**
* Sync the given directory changes to durable device.
* @throws IOException
*/
public void dirSync(@Nullable FsVolumeSpi volume, File dir)
throws IOException {
final long begin = profilingEventHook.beforeFileIo(volume, SYNC, 0);
try {
faultInjectorEventHook.beforeFileIo(volume, SYNC, 0);
IOUtils.... | 3.68 |
hbase_HRegion_dropMemStoreContents | /**
* Be careful, this method will drop all data in the memstore of this region. Currently, this
* method is used to drop memstore to prevent memory leak when replaying recovered.edits while
* opening region.
*/
private MemStoreSize dropMemStoreContents() throws IOException {
MemStoreSizing totalFreedSize = new N... | 3.68 |
hbase_HBaseZKTestingUtility_cleanupTestDir | /** Returns True if we removed the test dirs */
@Override
public boolean cleanupTestDir() {
boolean ret = super.cleanupTestDir();
if (deleteDir(this.clusterTestDir)) {
this.clusterTestDir = null;
return ret;
}
return false;
} | 3.68 |
querydsl_SerializerBase_serializeConstant | /**
* Serialize the constant as parameter to the query. The default implementation writes the
* label name for the constants. Some dialects may replace this by indexed based or
* positional parameterization.
* Dialects may also use this to prefix the parameter with for example ":" or "?".
*
* @param parameterInde... | 3.68 |
flink_RequestedGlobalProperties_parameterizeChannel | /**
* Parametrizes the ship strategy fields of a channel such that the channel produces the desired
* global properties.
*
* @param channel The channel to parametrize.
* @param globalDopChange Flag indicating whether the parallelism changes between sender and
* receiver.
* @param exchangeMode The mode of dat... | 3.68 |
morf_HumanReadableStatementHelper_generateInCriterionString | /**
* Generates a string describing the IN binary operator. This may generate either a list of literal
* values, such as "x in ('1', '2')", or a sub-select statement. For the single-field sub-select form
* a more readable "x in Foo table" form is produced rather than "x in (select x from foo)".
*
* @param criterio... | 3.68 |
zilla_HpackContext_staticIndex8 | // Index in static table for the given name of length 8
private static int staticIndex8(DirectBuffer name)
{
switch (name.getByte(7))
{
case 'e':
if (STATIC_TABLE[42].name.equals(name)) // if-range
{
return 42;
}
break;
case 'h':
if... | 3.68 |
flink_IntermediateResult_getPartitionById | /**
* Returns the partition with the given ID.
*
* @param resultPartitionId ID of the partition to look up
* @throws NullPointerException If partition ID <code>null</code>
* @throws IllegalArgumentException Thrown if unknown partition ID
* @return Intermediate result partition with the given ID
*/
public Interme... | 3.68 |
flink_WorkerResourceSpec_setExtendedResources | /**
* Add the given extended resources. This will discard all the previous added extended
* resources.
*/
public Builder setExtendedResources(Collection<ExternalResource> extendedResources) {
this.extendedResources =
extendedResources.stream()
.collect(
... | 3.68 |
flink_StreamTaskNetworkInputFactory_create | /**
* Factory method for {@link StreamTaskNetworkInput} or {@link RescalingStreamTaskNetworkInput}
* depending on {@link InflightDataRescalingDescriptor}.
*/
public static <T> StreamTaskInput<T> create(
CheckpointedInputGate checkpointedInputGate,
TypeSerializer<T> inputSerializer,
IOManager ... | 3.68 |
hudi_StreamerUtil_getPayloadConfig | /**
* Returns the payload config with given configuration.
*/
public static HoodiePayloadConfig getPayloadConfig(Configuration conf) {
return HoodiePayloadConfig.newBuilder()
.withPayloadClass(conf.getString(FlinkOptions.PAYLOAD_CLASS_NAME))
.withPayloadOrderingField(conf.getString(FlinkOptions.PRECOMBI... | 3.68 |
hbase_MetricsREST_incrementFailedScanRequests | /**
* @param inc How much to add to failedScanCount.
*/
public void incrementFailedScanRequests(final int inc) {
source.incrementFailedScanRequests(inc);
} | 3.68 |
hadoop_CredentialInitializationException_retryable | /**
* This exception is not going to go away if you try calling it again.
* @return false, always.
*/
@Override
public boolean retryable() {
return false;
} | 3.68 |
hbase_ThriftUtilities_addAttributes | /**
* Adds all the attributes into the Operation object
*/
private static void addAttributes(OperationWithAttributes op,
Map<ByteBuffer, ByteBuffer> attributes) {
if (attributes == null || attributes.isEmpty()) {
return;
}
for (Map.Entry<ByteBuffer, ByteBuffer> entry : attributes.entrySet()) {
String ... | 3.68 |
pulsar_AuthenticationSasl_isRoleTokenExpired | // role token exists but expired return true
private boolean isRoleTokenExpired(Map<String, String> responseHeaders) {
if ((saslRoleToken != null)
&& (responseHeaders != null)
// header type match
&& (responseHeaders.get(SASL_HEADER_TYPE) != null && responseHeaders.get(SASL_HEADER_TYPE)
... | 3.68 |
hadoop_Abfss_finalize | /**
* Close the file system; the FileContext API doesn't have an explicit close.
*/
@Override
protected void finalize() throws Throwable {
fsImpl.close();
super.finalize();
} | 3.68 |
framework_DefaultErrorHandler_findAbstractComponent | /**
* Returns the AbstractComponent associated with the given error if such can
* be found.
*
* @param event
* The error to investigate
* @return The {@link AbstractComponent} to error relates to or null if
* could not be determined or if the error does not relate to any
* AbstractCom... | 3.68 |
framework_DefaultItemSorter_compare | /*
* (non-Javadoc)
*
* @see com.vaadin.data.util.ItemSorter#compare(java.lang.Object,
* java.lang.Object)
*/
@Override
public int compare(Object o1, Object o2) {
Item item1 = container.getItem(o1);
Item item2 = container.getItem(o2);
/*
* Items can be null if the container is filtered. Null is co... | 3.68 |
flink_MemorySegment_size | /**
* Gets the size of the memory segment, in bytes.
*
* @return The size of the memory segment.
*/
public int size() {
return size;
} | 3.68 |
dubbo_SerializableClassRegistry_registerClass | /**
* only supposed to be called at startup time
*
* @param clazz object type
* @param serializer object serializer
*/
public static void registerClass(Class<?> clazz, Object serializer) {
if (clazz == null) {
throw new IllegalArgumentException("Class registered to kryo cannot be null!");
}
REG... | 3.68 |
flink_SpillingThread_getMergingIterator | /**
* Returns an iterator that iterates over the merged result from all given channels.
*
* @param channelIDs The channels that are to be merged and returned.
* @param inputSegments The buffers to be used for reading. The list contains for each channel
* one list of input segments. The size of the <code>inputS... | 3.68 |
hadoop_BlockStorageMovementCommand_getBlockPoolId | /**
* Returns block pool ID.
*/
public String getBlockPoolId() {
return blockPoolId;
} | 3.68 |
hbase_PrivateCellUtil_getRowAsInt | /**
* Converts the rowkey bytes of the given cell into an int value
* @return rowkey as int
*/
public static int getRowAsInt(Cell cell) {
if (cell instanceof ByteBufferExtendedCell) {
return ByteBufferUtils.toInt(((ByteBufferExtendedCell) cell).getRowByteBuffer(),
((ByteBufferExtendedCell) cell).getRowPo... | 3.68 |
flink_CheckpointStorageLoader_createDefaultCheckpointStorage | /**
* Creates a default checkpoint storage instance if none was explicitly configured. For
* backwards compatibility, the default storage will be {@link FileSystemCheckpointStorage} if a
* checkpoint directory was configured, {@link
* org.apache.flink.runtime.state.storage.JobManagerCheckpointStorage} otherwise.
*... | 3.68 |
hbase_ColumnRangeFilter_areSerializedFieldsEqual | /**
* Returns true if and only if the fields of the filter that are serialized are equal to the
* corresponding fields in other. Used for testing.
*/
@Override
boolean areSerializedFieldsEqual(Filter o) {
if (o == this) {
return true;
}
if (!(o instanceof ColumnRangeFilter)) {
return false;
}
Colum... | 3.68 |
hbase_RegionScannerImpl_populateResult | /**
* Fetches records with currentRow into results list, until next row, batchLimit (if not -1) is
* reached, or remainingResultSize (if not -1) is reaced
* @param heap KeyValueHeap to fetch data from.It must be positioned on correct row before call.
* @return state of last call to {@link KeyValueHeap#next()}
*/
p... | 3.68 |
hbase_TableRecordReader_restart | /**
* Restart from survivable exceptions by creating a new scanner.
* @param firstRow The first row to start at.
* @throws IOException When restarting fails.
*/
public void restart(byte[] firstRow) throws IOException {
this.recordReaderImpl.restart(firstRow);
} | 3.68 |
framework_TextArea_setWordwrap | /**
* Sets the text area's word-wrap mode on or off.
*
* @param wordwrap
* the boolean value specifying if the text area should be in
* word-wrap mode.
*/
public void setWordwrap(boolean wordwrap) {
getState().wordwrap = wordwrap;
} | 3.68 |
AreaShop_AreaShop_getFeatureManager | /**
* Get the FeatureManager.
* Manages region specific features.
* @return The FeatureManager
*/
public FeatureManager getFeatureManager() {
return featureManager;
} | 3.68 |
framework_BootstrapPageResponse_setDateHeader | /**
* Properly formats a timestamp as a date in a header that will be included
* in the HTTP response. If the header had already been set, the new value
* overwrites the previous one.
*
* @see #setHeader(String, String)
* @see VaadinResponse#setDateHeader(String, long)
*
* @param name
* the name of ... | 3.68 |
hadoop_YarnClient_getApplications | /**
* <p>
* Get a list of ApplicationReports that match the given
* {@link GetApplicationsRequest}.
*</p>
*
* <p>
* If the user does not have <code>VIEW_APP</code> access for an application
* then the corresponding report will be filtered as described in
* {@link #getApplicationReport(ApplicationId)}.
* </p>
... | 3.68 |
pulsar_RangeCache_removeRange | /**
*
* @param first
* @param last
* @param lastInclusive
* @return an pair of ints, containing the number of removed entries and the total size
*/
public Pair<Integer, Long> removeRange(Key first, Key last, boolean lastInclusive) {
Map<Key, Value> subMap = entries.subMap(first, true, last, lastInclusive);
... | 3.68 |
flink_TypeMappingUtils_getProctimeAttribute | /** Returns the proctime attribute of the [[TableSource]] if it is defined. */
private static Optional<String> getProctimeAttribute(TableSource<?> tableSource) {
if (tableSource instanceof DefinedProctimeAttribute) {
return Optional.ofNullable(
((DefinedProctimeAttribute) tableSource).getPro... | 3.68 |
hbase_HBaseTestingUtility_getSupportedCompressionAlgorithms | /**
* Get supported compression algorithms.
* @return supported compression algorithms.
*/
public static Compression.Algorithm[] getSupportedCompressionAlgorithms() {
String[] allAlgos = HFile.getSupportedCompressionAlgorithms();
List<Compression.Algorithm> supportedAlgos = new ArrayList<>();
for (String algoN... | 3.68 |
framework_VTwinColSelect_clearInternalWidths | /** For internal use only. May be removed or replaced in the future. */
public void clearInternalWidths() {
String colWidth = DEFAULT_COLUMN_COUNT + "em";
String containerWidth = 2 * DEFAULT_COLUMN_COUNT + 4 + "em";
// Caption wrapper width == optionsSelect + buttons +
// selectionsSelect
String cap... | 3.68 |
hadoop_FederationCache_buildSubClusterInfoResponse | /**
* Build SubClusterInfo Response.
*
* @param filterInactiveSubClusters whether to filter out inactive sub-clusters.
* @return SubClusterInfo Response.
* @throws YarnException exceptions from yarn servers.
*/
private CacheResponse<SubClusterInfo> buildSubClusterInfoResponse(
final boolean filterInactiveSubC... | 3.68 |
framework_VScrollTable_buildCaptionHtmlSnippet | /**
* Helper function to build html snippet for column or row headers.
*
* @param uidl
* possibly with values caption and icon
* @return html snippet containing possibly an icon + caption text
*/
protected String buildCaptionHtmlSnippet(UIDL uidl) {
String s = uidl.hasAttribute("caption")
... | 3.68 |
framework_AbsoluteLayout_writeDesign | /*
* (non-Javadoc)
*
* @see com.vaadin.ui.AbstractComponent#writeDesign(org.jsoup.nodes.Node,
* com.vaadin.ui.declarative.DesignContext)
*/
@Override
public void writeDesign(Element design, DesignContext designContext) {
super.writeDesign(design, designContext);
AbsoluteLayout def = designContext.getDefaul... | 3.68 |
hadoop_Cluster_getAllJobStatuses | /**
* Get job status for all jobs in the cluster.
* @return job status for all jobs in cluster
* @throws IOException
* @throws InterruptedException
*/
public JobStatus[] getAllJobStatuses() throws IOException, InterruptedException {
return client.getAllJobs();
} | 3.68 |
querydsl_AbstractFetchableMongodbQuery_fetch | /**
* Fetch with the specific fields
*
* @param paths fields to return
* @return results
*/
public List<K> fetch(Path<?>... paths) {
getQueryMixin().setProjection(paths);
return fetch();
} | 3.68 |
pulsar_ProtocolHandlerUtils_getProtocolHandlerDefinition | /**
* Retrieve the protocol handler definition from the provided handler nar package.
*
* @param narPath the path to the protocol handler NAR package
* @return the protocol handler definition
* @throws IOException when fail to load the protocol handler or get the definition
*/
public static ProtocolHandlerDefinit... | 3.68 |
dubbo_DynamicConfiguration_getProperties | /**
* This method are mostly used to get a compound config file, such as a complete dubbo.properties file.
*
* @revision 2.7.4
*/
default String getProperties(String key, String group, long timeout) throws IllegalStateException {
return getConfig(key, group, timeout);
} | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.