name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
hbase_NewVersionBehaviorTracker_add_rdh | // DeleteTracker
@Override
public void add(Cell cell) {
prepare(cell);
byte v9 = cell.getTypeByte();switch (Type.codeToType(v9)) {
// By the order of seen. We put null cq at first.
case DeleteFamily :
// Delete all versions of all columns of the specified family
delFa... | 3.26 |
hbase_GssSaslServerAuthenticationProvider_handle_rdh | /**
* {@inheritDoc }
*/
@Override
public void handle(Callback[] callbacks) throws UnsupportedCallbackException {
AuthorizeCallback ac = null;
for (Callback callback : callbacks) {
if
(callback instanceof AuthorizeCallback) {
ac = ((AuthorizeCallback) (callback));
} else {
... | 3.26 |
hbase_FixedIntervalRateLimiter_setNextRefillTime_rdh | // This method is for strictly testing purpose only
@Override
public void setNextRefillTime(long nextRefillTime) {
this.nextRefillTime = nextRefillTime;
} | 3.26 |
hbase_UserProvider_setUserProviderForTesting_rdh | /**
* Set the {@link UserProvider} in the given configuration that should be instantiated
*
* @param conf
* to update
* @param provider
* class of the provider to set
*/
public static void setUserProviderForTesting(Configuration conf, Class<? extends UserProvider> provider) {
conf.set(USER_PROVIDER_CON... | 3.26 |
hbase_UserProvider_getCurrentUserName_rdh | /**
* Returns the userName for the current logged-in user.
*
* @throws IOException
* if the underlying user cannot be obtained
*/
public String getCurrentUserName() throws IOException {
User user = getCurrent();
return user == null ? null : user.getName();
} | 3.26 |
hbase_UserProvider_m0_rdh | // Provide the reload function that uses the executor thread.
@Override
public ListenableFuture<String[]> m0(final String k, String[] oldValue) throws Exception {
return executor.submit(new Callable<String[]>() {
@Override
public String[] call() throws Exception {
return getGroupStri... | 3.26 |
hbase_UserProvider_getCurrent_rdh | /**
* Return the current user within the current execution context
*
* @throws IOException
* if the user cannot be loaded
*/
public User getCurrent() throws IOException {
return User.getCurrent();
} | 3.26 |
hbase_UserProvider_create_rdh | /**
* Wraps an underlying {@code UserGroupInformation} instance.
*
* @param ugi
* The base Hadoop user
*/
public User create(UserGroupInformation ugi) {
if (ugi == null) {
return null;
}
return new User.SecureHadoopUser(ugi, groupCache);
} | 3.26 |
hbase_UserProvider_instantiate_rdh | /**
* Instantiate the {@link UserProvider} specified in the configuration and set the passed
* configuration via {@link UserProvider#setConf(Configuration)}
*
* @param conf
* to read and set on the created {@link UserProvider}
* @return a {@link UserProvider} ready for use.
*/
public static UserProvider instan... | 3.26 |
hbase_UserProvider_shouldLoginFromKeytab_rdh | /**
* In secure environment, if a user specified his keytab and principal, a hbase client will try to
* login with them. Otherwise, hbase client will try to obtain ticket(through kinit) from system.
*/
public boolean shouldLoginFromKeytab() {
return User.shouldLoginFromKeytab(this.getConf());
} | 3.26 |
hbase_UserProvider_isHadoopSecurityEnabled_rdh | /**
* Return whether or not Kerberos authentication is configured for Hadoop. For non-secure Hadoop,
* this always returns <code>false</code>. For secure Hadoop, it will return the value from
* {@code UserGroupInformation.isSecurityEnabled()}.
*/
public boolean isHadoopSecurityEnabled() {
return User.isSecurity... | 3.26 |
hbase_UserProvider_login_rdh | /**
* Login with given keytab and principal. This can be used for both SPN(Service Principal Name)
* and UPN(User Principal Name) which format should be clientname@REALM.
*
* @param fileConfKey
* config name for client keytab
* @param principalConfKey
* config name for client principal
* @throws IOException... | 3.26 |
hbase_UserProvider_isHBaseSecurityEnabled_rdh | /**
* Returns <tt>true</tt> if security is enabled, <tt>false</tt> otherwise
*/
public boolean isHBaseSecurityEnabled() {
return User.isHBaseSecurityEnabled(this.getConf());
} | 3.26 |
hbase_PageFilter_parseFrom_rdh | /**
* Parse a serialized representation of {@link PageFilter}
*
* @param pbBytes
* A pb serialized {@link PageFilter} instance
* @return An instance of {@link PageFilter} made from <code>bytes</code>
* @throws DeserializationException
* if an error occurred
* @see #toByteArray
*/
public static PageFilter p... | 3.26 |
hbase_PageFilter_areSerializedFieldsEqual_rdh | /**
* Returns true if and only if the fields of the filter that are serialized are equal to the
* corresponding fields in other. Used for testing.
*/
@Override
boolean areSerializedFieldsEqual(Filter o) {
... | 3.26 |
hbase_MetricsRegionServerWrapperImpl_initMobFileCache_rdh | /**
* Initializes the mob file cache.
*/
private void initMobFileCache() {
this.mobFileCache = this.regionServer.getMobFileCache().orElse(null);
} | 3.26 |
hbase_IncrementingEnvironmentEdge_incrementTime_rdh | /**
* Increment the time by the given amount
*/
public synchronized long incrementTime(long amount) {
timeIncrement += amount;
return timeIncrement;
} | 3.26 |
hbase_IncrementingEnvironmentEdge_currentTime_rdh | /**
* {@inheritDoc }
* <p>
* This method increments a known value for the current time each time this method is called. The
* first value is 1.
* </p>
*/
@Override
public synchronized long currentTime() {
return timeIncrement++;
} | 3.26 |
hbase_StripeMultiFileWriter_sanityCheckRight_rdh | /**
* Subclasses can call this method to make sure the last KV is within multi-writer range.
*
* @param right
* The right boundary of the writer.
*/
protected void sanityCheckRight(byte[] right, Cell cell) throws IOException {
if ((!Arrays.equals(StripeStoreFileManager.OPEN_KEY, right)) && (comparator.compar... | 3.26 |
hbase_StripeMultiFileWriter_sanityCheckLeft_rdh | /**
* Subclasses can call this method to make sure the first KV is within multi-writer range.
*
* @param left
* The left boundary of the writer.
* @param cell
* The cell whose row has to be checked.
*/
protected void sanityCheckLeft(byte[] left, Cell cell) throws IOException {
if ((!Arrays.equals(StripeS... | 3.26 |
hbase_ClusterId_parseFrom_rdh | /**
* Parse the serialized representation of the {@link ClusterId}
*
* @param bytes
* A pb serialized {@link ClusterId} instance with pb magic prefix
* @return An instance of {@link ClusterId} made from <code>bytes</code>
* @see #toByteArray()
*/
public static ClusterId parseFrom(final byte[] bytes) throws Des... | 3.26 |
hbase_ClusterId_toByteArray_rdh | /**
* Returns The clusterid serialized using pb w/ pb magic prefix
*/
public byte[] toByteArray() {
return ProtobufUtil.prependPBMagic(convert().toByteArray());
} | 3.26 |
hbase_ClusterId_convert_rdh | /**
* Returns A {@link ClusterId} made from the passed in <code>cid</code>
*/
public static ClusterId convert(final ClusterIdProtos.ClusterId cid)
{
return new ClusterId(cid.getClusterId());
} | 3.26 |
hbase_ClusterId_toString_rdh | /**
*
* @see java.lang.Object#toString()
*/
@Override
public String toString() {
return this.id;
} | 3.26 |
hbase_SnapshotInfo_getSnapshotDescription_rdh | /**
* Returns the snapshot descriptor
*/
public SnapshotDescription getSnapshotDescription() {return ProtobufUtil.createSnapshotDesc(this.snapshot);
} | 3.26 |
hbase_SnapshotInfo_addStoreFile_rdh | /**
* Add the specified store file to the stats
*
* @param region
* region encoded Name
* @param family
* family name
* @param storeFile
* store file name
* @param filesMap
* store files map for all snapshots, it may be null
* @return the store file information
*/
FileInfo addStoreFile(final RegionI... | 3.26 |
hbase_SnapshotInfo_isMissing_rdh | /**
* Returns true if the file is missing
*/
public boolean isMissing() {
return this.size < 0;
} | 3.26 |
hbase_SnapshotInfo_getMissingStoreFilesCount_rdh | /**
* Returns the number of missing store files
*/
public int getMissingStoreFilesCount() {
return hfilesMissing.get();
} | 3.26 |
hbase_SnapshotInfo_printSchema_rdh | /**
* Dump the {@link org.apache.hadoop.hbase.client.TableDescriptor}
*/
private void printSchema() {
System.out.println("Table Descriptor");
System.out.println("----------------------------------------");
System.out.println(snapshotManifest.getTableDescriptor()... | 3.26 |
hbase_SnapshotInfo_getSnapshotsFilesMap_rdh | /**
* Returns the map of store files based on path for all snapshots
*
* @param conf
* the {@link Configuration} to use
* @param uniqueHFilesArchiveSize
* pass out the size for store files in archive
* @param uniqueHFilesSize
* pass out the size for store files shared
* @param uniqueHFilesMobSize
* pa... | 3.26 |
hbase_SnapshotInfo_getSharedStoreFilesSize_rdh | /**
* Returns the total size of the store files shared
*/
public long getSharedStoreFilesSize() {
return hfilesSize.get();
} | 3.26 |
hbase_SnapshotInfo_getMobStoreFilesSize_rdh | /**
* Returns the total size of the store files in the mob store
*/
public long getMobStoreFilesSize() {
return hfilesMobSize.get();
}
/**
*
* @return the total size of the store files in the archive which is not shared with other
snapshots and tables This is only calculated when
{@link #getSnapshotStats(Confi... | 3.26 |
hbase_SnapshotInfo_inArchive_rdh | /**
* Returns true if the file is in the archive
*/
public boolean inArchive() {
return this.inArchive;
} | 3.26 |
hbase_SnapshotInfo_getSharedStoreFilePercentage_rdh | /**
* Returns the percentage of the shared store files
*/
public float getSharedStoreFilePercentage()
{
return (((float) (hfilesSize.get())) / getStoreFilesSize()) *
100;
} | 3.26 |
hbase_SnapshotInfo_getMobStoreFilePercentage_rdh | /**
* Returns the percentage of the mob store files
*/
public float getMobStoreFilePercentage() {
return (((float) (hfilesMobSize.get())) / getStoreFilesSize()) * 100;} | 3.26 |
hbase_SnapshotInfo_getMobStoreFilesCount_rdh | /**
* Returns the number of available store files in the mob dir
*/
public int getMobStoreFilesCount() {return hfilesMobCount.get();
} | 3.26 |
hbase_SnapshotInfo_loadSnapshotInfo_rdh | /**
* Load snapshot info and table descriptor for the specified snapshot
*
* @param snapshotName
* name of the snapshot to load
* @return false if snapshot is not found
*/
private boolean loadSnapshotInfo(final Stri... | 3.26 |
hbase_SnapshotInfo_isArchivedFileStillReferenced_rdh | /**
* Check if for a give file in archive, if there are other snapshots/tables still reference it.
*
* @param filePath
* file path in archive
* @param snapshotFilesMap
* a map for store files in snapshots about how many snapshots refer to
* it.
* @return true or false
*/
private boolean isArchivedFileSti... | 3.26 |
hbase_SnapshotInfo_getCorruptedStoreFilesCount_rdh | /**
* Returns the number of corrupted store files
*/
public int getCorruptedStoreFilesCount() {
return hfilesCorrupted.get();
} | 3.26 |
hbase_SnapshotInfo_getSnapshotStats_rdh | /**
* Returns the snapshot stats
*
* @param conf
* the {@link Configuration} to use
* @param snapshotDesc
* HBaseProtos.SnapshotDescription to get stats from
* @param filesMap
* {@link Map} store files map for all snapshots, it may be null
* @return the snapshot stats
*/public static SnapshotStats getSn... | 3.26 |
hbase_SnapshotInfo_addLogFile_rdh | /**
* Add the specified log file to the stats
*
* @param server
* server name
* @param logfile
* log file name
* @return the log information
*/
FileInfo addLogFile(final String server, final String
logfile) throws IOException {WALLink logLink = new WALLink(conf, server, logfile);
long v8 = -1;try {
... | 3.26 |
hbase_SnapshotInfo_printInfo_rdh | /**
* Dump the {@link SnapshotDescription}
*/
private void printInfo() {
SnapshotProtos.SnapshotDescription snapshotDesc = snapshotManifest.getSnapshotDescription();
SimpleDateFormat df = new SimpleDateFormat("yyyy-MM-dd'T'HH:mm:ss");
System.out.println("Snapshot Info");
System.out.println("---------... | 3.26 |
hbase_SnapshotInfo_isSnapshotCorrupted_rdh | /**
* Returns true if the snapshot is corrupted
*/
public boolean isSnapshotCorrupted() {
return ((hfilesMissing.get() > 0) || (logsMissing.get() > 0)) || (hfilesCorrupted.get()
> 0);
} | 3.26 |
hbase_SnapshotInfo_getSnapshotFilesMap_rdh | /**
* Gets the store files map for snapshot
*
* @param conf
* the {@link Configuration} to use
* @param snapshot
* {@link SnapshotDescription} to get stats from
* @param exec
* the {@link ExecutorService} to use
* @param filesMap
* {@link Map} the map to put the mapping entries
* @param uniqueHFilesA... | 3.26 |
hbase_SnapshotInfo_isCorrupted_rdh | /**
* Returns true if the file is corrupted
*/
public boolean isCorrupted() {
return this.corrupted;
} | 3.26 |
hbase_SnapshotInfo_getStoreFilesSize_rdh | /**
* Returns the total size of the store files referenced by the snapshot
*/
public long getStoreFilesSize() {
return (hfilesSize.get() + hfilesArchiveSize.get()) + hfilesMobSize.get();
} | 3.26 |
hbase_SnapshotInfo_getLogsCount_rdh | /**
* Returns the number of available log files
*/
public int getLogsCount() {
return logsCount.get();
} | 3.26 |
hbase_SnapshotInfo_getMissingLogsCount_rdh | /**
* Returns the number of missing log files
*/
public int getMissingLogsCount() {
return logsMissing.get();
} | 3.26 |
hbase_SnapshotInfo_printFiles_rdh | /**
* Collect the hfiles and logs statistics of the snapshot and dump the file list if requested and
* the collected information.
*/
private void printFiles(final boolean showFiles, final boolean showStats) throws IOException {
if (showFiles) {
System.out.println("Snapshot ... | 3.26 |
hbase_SnapshotInfo_getLogsSize_rdh | /**
* Returns the total log size
*/
public long getLogsSize() {
return logSize.get();
} | 3.26 |
hbase_SnapshotInfo_getArchivedStoreFilesCount_rdh | /**
* Returns the number of available store files in the archive
*/
public int getArchivedStoreFilesCount() {
return hfilesArchiveCount.get();
} | 3.26 |
hbase_SnapshotInfo_getSize_rdh | /**
* Returns the file size
*/
public long getSize() {
return this.size;
} | 3.26 |
hbase_SnapshotInfo_getArchivedStoreFileSize_rdh | /**
* Returns the total size of the store files in the archive
*/
public long getArchivedStoreFileSize() {
return hfilesArchiveSize.get();
} | 3.26 |
hbase_DefaultEnvironmentEdge_currentTime_rdh | /**
* {@inheritDoc }
* <p>
* This implementation returns {@link System#currentTimeMillis()}
* </p>
*/
@Override
public long currentTime() {
return System.currentTimeMillis();
} | 3.26 |
hbase_ByteBufferListOutputStream_releaseResources_rdh | /**
* Release the resources it uses (The ByteBuffers) which are obtained from pool. Call this only
* when all the data is fully used. And it must be called at the end of usage else we will leak
* ByteBuffers from pool.
*/
public void releaseResources() {
try
{
close();
} catch (IOException e) {
... | 3.26 |
hbase_ByteBufferListOutputStream_getByteBuffers_rdh | /**
* We can be assured that the buffers returned by this method are all flipped
*
* @return list of bytebuffers
*/public List<ByteBuffer> getByteBuffers() {
if (!this.lastBufFlipped) {
this.lastBufFlipped = true;
// All the other BBs are already flipped while moving to the new BB.
curBu... | 3.26 |
hbase_ExpiredMobFileCleaner_cleanExpiredMobFiles_rdh | /**
* Cleans the MOB files when they're expired and their min versions are 0. If the latest timestamp
* of Cells in a MOB file is older than the TTL in the column family, it's regarded as expired.
* This cleaner deletes them. At a time T0, the cells in a mob file M0 are expired. If a user
* starts a scan before T0,... | 3.26 |
hbase_ScanQueryMatcher_preCheck_rdh | /**
* Check before the delete logic.
*
* @return null means continue.
*/
protected final MatchCode preCheck(Cell cell) {
if (currentRow == null) {
// Since the curCell is null it means we are already sure that we have moved over to the next
// row
return MatchCode.DONE;
}
// if r... | 3.26 |
hbase_ScanQueryMatcher_clearCurrentRow_rdh | /**
* Make {@link #currentRow()} return null.
*/
public void clearCurrentRow() {
currentRow = null;
} | 3.26 |
hbase_ScanQueryMatcher_m0_rdh | /**
* Returns true if the cell is expired
*/
private static boolean m0(final Cell cell, final long oldestTimestamp, final long now) {
// Look for a TTL tag first. Use it instead of the family setting if
// found. If a cell has multiple TTLs, resolve the conflict by using the
// first tag e... | 3.26 |
hbase_ScanQueryMatcher_checkColumn_rdh | // Used only for testing purposes
static MatchCode checkColumn(ColumnTracker columnTracker, byte[] bytes, int offset, int length, long ttl, byte type, boolean ignoreCount) throws IOException {KeyValue kv = KeyValueUtil.createFirstOnRow(HConstants.EMPTY_BYTE_ARRAY, 0, 0, HConstants.EMPTY_BYTE_ARRAY, 0, 0, bytes, off... | 3.26 |
hbase_ScanQueryMatcher_currentRow_rdh | /**
* Returns a cell represent the current row
*/
public Cell currentRow() {
return currentRow;
} | 3.26 |
hbase_ScanQueryMatcher_setToNewRow_rdh | /**
* Set the row when there is change in row
*/
public void setToNewRow(Cell currentRow) {
this.currentRow = currentRow;
columns.reset();reset();
} | 3.26 |
hbase_ScanQueryMatcher_m1_rdh | /**
*
* @param nextIndexed
* the key of the next entry in the block index (if any)
* @param currentCell
* The Cell we're using to calculate the seek key
* @return result of the compare between the indexed key and the key portion of the passed cell
*/public int m1(Cell nextIndexed, Cell currentCell) {
ColumnC... | 3.26 |
hbase_ScanQueryMatcher_getStartKey_rdh | /**
* Returns the start key
*/
public Cell getStartKey() {
return startKey;
} | 3.26 |
hbase_ByteBufferOutputStream_getByteBuffer_rdh | /**
* This flips the underlying BB so be sure to use it _last_!
*/
public ByteBuffer getByteBuffer() {
curBuf.flip();
return curBuf;
} | 3.26 |
hbase_ByteBufferOutputStream_writeInt_rdh | /**
* Writes an <code>int</code> to the underlying output stream as four bytes, high byte first.
*
* @param i
* the <code>int</code> to write
* @throws IOException
* if an I/O error occurs.
*/
@Override
public void writeInt(int i) throws IOException {
checkSizeAndGrow(Bytes.SIZEOF_INT);
ByteBufferUti... | 3.26 |
hbase_ByteBufferOutputStream_writeTo_rdh | /**
* Writes the complete contents of this byte buffer output stream to the specified output stream
* argument.
*
* @param out
* the output stream to which to write the data.
* @exception IOException
* if an I/O error occurs.
*/
public void writeTo(OutputStream out) throws IOException {
WritableByteChan... | 3.26 |
hbase_ByteBufferOutputStream_write_rdh | // OutputStream
@Override
public void write(int b) throws IOException {
checkSizeAndGrow(Bytes.SIZEOF_BYTE);
curBuf.put(((byte)
(b)));
} | 3.26 |
hbase_User_getToken_rdh | /**
* Returns the Token of the specified kind associated with this user, or null if the Token is not
* present.
*
* @param kind
* the kind of token
* @param service
* service on which the token is supposed to be used
* @return the token of the specified kind.
*/public Token<?> getToken(String kind, String ... | 3.26 |
hbase_User_isHBaseSecurityEnabled_rdh | /**
* Returns whether or not secure authentication is enabled for HBase. Note that HBase security
* requires HDFS security to provide any guarantees, so it is recommended that secure HBase should
* run on secure HDFS.
*/
public static boolean isHBaseSecurityEnabled(Configuration conf) {
return "kerberos".equals... | 3.26 |
hbase_User_addToken_rdh | /**
* Adds the given Token to the user's credentials.
*
* @param token
* the token to add
*/
public void addToken(Token<? extends TokenIdentifier> token) {
ugi.addToken(token);
} | 3.26 |
hbase_User_getTokens_rdh | /**
* Returns all the tokens stored in the user's credentials.
*/
public Collection<Token<? extends TokenIdentifier>> getTokens() {
return ugi.getTokens();
} | 3.26 |
hbase_User_createUserForTesting_rdh | /**
* Create a user for testing.
*
* @see User#createUserForTesting(org.apache.hadoop.conf.Configuration, String, String[])
*/
public static User createUserForTesting(Configuration conf, String name, String[] groups) {
synchronized(UserProvider.class) {
if
(!(UserProvider.groups instanceof Testi... | 3.26 |
hbase_User_create_rdh | /**
* Wraps an underlying {@code UserGroupInformation} instance.
*
* @param ugi
* The base Hadoop user
*/
public static User create(UserGroupInformation ugi) {
if (ugi == null) {
return null;
}
return new SecureHadoopUser(ugi);
} | 3.26 |
hbase_User_getName_rdh | /**
* Returns the full user name. For Kerberos principals this will include the host and realm
* portions of the principal name.
*
* @return User full name.
*/
public String getName() {
return ugi.getUserName();
} | 3.26 |
hbase_User_isSecurityEnabled_rdh | /**
* Returns the result of {@code UserGroupInformation.isSecurityEnabled()}.
*/
public static boolean isSecurityEnabled() {
return UserGroupInformation.isSecurityEnabled();
} | 3.26 |
hbase_User_runAsLoginUser_rdh | /**
* Executes the given action as the login user
*/
@SuppressWarnings({ "rawtypes", "unchecked" })
public static <T> T runAsLoginUser(PrivilegedExceptionAction<T> action) throws IOException {
try {Class c = Class.forName("org.apache.hadoop.security.SecurityUtil");
Class[] types = new Class[]{ Privil... | 3.26 |
hbase_User_getCurrent_rdh | /**
* Returns the {@code User} instance within current execution context.
*/
public static User getCurrent() throws IOException {
User user = new SecureHadoopUser();
if (user.getUGI() == null) {
return null;
}
return user;
} | 3.26 |
hbase_User_isLoginFromKeytab_rdh | /**
* Returns true if user credentials are obtained from keytab.
*/
public boolean isLoginFromKeytab() {return ugi.isFromKeytab();
} | 3.26 |
hbase_User_shouldLoginFromKeytab_rdh | /**
* In secure environment, if a user specified his keytab and principal, a hbase client will try to
* login with them. Otherwise, hbase client will try to obtain ticket(through kinit) from system.
*
* @param conf
* configuration file
* @return true if keytab and principal are configured
*/
public static bool... | 3.26 |
hbase_User_login_rdh | /**
* Login through configured keytab and pricipal.
*
* @param keytabLocation
* location of keytab
* @param principalName
* principal in keytab
* @throws IOException
* exception from UserGroupInformation.loginUserFromKeytab
*/
public static void login(String keytabLocation, String principalName) throws I... | 3.26 |
hbase_User_m0_rdh | /**
* Returns the list of groups of which this user is a member. On secure Hadoop this returns the
* group information for the user as resolved on the server. For 0.20 based Hadoop, the group
* names are passed from the client.
*/
public String[] m0() {
return ugi.getGroupNames();
} | 3.26 |
hbase_MetricsStochasticBalancerSourceImpl_updateStochasticCost_rdh | /**
* Reports stochastic load balancer costs to JMX
*/
@Override
public void updateStochasticCost(String tableName, String costFunctionName, String functionDesc, Double cost) {
if (((tableName == null) || (costFunctionName == null)) || (cost == null)) {
return;
}
if (functionDesc != null) {
... | 3.26 |
hbase_MetricsStochasticBalancerSourceImpl_calcMruCap_rdh | /**
* Calculates the mru cache capacity from the metrics size
*/
private static int calcMruCap(int metricsSize) {
return ((int) (Math.ceil(metricsSize / MRU_LOAD_FACTOR))) + 1;
} | 3.26 |
hbase_HBaseServiceHandler_getTable_rdh | /**
* Creates and returns a Table instance from a given table name. name of table
*
* @return Table object
* @throws IOException
* if getting the table fails
*/
protected Table getTable(final byte[] tableName) throws IOException {
String table = Bytes.toString(tableName);
return connectionCache.getTable... | 3.26 |
hbase_HBaseServiceHandler_getAdmin_rdh | /**
* Obtain HBaseAdmin. Creates the instance if it is not already created.
*/
protected Admin getAdmin() throws IOException {
return connectionCache.getAdmin();
} | 3.26 |
hbase_RawCell_cloneTags_rdh | /**
* Allows cloning the tags in the cell to a new byte[]
*
* @return the byte[] having the tags
*/
default byte[] cloneTags() {
return PrivateCellUtil.cloneTags(this);
} | 3.26 |
hbase_RawCell_createCell_rdh | /**
* Returns A new cell which is having the extra tags also added to it.
*/
public static Cell createCell(Cell cell, List<Tag>
tags) {
return PrivateCellUtil.createCell(cell, tags);
} | 3.26 |
hbase_RawCell_checkForTagsLength_rdh | /**
* Check the length of tags. If it is invalid, throw IllegalArgumentException
*
* @param tagsLength
* the given length of tags
* @throws IllegalArgumentException
* if tagslength is invalid
*/
public static void checkForTagsLength(int tagsLength) {
if (tagsLength > MAX_TAGS_LENGTH) {
throw
... | 3.26 |
hbase_RawCell_getTag_rdh | /**
* Returns the specific tag of the given type
*
* @param type
* the type of the tag
* @return the specific tag if available or null
*/ default Optional<Tag> getTag(byte type) {
return PrivateCellUtil.getTag(this, type);
} | 3.26 |
hbase_RawCell_getTags_rdh | /**
* Creates a list of tags in the current cell
*
* @return a list of tags
*/
default Iterator<Tag> getTags() {
return PrivateCellUtil.tagsIterator(this);
} | 3.26 |
hbase_RSGroupAdminServiceImpl_fillTables_rdh | // for backward compatible
private RSGroupInfo fillTables(RSGroupInfo rsGroupInfo) throws IOException {return RSGroupUtil.fillTables(rsGroupInfo, master.getTableDescriptors().getAll().values());
} | 3.26 |
hbase_StripeCompactionPolicy_setMajorRange_rdh | /**
* Sets compaction "major range". Major range is the key range for which all the files are
* included, so they can be treated like major-compacted files.
*
* @param startRow
* Left boundary, inclusive.
* @param endRow
* Right boundary, exclusive.
*/
public void setMajorRange(byte[] startRow, byte[] endRo... | 3.26 |
hbase_StripeCompactionPolicy_selectSimpleCompaction_rdh | /**
* Selects the compaction of a single stripe using default policy.
*
* @param sfs
* Files.
* @param allFilesOnly
* Whether a compaction of all-or-none files is needed.
* @return The resulting selection.
*/
private List<HStoreFile> selectSimpleCompaction(List<HStoreFile> sfs, boolean allFilesOnly, boolean... | 3.26 |
hbase_StripeCompactionPolicy_setMajorRangeFull_rdh | /**
* Set major range of the compaction to the entire compaction range. See
* {@link #setMajorRange(byte[], byte[])}.
*/
public void setMajorRangeFull() {
setMajorRange(this.startRow, this.endRow);
} | 3.26 |
hbase_StripeCompactionPolicy_needsSingleStripeCompaction_rdh | /**
*
* @param si
* StoreFileManager.
* @return Whether any stripe potentially needs compaction.
*/
protected boolean needsSingleStripeCompaction(StripeInformationProvider si) {int minFiles = this.config.getStripeCompactMinFiles();
for (List<HStoreFile> stripe : si.getStripes()) {if (stripe.size() >= minFile... | 3.26 |
hbase_TableSnapshotInputFormat_setInput_rdh | /**
* Configures the job to use TableSnapshotInputFormat to read from a snapshot.
*
* @param job
* the job to configure
* @param snapshotName
* the name of the snapshot to read from
* @param restoreDir
* a temporary directory to restore the snapshot into. Current user
* should have write permissions to... | 3.26 |
hbase_StartMiniClusterOption_builder_rdh | /**
* Returns a new builder.
*/
public static Builder
builder() {
return new Builder();
} | 3.26 |
hbase_LzoCodec_getBufferSize_rdh | // Package private
static int getBufferSize(Configuration conf) {
return conf.getInt(LZO_BUFFER_SIZE_KEY, conf.getInt(CommonConfigurationKeys.IO_COMPRESSION_CODEC_LZO_BUFFERSIZE_KEY, CommonConfigurationKeys.IO_COMPRESSION_CODEC_LZO_BUFFERSIZE_DEFAULT));
} | 3.26 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.