name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
hbase_NewVersionBehaviorTracker_add_rdh
// DeleteTracker @Override public void add(Cell cell) { prepare(cell); byte v9 = cell.getTypeByte();switch (Type.codeToType(v9)) { // By the order of seen. We put null cq at first. case DeleteFamily : // Delete all versions of all columns of the specified family delFa...
3.26
hbase_GssSaslServerAuthenticationProvider_handle_rdh
/** * {@inheritDoc } */ @Override public void handle(Callback[] callbacks) throws UnsupportedCallbackException { AuthorizeCallback ac = null; for (Callback callback : callbacks) { if (callback instanceof AuthorizeCallback) { ac = ((AuthorizeCallback) (callback)); } else { ...
3.26
hbase_FixedIntervalRateLimiter_setNextRefillTime_rdh
// This method is for strictly testing purpose only @Override public void setNextRefillTime(long nextRefillTime) { this.nextRefillTime = nextRefillTime; }
3.26
hbase_UserProvider_setUserProviderForTesting_rdh
/** * Set the {@link UserProvider} in the given configuration that should be instantiated * * @param conf * to update * @param provider * class of the provider to set */ public static void setUserProviderForTesting(Configuration conf, Class<? extends UserProvider> provider) { conf.set(USER_PROVIDER_CON...
3.26
hbase_UserProvider_getCurrentUserName_rdh
/** * Returns the userName for the current logged-in user. * * @throws IOException * if the underlying user cannot be obtained */ public String getCurrentUserName() throws IOException { User user = getCurrent(); return user == null ? null : user.getName(); }
3.26
hbase_UserProvider_m0_rdh
// Provide the reload function that uses the executor thread. @Override public ListenableFuture<String[]> m0(final String k, String[] oldValue) throws Exception { return executor.submit(new Callable<String[]>() { @Override public String[] call() throws Exception { return getGroupStri...
3.26
hbase_UserProvider_getCurrent_rdh
/** * Return the current user within the current execution context * * @throws IOException * if the user cannot be loaded */ public User getCurrent() throws IOException { return User.getCurrent(); }
3.26
hbase_UserProvider_create_rdh
/** * Wraps an underlying {@code UserGroupInformation} instance. * * @param ugi * The base Hadoop user */ public User create(UserGroupInformation ugi) { if (ugi == null) { return null; } return new User.SecureHadoopUser(ugi, groupCache); }
3.26
hbase_UserProvider_instantiate_rdh
/** * Instantiate the {@link UserProvider} specified in the configuration and set the passed * configuration via {@link UserProvider#setConf(Configuration)} * * @param conf * to read and set on the created {@link UserProvider} * @return a {@link UserProvider} ready for use. */ public static UserProvider instan...
3.26
hbase_UserProvider_shouldLoginFromKeytab_rdh
/** * In secure environment, if a user specified his keytab and principal, a hbase client will try to * login with them. Otherwise, hbase client will try to obtain ticket(through kinit) from system. */ public boolean shouldLoginFromKeytab() { return User.shouldLoginFromKeytab(this.getConf()); }
3.26
hbase_UserProvider_isHadoopSecurityEnabled_rdh
/** * Return whether or not Kerberos authentication is configured for Hadoop. For non-secure Hadoop, * this always returns <code>false</code>. For secure Hadoop, it will return the value from * {@code UserGroupInformation.isSecurityEnabled()}. */ public boolean isHadoopSecurityEnabled() { return User.isSecurity...
3.26
hbase_UserProvider_login_rdh
/** * Login with given keytab and principal. This can be used for both SPN(Service Principal Name) * and UPN(User Principal Name) which format should be clientname@REALM. * * @param fileConfKey * config name for client keytab * @param principalConfKey * config name for client principal * @throws IOException...
3.26
hbase_UserProvider_isHBaseSecurityEnabled_rdh
/** * Returns <tt>true</tt> if security is enabled, <tt>false</tt> otherwise */ public boolean isHBaseSecurityEnabled() { return User.isHBaseSecurityEnabled(this.getConf()); }
3.26
hbase_PageFilter_parseFrom_rdh
/** * Parse a serialized representation of {@link PageFilter} * * @param pbBytes * A pb serialized {@link PageFilter} instance * @return An instance of {@link PageFilter} made from <code>bytes</code> * @throws DeserializationException * if an error occurred * @see #toByteArray */ public static PageFilter p...
3.26
hbase_PageFilter_areSerializedFieldsEqual_rdh
/** * Returns true if and only if the fields of the filter that are serialized are equal to the * corresponding fields in other. Used for testing. */ @Override boolean areSerializedFieldsEqual(Filter o) { ...
3.26
hbase_MetricsRegionServerWrapperImpl_initMobFileCache_rdh
/** * Initializes the mob file cache. */ private void initMobFileCache() { this.mobFileCache = this.regionServer.getMobFileCache().orElse(null); }
3.26
hbase_IncrementingEnvironmentEdge_incrementTime_rdh
/** * Increment the time by the given amount */ public synchronized long incrementTime(long amount) { timeIncrement += amount; return timeIncrement; }
3.26
hbase_IncrementingEnvironmentEdge_currentTime_rdh
/** * {@inheritDoc } * <p> * This method increments a known value for the current time each time this method is called. The * first value is 1. * </p> */ @Override public synchronized long currentTime() { return timeIncrement++; }
3.26
hbase_StripeMultiFileWriter_sanityCheckRight_rdh
/** * Subclasses can call this method to make sure the last KV is within multi-writer range. * * @param right * The right boundary of the writer. */ protected void sanityCheckRight(byte[] right, Cell cell) throws IOException { if ((!Arrays.equals(StripeStoreFileManager.OPEN_KEY, right)) && (comparator.compar...
3.26
hbase_StripeMultiFileWriter_sanityCheckLeft_rdh
/** * Subclasses can call this method to make sure the first KV is within multi-writer range. * * @param left * The left boundary of the writer. * @param cell * The cell whose row has to be checked. */ protected void sanityCheckLeft(byte[] left, Cell cell) throws IOException { if ((!Arrays.equals(StripeS...
3.26
hbase_ClusterId_parseFrom_rdh
/** * Parse the serialized representation of the {@link ClusterId} * * @param bytes * A pb serialized {@link ClusterId} instance with pb magic prefix * @return An instance of {@link ClusterId} made from <code>bytes</code> * @see #toByteArray() */ public static ClusterId parseFrom(final byte[] bytes) throws Des...
3.26
hbase_ClusterId_toByteArray_rdh
/** * Returns The clusterid serialized using pb w/ pb magic prefix */ public byte[] toByteArray() { return ProtobufUtil.prependPBMagic(convert().toByteArray()); }
3.26
hbase_ClusterId_convert_rdh
/** * Returns A {@link ClusterId} made from the passed in <code>cid</code> */ public static ClusterId convert(final ClusterIdProtos.ClusterId cid) { return new ClusterId(cid.getClusterId()); }
3.26
hbase_ClusterId_toString_rdh
/** * * @see java.lang.Object#toString() */ @Override public String toString() { return this.id; }
3.26
hbase_SnapshotInfo_getSnapshotDescription_rdh
/** * Returns the snapshot descriptor */ public SnapshotDescription getSnapshotDescription() {return ProtobufUtil.createSnapshotDesc(this.snapshot); }
3.26
hbase_SnapshotInfo_addStoreFile_rdh
/** * Add the specified store file to the stats * * @param region * region encoded Name * @param family * family name * @param storeFile * store file name * @param filesMap * store files map for all snapshots, it may be null * @return the store file information */ FileInfo addStoreFile(final RegionI...
3.26
hbase_SnapshotInfo_isMissing_rdh
/** * Returns true if the file is missing */ public boolean isMissing() { return this.size < 0; }
3.26
hbase_SnapshotInfo_getMissingStoreFilesCount_rdh
/** * Returns the number of missing store files */ public int getMissingStoreFilesCount() { return hfilesMissing.get(); }
3.26
hbase_SnapshotInfo_printSchema_rdh
/** * Dump the {@link org.apache.hadoop.hbase.client.TableDescriptor} */ private void printSchema() { System.out.println("Table Descriptor"); System.out.println("----------------------------------------"); System.out.println(snapshotManifest.getTableDescriptor()...
3.26
hbase_SnapshotInfo_getSnapshotsFilesMap_rdh
/** * Returns the map of store files based on path for all snapshots * * @param conf * the {@link Configuration} to use * @param uniqueHFilesArchiveSize * pass out the size for store files in archive * @param uniqueHFilesSize * pass out the size for store files shared * @param uniqueHFilesMobSize * pa...
3.26
hbase_SnapshotInfo_getSharedStoreFilesSize_rdh
/** * Returns the total size of the store files shared */ public long getSharedStoreFilesSize() { return hfilesSize.get(); }
3.26
hbase_SnapshotInfo_getMobStoreFilesSize_rdh
/** * Returns the total size of the store files in the mob store */ public long getMobStoreFilesSize() { return hfilesMobSize.get(); } /** * * @return the total size of the store files in the archive which is not shared with other snapshots and tables This is only calculated when {@link #getSnapshotStats(Confi...
3.26
hbase_SnapshotInfo_inArchive_rdh
/** * Returns true if the file is in the archive */ public boolean inArchive() { return this.inArchive; }
3.26
hbase_SnapshotInfo_getSharedStoreFilePercentage_rdh
/** * Returns the percentage of the shared store files */ public float getSharedStoreFilePercentage() { return (((float) (hfilesSize.get())) / getStoreFilesSize()) * 100; }
3.26
hbase_SnapshotInfo_getMobStoreFilePercentage_rdh
/** * Returns the percentage of the mob store files */ public float getMobStoreFilePercentage() { return (((float) (hfilesMobSize.get())) / getStoreFilesSize()) * 100;}
3.26
hbase_SnapshotInfo_getMobStoreFilesCount_rdh
/** * Returns the number of available store files in the mob dir */ public int getMobStoreFilesCount() {return hfilesMobCount.get(); }
3.26
hbase_SnapshotInfo_loadSnapshotInfo_rdh
/** * Load snapshot info and table descriptor for the specified snapshot * * @param snapshotName * name of the snapshot to load * @return false if snapshot is not found */ private boolean loadSnapshotInfo(final Stri...
3.26
hbase_SnapshotInfo_isArchivedFileStillReferenced_rdh
/** * Check if for a give file in archive, if there are other snapshots/tables still reference it. * * @param filePath * file path in archive * @param snapshotFilesMap * a map for store files in snapshots about how many snapshots refer to * it. * @return true or false */ private boolean isArchivedFileSti...
3.26
hbase_SnapshotInfo_getCorruptedStoreFilesCount_rdh
/** * Returns the number of corrupted store files */ public int getCorruptedStoreFilesCount() { return hfilesCorrupted.get(); }
3.26
hbase_SnapshotInfo_getSnapshotStats_rdh
/** * Returns the snapshot stats * * @param conf * the {@link Configuration} to use * @param snapshotDesc * HBaseProtos.SnapshotDescription to get stats from * @param filesMap * {@link Map} store files map for all snapshots, it may be null * @return the snapshot stats */public static SnapshotStats getSn...
3.26
hbase_SnapshotInfo_addLogFile_rdh
/** * Add the specified log file to the stats * * @param server * server name * @param logfile * log file name * @return the log information */ FileInfo addLogFile(final String server, final String logfile) throws IOException {WALLink logLink = new WALLink(conf, server, logfile); long v8 = -1;try { ...
3.26
hbase_SnapshotInfo_printInfo_rdh
/** * Dump the {@link SnapshotDescription} */ private void printInfo() { SnapshotProtos.SnapshotDescription snapshotDesc = snapshotManifest.getSnapshotDescription(); SimpleDateFormat df = new SimpleDateFormat("yyyy-MM-dd'T'HH:mm:ss"); System.out.println("Snapshot Info"); System.out.println("---------...
3.26
hbase_SnapshotInfo_isSnapshotCorrupted_rdh
/** * Returns true if the snapshot is corrupted */ public boolean isSnapshotCorrupted() { return ((hfilesMissing.get() > 0) || (logsMissing.get() > 0)) || (hfilesCorrupted.get() > 0); }
3.26
hbase_SnapshotInfo_getSnapshotFilesMap_rdh
/** * Gets the store files map for snapshot * * @param conf * the {@link Configuration} to use * @param snapshot * {@link SnapshotDescription} to get stats from * @param exec * the {@link ExecutorService} to use * @param filesMap * {@link Map} the map to put the mapping entries * @param uniqueHFilesA...
3.26
hbase_SnapshotInfo_isCorrupted_rdh
/** * Returns true if the file is corrupted */ public boolean isCorrupted() { return this.corrupted; }
3.26
hbase_SnapshotInfo_getStoreFilesSize_rdh
/** * Returns the total size of the store files referenced by the snapshot */ public long getStoreFilesSize() { return (hfilesSize.get() + hfilesArchiveSize.get()) + hfilesMobSize.get(); }
3.26
hbase_SnapshotInfo_getLogsCount_rdh
/** * Returns the number of available log files */ public int getLogsCount() { return logsCount.get(); }
3.26
hbase_SnapshotInfo_getMissingLogsCount_rdh
/** * Returns the number of missing log files */ public int getMissingLogsCount() { return logsMissing.get(); }
3.26
hbase_SnapshotInfo_printFiles_rdh
/** * Collect the hfiles and logs statistics of the snapshot and dump the file list if requested and * the collected information. */ private void printFiles(final boolean showFiles, final boolean showStats) throws IOException { if (showFiles) { System.out.println("Snapshot ...
3.26
hbase_SnapshotInfo_getLogsSize_rdh
/** * Returns the total log size */ public long getLogsSize() { return logSize.get(); }
3.26
hbase_SnapshotInfo_getArchivedStoreFilesCount_rdh
/** * Returns the number of available store files in the archive */ public int getArchivedStoreFilesCount() { return hfilesArchiveCount.get(); }
3.26
hbase_SnapshotInfo_getSize_rdh
/** * Returns the file size */ public long getSize() { return this.size; }
3.26
hbase_SnapshotInfo_getArchivedStoreFileSize_rdh
/** * Returns the total size of the store files in the archive */ public long getArchivedStoreFileSize() { return hfilesArchiveSize.get(); }
3.26
hbase_DefaultEnvironmentEdge_currentTime_rdh
/** * {@inheritDoc } * <p> * This implementation returns {@link System#currentTimeMillis()} * </p> */ @Override public long currentTime() { return System.currentTimeMillis(); }
3.26
hbase_ByteBufferListOutputStream_releaseResources_rdh
/** * Release the resources it uses (The ByteBuffers) which are obtained from pool. Call this only * when all the data is fully used. And it must be called at the end of usage else we will leak * ByteBuffers from pool. */ public void releaseResources() { try { close(); } catch (IOException e) { ...
3.26
hbase_ByteBufferListOutputStream_getByteBuffers_rdh
/** * We can be assured that the buffers returned by this method are all flipped * * @return list of bytebuffers */public List<ByteBuffer> getByteBuffers() { if (!this.lastBufFlipped) { this.lastBufFlipped = true; // All the other BBs are already flipped while moving to the new BB. curBu...
3.26
hbase_ExpiredMobFileCleaner_cleanExpiredMobFiles_rdh
/** * Cleans the MOB files when they're expired and their min versions are 0. If the latest timestamp * of Cells in a MOB file is older than the TTL in the column family, it's regarded as expired. * This cleaner deletes them. At a time T0, the cells in a mob file M0 are expired. If a user * starts a scan before T0,...
3.26
hbase_ScanQueryMatcher_preCheck_rdh
/** * Check before the delete logic. * * @return null means continue. */ protected final MatchCode preCheck(Cell cell) { if (currentRow == null) { // Since the curCell is null it means we are already sure that we have moved over to the next // row return MatchCode.DONE; } // if r...
3.26
hbase_ScanQueryMatcher_clearCurrentRow_rdh
/** * Make {@link #currentRow()} return null. */ public void clearCurrentRow() { currentRow = null; }
3.26
hbase_ScanQueryMatcher_m0_rdh
/** * Returns true if the cell is expired */ private static boolean m0(final Cell cell, final long oldestTimestamp, final long now) { // Look for a TTL tag first. Use it instead of the family setting if // found. If a cell has multiple TTLs, resolve the conflict by using the // first tag e...
3.26
hbase_ScanQueryMatcher_checkColumn_rdh
// Used only for testing purposes static MatchCode checkColumn(ColumnTracker columnTracker, byte[] bytes, int offset, int length, long ttl, byte type, boolean ignoreCount) throws IOException {KeyValue kv = KeyValueUtil.createFirstOnRow(HConstants.EMPTY_BYTE_ARRAY, 0, 0, HConstants.EMPTY_BYTE_ARRAY, 0, 0, bytes, off...
3.26
hbase_ScanQueryMatcher_currentRow_rdh
/** * Returns a cell represent the current row */ public Cell currentRow() { return currentRow; }
3.26
hbase_ScanQueryMatcher_setToNewRow_rdh
/** * Set the row when there is change in row */ public void setToNewRow(Cell currentRow) { this.currentRow = currentRow; columns.reset();reset(); }
3.26
hbase_ScanQueryMatcher_m1_rdh
/** * * @param nextIndexed * the key of the next entry in the block index (if any) * @param currentCell * The Cell we're using to calculate the seek key * @return result of the compare between the indexed key and the key portion of the passed cell */public int m1(Cell nextIndexed, Cell currentCell) { ColumnC...
3.26
hbase_ScanQueryMatcher_getStartKey_rdh
/** * Returns the start key */ public Cell getStartKey() { return startKey; }
3.26
hbase_ByteBufferOutputStream_getByteBuffer_rdh
/** * This flips the underlying BB so be sure to use it _last_! */ public ByteBuffer getByteBuffer() { curBuf.flip(); return curBuf; }
3.26
hbase_ByteBufferOutputStream_writeInt_rdh
/** * Writes an <code>int</code> to the underlying output stream as four bytes, high byte first. * * @param i * the <code>int</code> to write * @throws IOException * if an I/O error occurs. */ @Override public void writeInt(int i) throws IOException { checkSizeAndGrow(Bytes.SIZEOF_INT); ByteBufferUti...
3.26
hbase_ByteBufferOutputStream_writeTo_rdh
/** * Writes the complete contents of this byte buffer output stream to the specified output stream * argument. * * @param out * the output stream to which to write the data. * @exception IOException * if an I/O error occurs. */ public void writeTo(OutputStream out) throws IOException { WritableByteChan...
3.26
hbase_ByteBufferOutputStream_write_rdh
// OutputStream @Override public void write(int b) throws IOException { checkSizeAndGrow(Bytes.SIZEOF_BYTE); curBuf.put(((byte) (b))); }
3.26
hbase_User_getToken_rdh
/** * Returns the Token of the specified kind associated with this user, or null if the Token is not * present. * * @param kind * the kind of token * @param service * service on which the token is supposed to be used * @return the token of the specified kind. */public Token<?> getToken(String kind, String ...
3.26
hbase_User_isHBaseSecurityEnabled_rdh
/** * Returns whether or not secure authentication is enabled for HBase. Note that HBase security * requires HDFS security to provide any guarantees, so it is recommended that secure HBase should * run on secure HDFS. */ public static boolean isHBaseSecurityEnabled(Configuration conf) { return "kerberos".equals...
3.26
hbase_User_addToken_rdh
/** * Adds the given Token to the user's credentials. * * @param token * the token to add */ public void addToken(Token<? extends TokenIdentifier> token) { ugi.addToken(token); }
3.26
hbase_User_getTokens_rdh
/** * Returns all the tokens stored in the user's credentials. */ public Collection<Token<? extends TokenIdentifier>> getTokens() { return ugi.getTokens(); }
3.26
hbase_User_createUserForTesting_rdh
/** * Create a user for testing. * * @see User#createUserForTesting(org.apache.hadoop.conf.Configuration, String, String[]) */ public static User createUserForTesting(Configuration conf, String name, String[] groups) { synchronized(UserProvider.class) { if (!(UserProvider.groups instanceof Testi...
3.26
hbase_User_create_rdh
/** * Wraps an underlying {@code UserGroupInformation} instance. * * @param ugi * The base Hadoop user */ public static User create(UserGroupInformation ugi) { if (ugi == null) { return null; } return new SecureHadoopUser(ugi); }
3.26
hbase_User_getName_rdh
/** * Returns the full user name. For Kerberos principals this will include the host and realm * portions of the principal name. * * @return User full name. */ public String getName() { return ugi.getUserName(); }
3.26
hbase_User_isSecurityEnabled_rdh
/** * Returns the result of {@code UserGroupInformation.isSecurityEnabled()}. */ public static boolean isSecurityEnabled() { return UserGroupInformation.isSecurityEnabled(); }
3.26
hbase_User_runAsLoginUser_rdh
/** * Executes the given action as the login user */ @SuppressWarnings({ "rawtypes", "unchecked" }) public static <T> T runAsLoginUser(PrivilegedExceptionAction<T> action) throws IOException { try {Class c = Class.forName("org.apache.hadoop.security.SecurityUtil"); Class[] types = new Class[]{ Privil...
3.26
hbase_User_getCurrent_rdh
/** * Returns the {@code User} instance within current execution context. */ public static User getCurrent() throws IOException { User user = new SecureHadoopUser(); if (user.getUGI() == null) { return null; } return user; }
3.26
hbase_User_isLoginFromKeytab_rdh
/** * Returns true if user credentials are obtained from keytab. */ public boolean isLoginFromKeytab() {return ugi.isFromKeytab(); }
3.26
hbase_User_shouldLoginFromKeytab_rdh
/** * In secure environment, if a user specified his keytab and principal, a hbase client will try to * login with them. Otherwise, hbase client will try to obtain ticket(through kinit) from system. * * @param conf * configuration file * @return true if keytab and principal are configured */ public static bool...
3.26
hbase_User_login_rdh
/** * Login through configured keytab and pricipal. * * @param keytabLocation * location of keytab * @param principalName * principal in keytab * @throws IOException * exception from UserGroupInformation.loginUserFromKeytab */ public static void login(String keytabLocation, String principalName) throws I...
3.26
hbase_User_m0_rdh
/** * Returns the list of groups of which this user is a member. On secure Hadoop this returns the * group information for the user as resolved on the server. For 0.20 based Hadoop, the group * names are passed from the client. */ public String[] m0() { return ugi.getGroupNames(); }
3.26
hbase_MetricsStochasticBalancerSourceImpl_updateStochasticCost_rdh
/** * Reports stochastic load balancer costs to JMX */ @Override public void updateStochasticCost(String tableName, String costFunctionName, String functionDesc, Double cost) { if (((tableName == null) || (costFunctionName == null)) || (cost == null)) { return; } if (functionDesc != null) { ...
3.26
hbase_MetricsStochasticBalancerSourceImpl_calcMruCap_rdh
/** * Calculates the mru cache capacity from the metrics size */ private static int calcMruCap(int metricsSize) { return ((int) (Math.ceil(metricsSize / MRU_LOAD_FACTOR))) + 1; }
3.26
hbase_HBaseServiceHandler_getTable_rdh
/** * Creates and returns a Table instance from a given table name. name of table * * @return Table object * @throws IOException * if getting the table fails */ protected Table getTable(final byte[] tableName) throws IOException { String table = Bytes.toString(tableName); return connectionCache.getTable...
3.26
hbase_HBaseServiceHandler_getAdmin_rdh
/** * Obtain HBaseAdmin. Creates the instance if it is not already created. */ protected Admin getAdmin() throws IOException { return connectionCache.getAdmin(); }
3.26
hbase_RawCell_cloneTags_rdh
/** * Allows cloning the tags in the cell to a new byte[] * * @return the byte[] having the tags */ default byte[] cloneTags() { return PrivateCellUtil.cloneTags(this); }
3.26
hbase_RawCell_createCell_rdh
/** * Returns A new cell which is having the extra tags also added to it. */ public static Cell createCell(Cell cell, List<Tag> tags) { return PrivateCellUtil.createCell(cell, tags); }
3.26
hbase_RawCell_checkForTagsLength_rdh
/** * Check the length of tags. If it is invalid, throw IllegalArgumentException * * @param tagsLength * the given length of tags * @throws IllegalArgumentException * if tagslength is invalid */ public static void checkForTagsLength(int tagsLength) { if (tagsLength > MAX_TAGS_LENGTH) { throw ...
3.26
hbase_RawCell_getTag_rdh
/** * Returns the specific tag of the given type * * @param type * the type of the tag * @return the specific tag if available or null */ default Optional<Tag> getTag(byte type) { return PrivateCellUtil.getTag(this, type); }
3.26
hbase_RawCell_getTags_rdh
/** * Creates a list of tags in the current cell * * @return a list of tags */ default Iterator<Tag> getTags() { return PrivateCellUtil.tagsIterator(this); }
3.26
hbase_RSGroupAdminServiceImpl_fillTables_rdh
// for backward compatible private RSGroupInfo fillTables(RSGroupInfo rsGroupInfo) throws IOException {return RSGroupUtil.fillTables(rsGroupInfo, master.getTableDescriptors().getAll().values()); }
3.26
hbase_StripeCompactionPolicy_setMajorRange_rdh
/** * Sets compaction "major range". Major range is the key range for which all the files are * included, so they can be treated like major-compacted files. * * @param startRow * Left boundary, inclusive. * @param endRow * Right boundary, exclusive. */ public void setMajorRange(byte[] startRow, byte[] endRo...
3.26
hbase_StripeCompactionPolicy_selectSimpleCompaction_rdh
/** * Selects the compaction of a single stripe using default policy. * * @param sfs * Files. * @param allFilesOnly * Whether a compaction of all-or-none files is needed. * @return The resulting selection. */ private List<HStoreFile> selectSimpleCompaction(List<HStoreFile> sfs, boolean allFilesOnly, boolean...
3.26
hbase_StripeCompactionPolicy_setMajorRangeFull_rdh
/** * Set major range of the compaction to the entire compaction range. See * {@link #setMajorRange(byte[], byte[])}. */ public void setMajorRangeFull() { setMajorRange(this.startRow, this.endRow); }
3.26
hbase_StripeCompactionPolicy_needsSingleStripeCompaction_rdh
/** * * @param si * StoreFileManager. * @return Whether any stripe potentially needs compaction. */ protected boolean needsSingleStripeCompaction(StripeInformationProvider si) {int minFiles = this.config.getStripeCompactMinFiles(); for (List<HStoreFile> stripe : si.getStripes()) {if (stripe.size() >= minFile...
3.26
hbase_TableSnapshotInputFormat_setInput_rdh
/** * Configures the job to use TableSnapshotInputFormat to read from a snapshot. * * @param job * the job to configure * @param snapshotName * the name of the snapshot to read from * @param restoreDir * a temporary directory to restore the snapshot into. Current user * should have write permissions to...
3.26
hbase_StartMiniClusterOption_builder_rdh
/** * Returns a new builder. */ public static Builder builder() { return new Builder(); }
3.26
hbase_LzoCodec_getBufferSize_rdh
// Package private static int getBufferSize(Configuration conf) { return conf.getInt(LZO_BUFFER_SIZE_KEY, conf.getInt(CommonConfigurationKeys.IO_COMPRESSION_CODEC_LZO_BUFFERSIZE_KEY, CommonConfigurationKeys.IO_COMPRESSION_CODEC_LZO_BUFFERSIZE_DEFAULT)); }
3.26