name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
hadoop_RenameOperation_renameFileToDest
/** * The source is a file: rename it to the destination, which * will be under the current destination path if that is a directory. * @return the path of the object created. * @throws IOException failure */ protected Path renameFileToDest() throws IOException { final StoreContext storeContext = getStoreContext(...
3.68
flink_BlockingBackChannelBroker_instance
/** Retrieve singleton instance. */ public static Broker<BlockingBackChannel> instance() { return INSTANCE; }
3.68
shardingsphere-elasticjob_JobAnnotationBuilder_generateJobConfiguration
/** * Generate job configuration from @ElasticJobConfiguration. * * @param type The job of @ElasticJobConfiguration annotation class * @return job configuration */ public static JobConfiguration generateJobConfiguration(final Class<?> type) { ElasticJobConfiguration annotation = type.getAnnotation(ElasticJobC...
3.68
hbase_SyncTable_main
/** * Main entry point. */ public static void main(String[] args) throws Exception { int ret = ToolRunner.run(new SyncTable(HBaseConfiguration.create()), args); System.exit(ret); }
3.68
morf_SqlServerDialect_getSqlForDateToYyyymmddHHmmss
/** * @see org.alfasoftware.morf.jdbc.SqlDialect#getSqlForDateToYyyymmddHHmmss(org.alfasoftware.morf.sql.element.Function) */ @Override protected String getSqlForDateToYyyymmddHHmmss(Function function) { return String.format("REPLACE(REPLACE(REPLACE(CONVERT(VARCHAR(19),%s, 120),'-',''), ':', ''), ' ', '')", getSqlF...
3.68
querydsl_MathExpressions_cosh
/** * Create a {@code cosh(num)} expression * * <p>Returns the hyperbolic cosine of num radians.</p> * * @param num numeric expression * @return cosh(num) */ public static <A extends Number & Comparable<?>> NumberExpression<Double> cosh(Expression<A> num) { return Expressions.numberOperation(Double.class, Op...
3.68
hbase_AvlUtil_seekTo
/** * Reset the iterator, and seeks to the specified key * @param root the current root of the tree * @param key the key for the node we are trying to find * @param keyComparator the comparator to use to match node and key */ public void seekTo(final TNode root, final Object key, final AvlKeyC...
3.68
framework_WindowElement_close
/** * Clicks the close button of the window. */ public void close() { getCloseButton().click(); }
3.68
hbase_HFileBlockIndex_writeIndexBlocks
/** * Writes the root level and intermediate levels of the block index into the output stream, * generating the tree from bottom up. Assumes that the leaf level has been inline-written to * the disk if there is enough data for more than one leaf block. We iterate by breaking the * current level of the block index, ...
3.68
graphhopper_VectorTile_getFeaturesCount
/** * <pre> * The actual features in this tile. * </pre> * * <code>repeated .vector_tile.Tile.Feature features = 2;</code> */ public int getFeaturesCount() { if (featuresBuilder_ == null) { return features_.size(); } else { return featuresBuilder_.getCount(); } }
3.68
hadoop_ProcessIdFileReader_getProcessId
/** * Get the process id from specified file path. * Parses each line to find a valid number * and returns the first one found. * @return Process Id if obtained from path specified else null * @throws IOException */ public static String getProcessId(Path path) throws IOException { if (path == null) { throw ...
3.68
flink_TaskStateManagerImpl_notifyCheckpointAborted
/** Tracking when some local state can be disposed. */ @Override public void notifyCheckpointAborted(long checkpointId) { localStateStore.abortCheckpoint(checkpointId); }
3.68
druid_Lexer_lexError
/** * Report an error at the given position using the provided arguments. */ protected void lexError(String key, Object... args) { token = ERROR; }
3.68
rocketmq-connect_RocketMQScheduledReporter_reportHistograms
/** * report histograms * * @param histograms */ private void reportHistograms(SortedMap<MetricName, Double> histograms) { histograms.forEach((name, value) -> { send(name, value); }); }
3.68
morf_XmlDataSetProducer_tableExists
/** * @see org.alfasoftware.morf.metadata.Schema#tableExists(java.lang.String) */ @Override public boolean tableExists(String name) { return xmlStreamProvider.tableExists(name); }
3.68
hadoop_MountTableRefresherService_refresh
/** * Refresh mount table cache of this router as well as all other routers. * * @throws StateStoreUnavailableException if the state store is not available. */ public void refresh() throws StateStoreUnavailableException { RouterStore routerStore = router.getRouterStateManager(); try { routerStore.loadCache...
3.68
framework_PushRequestHandler_createPushHandler
/** * Creates a push handler for this request handler. * <p> * Create your own request handler and override this method if you want to * customize the {@link PushHandler}, e.g. to dynamically decide the suspend * timeout. * * @since 7.6 * @param service * the vaadin service * @return the push handl...
3.68
framework_HierarchyMapper_doExpand
/** * Expands the given item if it is collapsed and has children, and returns * whether this method expanded the item. * * @param item * the item to expand * @return {@code true} if this method expanded the item, {@code false} * otherwise */ private boolean doExpand(T item) { boolean expa...
3.68
AreaShop_CancellableRegionEvent_isCancelled
/** * Check if the event has been cancelled. * @return true if the event has been cancelled, otherwise false */ public boolean isCancelled() { return cancelled; }
3.68
hadoop_OBSFileSystem_getMaxEntriesToDelete
/** * Return maximum number of entries in one multi-object delete call. * * @return the maximum number of entries in one multi-object delete call */ int getMaxEntriesToDelete() { return maxEntriesToDelete; }
3.68
hadoop_SnapshotDiffReportListing_getLastPath
/** * @return {@link #lastPath} */ public byte[] getLastPath() { return lastPath; }
3.68
hadoop_AwsStatisticsCollector_publish
/** * This is the callback from the AWS SDK where metrics * can be collected. * @param metricCollection metrics collection */ @Override public void publish(MetricCollection metricCollection) { // MetricCollections are nested, so we need to traverse through their // "children" to collect the desired metrics. E.g...
3.68
hbase_Scan_setAllowPartialResults
/** * Setting whether the caller wants to see the partial results when server returns * less-than-expected cells. It is helpful while scanning a huge row to prevent OOM at client. By * default this value is false and the complete results will be assembled client side before being * delivered to the caller. * @see ...
3.68
pulsar_Schema_AVRO
/** * Create a Avro schema type with schema definition. * * @param schemaDefinition the definition of the schema * @return a Schema instance */ static <T> Schema<T> AVRO(SchemaDefinition<T> schemaDefinition) { return DefaultImplementation.getDefaultImplementation().newAvroSchema(schemaDefinition); }
3.68
hbase_FirstKeyOnlyFilter_hasFoundKV
/** Returns true if first KV has been found. */ protected boolean hasFoundKV() { return this.foundKV; }
3.68
framework_UIConnector_loadTheme
/** * Loads the given theme and replaces the given link element with the new * theme link element. * * @param newTheme * The name of the new theme * @param newThemeUrl * The url of the new theme * @param tagToReplace * The link element to replace. If null, then the new link * ...
3.68
framework_TabSheet_setTabPosition
/** * Sets the position of the tab. * * @param tab * The tab * @param position * The new position of the tab */ public void setTabPosition(Tab tab, int position) { int oldPosition = getTabPosition(tab); components.remove(oldPosition); components.add(position, tab.getComponent())...
3.68
framework_VPopupTimeCalendar_makeDate
/** * Creates a date based on the provided date values map. * * @param dateValues * a map with date values to convert into a date * @return the date based on the dateValues map */ @SuppressWarnings("deprecation") public static Date makeDate(Map<DateTimeResolution, Integer> dateValues) { if (dateVal...
3.68
framework_FileDropTarget_getParent
/** * Returns the component this extension is attached to. * * @return Extended component. */ @Override public T getParent() { return super.getParent(); }
3.68
framework_VComboBox_getItemOffsetHeight
/* * Gets the height of one menu item. */ int getItemOffsetHeight() { List<MenuItem> items = getItems(); return items != null && !items.isEmpty() ? items.get(0).getOffsetHeight() : 0; }
3.68
flink_Task_transitionState
/** * Try to transition the execution state from the current state to the new state. * * @param currentState of the execution * @param newState of the execution * @param cause of the transition change or null * @return true if the transition was successful, otherwise false */ private boolean transitionState( ...
3.68
zxing_PDF417_getNumberOfPadCodewords
/** * Calculates the number of pad codewords as described in 4.9.2 of ISO/IEC 15438:2001(E). * * @param m the number of source codewords prior to the additional of the Symbol Length * Descriptor and any pad codewords * @param k the number of error correction codewords * @param c the number of columns in ...
3.68
flink_JobVertexBackPressureHandler_getBackPressureLevel
/** * Returns the back pressure level as a String. * * @param backPressureRatio Ratio of back pressures samples to total number of samples. * @return Back pressure level ('ok', 'low', or 'high') */ private static JobVertexBackPressureInfo.VertexBackPressureLevel getBackPressureLevel( double backPressureRat...
3.68
flink_TwoPhaseCommitSinkFunction_recoverAndCommitInternal
/** * This method must be the only place to call {@link #recoverAndCommit(Object)} to ensure that * the configuration parameters {@link #transactionTimeout} and {@link * #ignoreFailuresAfterTransactionTimeout} are respected. */ private void recoverAndCommitInternal(TransactionHolder<TXN> transactionHolder) { tr...
3.68
hadoop_ListResultEntrySchema_name
/** * Get the name value. * * @return the name value */ public String name() { return name; }
3.68
framework_SQLContainer_commit
/** * Commits all the changes, additions and removals made to the items of this * container. * * @throws UnsupportedOperationException * @throws SQLException */ public void commit() throws UnsupportedOperationException, SQLException { try { getLogger().log(Level.FINER, "Commiting chang...
3.68
flink_BlobServer_moveTempFileToStore
/** * Moves the temporary <tt>incomingFile</tt> to its permanent location where it is available for * use. * * @param incomingFile temporary file created during transfer * @param jobId ID of the job this blob belongs to or <tt>null</tt> if job-unrelated * @param digest BLOB content digest, i.e. hash * @param blo...
3.68
rocketmq-connect_RocketMQScheduledReporter_reportCounters
/** * report counters * * @param counters */ private void reportCounters(SortedMap<MetricName, Long> counters) { counters.forEach((name, value) -> { send(name, Double.parseDouble(value.toString())); }); }
3.68
hudi_HoodieRealtimeRecordReaderUtils_avroToArrayWritable
/** * Convert the projected read from delta record into an array writable. */ public static Writable avroToArrayWritable(Object value, Schema schema) { return avroToArrayWritable(value, schema, false); }
3.68
hudi_CompactionUtils_getEarliestInstantToRetainForCompaction
/** * Gets the earliest instant to retain for MOR compaction. * If there is no completed compaction, * num delta commits >= "hoodie.compact.inline.max.delta.commits" * If there is a completed compaction, * num delta commits after latest completed compaction >= "hoodie.compact.inline.max.delta.commits" * * @param...
3.68
hadoop_BatchedRequests_getIteratorType
/** * Get Iterator type. * @return Iterator type. */ public IteratorType getIteratorType() { return iteratorType; }
3.68
framework_ColorPickerSelect_createAllColors
/** * Creates the all colors. * * @param rows * the rows * @param columns * the columns * * @return the color[][] */ private Color[][] createAllColors(int rows, int columns) { Color[][] colors = new Color[rows][columns]; for (int row = 0; row < rows; row++) { for (int col...
3.68
pulsar_ConsumerConfiguration_setSubscriptionType
/** * Select the subscription type to be used when subscribing to the topic. * <p> * Default is {@link SubscriptionType#Exclusive} * * @param subscriptionType * the subscription type value */ public ConsumerConfiguration setSubscriptionType(SubscriptionType subscriptionType) { Objects.requireNonNu...
3.68
hudi_CleanPlanner_getLastCompletedCommitTimestamp
/** * Returns the last completed commit timestamp before clean. */ public String getLastCompletedCommitTimestamp() { if (commitTimeline.lastInstant().isPresent()) { return commitTimeline.lastInstant().get().getTimestamp(); } else { return ""; } }
3.68
hadoop_ManifestSuccessData_dumpDiagnostics
/** * Dump the diagnostics (if any) to a string. * @param prefix prefix before every entry * @param middle string between key and value * @param suffix suffix to each entry * @return the dumped string */ public String dumpDiagnostics(String prefix, String middle, String suffix) { return joinMap(diagnostics, pre...
3.68
hudi_BaseHoodieClient_close
/** * Releases any resources used by the client. */ @Override public void close() { stopEmbeddedServerView(true); this.context.setJobStatus("", ""); this.heartbeatClient.close(); this.txnManager.close(); }
3.68
graphhopper_DijkstraOneToMany_getMemoryUsageAsString
/** * List currently used memory in MB (approximately) */ public String getMemoryUsageAsString() { long len = weights.length; return ((8L + 4L + 4L) * len + changedNodes.getCapacity() * 4L + heap.getCapacity() * (4L + 4L)) / Helper.MB + "MB"; }
3.68
hadoop_LoggingAuditor_prepareActiveContext
/** * Get/Prepare the active context for a span. * @return the common audit context. */ private CommonAuditContext prepareActiveContext() { return currentAuditContext(); }
3.68
framework_UIDL_getIntVariable
/** * Gets the value of the named variable. * * @param name * the name of the variable * @return the value of the variable */ public int getIntVariable(String name) { return var().getInt(name); }
3.68
hmily_SpringCloudHmilyOrderApplication_main
/** * main. * * @param args args */ public static void main(final String[] args) { SpringApplication.run(SpringCloudHmilyOrderApplication.class, args); }
3.68
hadoop_GetContentSummaryOperation_probePathStatusOrNull
/** * Get the status of a path, downgrading FNFE to null result. * @param p path to probe. * @param probes probes to exec * @return the status or null * @throws IOException failure other than FileNotFound */ private S3AFileStatus probePathStatusOrNull(final Path p, final Set<StatusProbeEnum> probes) throws IO...
3.68
hbase_AsyncTableRegionLocator_getEndKeys
/** * Gets the ending row key for every region in the currently open table. * <p> * This is mainly useful for the MapReduce integration. * @return Array of region ending row keys */ default CompletableFuture<List<byte[]>> getEndKeys() { return getStartEndKeys().thenApply( startEndKeys -> startEndKeys.stream(...
3.68
hadoop_AHSLogsPage_preHead
/* * (non-Javadoc) * * @see * org.apache.hadoop.yarn.server.applicationhistoryservice.webapp.AHSView# * preHead(org.apache.hadoop .yarn.webapp.hamlet.Hamlet.HTML) */ @Override protected void preHead(Page.HTML<__> html) { String logEntity = $(ENTITY_STRING); if (logEntity == null || logEntity.isEmpty()) { ...
3.68
hbase_CommonFSUtils_checkShortCircuitReadBufferSize
/** * Check if short circuit read buffer size is set and if not, set it to hbase value. * @param conf must not be null */ public static void checkShortCircuitReadBufferSize(final Configuration conf) { final int defaultSize = HConstants.DEFAULT_BLOCKSIZE * 2; final int notSet = -1; // DFSConfigKeys.DFS_CLIENT_R...
3.68
hudi_EmbeddedTimelineService_stopForBasePath
/** * Stops the embedded timeline service for the given base path. If a timeline service is managing multiple tables, it will only be shutdown once all tables have been stopped. * @param basePath For the table to stop the service for */ public void stopForBasePath(String basePath) { synchronized (SERVICE_LOCK) { ...
3.68
dubbo_CodecSupport_getNullBytesOf
/** * Get the null object serialize result byte[] of Serialization from the cache, * if not, generate it first. * * @param s Serialization Instances * @return serialize result of null object */ public static byte[] getNullBytesOf(Serialization s) { return ConcurrentHashMapUtils.computeIfAbsent(ID_NULLBYTES_MA...
3.68
hbase_HBaseServerException_setServerOverloaded
/** * Necessary for parsing RemoteException on client side * @param serverOverloaded True if server was overloaded when exception was thrown */ public void setServerOverloaded(boolean serverOverloaded) { this.serverOverloaded = serverOverloaded; }
3.68
framework_Html5File_setStreamVariable
/** * Sets the {@link StreamVariable} that into which the file contents will be * written. Usage of StreamVariable is similar to {@link Upload} component. * <p> * If the {@link StreamVariable} is not set in the {@link DropHandler} the * file contents will not be sent to server. * <p> * <em>Note!</em> receiving f...
3.68
framework_DropTargetExtensionConnector_setDropEffect
/** * Set the drop effect for the dragenter / dragover event, if one has been * set from server side. * <p> * From Moz Foundation: "You can modify the dropEffect property during the * dragenter or dragover events, if for example, a particular drop target * only supports certain operations. You can modify the drop...
3.68
flink_TumbleWithSize_on
/** * Specifies the time attribute on which rows are grouped. * * <p>For streaming tables you can specify grouping by a event-time or processing-time * attribute. * * <p>For batch tables you can specify grouping on a timestamp or long attribute. * * @param timeField time attribute for streaming and batch tables...
3.68
hbase_HFileReaderImpl_positionForNextBlock
/** * Set our selves up for the next 'next' invocation, set up next block. * @return True is more to read else false if at the end. */ private boolean positionForNextBlock() throws IOException { // Methods are small so they get inlined because they are 'hot'. long lastDataBlockOffset = reader.getTrailer().getLas...
3.68
hudi_SparkRDDReadClient_addHoodieSupport
/** * Adds support for accessing Hoodie built tables from SparkSQL, as you normally would. * * @return SparkConf object to be used to construct the SparkContext by caller */ public static SparkConf addHoodieSupport(SparkConf conf) { conf.set("spark.sql.hive.convertMetastoreParquet", "false"); return conf; }
3.68
flink_IterativeStream_closeWith
/** * Closes the iteration. This method defines the end of the iterative program part that will * be fed back to the start of the iteration as the second input in the {@link * ConnectedStreams}. * * @param feedbackStream {@link DataStream} that will be used as second input to the * iteration head. * @return ...
3.68
flink_HiveParserSemanticAnalyzer_processLateralView
/** * Given the AST with TOK_LATERAL_VIEW as the root, get the alias for the table or subquery in * the lateral view and also make a mapping from the alias to all the lateral view AST's. */ private String processLateralView(HiveParserQB qb, HiveParserASTNode lateralView) throws SemanticException { int nu...
3.68
hbase_HFileArchiveUtil_getTableName
/* * @return table name given archive file path */ public static TableName getTableName(Path archivePath) { Path p = archivePath; String tbl = null; // namespace is the 4th parent of file for (int i = 0; i < 5; i++) { if (p == null) return null; if (i == 3) tbl = p.getName(); p = p.getParent(); ...
3.68
framework_VUI_windowSizeMaybeChanged
/** * Called when the window or parent div might have been resized. * * This immediately checks the sizes of the window and the parent div (if * monitoring it) and triggers layout recalculation if they have changed. * * @param newWindowWidth * The new width of the window * @param newWindowHeight * ...
3.68
querydsl_MapExpressionBase_containsKey
/** * Create a {@code key in keys(this)} expression * * @param key key * @return expression */ public final BooleanExpression containsKey(K key) { return Expressions.booleanOperation(Ops.CONTAINS_KEY, mixin, ConstantImpl.create(key)); }
3.68
pulsar_PerfClientUtils_printJVMInformation
/** * Print useful JVM information, you need this information in order to be able * to compare the results of executions in different environments. * @param log */ public static void printJVMInformation(Logger log) { log.info("JVM args {}", ManagementFactory.getRuntimeMXBean().getInputArguments()); log.info...
3.68
morf_TableLoader_truncate
/** * Empties the specified table. * * @param table */ private void truncate(Table table) { // Get our own table definition based on the name to avoid case sensitivity bugs on some database vendors log.debug("Clearing table [" + table.getName() + "]"); // Try to use a truncate try { runRecoverably(() -...
3.68
zxing_PDF417Common_getBitCountSum
/** * @param moduleBitCount values to sum * @return sum of values * @deprecated call {@link MathUtils#sum(int[])} */ @Deprecated public static int getBitCountSum(int[] moduleBitCount) { return MathUtils.sum(moduleBitCount); }
3.68
hbase_RegionServerSpaceQuotaManager_getActiveEnforcements
/** * Creates an object well-suited for the RegionServer to use in verifying active policies. */ public ActivePolicyEnforcement getActiveEnforcements() { return new ActivePolicyEnforcement(copyActiveEnforcements(), copyQuotaSnapshots(), rsServices); }
3.68
hadoop_StageConfig_withTaskAttemptDir
/** * Set the Task attempt directory. * @param value new value * @return this */ public StageConfig withTaskAttemptDir(final Path value) { checkOpen(); taskAttemptDir = value; return this; }
3.68
flink_HardwareDescription_getSizeOfManagedMemory
/** * Returns the size of the memory managed by the system for caching, hashing, sorting, ... * * @return The size of the memory managed by the system. */ public long getSizeOfManagedMemory() { return this.sizeOfManagedMemory; }
3.68
framework_PropertysetItem_getItemPropertyIds
/** * Gets the collection of IDs of all Properties stored in the Item. * * @return unmodifiable collection containing IDs of the Properties stored * the Item */ @Override public Collection<?> getItemPropertyIds() { return Collections.unmodifiableCollection(list); }
3.68
hudi_AbstractStreamWriteFunction_instantToWrite
/** * Prepares the instant time to write with for next checkpoint. * * @param hasData Whether the task has buffering data * @return The instant time */ protected String instantToWrite(boolean hasData) { String instant = lastPendingInstant(); // if exactly-once semantics turns on, // waits for the checkpoint ...
3.68
flink_GSCommitRecoverableSerializer_getVersion
/** * The serializer version. Note that, if this changes, then the version of {@link * GSResumeRecoverableSerializer} must also change, because it uses this class to serialize * itself, in part. * * @return The serializer version. */ @Override public int getVersion() { return SERIALIZER_VERSION; }
3.68
hbase_MetricsSource_setWALReaderEditsBufferUsage
/** * Sets the amount of memory in bytes used in this RegionServer by edits pending replication. */ public void setWALReaderEditsBufferUsage(long usageInBytes) { globalSourceSource.setWALReaderEditsBufferBytes(usageInBytes); }
3.68
hadoop_SinglePendingCommit_getUploadId
/** @return ID of the upload. */ public String getUploadId() { return uploadId; }
3.68
hudi_HoodieDataBlock_getRecordIterator
/** * Batch get of keys of interest. Implementation can choose to either do full scan and return matched entries or * do a seek based parsing and return matched entries. * * @param keys keys of interest. * @return List of IndexedRecords for the keys of interest. * @throws IOException in case of failures encounter...
3.68
morf_IdTableTracker_activeNameFor
/** * @see org.alfasoftware.morf.upgrade.TableNameResolver#activeNameFor(java.lang.String) */ @Override public String activeNameFor(String tableName) { if ("IDTABLE".equals(tableName)) return idTableName; return tableName; }
3.68
rocketmq-connect_AbstractStateManagementService_put
/** * Set the state of the connector to the given value. * * @param status the status of the task */ @Override public void put(TaskStatus status) { sendTaskStatus(status, false); }
3.68
flink_HadoopBlockLocation_stripHostname
/** * Looks for a domain suffix in a FQDN and strips it if present. * * @param originalHostname the original hostname, possibly an FQDN * @return the stripped hostname without the domain suffix */ private static String stripHostname(final String originalHostname) { // Check if the hostname domains the domain ...
3.68
querydsl_GenericExporter_setEntityAnnotation
/** * Set the entity annotation * * @param entityAnnotation entity annotation */ public void setEntityAnnotation(Class<? extends Annotation> entityAnnotation) { this.entityAnnotation = entityAnnotation; }
3.68
pulsar_AdditionalServlets_load
/** * Load the additional servlet for the given <tt>servlet name</tt> list. * * @param conf the pulsar service configuration * @return the collection of additional servlet */ public static AdditionalServlets load(PulsarConfiguration conf) throws IOException { String additionalServletDirectory = conf.getPropert...
3.68
hbase_KeyValue_getKeyDataStructureSize
/** * Computes the number of bytes that a <code>KeyValue</code> instance with the provided * characteristics would take up in its underlying data structure for the key. * @param rlength row length * @param flength family length * @param qlength qualifier length * @return the key data structure length */ public s...
3.68
querydsl_SQLExpressions_addMonths
/** * Add the given amount of months to the date * * @param date date * @param months months to add * @return converted date */ public static <D extends Comparable> DateExpression<D> addMonths(DateExpression<D> date, int months) { return Expressions.dateOperation(date.getType(), Ops.DateTimeOps.ADD_MONTHS, da...
3.68
hbase_ServerRegionReplicaUtil_isRegionReplicaWaitForPrimaryFlushEnabled
/** Returns True if wait for primary to flush is enabled for user-space tables. */ public static boolean isRegionReplicaWaitForPrimaryFlushEnabled(Configuration conf) { return conf.getBoolean(REGION_REPLICA_WAIT_FOR_PRIMARY_FLUSH_CONF_KEY, DEFAULT_REGION_REPLICA_WAIT_FOR_PRIMARY_FLUSH); }
3.68
hadoop_ExecutorServiceFuturePool_executeFunction
/** * @param f function to run in future on executor pool * @return future * @throws java.util.concurrent.RejectedExecutionException can be thrown * @throws NullPointerException if f param is null */ public Future<Void> executeFunction(final Supplier<Void> f) { return executor.submit(f::get); }
3.68
flink_FreeingBufferRecycler_recycle
/** * Frees the given memory segment. * * @param memorySegment The memory segment to be recycled. */ @Override public void recycle(MemorySegment memorySegment) { memorySegment.free(); }
3.68
hadoop_IdentifierResolver_setInputWriterClass
/** * Sets the {@link InputWriter} class. */ protected void setInputWriterClass(Class<? extends InputWriter> inputWriterClass) { this.inputWriterClass = inputWriterClass; }
3.68
morf_SqlDialect_getSqlForRowNumber
/** * Produce SQL for getting the row number of the row in the partition * * @return a string representation of the SQL for finding the last day of the month. */ protected String getSqlForRowNumber(){ return "ROW_NUMBER()"; }
3.68
hadoop_BlockStorageMovementNeeded_size
/** * Returns queue size. */ public synchronized int size() { return storageMovementNeeded.size(); }
3.68
framework_RpcDataProviderExtension_beforeClientResponse
/** * {@inheritDoc} * <p> * RpcDataProviderExtension makes all actual RPC calls from this function * based on changes in the container. */ @Override public void beforeClientResponse(boolean initial) { if (initial || bareItemSetTriggeredSizeChange) { /* * Push initial set of rows, assuming Grid...
3.68
morf_SqlDialect_getDeleteLimitSuffix
/** * Returns the SQL that specifies the deletion limit as a suffix, if any, for the dialect. * * @param limit The delete limit. * @return The SQL fragment. */ protected Optional<String> getDeleteLimitSuffix(@SuppressWarnings("unused") int limit) { return Optional.empty(); }
3.68
streampipes_UnicodeTokenizer_tokenize
/** * Tokenizes the text and returns an array of tokens. * * @param text The text * @return The tokens */ public static String[] tokenize(final CharSequence text) { return PAT_NOT_WORD_BOUNDARY.matcher(PAT_WORD_BOUNDARY.matcher(text).replaceAll("\u2063")) .replaceAll("$1").replaceAll("[ \u2063]+", " ").tri...
3.68
framework_Calendar_setTimeZone
/** * Set time zone that this component will use. Null value sets the default * time zone. * * @param zone * Time zone to use */ public void setTimeZone(TimeZone zone) { timezone = zone; if (!currentCalendar.getTimeZone().equals(zone)) { if (zone == null) { zone = TimeZone.g...
3.68
hbase_MutableRegionInfo_getRegionId
/** Returns the regionId */ @Override public long getRegionId() { return regionId; }
3.68
hadoop_CopyOutputFormat_getOutputCommitter
/** {@inheritDoc} */ @Override public OutputCommitter getOutputCommitter(TaskAttemptContext context) throws IOException { return new CopyCommitter(getOutputPath(context), context); }
3.68
morf_ExceptSetOperator_accept
/** * @see org.alfasoftware.morf.sql.SchemaAndDataChangeVisitable#accept(org.alfasoftware.morf.upgrade.SchemaAndDataChangeVisitor) */ @Override public void accept(SchemaAndDataChangeVisitor visitor) { visitor.visit(this); selectStatement.accept(visitor); }
3.68
hudi_HoodieAvroIndexedRecord_readRecordPayload
/** * NOTE: This method is declared final to make sure there's no polymorphism and therefore * JIT compiler could perform more aggressive optimizations */ @SuppressWarnings("unchecked") @Override protected final IndexedRecord readRecordPayload(Kryo kryo, Input input) { // NOTE: We're leveraging Spark's defau...
3.68