name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
hadoop_RenameOperation_renameFileToDest | /**
* The source is a file: rename it to the destination, which
* will be under the current destination path if that is a directory.
* @return the path of the object created.
* @throws IOException failure
*/
protected Path renameFileToDest() throws IOException {
final StoreContext storeContext = getStoreContext(... | 3.68 |
flink_BlockingBackChannelBroker_instance | /** Retrieve singleton instance. */
public static Broker<BlockingBackChannel> instance() {
return INSTANCE;
} | 3.68 |
shardingsphere-elasticjob_JobAnnotationBuilder_generateJobConfiguration | /**
* Generate job configuration from @ElasticJobConfiguration.
*
* @param type The job of @ElasticJobConfiguration annotation class
* @return job configuration
*/
public static JobConfiguration generateJobConfiguration(final Class<?> type) {
ElasticJobConfiguration annotation = type.getAnnotation(ElasticJobC... | 3.68 |
hbase_SyncTable_main | /**
* Main entry point.
*/
public static void main(String[] args) throws Exception {
int ret = ToolRunner.run(new SyncTable(HBaseConfiguration.create()), args);
System.exit(ret);
} | 3.68 |
morf_SqlServerDialect_getSqlForDateToYyyymmddHHmmss | /**
* @see org.alfasoftware.morf.jdbc.SqlDialect#getSqlForDateToYyyymmddHHmmss(org.alfasoftware.morf.sql.element.Function)
*/
@Override
protected String getSqlForDateToYyyymmddHHmmss(Function function) {
return String.format("REPLACE(REPLACE(REPLACE(CONVERT(VARCHAR(19),%s, 120),'-',''), ':', ''), ' ', '')", getSqlF... | 3.68 |
querydsl_MathExpressions_cosh | /**
* Create a {@code cosh(num)} expression
*
* <p>Returns the hyperbolic cosine of num radians.</p>
*
* @param num numeric expression
* @return cosh(num)
*/
public static <A extends Number & Comparable<?>> NumberExpression<Double> cosh(Expression<A> num) {
return Expressions.numberOperation(Double.class, Op... | 3.68 |
hbase_AvlUtil_seekTo | /**
* Reset the iterator, and seeks to the specified key
* @param root the current root of the tree
* @param key the key for the node we are trying to find
* @param keyComparator the comparator to use to match node and key
*/
public void seekTo(final TNode root, final Object key,
final AvlKeyC... | 3.68 |
framework_WindowElement_close | /**
* Clicks the close button of the window.
*/
public void close() {
getCloseButton().click();
} | 3.68 |
hbase_HFileBlockIndex_writeIndexBlocks | /**
* Writes the root level and intermediate levels of the block index into the output stream,
* generating the tree from bottom up. Assumes that the leaf level has been inline-written to
* the disk if there is enough data for more than one leaf block. We iterate by breaking the
* current level of the block index, ... | 3.68 |
graphhopper_VectorTile_getFeaturesCount | /**
* <pre>
* The actual features in this tile.
* </pre>
*
* <code>repeated .vector_tile.Tile.Feature features = 2;</code>
*/
public int getFeaturesCount() {
if (featuresBuilder_ == null) {
return features_.size();
} else {
return featuresBuilder_.getCount();
}
} | 3.68 |
hadoop_ProcessIdFileReader_getProcessId | /**
* Get the process id from specified file path.
* Parses each line to find a valid number
* and returns the first one found.
* @return Process Id if obtained from path specified else null
* @throws IOException
*/
public static String getProcessId(Path path) throws IOException {
if (path == null) {
throw ... | 3.68 |
flink_TaskStateManagerImpl_notifyCheckpointAborted | /** Tracking when some local state can be disposed. */
@Override
public void notifyCheckpointAborted(long checkpointId) {
localStateStore.abortCheckpoint(checkpointId);
} | 3.68 |
druid_Lexer_lexError | /**
* Report an error at the given position using the provided arguments.
*/
protected void lexError(String key, Object... args) {
token = ERROR;
} | 3.68 |
rocketmq-connect_RocketMQScheduledReporter_reportHistograms | /**
* report histograms
*
* @param histograms
*/
private void reportHistograms(SortedMap<MetricName, Double> histograms) {
histograms.forEach((name, value) -> {
send(name, value);
});
} | 3.68 |
morf_XmlDataSetProducer_tableExists | /**
* @see org.alfasoftware.morf.metadata.Schema#tableExists(java.lang.String)
*/
@Override
public boolean tableExists(String name) {
return xmlStreamProvider.tableExists(name);
} | 3.68 |
hadoop_MountTableRefresherService_refresh | /**
* Refresh mount table cache of this router as well as all other routers.
*
* @throws StateStoreUnavailableException if the state store is not available.
*/
public void refresh() throws StateStoreUnavailableException {
RouterStore routerStore = router.getRouterStateManager();
try {
routerStore.loadCache... | 3.68 |
framework_PushRequestHandler_createPushHandler | /**
* Creates a push handler for this request handler.
* <p>
* Create your own request handler and override this method if you want to
* customize the {@link PushHandler}, e.g. to dynamically decide the suspend
* timeout.
*
* @since 7.6
* @param service
* the vaadin service
* @return the push handl... | 3.68 |
framework_HierarchyMapper_doExpand | /**
* Expands the given item if it is collapsed and has children, and returns
* whether this method expanded the item.
*
* @param item
* the item to expand
* @return {@code true} if this method expanded the item, {@code false}
* otherwise
*/
private boolean doExpand(T item) {
boolean expa... | 3.68 |
AreaShop_CancellableRegionEvent_isCancelled | /**
* Check if the event has been cancelled.
* @return true if the event has been cancelled, otherwise false
*/
public boolean isCancelled() {
return cancelled;
} | 3.68 |
hadoop_OBSFileSystem_getMaxEntriesToDelete | /**
* Return maximum number of entries in one multi-object delete call.
*
* @return the maximum number of entries in one multi-object delete call
*/
int getMaxEntriesToDelete() {
return maxEntriesToDelete;
} | 3.68 |
hadoop_SnapshotDiffReportListing_getLastPath | /**
* @return {@link #lastPath}
*/
public byte[] getLastPath() {
return lastPath;
} | 3.68 |
hadoop_AwsStatisticsCollector_publish | /**
* This is the callback from the AWS SDK where metrics
* can be collected.
* @param metricCollection metrics collection
*/
@Override
public void publish(MetricCollection metricCollection) {
// MetricCollections are nested, so we need to traverse through their
// "children" to collect the desired metrics. E.g... | 3.68 |
hbase_Scan_setAllowPartialResults | /**
* Setting whether the caller wants to see the partial results when server returns
* less-than-expected cells. It is helpful while scanning a huge row to prevent OOM at client. By
* default this value is false and the complete results will be assembled client side before being
* delivered to the caller.
* @see ... | 3.68 |
pulsar_Schema_AVRO | /**
* Create a Avro schema type with schema definition.
*
* @param schemaDefinition the definition of the schema
* @return a Schema instance
*/
static <T> Schema<T> AVRO(SchemaDefinition<T> schemaDefinition) {
return DefaultImplementation.getDefaultImplementation().newAvroSchema(schemaDefinition);
} | 3.68 |
hbase_FirstKeyOnlyFilter_hasFoundKV | /** Returns true if first KV has been found. */
protected boolean hasFoundKV() {
return this.foundKV;
} | 3.68 |
framework_UIConnector_loadTheme | /**
* Loads the given theme and replaces the given link element with the new
* theme link element.
*
* @param newTheme
* The name of the new theme
* @param newThemeUrl
* The url of the new theme
* @param tagToReplace
* The link element to replace. If null, then the new link
* ... | 3.68 |
framework_TabSheet_setTabPosition | /**
* Sets the position of the tab.
*
* @param tab
* The tab
* @param position
* The new position of the tab
*/
public void setTabPosition(Tab tab, int position) {
int oldPosition = getTabPosition(tab);
components.remove(oldPosition);
components.add(position, tab.getComponent())... | 3.68 |
framework_VPopupTimeCalendar_makeDate | /**
* Creates a date based on the provided date values map.
*
* @param dateValues
* a map with date values to convert into a date
* @return the date based on the dateValues map
*/
@SuppressWarnings("deprecation")
public static Date makeDate(Map<DateTimeResolution, Integer> dateValues) {
if (dateVal... | 3.68 |
framework_FileDropTarget_getParent | /**
* Returns the component this extension is attached to.
*
* @return Extended component.
*/
@Override
public T getParent() {
return super.getParent();
} | 3.68 |
framework_VComboBox_getItemOffsetHeight | /*
* Gets the height of one menu item.
*/
int getItemOffsetHeight() {
List<MenuItem> items = getItems();
return items != null && !items.isEmpty()
? items.get(0).getOffsetHeight()
: 0;
} | 3.68 |
flink_Task_transitionState | /**
* Try to transition the execution state from the current state to the new state.
*
* @param currentState of the execution
* @param newState of the execution
* @param cause of the transition change or null
* @return true if the transition was successful, otherwise false
*/
private boolean transitionState(
... | 3.68 |
zxing_PDF417_getNumberOfPadCodewords | /**
* Calculates the number of pad codewords as described in 4.9.2 of ISO/IEC 15438:2001(E).
*
* @param m the number of source codewords prior to the additional of the Symbol Length
* Descriptor and any pad codewords
* @param k the number of error correction codewords
* @param c the number of columns in ... | 3.68 |
flink_JobVertexBackPressureHandler_getBackPressureLevel | /**
* Returns the back pressure level as a String.
*
* @param backPressureRatio Ratio of back pressures samples to total number of samples.
* @return Back pressure level ('ok', 'low', or 'high')
*/
private static JobVertexBackPressureInfo.VertexBackPressureLevel getBackPressureLevel(
double backPressureRat... | 3.68 |
flink_TwoPhaseCommitSinkFunction_recoverAndCommitInternal | /**
* This method must be the only place to call {@link #recoverAndCommit(Object)} to ensure that
* the configuration parameters {@link #transactionTimeout} and {@link
* #ignoreFailuresAfterTransactionTimeout} are respected.
*/
private void recoverAndCommitInternal(TransactionHolder<TXN> transactionHolder) {
tr... | 3.68 |
hadoop_ListResultEntrySchema_name | /**
* Get the name value.
*
* @return the name value
*/
public String name() {
return name;
} | 3.68 |
framework_SQLContainer_commit | /**
* Commits all the changes, additions and removals made to the items of this
* container.
*
* @throws UnsupportedOperationException
* @throws SQLException
*/
public void commit() throws UnsupportedOperationException, SQLException {
try {
getLogger().log(Level.FINER,
"Commiting chang... | 3.68 |
flink_BlobServer_moveTempFileToStore | /**
* Moves the temporary <tt>incomingFile</tt> to its permanent location where it is available for
* use.
*
* @param incomingFile temporary file created during transfer
* @param jobId ID of the job this blob belongs to or <tt>null</tt> if job-unrelated
* @param digest BLOB content digest, i.e. hash
* @param blo... | 3.68 |
rocketmq-connect_RocketMQScheduledReporter_reportCounters | /**
* report counters
*
* @param counters
*/
private void reportCounters(SortedMap<MetricName, Long> counters) {
counters.forEach((name, value) -> {
send(name, Double.parseDouble(value.toString()));
});
} | 3.68 |
hudi_HoodieRealtimeRecordReaderUtils_avroToArrayWritable | /**
* Convert the projected read from delta record into an array writable.
*/
public static Writable avroToArrayWritable(Object value, Schema schema) {
return avroToArrayWritable(value, schema, false);
} | 3.68 |
hudi_CompactionUtils_getEarliestInstantToRetainForCompaction | /**
* Gets the earliest instant to retain for MOR compaction.
* If there is no completed compaction,
* num delta commits >= "hoodie.compact.inline.max.delta.commits"
* If there is a completed compaction,
* num delta commits after latest completed compaction >= "hoodie.compact.inline.max.delta.commits"
*
* @param... | 3.68 |
hadoop_BatchedRequests_getIteratorType | /**
* Get Iterator type.
* @return Iterator type.
*/
public IteratorType getIteratorType() {
return iteratorType;
} | 3.68 |
framework_ColorPickerSelect_createAllColors | /**
* Creates the all colors.
*
* @param rows
* the rows
* @param columns
* the columns
*
* @return the color[][]
*/
private Color[][] createAllColors(int rows, int columns) {
Color[][] colors = new Color[rows][columns];
for (int row = 0; row < rows; row++) {
for (int col... | 3.68 |
pulsar_ConsumerConfiguration_setSubscriptionType | /**
* Select the subscription type to be used when subscribing to the topic.
* <p>
* Default is {@link SubscriptionType#Exclusive}
*
* @param subscriptionType
* the subscription type value
*/
public ConsumerConfiguration setSubscriptionType(SubscriptionType subscriptionType) {
Objects.requireNonNu... | 3.68 |
hudi_CleanPlanner_getLastCompletedCommitTimestamp | /**
* Returns the last completed commit timestamp before clean.
*/
public String getLastCompletedCommitTimestamp() {
if (commitTimeline.lastInstant().isPresent()) {
return commitTimeline.lastInstant().get().getTimestamp();
} else {
return "";
}
} | 3.68 |
hadoop_ManifestSuccessData_dumpDiagnostics | /**
* Dump the diagnostics (if any) to a string.
* @param prefix prefix before every entry
* @param middle string between key and value
* @param suffix suffix to each entry
* @return the dumped string
*/
public String dumpDiagnostics(String prefix, String middle, String suffix) {
return joinMap(diagnostics, pre... | 3.68 |
hudi_BaseHoodieClient_close | /**
* Releases any resources used by the client.
*/
@Override
public void close() {
stopEmbeddedServerView(true);
this.context.setJobStatus("", "");
this.heartbeatClient.close();
this.txnManager.close();
} | 3.68 |
graphhopper_DijkstraOneToMany_getMemoryUsageAsString | /**
* List currently used memory in MB (approximately)
*/
public String getMemoryUsageAsString() {
long len = weights.length;
return ((8L + 4L + 4L) * len
+ changedNodes.getCapacity() * 4L
+ heap.getCapacity() * (4L + 4L)) / Helper.MB
+ "MB";
} | 3.68 |
hadoop_LoggingAuditor_prepareActiveContext | /**
* Get/Prepare the active context for a span.
* @return the common audit context.
*/
private CommonAuditContext prepareActiveContext() {
return currentAuditContext();
} | 3.68 |
framework_UIDL_getIntVariable | /**
* Gets the value of the named variable.
*
* @param name
* the name of the variable
* @return the value of the variable
*/
public int getIntVariable(String name) {
return var().getInt(name);
} | 3.68 |
hmily_SpringCloudHmilyOrderApplication_main | /**
* main.
*
* @param args args
*/
public static void main(final String[] args) {
SpringApplication.run(SpringCloudHmilyOrderApplication.class, args);
} | 3.68 |
hadoop_GetContentSummaryOperation_probePathStatusOrNull | /**
* Get the status of a path, downgrading FNFE to null result.
* @param p path to probe.
* @param probes probes to exec
* @return the status or null
* @throws IOException failure other than FileNotFound
*/
private S3AFileStatus probePathStatusOrNull(final Path p,
final Set<StatusProbeEnum> probes) throws IO... | 3.68 |
hbase_AsyncTableRegionLocator_getEndKeys | /**
* Gets the ending row key for every region in the currently open table.
* <p>
* This is mainly useful for the MapReduce integration.
* @return Array of region ending row keys
*/
default CompletableFuture<List<byte[]>> getEndKeys() {
return getStartEndKeys().thenApply(
startEndKeys -> startEndKeys.stream(... | 3.68 |
hadoop_AHSLogsPage_preHead | /*
* (non-Javadoc)
*
* @see
* org.apache.hadoop.yarn.server.applicationhistoryservice.webapp.AHSView#
* preHead(org.apache.hadoop .yarn.webapp.hamlet.Hamlet.HTML)
*/
@Override
protected void preHead(Page.HTML<__> html) {
String logEntity = $(ENTITY_STRING);
if (logEntity == null || logEntity.isEmpty()) {
... | 3.68 |
hbase_CommonFSUtils_checkShortCircuitReadBufferSize | /**
* Check if short circuit read buffer size is set and if not, set it to hbase value.
* @param conf must not be null
*/
public static void checkShortCircuitReadBufferSize(final Configuration conf) {
final int defaultSize = HConstants.DEFAULT_BLOCKSIZE * 2;
final int notSet = -1;
// DFSConfigKeys.DFS_CLIENT_R... | 3.68 |
hudi_EmbeddedTimelineService_stopForBasePath | /**
* Stops the embedded timeline service for the given base path. If a timeline service is managing multiple tables, it will only be shutdown once all tables have been stopped.
* @param basePath For the table to stop the service for
*/
public void stopForBasePath(String basePath) {
synchronized (SERVICE_LOCK) {
... | 3.68 |
dubbo_CodecSupport_getNullBytesOf | /**
* Get the null object serialize result byte[] of Serialization from the cache,
* if not, generate it first.
*
* @param s Serialization Instances
* @return serialize result of null object
*/
public static byte[] getNullBytesOf(Serialization s) {
return ConcurrentHashMapUtils.computeIfAbsent(ID_NULLBYTES_MA... | 3.68 |
hbase_HBaseServerException_setServerOverloaded | /**
* Necessary for parsing RemoteException on client side
* @param serverOverloaded True if server was overloaded when exception was thrown
*/
public void setServerOverloaded(boolean serverOverloaded) {
this.serverOverloaded = serverOverloaded;
} | 3.68 |
framework_Html5File_setStreamVariable | /**
* Sets the {@link StreamVariable} that into which the file contents will be
* written. Usage of StreamVariable is similar to {@link Upload} component.
* <p>
* If the {@link StreamVariable} is not set in the {@link DropHandler} the
* file contents will not be sent to server.
* <p>
* <em>Note!</em> receiving f... | 3.68 |
framework_DropTargetExtensionConnector_setDropEffect | /**
* Set the drop effect for the dragenter / dragover event, if one has been
* set from server side.
* <p>
* From Moz Foundation: "You can modify the dropEffect property during the
* dragenter or dragover events, if for example, a particular drop target
* only supports certain operations. You can modify the drop... | 3.68 |
flink_TumbleWithSize_on | /**
* Specifies the time attribute on which rows are grouped.
*
* <p>For streaming tables you can specify grouping by a event-time or processing-time
* attribute.
*
* <p>For batch tables you can specify grouping on a timestamp or long attribute.
*
* @param timeField time attribute for streaming and batch tables... | 3.68 |
hbase_HFileReaderImpl_positionForNextBlock | /**
* Set our selves up for the next 'next' invocation, set up next block.
* @return True is more to read else false if at the end.
*/
private boolean positionForNextBlock() throws IOException {
// Methods are small so they get inlined because they are 'hot'.
long lastDataBlockOffset = reader.getTrailer().getLas... | 3.68 |
hudi_SparkRDDReadClient_addHoodieSupport | /**
* Adds support for accessing Hoodie built tables from SparkSQL, as you normally would.
*
* @return SparkConf object to be used to construct the SparkContext by caller
*/
public static SparkConf addHoodieSupport(SparkConf conf) {
conf.set("spark.sql.hive.convertMetastoreParquet", "false");
return conf;
} | 3.68 |
flink_IterativeStream_closeWith | /**
* Closes the iteration. This method defines the end of the iterative program part that will
* be fed back to the start of the iteration as the second input in the {@link
* ConnectedStreams}.
*
* @param feedbackStream {@link DataStream} that will be used as second input to the
* iteration head.
* @return ... | 3.68 |
flink_HiveParserSemanticAnalyzer_processLateralView | /**
* Given the AST with TOK_LATERAL_VIEW as the root, get the alias for the table or subquery in
* the lateral view and also make a mapping from the alias to all the lateral view AST's.
*/
private String processLateralView(HiveParserQB qb, HiveParserASTNode lateralView)
throws SemanticException {
int nu... | 3.68 |
hbase_HFileArchiveUtil_getTableName | /*
* @return table name given archive file path
*/
public static TableName getTableName(Path archivePath) {
Path p = archivePath;
String tbl = null;
// namespace is the 4th parent of file
for (int i = 0; i < 5; i++) {
if (p == null) return null;
if (i == 3) tbl = p.getName();
p = p.getParent();
... | 3.68 |
framework_VUI_windowSizeMaybeChanged | /**
* Called when the window or parent div might have been resized.
*
* This immediately checks the sizes of the window and the parent div (if
* monitoring it) and triggers layout recalculation if they have changed.
*
* @param newWindowWidth
* The new width of the window
* @param newWindowHeight
* ... | 3.68 |
querydsl_MapExpressionBase_containsKey | /**
* Create a {@code key in keys(this)} expression
*
* @param key key
* @return expression
*/
public final BooleanExpression containsKey(K key) {
return Expressions.booleanOperation(Ops.CONTAINS_KEY, mixin, ConstantImpl.create(key));
} | 3.68 |
pulsar_PerfClientUtils_printJVMInformation | /**
* Print useful JVM information, you need this information in order to be able
* to compare the results of executions in different environments.
* @param log
*/
public static void printJVMInformation(Logger log) {
log.info("JVM args {}", ManagementFactory.getRuntimeMXBean().getInputArguments());
log.info... | 3.68 |
morf_TableLoader_truncate | /**
* Empties the specified table.
*
* @param table
*/
private void truncate(Table table) {
// Get our own table definition based on the name to avoid case sensitivity bugs on some database vendors
log.debug("Clearing table [" + table.getName() + "]");
// Try to use a truncate
try {
runRecoverably(() -... | 3.68 |
zxing_PDF417Common_getBitCountSum | /**
* @param moduleBitCount values to sum
* @return sum of values
* @deprecated call {@link MathUtils#sum(int[])}
*/
@Deprecated
public static int getBitCountSum(int[] moduleBitCount) {
return MathUtils.sum(moduleBitCount);
} | 3.68 |
hbase_RegionServerSpaceQuotaManager_getActiveEnforcements | /**
* Creates an object well-suited for the RegionServer to use in verifying active policies.
*/
public ActivePolicyEnforcement getActiveEnforcements() {
return new ActivePolicyEnforcement(copyActiveEnforcements(), copyQuotaSnapshots(), rsServices);
} | 3.68 |
hadoop_StageConfig_withTaskAttemptDir | /**
* Set the Task attempt directory.
* @param value new value
* @return this
*/
public StageConfig withTaskAttemptDir(final Path value) {
checkOpen();
taskAttemptDir = value;
return this;
} | 3.68 |
flink_HardwareDescription_getSizeOfManagedMemory | /**
* Returns the size of the memory managed by the system for caching, hashing, sorting, ...
*
* @return The size of the memory managed by the system.
*/
public long getSizeOfManagedMemory() {
return this.sizeOfManagedMemory;
} | 3.68 |
framework_PropertysetItem_getItemPropertyIds | /**
* Gets the collection of IDs of all Properties stored in the Item.
*
* @return unmodifiable collection containing IDs of the Properties stored
* the Item
*/
@Override
public Collection<?> getItemPropertyIds() {
return Collections.unmodifiableCollection(list);
} | 3.68 |
hudi_AbstractStreamWriteFunction_instantToWrite | /**
* Prepares the instant time to write with for next checkpoint.
*
* @param hasData Whether the task has buffering data
* @return The instant time
*/
protected String instantToWrite(boolean hasData) {
String instant = lastPendingInstant();
// if exactly-once semantics turns on,
// waits for the checkpoint ... | 3.68 |
flink_GSCommitRecoverableSerializer_getVersion | /**
* The serializer version. Note that, if this changes, then the version of {@link
* GSResumeRecoverableSerializer} must also change, because it uses this class to serialize
* itself, in part.
*
* @return The serializer version.
*/
@Override
public int getVersion() {
return SERIALIZER_VERSION;
} | 3.68 |
hbase_MetricsSource_setWALReaderEditsBufferUsage | /**
* Sets the amount of memory in bytes used in this RegionServer by edits pending replication.
*/
public void setWALReaderEditsBufferUsage(long usageInBytes) {
globalSourceSource.setWALReaderEditsBufferBytes(usageInBytes);
} | 3.68 |
hadoop_SinglePendingCommit_getUploadId | /** @return ID of the upload. */
public String getUploadId() {
return uploadId;
} | 3.68 |
hudi_HoodieDataBlock_getRecordIterator | /**
* Batch get of keys of interest. Implementation can choose to either do full scan and return matched entries or
* do a seek based parsing and return matched entries.
*
* @param keys keys of interest.
* @return List of IndexedRecords for the keys of interest.
* @throws IOException in case of failures encounter... | 3.68 |
morf_IdTableTracker_activeNameFor | /**
* @see org.alfasoftware.morf.upgrade.TableNameResolver#activeNameFor(java.lang.String)
*/
@Override
public String activeNameFor(String tableName) {
if ("IDTABLE".equals(tableName))
return idTableName;
return tableName;
} | 3.68 |
rocketmq-connect_AbstractStateManagementService_put | /**
* Set the state of the connector to the given value.
*
* @param status the status of the task
*/
@Override
public void put(TaskStatus status) {
sendTaskStatus(status, false);
} | 3.68 |
flink_HadoopBlockLocation_stripHostname | /**
* Looks for a domain suffix in a FQDN and strips it if present.
*
* @param originalHostname the original hostname, possibly an FQDN
* @return the stripped hostname without the domain suffix
*/
private static String stripHostname(final String originalHostname) {
// Check if the hostname domains the domain ... | 3.68 |
querydsl_GenericExporter_setEntityAnnotation | /**
* Set the entity annotation
*
* @param entityAnnotation entity annotation
*/
public void setEntityAnnotation(Class<? extends Annotation> entityAnnotation) {
this.entityAnnotation = entityAnnotation;
} | 3.68 |
pulsar_AdditionalServlets_load | /**
* Load the additional servlet for the given <tt>servlet name</tt> list.
*
* @param conf the pulsar service configuration
* @return the collection of additional servlet
*/
public static AdditionalServlets load(PulsarConfiguration conf) throws IOException {
String additionalServletDirectory = conf.getPropert... | 3.68 |
hbase_KeyValue_getKeyDataStructureSize | /**
* Computes the number of bytes that a <code>KeyValue</code> instance with the provided
* characteristics would take up in its underlying data structure for the key.
* @param rlength row length
* @param flength family length
* @param qlength qualifier length
* @return the key data structure length
*/
public s... | 3.68 |
querydsl_SQLExpressions_addMonths | /**
* Add the given amount of months to the date
*
* @param date date
* @param months months to add
* @return converted date
*/
public static <D extends Comparable> DateExpression<D> addMonths(DateExpression<D> date, int months) {
return Expressions.dateOperation(date.getType(), Ops.DateTimeOps.ADD_MONTHS, da... | 3.68 |
hbase_ServerRegionReplicaUtil_isRegionReplicaWaitForPrimaryFlushEnabled | /** Returns True if wait for primary to flush is enabled for user-space tables. */
public static boolean isRegionReplicaWaitForPrimaryFlushEnabled(Configuration conf) {
return conf.getBoolean(REGION_REPLICA_WAIT_FOR_PRIMARY_FLUSH_CONF_KEY,
DEFAULT_REGION_REPLICA_WAIT_FOR_PRIMARY_FLUSH);
} | 3.68 |
hadoop_ExecutorServiceFuturePool_executeFunction | /**
* @param f function to run in future on executor pool
* @return future
* @throws java.util.concurrent.RejectedExecutionException can be thrown
* @throws NullPointerException if f param is null
*/
public Future<Void> executeFunction(final Supplier<Void> f) {
return executor.submit(f::get);
} | 3.68 |
flink_FreeingBufferRecycler_recycle | /**
* Frees the given memory segment.
*
* @param memorySegment The memory segment to be recycled.
*/
@Override
public void recycle(MemorySegment memorySegment) {
memorySegment.free();
} | 3.68 |
hadoop_IdentifierResolver_setInputWriterClass | /**
* Sets the {@link InputWriter} class.
*/
protected void setInputWriterClass(Class<? extends InputWriter>
inputWriterClass) {
this.inputWriterClass = inputWriterClass;
} | 3.68 |
morf_SqlDialect_getSqlForRowNumber | /**
* Produce SQL for getting the row number of the row in the partition
*
* @return a string representation of the SQL for finding the last day of the month.
*/
protected String getSqlForRowNumber(){
return "ROW_NUMBER()";
} | 3.68 |
hadoop_BlockStorageMovementNeeded_size | /**
* Returns queue size.
*/
public synchronized int size() {
return storageMovementNeeded.size();
} | 3.68 |
framework_RpcDataProviderExtension_beforeClientResponse | /**
* {@inheritDoc}
* <p>
* RpcDataProviderExtension makes all actual RPC calls from this function
* based on changes in the container.
*/
@Override
public void beforeClientResponse(boolean initial) {
if (initial || bareItemSetTriggeredSizeChange) {
/*
* Push initial set of rows, assuming Grid... | 3.68 |
morf_SqlDialect_getDeleteLimitSuffix | /**
* Returns the SQL that specifies the deletion limit as a suffix, if any, for the dialect.
*
* @param limit The delete limit.
* @return The SQL fragment.
*/
protected Optional<String> getDeleteLimitSuffix(@SuppressWarnings("unused") int limit) {
return Optional.empty();
} | 3.68 |
streampipes_UnicodeTokenizer_tokenize | /**
* Tokenizes the text and returns an array of tokens.
*
* @param text The text
* @return The tokens
*/
public static String[] tokenize(final CharSequence text) {
return PAT_NOT_WORD_BOUNDARY.matcher(PAT_WORD_BOUNDARY.matcher(text).replaceAll("\u2063"))
.replaceAll("$1").replaceAll("[ \u2063]+", " ").tri... | 3.68 |
framework_Calendar_setTimeZone | /**
* Set time zone that this component will use. Null value sets the default
* time zone.
*
* @param zone
* Time zone to use
*/
public void setTimeZone(TimeZone zone) {
timezone = zone;
if (!currentCalendar.getTimeZone().equals(zone)) {
if (zone == null) {
zone = TimeZone.g... | 3.68 |
hbase_MutableRegionInfo_getRegionId | /** Returns the regionId */
@Override
public long getRegionId() {
return regionId;
} | 3.68 |
hadoop_CopyOutputFormat_getOutputCommitter | /** {@inheritDoc} */
@Override
public OutputCommitter getOutputCommitter(TaskAttemptContext context) throws IOException {
return new CopyCommitter(getOutputPath(context), context);
} | 3.68 |
morf_ExceptSetOperator_accept | /**
* @see org.alfasoftware.morf.sql.SchemaAndDataChangeVisitable#accept(org.alfasoftware.morf.upgrade.SchemaAndDataChangeVisitor)
*/
@Override
public void accept(SchemaAndDataChangeVisitor visitor) {
visitor.visit(this);
selectStatement.accept(visitor);
} | 3.68 |
hudi_HoodieAvroIndexedRecord_readRecordPayload | /**
* NOTE: This method is declared final to make sure there's no polymorphism and therefore
* JIT compiler could perform more aggressive optimizations
*/
@SuppressWarnings("unchecked")
@Override
protected final IndexedRecord readRecordPayload(Kryo kryo, Input input) {
// NOTE: We're leveraging Spark's defau... | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.