name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
morf_SqlDialect_getSqlForWindowFunction | /**
* Generates standards compliant SQL from the function within a window function
* @param function The field to convert
* @return The resulting SQL
*/
protected String getSqlForWindowFunction(Function function) {
return getSqlFrom(function);
} | 3.68 |
framework_GridSingleSelect_getSelectedItem | /**
* Returns the currently selected item, or an empty optional if no item is
* selected.
*
* @return an optional of the selected item if any, an empty optional
* otherwise
*/
public Optional<T> getSelectedItem() {
return model.getSelectedItem();
} | 3.68 |
framework_GridDragSource_addGridDragStartListener | /**
* Attaches dragstart listener for the current drag source grid.
*
* @param listener
* Listener to handle the dragstart event.
* @return Handle to be used to remove this listener.
* @see GridDragStartEvent
*/
public Registration addGridDragStartListener(
GridDragStartListener<T> listener) {... | 3.68 |
hadoop_S3APrefetchingInputStream_seek | /**
* Updates internal data such that the next read will take place at the given {@code pos}.
*
* @param pos new read position.
* @throws IOException if there is an IO error during this operation.
*/
@Override
public synchronized void seek(long pos) throws IOException {
throwIfClosed();
inputStream.seek(pos);
... | 3.68 |
hbase_MetricsSource_getReplicableEdits | /**
* Gets the number of edits eligible for replication read from this source queue logs so far.
* @return replicableEdits total number of replicable edits read from this queue logs.
*/
public long getReplicableEdits() {
return this.singleSourceSource.getWALEditsRead() - this.singleSourceSource.getEditsFiltered();... | 3.68 |
rocketmq-connect_RetryWithToleranceOperator_execute | /**
* Execute the recoverable operation. If the operation is already in a failed state, then simply return
* with the existing failure.
*/
public <V> V execute(Operation<V> operation, ErrorReporter.Stage stage, Class<?> executingClass) {
context.currentContext(stage, executingClass);
if (context.failed()) {
... | 3.68 |
morf_H2Dialect_getSqlForWindowFunction | /**
* @see org.alfasoftware.morf.jdbc.SqlDialect#getSqlForWindowFunction(Function)
*/
@Override
protected String getSqlForWindowFunction(Function function) {
FunctionType functionType = function.getType();
switch (functionType) {
case ROW_NUMBER:
return "ROW_NUMBER()";
default:
return super.... | 3.68 |
hbase_SaslServerAuthenticationProviders_reset | /**
* Removes the cached singleton instance of {@link SaslServerAuthenticationProviders}.
*/
public static void reset() {
synchronized (holder) {
holder.set(null);
}
} | 3.68 |
pulsar_HeapDumpUtil_getHotSpotDiagnosticMXBean | // Utility method to get the HotSpotDiagnosticMXBean
private static HotSpotDiagnosticMXBean getHotSpotDiagnosticMXBean() {
try {
MBeanServer server = ManagementFactory.getPlatformMBeanServer();
return ManagementFactory.newPlatformMXBeanProxy(server, HOTSPOT_BEAN_NAME, HotSpotDiagnosticMXBean.class);... | 3.68 |
graphhopper_ArrayUtil_permutation | /**
* Creates an IntArrayList filled with a permutation of the numbers 0,1,2,...,size-1
*/
public static IntArrayList permutation(int size, Random rnd) {
IntArrayList result = iota(size);
shuffle(result, rnd);
return result;
} | 3.68 |
flink_SkipListUtils_putValueData | /**
* Puts the value data into value space.
*
* @param memorySegment memory segment for value space.
* @param offset offset of value space in memory segment.
* @param value value data.
*/
public static void putValueData(MemorySegment memorySegment, int offset, byte[] value) {
MemorySegment valueSegment = Memo... | 3.68 |
hbase_OrderedBlobVar_encode | /**
* Write a subset of {@code val} to {@code dst}.
* @param dst the {@link PositionedByteRange} to write to
* @param val the value to write to {@code dst}
* @param voff the offset in {@code dst} where to write {@code val} to
* @param vlen the lenght of {@code val}
* @return the number of bytes written
*/
publ... | 3.68 |
morf_SqlDialect_buildSQLToStartTracing | /**
* @param identifier Unique identifier for trace file name, can be null.
* @return Sql required to turn on tracing, or null if tracing is not
* supported.
*/
public List<String> buildSQLToStartTracing(@SuppressWarnings("unused") String identifier) {
return null;
} | 3.68 |
dubbo_DubboBootstrap_initialize | /**
* Initialize
*/
public void initialize() {
applicationDeployer.initialize();
} | 3.68 |
hbase_FileIOEngine_shutdown | /**
* Close the file
*/
@Override
public void shutdown() {
for (int i = 0; i < filePaths.length; i++) {
try {
if (fileChannels[i] != null) {
fileChannels[i].close();
}
if (rafs[i] != null) {
rafs[i].close();
}
} catch (IOException ex) {
LOG.error("Failed closing... | 3.68 |
framework_VaadinService_ensureAccessQueuePurged | /**
* Makes sure the pending access queue is purged for the provided session.
* If the session is currently locked by the current thread or some other
* thread, the queue will be purged when the session is unlocked. If the
* lock is not held by any thread, it is acquired and the queue is purged
* right away.
*
*... | 3.68 |
hadoop_HdfsFileStatus_owner | /**
* Set the owner for this entity (default = null).
* @param owner Owner
* @return This Builder instance
*/
public Builder owner(String owner) {
this.owner = owner;
return this;
} | 3.68 |
framework_ComplexRenderer_destroy | /**
* Called when the renderer is deemed to be destroyed and no longer used by
* the Grid.
*/
public void destroy() {
// Implement if needed
} | 3.68 |
flink_SqlJsonValueFunctionWrapper_explicitTypeSpec | /**
* Copied and modified from the original {@link SqlJsonValueFunction}.
*
* <p>Changes: Instead of returning {@link Optional} this method returns null directly.
*/
private static RelDataType explicitTypeSpec(SqlOperatorBinding opBinding) {
if (opBinding.getOperandCount() > 2
&& opBinding.isOperand... | 3.68 |
framework_ReflectTools_checkClassAccessibility | /**
* Makes a check whether the provided class is externally accessible for
* instantiation (e.g. it's not inner class (nested and not static) and is
* not a local class).
*
* @param cls
* type to check
*/
private static void checkClassAccessibility(Class<?> cls) {
if (cls.isMemberClass() && !Modi... | 3.68 |
flink_HiveConfUtils_create | /**
* Create HiveConf instance via Hadoop configuration. Since {@link
* HiveConf#HiveConf(org.apache.hadoop.conf.Configuration, java.lang.Class)} will override
* properties in Hadoop configuration with Hive default values ({@link org.apache
* .hadoop.hive.conf.HiveConf.ConfVars}), so we should use this method to cr... | 3.68 |
hudi_HoodieLogFileReader_readBlock | // TODO : convert content and block length to long by using ByteBuffer, raw byte [] allows
// for max of Integer size
private HoodieLogBlock readBlock() throws IOException {
int blockSize;
long blockStartPos = inputStream.getPos();
try {
// 1 Read the total size of the block
blockSize = (int) inputStream.... | 3.68 |
pulsar_ProducerConfiguration_setMaxPendingMessagesAcrossPartitions | /**
* Set the number of max pending messages across all the partitions
* <p>
* This setting will be used to lower the max pending messages for each partition
* ({@link #setMaxPendingMessages(int)}), if the total exceeds the configured value.
*
* @param maxPendingMessagesAcrossPartitions
*/
public void setMaxPend... | 3.68 |
framework_VAccordion_getWidgetWidth | /**
* Returns the offset width of the wrapped widget.
*
* @return the offset width in pixels, or zero if no widget is set
*/
public int getWidgetWidth() {
if (widget == null) {
return 0;
}
return widget.getOffsetWidth();
} | 3.68 |
hbase_RegionReplicationSink_replicated | /**
* Should be called regardless of the result of the replicating operation. Unless you still want
* to reuse this entry, otherwise you must call this method to release the possible off heap
* memories.
*/
void replicated() {
if (rpcCall != null) {
rpcCall.releaseByWAL();
}
} | 3.68 |
hadoop_AsyncDataService_execute | /**
* Execute the task sometime in the future.
*/
synchronized void execute(Runnable task) {
if (executor == null) {
throw new RuntimeException("AsyncDataService is already shutdown");
}
if (LOG.isDebugEnabled()) {
LOG.debug("Current active thread number: " + executor.getActiveCount()
+ " queue ... | 3.68 |
hadoop_UriUtils_containsAbfsUrl | /**
* Checks whether a string includes abfs url.
* @param string the string to check.
* @return true if string has abfs url.
*/
public static boolean containsAbfsUrl(final String string) {
if (string == null || string.isEmpty()) {
return false;
}
return ABFS_URI_PATTERN.matcher(string).matches();
} | 3.68 |
graphhopper_FindMinMax_findMinMax | /**
* This method returns the smallest value possible in "min" and the smallest value that cannot be
* exceeded by any edge in max.
*/
static MinMax findMinMax(Set<String> createdObjects, MinMax minMax, List<Statement> statements, EncodedValueLookup lookup) {
// 'blocks' of the statements are applied one after t... | 3.68 |
Activiti_DefaultDeploymentCache_size | // For testing purposes only
public int size() {
return cache.size();
} | 3.68 |
flink_OneInputStateTransformation_keyBy | /**
* Partitions the operator state of a {@link OperatorTransformation} using field expressions. A
* field expression is either the name of a public field or a getter method with parentheses of
* the {@code OperatorTransformation}'s underlying type. A dot can be used to drill down into
* objects, as in {@code "fiel... | 3.68 |
hadoop_KerberosAuthException_getKeytabFile | /** @return The keytab file path, or null if not set. */
public String getKeytabFile() {
return keytabFile;
} | 3.68 |
flink_FileCache_shutdown | /** Shuts down the file cache by cancelling all. */
public void shutdown() {
synchronized (lock) {
// first shutdown the thread pool
ScheduledExecutorService es = this.executorService;
if (es != null) {
es.shutdown();
try {
es.awaitTermination(cleanupI... | 3.68 |
hbase_ServerName_parseVersionedServerName | /**
* Use this method instantiating a {@link ServerName} from bytes gotten from a call to
* {@link #getVersionedBytes()}. Will take care of the case where bytes were written by an earlier
* version of hbase.
* @param versionedBytes Pass bytes gotten from a call to {@link #getVersionedBytes()}
* @return A ServerNam... | 3.68 |
flink_ExtractionUtils_isStructuredFieldDirectlyWritable | /** Checks whether a field is directly writable without a setter or constructor. */
public static boolean isStructuredFieldDirectlyWritable(Field field) {
final int m = field.getModifiers();
// field is immutable
if (Modifier.isFinal(m)) {
return false;
}
// field is directly writable
... | 3.68 |
hbase_PrivateCellUtil_createFirstOnRowCol | /**
* Create a Cell that is smaller than all other possible Cells for the given Cell's rk:cf and
* passed qualifier.
* @return Last possible Cell on passed Cell's rk:cf and passed qualifier.
*/
public static Cell createFirstOnRowCol(final Cell cell, byte[] qArray, int qoffest, int qlength) {
if (cell instanceof B... | 3.68 |
graphhopper_MatrixResponse_getDistance | /**
* Returns the distance for the specific entry (from -> to) in meter or {@link Double#MAX_VALUE} in case
* no connection was found (and {@link GHMRequest#setFailFast(boolean)} was set to true).
*/
public double getDistance(int from, int to) {
if (hasErrors()) {
throw new IllegalStateException("Cann... | 3.68 |
framework_VTwinColSelect_getCaptionWrapper | /**
* For internal use only. May be removed or replaced in the future.
*
* @return the caption wrapper widget
*/
public Widget getCaptionWrapper() {
return captionWrapper;
} | 3.68 |
framework_Upload_addProgressListener | /**
* Adds the upload progress event listener.
*
* @param listener
* the progress listener to be added
* @since 8.0
*/
public Registration addProgressListener(ProgressListener listener) {
Objects.requireNonNull(listener, "Listener must not be null.");
if (progressListeners == null) {
pr... | 3.68 |
hbase_TableInputFormatBase_getTable | /**
* Allows subclasses to get the {@link Table}.
*/
protected Table getTable() {
if (table == null) {
throw new IllegalStateException(NOT_INITIALIZED);
}
return this.table;
} | 3.68 |
flink_SerializedCompositeKeyBuilder_buildCompositeKeyUserKey | /**
* Returns a serialized composite key, from the key and key-group provided in a previous call to
* {@link #setKeyAndKeyGroup(Object, int)} and the namespace provided in {@link
* #setNamespace(Object, TypeSerializer)}, followed by the given user-key.
*
* @param userKey the user-key to concatenate for the seriali... | 3.68 |
hadoop_FederationRegistryClient_loadStateFromRegistry | /**
* Load the information of one application from registry.
*
* @param appId application id
* @return the sub-cluster to UAM token mapping
*/
public synchronized Map<String, Token<AMRMTokenIdentifier>>
loadStateFromRegistry(ApplicationId appId) {
Map<String, Token<AMRMTokenIdentifier>> retMap = new HashMap<... | 3.68 |
flink_HiveParserSemanticAnalyzer_addCTEAsSubQuery | /*
* If a CTE is referenced in a QueryBlock:
* - add it as a SubQuery for now.
* - SQ.alias is the alias used in HiveParserQB. (if no alias is specified,
* it used the CTE name. Works just like table references)
* - Adding SQ done by:
* - copying AST of CTE
* - setting ASTOrigin on cloned AST.
*... | 3.68 |
hadoop_WordMedian_map | /**
* Emits a key-value pair for counting the word. Outputs are (IntWritable,
* IntWritable).
*
* @param value
* This will be a line of text coming in from our input file.
*/
public void map(Object key, Text value, Context context)
throws IOException, InterruptedException {
StringTokenizer itr = n... | 3.68 |
flink_StreamingJobGraphGenerator_buildVertexRegionSlotSharingGroups | /**
* Maps a vertex to its region slot sharing group. If {@link
* StreamGraph#isAllVerticesInSameSlotSharingGroupByDefault()} returns true, all regions will be
* in the same slot sharing group.
*/
private Map<JobVertexID, SlotSharingGroup> buildVertexRegionSlotSharingGroups() {
final Map<JobVertexID, SlotSharin... | 3.68 |
flink_NettyShuffleEnvironmentConfiguration_fromConfiguration | /**
* Utility method to extract network related parameters from the configuration and to sanity
* check them.
*
* @param configuration configuration object
* @param networkMemorySize the size of memory reserved for shuffle environment
* @param localTaskManagerCommunication true, to skip initializing the network s... | 3.68 |
flink_BufferBuilder_commit | /**
* Make the change visible to the readers. This is costly operation (volatile access) thus in
* case of bulk writes it's better to commit them all together instead one by one.
*/
public void commit() {
positionMarker.commit();
} | 3.68 |
framework_AutomaticImmediate_fireValueChange | /*
* (non-Javadoc)
*
* @see com.vaadin.ui.AbstractField#fireValueChange(boolean)
*/
@Override
protected void fireValueChange(boolean repaintIsNotNeeded) {
log("fireValueChange");
super.fireValueChange(repaintIsNotNeeded);
} | 3.68 |
hbase_HFileSystem_getStoragePolicyForOldHDFSVersion | /**
* Before Hadoop 2.8.0, there's no getStoragePolicy method for FileSystem interface, and we need
* to keep compatible with it. See HADOOP-12161 for more details.
* @param path Path to get storage policy against
* @return the storage policy name
*/
private String getStoragePolicyForOldHDFSVersion(Path path) {
... | 3.68 |
hadoop_FilePosition_bufferFullyRead | /**
* Determines whether the current buffer has been fully read.
*
* @return true if the current buffer has been fully read, false otherwise.
*/
public boolean bufferFullyRead() {
throwIfInvalidBuffer();
return (bufferStartOffset == readStartOffset)
&& (relative() == buffer.limit())
&& (numBytesRead... | 3.68 |
querydsl_PathBuilder_getString | /**
* Create a new String typed path
*
* @param property property name
* @return property path
*/
public StringPath getString(String property) {
validate(property, String.class);
return super.createString(property);
} | 3.68 |
framework_GridElement_getTableWrapper | /**
* Get the element wrapping the table element.
*
* @return The element that wraps the table element
*/
public TestBenchElement getTableWrapper() {
List<WebElement> rootElements = findElements(By.xpath("./div"));
return (TestBenchElement) rootElements.get(2);
} | 3.68 |
hadoop_WriteOperationHelper_retry | /**
* Execute a function with retry processing.
* Also activates the current span.
* @param <T> type of return value
* @param action action to execute (used in error messages)
* @param path path of work (used in error messages)
* @param idempotent does the operation have semantics
* which mean that it can be ret... | 3.68 |
hbase_LeaseManager_removeLease | /**
* Remove named lease. Lease is removed from the map of leases.
* @param leaseName name of lease
* @return Removed lease
*/
Lease removeLease(final String leaseName) throws LeaseException {
Lease lease = leases.remove(leaseName);
if (lease == null) {
throw new LeaseException("lease '" + leaseName + "' do... | 3.68 |
dubbo_ScopeModelAware_setFrameworkModel | /**
* Override this method if you just need framework model
* @param frameworkModel
*/
default void setFrameworkModel(FrameworkModel frameworkModel) {} | 3.68 |
hadoop_Chain_addReducer | /**
* Add reducer that reads from context and writes to a queue
*/
@SuppressWarnings("unchecked")
void addReducer(TaskInputOutputContext inputContext,
ChainBlockingQueue<KeyValuePair<?, ?>> outputQueue) throws IOException,
InterruptedException {
Class<?> keyOutClass = rConf.getClass(REDUCER_OUTPUT_KEY_CLAS... | 3.68 |
hudi_Registry_getRegistry | /**
* Get (or create) the registry for a provided name and given class.
*
* @param registryName Name of the registry.
* @param clazz The fully qualified name of the registry class to create.
*/
static Registry getRegistry(String registryName, String clazz) {
synchronized (Registry.class) {
if (!REGISTRY_MAP.... | 3.68 |
hmily_HmilyColumnSegment_getQualifiedName | /**
* Get qualified name with quote characters.
* i.e. `field1`, `table1`, field1, table1, `table1`.`field1`, `table1`.field1, table1.`field1` or table1.field1
*
* @return qualified name with quote characters
*/
public String getQualifiedName() {
return null == owner
? identifier.getValueWithQuoteC... | 3.68 |
pulsar_ReaderConfiguration_getSubscriptionRolePrefix | /**
* @return the subscription role prefix for subscription auth
*/
public String getSubscriptionRolePrefix() {
return conf.getSubscriptionRolePrefix();
} | 3.68 |
hibernate-validator_ConstraintDescriptorImpl_getCompositionType | /**
* @return the compositionType
*/
public CompositionType getCompositionType() {
return compositionType;
} | 3.68 |
dubbo_DefaultExecutorRepository_createExecutorIfAbsent | /**
* Get called when the server or client instance initiating.
*
* @param url
* @return
*/
@Override
public synchronized ExecutorService createExecutorIfAbsent(URL url) {
String executorKey = getExecutorKey(url);
ConcurrentMap<String, ExecutorService> executors =
ConcurrentHashMapUtils.compute... | 3.68 |
morf_AbstractSqlDialectTest_testRenameIndexStatements | /**
* Tests that the syntax is correct for renaming an index.
*/
@SuppressWarnings("unchecked")
@Test
public void testRenameIndexStatements() {
AlteredTable alteredTable = new AlteredTable(testTable, null, null,
FluentIterable.from(testTable.indexes()).transform(Index::getName).filter(i -> !i.equals(TEST_1)).ap... | 3.68 |
pulsar_ManagedLedgerImpl_asyncCreateLedger | /**
* Create ledger async and schedule a timeout task to check ledger-creation is complete else it fails the callback
* with TimeoutException.
*
* @param bookKeeper
* @param config
* @param digestType
* @param cb
* @param metadata
*/
protected void asyncCreateLedger(BookKeeper bookKeeper, ManagedLedgerConfig c... | 3.68 |
flink_BlockerSync_awaitBlocker | /**
* Waits until the blocking thread has entered the method {@link #block()} or {@link
* #blockNonInterruptible()}.
*/
public void awaitBlocker() throws InterruptedException {
synchronized (lock) {
while (!blockerReady) {
lock.wait();
}
}
} | 3.68 |
graphhopper_VectorTile_addKeys | /**
* <pre>
* Dictionary encoding for keys
* </pre>
*
* <code>repeated string keys = 3;</code>
*/
public Builder addKeys(
java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
ensureKeysIsMut... | 3.68 |
hibernate-validator_StringHelper_join | /**
* Joins the elements of the given iterable to a string, separated by the given separator string.
*
* @param iterable the iterable to join
* @param separator the separator string
*
* @return a string made up of the string representations of the given iterable members, separated by the given separator
* ... | 3.68 |
hadoop_EditLogInputStream_getCachedOp | /**
* return the cachedOp, and reset it to null.
*/
FSEditLogOp getCachedOp() {
FSEditLogOp op = this.cachedOp;
cachedOp = null;
return op;
} | 3.68 |
open-banking-gateway_FintechConsentSpecSecureStorage_registerFintechUser | /**
* Registers FinTech user
* @param user User entity
* @param password Datasafe password for the user
*/
public void registerFintechUser(FintechUser user, Supplier<char[]> password) {
this.userProfile()
.createDocumentKeystore(
user.getUserIdAuth(password),
... | 3.68 |
flink_CoGroupOperatorBase_getGroupOrder | /**
* Gets the value order for an input, i.e. the order of elements within a group. If no such
* order has been set, this method returns null.
*
* @param inputNum The number of the input (here either <i>0</i> or <i>1</i>).
* @return The group order.
*/
public Ordering getGroupOrder(int inputNum) {
if (inputNu... | 3.68 |
hbase_CompactingMemStore_preUpdate | /**
* Issue any synchronization and test needed before applying the update For compacting memstore
* this means checking the update can increase the size without overflow
* @param currentActive the segment to be updated
* @param cell the cell to be added
* @param memstoreSizing object to accumulate regi... | 3.68 |
hadoop_TimelineDomains_setDomains | /**
* Set the domain list to the given list of domains
*
* @param domains
* a list of domains
*/
public void setDomains(List<TimelineDomain> domains) {
this.domains = domains;
} | 3.68 |
hadoop_SnappyCodec_getDefaultExtension | /**
* Get the default filename extension for this kind of compression.
*
* @return <code>.snappy</code>.
*/
@Override
public String getDefaultExtension() {
return CodecConstants.SNAPPY_CODEC_EXTENSION;
} | 3.68 |
flink_AsyncDataStream_addOperator | /**
* Add an AsyncWaitOperator.
*
* @param in The {@link DataStream} where the {@link AsyncWaitOperator} will be added.
* @param func {@link AsyncFunction} wrapped inside {@link AsyncWaitOperator}.
* @param timeout for the asynchronous operation to complete
* @param bufSize The max number of inputs the {@link Asy... | 3.68 |
hudi_FiveToSixUpgradeHandler_deleteCompactionRequestedFileFromAuxiliaryFolder | /**
* See HUDI-6040.
*/
private void deleteCompactionRequestedFileFromAuxiliaryFolder(HoodieTable table) {
HoodieTableMetaClient metaClient = table.getMetaClient();
HoodieTimeline compactionTimeline = metaClient.getActiveTimeline().filterPendingCompactionTimeline()
.filter(instant -> instant.getState() == H... | 3.68 |
flink_Types_GENERIC | /**
* Returns generic type information for any Java object. The serialization logic will use the
* general purpose serializer Kryo.
*
* <p>Generic types are black-boxes for Flink, but allow any object and null values in fields.
*
* <p>By default, serialization of this type is not very efficient. Please read the
... | 3.68 |
hbase_DataBlockEncoding_getEncoder | /**
* Return new data block encoder for given algorithm type.
* @return data block encoder if algorithm is specified, null if none is selected.
*/
public DataBlockEncoder getEncoder() {
if (encoder == null && id != 0) {
// lazily create the encoder
encoder = createEncoder(encoderCls);
}
return encoder;... | 3.68 |
shardingsphere-elasticjob_JobAPIFactory_createJobConfigurationAPI | /**
* Create job configuration API.
*
* @param connectString registry center connect string
* @param namespace registry center namespace
* @param digest registry center digest
* @return job configuration API
*/
public static JobConfigurationAPI createJobConfigurationAPI(final String connectString, final String n... | 3.68 |
hadoop_SerialJobFactory_update | /**
* SERIAL. Once you get notification from StatsCollector about the job
* completion ,simply notify the waiting thread.
*
* @param item
*/
@Override
public void update(Statistics.JobStats item) {
//simply notify in case of serial submissions. We are just bothered
//if submitted job is completed or not.
loc... | 3.68 |
morf_DataSetProducerBuilderImpl_records | /**
* @see org.alfasoftware.morf.dataset.DataSetProducer#records(java.lang.String)
*/
@Override
public List<Record> records(String tableName) {
List<Record> records = recordMap.get(tableName.toUpperCase());
if (records == null) {
throw new IllegalStateException("No record data has been provided for table [" +... | 3.68 |
hadoop_PutTracker_getDestKey | /**
* get the destination key. The default implementation returns the
* key passed in: there is no adjustment of the destination.
* @return the destination to use in PUT requests.
*/
public String getDestKey() {
return destKey;
} | 3.68 |
dubbo_ServiceInstancesChangedListener_onEvent | /**
* On {@link ServiceInstancesChangedEvent the service instances change event}
*
* @param event {@link ServiceInstancesChangedEvent}
*/
public void onEvent(ServiceInstancesChangedEvent event) {
if (destroyed.get() || !accept(event) || isRetryAndExpired(event)) {
return;
}
doOnEvent(event);
} | 3.68 |
framework_DragSourceExtensionConnector_fixDragImageOffsetsForDesktop | /**
* Fixes missing or offset drag image caused by using css transform:
* translate (or such) by using a cloned drag image element, for which the
* property has been cleared.
* <p>
* This bug only occurs on Desktop with Safari (gets offset and clips the
* element for the parts that are not inside the element star... | 3.68 |
streampipes_PrimitivePropertyBuilder_description | /**
* Assigns a human-readable description to the event property. The description is used in the StreamPipes UI for
* better explaining users the meaning of the property.
*
* @param description
* @return this
*/
public PrimitivePropertyBuilder description(String description) {
this.eventProperty.setDescription(... | 3.68 |
framework_VTabsheet_updateContentNodeHeight | /** For internal use only. May be removed or replaced in the future. */
public void updateContentNodeHeight() {
if (!isDynamicHeight()) {
int contentHeight = getOffsetHeight();
contentHeight -= deco.getOffsetHeight();
contentHeight -= tb.getOffsetHeight();
ComputedStyle cs = new Com... | 3.68 |
hadoop_AbstractTask_getTaskCmd | /**
* Get TaskCmd for a Task.
* @return TaskCMD: Its a task command line such as sleep 10
*/
@Override
public final String getTaskCmd() {
return taskCmd;
} | 3.68 |
hadoop_TwoColumnLayout_content | /**
* @return the class that will render the content of the page.
*/
protected Class<? extends SubView> content() {
return LipsumBlock.class;
} | 3.68 |
flink_JobEdge_getShipStrategyName | /**
* Gets the name of the ship strategy for the represented input, like "forward", "partition
* hash", "rebalance", "broadcast", ...
*
* @return The name of the ship strategy for the represented input, or null, if none was set.
*/
public String getShipStrategyName() {
return shipStrategyName;
} | 3.68 |
flink_LegacySinkTransformation_setStateKeySelector | /**
* Sets the {@link KeySelector} that must be used for partitioning keyed state of this Sink.
*
* @param stateKeySelector The {@code KeySelector} to set
*/
public void setStateKeySelector(KeySelector<T, ?> stateKeySelector) {
this.stateKeySelector = stateKeySelector;
updateManagedMemoryStateBackendUseCase... | 3.68 |
framework_ContainerHierarchicalWrapper_addListener | /**
* @deprecated As of 7.0, replaced by
* {@link #addPropertySetChangeListener(Container.PropertySetChangeListener)}
*/
@Override
@Deprecated
public void addListener(Container.PropertySetChangeListener listener) {
addPropertySetChangeListener(listener);
} | 3.68 |
hadoop_ServiceLauncher_getServiceException | /**
* Get the exit exception used to end this service.
* @return an exception, which will be null until the service
* has exited (and {@code System.exit} has not been called)
*/
public final ExitUtil.ExitException getServiceException() {
return serviceException;
} | 3.68 |
morf_SqlServerDialect_getSqlForLeftPad | /**
* @see org.alfasoftware.morf.jdbc.SqlDialect#getSqlForLeftPad(org.alfasoftware.morf.sql.element.AliasedField, org.alfasoftware.morf.sql.element.AliasedField, org.alfasoftware.morf.sql.element.AliasedField)
*/
@Override
protected String getSqlForLeftPad(AliasedField field, AliasedField length, AliasedField charact... | 3.68 |
graphhopper_AbstractPathDetailsBuilder_endInterval | /**
* Ending intervals multiple times is safe, we only write the interval if it was open and not empty.
* Writes the interval to the pathDetails
*
* @param lastIndex the index the PathDetail ends
*/
public void endInterval(int lastIndex) {
if (isOpen) {
currentDetail.setLast(lastIndex);
pathDet... | 3.68 |
flink_ThreadInfoSamplesRequest_getMaxStackTraceDepth | /**
* Returns the configured maximum depth of the collected stack traces.
*
* @return the maximum depth of the collected stack traces.
*/
public int getMaxStackTraceDepth() {
return maxStackTraceDepth;
} | 3.68 |
hadoop_AHSLogsPage_content | /**
* The content of this page is the AggregatedLogsBlock
*
* @return AggregatedLogsBlock.class
*/
@Override
protected Class<? extends SubView> content() {
return AggregatedLogsBlock.class;
} | 3.68 |
framework_ApplicationConfiguration_startApplication | /**
* Starts the application with a given id by reading the configuration
* options stored by the bootstrap javascript.
*
* @param applicationId
* id of the application to load, this is also the id of the html
* element into which the application should be rendered.
*/
public static void st... | 3.68 |
hbase_AbstractFSWALProvider_findArchivedLog | /**
* Find the archived WAL file path if it is not able to locate in WALs dir.
* @param path - active WAL file path
* @param conf - configuration
* @return archived path if exists, null - otherwise
* @throws IOException exception
*/
public static Path findArchivedLog(Path path, Configuration conf) throws IOExcept... | 3.68 |
flink_PatternStream_sideOutputLateData | /**
* Send late arriving data to the side output identified by the given {@link OutputTag}. A
* record is considered late after the watermark has passed its timestamp.
*
* <p>You can get the stream of late data using {@link
* SingleOutputStreamOperator#getSideOutput(OutputTag)} on the {@link
* SingleOutputStreamO... | 3.68 |
rocketmq-connect_ConnectUtil_searchOffsetsByTimestamp | /** Search offsets by timestamp */
public static Map<MessageQueue, Long> searchOffsetsByTimestamp(
WorkerConfig config,
Collection<MessageQueue> messageQueues,
Long timestamp) {
Map<MessageQueue, Long> offsets = Maps.newConcurrentMap();
DefaultMQAdminExt adminClient = null;
try {
adminCl... | 3.68 |
hudi_LockManager_unlock | /**
* We need to take care of the scenarios that current thread may not be the holder of this lock
* and tries to call unlock()
*/
public void unlock() {
getLockProvider().unlock();
metrics.updateLockHeldTimerMetrics();
close();
} | 3.68 |
hbase_WALEdit_getFamilies | /**
* For use by FSWALEntry ONLY. An optimization.
* @return All families in {@link #getCells()}; may be null.
*/
public Set<byte[]> getFamilies() {
return this.families;
} | 3.68 |
framework_AbstractTextField_setPlaceholder | /**
* Sets the placeholder text. The placeholder is text that is displayed when
* the field would otherwise be empty, to prompt the user for input.
*
* @param placeholder
* the placeholder text to set
* @since 8.0
*/
public void setPlaceholder(String placeholder) {
getState().placeholder = placeho... | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.