name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
morf_SchemaModificationAdapter_dropExistingIndexesIfNecessary | /**
* Drop all existing indexes of the table that's about to be deployed.
*/
private void dropExistingIndexesIfNecessary(Table tableToDeploy) {
tableToDeploy.indexes().forEach(index -> {
Table existingTableWithSameIndex = existingIndexNamesAndTables.remove(index.getName().toUpperCase());
if (existingTableWi... | 3.68 |
hadoop_FederationStateStoreUtils_logAndThrowRetriableException | /**
* Throws an <code>FederationStateStoreRetriableException</code> due to an
* error in <code>FederationStateStore</code>.
*
* @param log the logger interface.
* @param errMsgFormat the error message format string.
* @param args referenced by the format specifiers in the format string.
* @throws YarnException o... | 3.68 |
framework_ConnectorFocusAndBlurHandler_removeHandlers | /**
* Remove all handlers from the widget and the connector.
*/
public void removeHandlers() {
if (focusRegistration != null) {
focusRegistration.removeHandler();
}
if (blurRegistration != null) {
blurRegistration.removeHandler();
}
if (stateChangeRegistration != null) {
st... | 3.68 |
hbase_SingleColumnValueFilter_setFilterIfMissing | /**
* Set whether entire row should be filtered if column is not found.
* <p>
* If true, the entire row will be skipped if the column is not found.
* <p>
* If false, the row will pass if the column is not found. This is default.
* @param filterIfMissing flag
*/
public void setFilterIfMissing(boolean filterIfMiss... | 3.68 |
hadoop_CloseableReferenceCount_unreference | /**
* Decrement the reference count.
*
* @return True if the object is closed and has no outstanding
* references.
*/
public boolean unreference() {
int newVal = status.decrementAndGet();
Preconditions.checkState(newVal != 0xffffffff,
"called unreference when the reference count... | 3.68 |
hadoop_RegisterApplicationMasterRequest_setPlacementConstraints | /**
* Set Placement Constraints applicable to the
* {@link org.apache.hadoop.yarn.api.records.SchedulingRequest}s
* of this application.
* The mapping is from a set of allocation tags to a
* <code>PlacementConstraint</code> associated with the tags.
* For example:
* Map <
* <hb_regionserver> -> n... | 3.68 |
framework_ServerRpcQueue_getAll | /**
* Returns a collection of all queued method invocations.
* <p>
* The returned collection must not be modified in any way
*
* @return a collection of all queued method invocations
*/
public Collection<MethodInvocation> getAll() {
return pendingInvocations.values();
} | 3.68 |
hadoop_NativeAzureFileSystemHelper_logAllLiveStackTraces | /*
* Helper method that logs stack traces from all live threads.
*/
public static void logAllLiveStackTraces() {
for (Map.Entry<Thread, StackTraceElement[]> entry : Thread.getAllStackTraces().entrySet()) {
LOG.debug("Thread " + entry.getKey().getName());
StackTraceElement[] trace = entry.getValue();
fo... | 3.68 |
framework_VaadinFinderLocatorStrategy_validatePath | /*
* (non-Javadoc)
*
* @see
* com.vaadin.client.componentlocator.LocatorStrategy#validatePath(java.
* lang.String)
*/
@Override
public boolean validatePath(String path) {
// This syntax is so difficult to regexp properly, that we'll just try
// to find something with it regardless of the correctness of th... | 3.68 |
hadoop_LoggingAuditor_addAttribute | /**
* Add an attribute.
* @param key key
* @param value value
*/
public final void addAttribute(String key, String value) {
attributes.put(key, value);
} | 3.68 |
hbase_MetricsStochasticBalancerSourceImpl_calcMruCap | /**
* Calculates the mru cache capacity from the metrics size
*/
private static int calcMruCap(int metricsSize) {
return (int) Math.ceil(metricsSize / MRU_LOAD_FACTOR) + 1;
} | 3.68 |
flink_Ordering_appendOrdering | /**
* Extends this ordering by appending an additional order requirement. If the index has been
* previously appended then the unmodified Ordering is returned.
*
* @param index Field index of the appended order requirement.
* @param type Type of the appended order requirement.
* @param order Order of the appended... | 3.68 |
framework_Slider_setValue | /**
* Sets the value of this object. If the new value is not equal to
* {@code getValue()}, fires a {@link ValueChangeEvent}. Throws
* {@code NullPointerException} if the value is null.
*
* @param value
* the new value, not {@code null}
* @throws NullPointerException
* if {@code value} is... | 3.68 |
framework_ListSet_contains | // Delegate contains operations to the set
@Override
public boolean contains(Object o) {
return itemSet.contains(o);
} | 3.68 |
flink_BinaryHashBucketArea_appendRecordAndInsert | /** Append record and insert to bucket. */
boolean appendRecordAndInsert(BinaryRowData record, int hashCode) throws IOException {
final int posHashCode = findBucket(hashCode);
// get the bucket for the given hash code
final int bucketArrayPos = posHashCode >> table.bucketsPerSegmentBits;
final int bucke... | 3.68 |
morf_AbstractSqlDialectTest_expectedSelectOrderByNullsFirstDesc | /**
* @return Expected SQL for {@link #testSelectOrderByNullsFirstDescendingScript()}
*/
protected String expectedSelectOrderByNullsFirstDesc() {
return "SELECT stringField FROM " + tableName(ALTERNATE_TABLE) + " ORDER BY stringField DESC NULLS FIRST";
} | 3.68 |
morf_Criterion_not | /**
* Helper method to create a new "NOT" expression.
*
* <blockquote><pre>
* Criterion.not(Criterion.eq(new Field("agreementnumber"), "A0001"));</pre></blockquote>
*
* @param criterion the first in the list of criteria
* @return a new Criterion object
*/
public static Criterion not(Criterion criterion) {
... | 3.68 |
MagicPlugin_Mage_getConversations | /**
* This isa non-API method that returns the live version of the conversation map
*/
@Nonnull
public Map<Player, MageConversation> getConversations() {
return conversations;
} | 3.68 |
framework_SelectorPath_generateJavaVariable | /**
* Generate Java variable assignment from given selector fragment
*
* @param pathFragment
* Selector fragment
* @return piece of java code
*/
private String generateJavaVariable(String pathFragment) {
// Get element type and predicates from fragment
List<SelectorPredicate> predicates = Selec... | 3.68 |
flink_ExecutionConfig_isForceAvroEnabled | /** Returns whether the Apache Avro is the default serializer for POJOs. */
public boolean isForceAvroEnabled() {
return configuration.get(PipelineOptions.FORCE_AVRO);
} | 3.68 |
streampipes_Aggregation_process | // Gets called every time a new event is fired, i.e. when an aggregation has to be calculated
protected void process(Iterable<Event> input, Collector<Event> out) {
List<Double> values = new ArrayList<>();
Event lastEvent = new Event();
// Adds the values of all recent events in input to aggregate them later
//... | 3.68 |
flink_Time_days | /** Creates a new {@link Time} that represents the given number of days. */
public static Time days(long days) {
return of(days, TimeUnit.DAYS);
} | 3.68 |
graphhopper_AngleCalc_convertAzimuth2xaxisAngle | /**
* convert north based clockwise azimuth (0, 360) into x-axis/east based angle (-Pi, Pi)
*/
public double convertAzimuth2xaxisAngle(double azimuth) {
if (Double.compare(azimuth, 360) > 0 || Double.compare(azimuth, 0) < 0) {
throw new IllegalArgumentException("Azimuth " + azimuth + " must be in (0, 360)... | 3.68 |
flink_DateTimeUtils_timestampMillisToTime | /**
* Get time from a timestamp.
*
* @param ts the timestamp in milliseconds.
* @return the time in milliseconds.
*/
public static int timestampMillisToTime(long ts) {
return (int) (ts % MILLIS_PER_DAY);
} | 3.68 |
hbase_Encryption_decrypt | /**
* Decrypt a stream of ciphertext given a context and IV
*/
public static void decrypt(OutputStream out, InputStream in, int outLen, Context context,
byte[] iv) throws IOException {
Decryptor d = context.getCipher().getDecryptor();
d.setKey(context.getKey());
d.setIv(iv); // can be null
decrypt(out, in, ... | 3.68 |
flink_MemorySize_parseBytes | /**
* Parses the given string as bytes. The supported expressions are listed under {@link
* MemorySize}.
*
* @param text The string to parse
* @return The parsed size, in bytes.
* @throws IllegalArgumentException Thrown, if the expression cannot be parsed.
*/
public static long parseBytes(String text) throws Ill... | 3.68 |
framework_StringToShortConverter_getModelType | /*
* (non-Javadoc)
*
* @see com.vaadin.data.util.converter.Converter#getModelType()
*/
@Override
public Class<Short> getModelType() {
return Short.class;
} | 3.68 |
hadoop_PublishedConfiguration_asJson | /**
* Return the values as json string
* @return the JSON representation
* @throws IOException marshalling failure
*/
public String asJson() throws IOException {
ObjectMapper mapper = new ObjectMapper();
mapper.configure(SerializationFeature.INDENT_OUTPUT, true);
String json = mapper.writeValueAsString(entrie... | 3.68 |
graphhopper_Entity_writeTimeField | /**
* Take a time expressed in seconds since noon - 12h (midnight, usually) and write it in HH:MM:SS format.
*/
protected void writeTimeField (int secsSinceMidnight) throws IOException {
if (secsSinceMidnight == INT_MISSING) {
writeStringField("");
return;
}
writeStringField(convertTo... | 3.68 |
hadoop_ManifestCommitter_getTaskAttemptDir | /**
* Get the task attempt dir.
* May be null.
* @return a path or null.
*/
private Path getTaskAttemptDir() {
return taskAttemptDir;
} | 3.68 |
hadoop_TaskInfo_getResourceUsageMetrics | /**
* @return Resource usage metrics
*/
public ResourceUsageMetrics getResourceUsageMetrics() {
return metrics;
} | 3.68 |
hmily_HmilyRepositoryFacade_writeHmilyLocks | /**
* Write hmily locks.
*
* @param locks locks
*/
public void writeHmilyLocks(final Collection<HmilyLock> locks) {
int count = hmilyRepository.writeHmilyLocks(locks);
if (count != locks.size()) {
HmilyLock lock = locks.iterator().next();
throw new HmilyLockConflictException(String.format("c... | 3.68 |
streampipes_SpGeometryBuilder_isInWGSCoordinateRange | /**
* Is in wgs coordinate range boolean.
*
* @param valueToCheck Any Value
* @param min Min value to check
* @param max max value to check
* @return true if value is in min max range
*/
private static boolean isInWGSCoordinateRange(double valueToCheck, double min, double max) {
return valueT... | 3.68 |
hadoop_CredentialProviderListFactory_loadAWSProviderClasses | /**
* Load list of AWS credential provider/credential provider factory classes.
* @param conf configuration
* @param key key
* @param defaultValue list of default values
* @return the list of classes, empty if the default list is empty and
* there was no match for the key in the configuration.
* @throws IOExcept... | 3.68 |
hbase_StorageClusterStatusModel_setRootIndexSizeKB | /**
* @param rootIndexSizeKB The current total size of root-level indexes for the region, in KB
*/
public void setRootIndexSizeKB(int rootIndexSizeKB) {
this.rootIndexSizeKB = rootIndexSizeKB;
} | 3.68 |
framework_AbstractComponent_getState | /**
* Returns the shared state bean with information to be sent from the server
* to the client.
*
* Subclasses should override this method and set any relevant fields of the
* state returned by super.getState().
*
* @since 7.0
*
* @return updated component shared state
*/
@Override
protected AbstractComponen... | 3.68 |
flink_BinarySegmentUtils_find | /**
* Find equal segments2 in segments1.
*
* @param segments1 segs to find.
* @param segments2 sub segs.
* @return Return the found offset, return -1 if not find.
*/
public static int find(
MemorySegment[] segments1,
int offset1,
int numBytes1,
MemorySegment[] segments2,
in... | 3.68 |
flink_TaskSlot_markActive | /**
* Mark this slot as active. A slot can only be marked active if it's in state allocated.
*
* <p>The method returns true if the slot was set to active. Otherwise it returns false.
*
* @return True if the new state of the slot is active; otherwise false
*/
public boolean markActive() {
if (TaskSlotState.ALL... | 3.68 |
framework_AbstractRemoteDataSource_insertRowData | /**
* Informs this data source that new data has been inserted from the server.
*
* @param firstRowIndex
* the destination index of the new row data
* @param count
* the number of rows inserted
*/
protected void insertRowData(int firstRowIndex, int count) {
Profiler.enter("AbstractRemot... | 3.68 |
flink_TypeTransformations_toNullable | /**
* Returns a type transformation that transforms data type to nullable data type but keeps other
* information unchanged.
*/
public static TypeTransformation toNullable() {
return DataType::nullable;
} | 3.68 |
querydsl_AbstractCollQuery_from | /**
* Add a query source
*
* @param <A> type of expression
* @param entity Path for the source
* @param col content of the source
* @return current object
*/
public <A> Q from(Path<A> entity, Iterable<? extends A> col) {
iterables.put(entity, col);
getMetadata().addJoin(JoinType.DEFAULT, entity);
ret... | 3.68 |
hadoop_RouterFedBalance_continueJob | /**
* Recover and continue the unfinished jobs.
*/
private int continueJob() throws InterruptedException {
BalanceProcedureScheduler scheduler =
new BalanceProcedureScheduler(getConf());
try {
scheduler.init(true);
while (true) {
Collection<BalanceJob> jobs = scheduler.getAllJobs();
int ... | 3.68 |
pulsar_MultiRolesTokenAuthorizationProvider_canProduceAsync | /**
* Check if the specified role has permission to send messages to the specified fully qualified topic name.
*
* @param topicName the fully qualified topic name associated with the topic.
* @param role the app id used to send messages to the topic.
*/
@Override
public CompletableFuture<Boolean> canProduceAs... | 3.68 |
hudi_HoodieTable_getIndexingMetadataWriter | /**
* Gets the metadata writer for async indexer.
*
* @param triggeringInstantTimestamp The instant that is triggering this metadata write.
* @return An instance of {@link HoodieTableMetadataWriter}.
*/
public Option<HoodieTableMetadataWriter> getIndexingMetadataWriter(String triggeringInstantTimestamp) {
return... | 3.68 |
hadoop_S3LogParser_e | /**
* Simple entry using the {@link #SIMPLE} pattern.
* @param name name of the element (for code clarity only)
* @return the pattern for the regexp
*/
private static String e(String name) {
return e(name, SIMPLE);
} | 3.68 |
hbase_ServerName_getVersionedBytes | /**
* Return {@link #getServerName()} as bytes with a short-sized prefix with the {@link #VERSION} of
* this class.
*/
public synchronized byte[] getVersionedBytes() {
if (this.bytes == null) {
this.bytes = Bytes.add(VERSION_BYTES, Bytes.toBytes(getServerName()));
}
return this.bytes;
} | 3.68 |
framework_Table_addGeneratedColumn | /**
* Adds a generated column to the Table.
* <p>
* A generated column is a column that exists only in the Table, not as a
* property in the underlying Container. It shows up just as a regular
* column.
* </p>
* <p>
* A generated column will override a property with the same id, so that the
* generated column ... | 3.68 |
flink_FieldReferenceLookup_getInputFields | /**
* Gives matching fields of underlying inputs in order of those inputs and order of fields
* within input.
*
* @return concatenated list of matching fields of all inputs.
*/
public List<FieldReferenceExpression> getInputFields(
List<ColumnExpansionStrategy> expansionStrategies) {
return fieldReferen... | 3.68 |
shardingsphere-elasticjob_GuaranteeService_clearAllStartedInfo | /**
* Clear all started job's info.
*/
public void clearAllStartedInfo() {
jobNodeStorage.removeJobNodeIfExisted(GuaranteeNode.STARTED_ROOT);
} | 3.68 |
pulsar_Record_getRecordSequence | /**
* Retrieves the sequence of the record from a source partition.
*
* @return Sequence Id associated with the record
*/
default Optional<Long> getRecordSequence() {
return Optional.empty();
} | 3.68 |
hbase_Chunk_getData | /** Returns This chunk's backing data. */
ByteBuffer getData() {
return this.data;
} | 3.68 |
morf_SchemaHomology_checkTable | /**
* @param table1 First table to compare.
* @param table2 Second table to compare.
*/
private void checkTable(Table table1, Table table2) {
matches("Table name", table1.getName().toUpperCase(), table2.getName().toUpperCase());
checkColumns(table1.getName(), table1.columns(), table2.columns());
checkIndexes... | 3.68 |
flink_FlinkDatabaseMetaData_getSchemas | // TODO Flink will support SHOW DATABASES LIKE statement in FLIP-297, this method will be
// supported after that issue.
@Override
public ResultSet getSchemas(String catalog, String schemaPattern) throws SQLException {
throw new UnsupportedOperationException();
} | 3.68 |
open-banking-gateway_FintechConsentAccessImpl_createAnonymousConsentNotPersist | /**
* Creates consent template, but does not persist it
*/
public ProtocolFacingConsent createAnonymousConsentNotPersist() {
return anonymousPsuConsentAccess.createDoNotPersist();
} | 3.68 |
framework_FieldGroup_discard | /**
* Discards all changes done to the bound fields.
* <p>
* Only has effect if buffered mode is used.
*
*/
public void discard() {
for (Field<?> f : fieldToPropertyId.keySet()) {
try {
f.discard();
} catch (Exception e) {
// TODO: handle exception
// What ca... | 3.68 |
flink_ExecutionJobVertex_cancelWithFuture | /**
* Cancels all currently running vertex executions.
*
* @return A future that is complete once all tasks have canceled.
*/
public CompletableFuture<Void> cancelWithFuture() {
return FutureUtils.waitForAll(mapExecutionVertices(ExecutionVertex::cancel));
} | 3.68 |
pulsar_AuthenticationDataProvider_hasDataFromCommand | /**
* Check if data from Pulsar protocol are available.
*
* @return true if this authentication data contain data from Pulsar protocol
*/
default boolean hasDataFromCommand() {
return false;
} | 3.68 |
hbase_StorageClusterStatusModel_setTotalCompactingKVs | /**
* @param totalCompactingKVs The total compacting key values in currently running compaction
*/
public void setTotalCompactingKVs(long totalCompactingKVs) {
this.totalCompactingKVs = totalCompactingKVs;
} | 3.68 |
hadoop_TimelineReaderWebServicesUtils_parseStr | /**
* Trims the passed string if its not null.
* @param str Passed string.
* @return trimmed string if string is not null, null otherwise.
*/
static String parseStr(String str) {
return StringUtils.trimToNull(str);
} | 3.68 |
hadoop_TFile_end | /**
* Get the end location of the TFile.
*
* @return The location right after the last key-value pair in TFile.
*/
Location end() {
return end;
} | 3.68 |
cron-utils_FieldConstraintsBuilder_withValidRange | /**
* Allows to set a range of valid values for field.
*
* @param startRange - start range value
* @param endRange - end range value
* @return same FieldConstraintsBuilder instance
*/
public FieldConstraintsBuilder withValidRange(final int startRange, final int endRange) {
this.startRange = startRange;
... | 3.68 |
hbase_DefaultMemStore_getNextRow | /**
* @param cell Find the row that comes after this one. If null, we return the first.
* @return Next row or null if none found.
*/
Cell getNextRow(final Cell cell) {
return getLowest(getNextRow(cell, this.getActive().getCellSet()),
getNextRow(cell, this.snapshot.getCellSet()));
} | 3.68 |
hbase_BitComparator_getOperator | /** Returns the bitwise operator */
public BitwiseOp getOperator() {
return bitOperator;
} | 3.68 |
framework_AbstractSelect_isEmpty | /**
* For multi-selectable fields, also an empty collection of values is
* considered to be an empty field.
*
* @see LegacyAbstractField#isEmpty().
*/
@Override
public boolean isEmpty() {
if (!multiSelect) {
return super.isEmpty();
} else {
Object value = getValue();
return super.is... | 3.68 |
hbase_TableName_isLegalFullyQualifiedTableName | /**
* Check passed byte array, "tableName", is legal user-space table name.
* @return Returns passed <code>tableName</code> param
* @throws IllegalArgumentException if passed a tableName is null or is made of other than 'word'
* characters or underscores: i.e.
* ... | 3.68 |
hbase_VersionModel_getOSVersion | /** Returns the OS name, version, and hardware architecture */
@XmlAttribute(name = "OS")
public String getOSVersion() {
return osVersion;
} | 3.68 |
hudi_PartitionAwareClusteringPlanStrategy_buildClusteringGroupsForPartition | /**
* Create Clustering group based on files eligible for clustering in the partition.
*/
protected Stream<HoodieClusteringGroup> buildClusteringGroupsForPartition(String partitionPath, List<FileSlice> fileSlices) {
HoodieWriteConfig writeConfig = getWriteConfig();
List<Pair<List<FileSlice>, Integer>> fileSliceG... | 3.68 |
hadoop_TimelineEntityGroupId_getTimelineEntityGroupId | /**
* Get the <code>timelineEntityGroupId</code>.
*
* @return <code>timelineEntityGroupId</code>
*/
public String getTimelineEntityGroupId() {
return this.id;
} | 3.68 |
hbase_FileIOEngine_read | /**
* Transfers data from file to the given byte buffer
* @param be an {@link BucketEntry} which maintains an (offset, len, refCnt)
* @return the {@link Cacheable} with block data inside.
* @throws IOException if any IO error happen.
*/
@Override
public Cacheable read(BucketEntry be) throws IOException {
long of... | 3.68 |
pulsar_PulsarClientImplementationBindingImpl_convertKeyValueSchemaInfoDataToString | /**
* Convert the key/value schema data.
*
* @param kvSchemaInfo the key/value schema info
* @return the convert key/value schema data string
*/
public String convertKeyValueSchemaInfoDataToString(KeyValue<SchemaInfo, SchemaInfo> kvSchemaInfo)
throws IOException {
return SchemaUtils.convertKeyValueSche... | 3.68 |
framework_VaadinService_setDefaultClassLoader | /**
* Tries to acquire default class loader and sets it as a class loader for
* this {@link VaadinService} if found. If current security policy disallows
* acquiring class loader instance it will log a message and re-throw
* {@link SecurityException}
*
* @throws SecurityException
* If current securit... | 3.68 |
framework_DesignContext_getPackagePrefixes | /**
* Gets all registered package prefixes.
*
*
* @since 7.5.0
* @see #getPackage(String)
* @return a collection of package prefixes
*/
public Collection<String> getPackagePrefixes() {
return Collections.unmodifiableCollection(prefixToPackage.keySet());
} | 3.68 |
framework_VCalendarPanel_getSubmitListener | /**
* Returns the submit listener that listens to selection made from the
* panel.
*
* @return The listener or NULL if no listener has been set
*/
public SubmitListener getSubmitListener() {
return submitListener;
} | 3.68 |
dubbo_ServiceModel_getServiceConfig | /**
* ServiceModel should be decoupled from AbstractInterfaceConfig and removed in a future version
* @return
*/
@Deprecated
public ServiceConfigBase<?> getServiceConfig() {
if (config == null) {
return null;
}
if (config instanceof ServiceConfigBase) {
return (ServiceConfigBase<?>) confi... | 3.68 |
flink_SinkFunction_invoke | /**
* Writes the given value to the sink. This function is called for every record.
*
* <p>You have to override this method when implementing a {@code SinkFunction}, this is a
* {@code default} method for backward compatibility with the old-style method only.
*
* @param value The input record.
* @param context A... | 3.68 |
hudi_AbstractStreamWriteFunction_invalidInstant | /**
* Returns whether the pending instant is invalid to write with.
*/
private boolean invalidInstant(String instant, boolean hasData) {
return instant.equals(this.currentInstant) && hasData;
} | 3.68 |
hbase_HFileBlockDefaultEncodingContext_close | /**
* Releases the compressor this writer uses to compress blocks into the compressor pool.
*/
@Override
public void close() {
if (compressor != null) {
this.fileContext.getCompression().returnCompressor(compressor);
compressor = null;
}
} | 3.68 |
hadoop_ApplicationServiceRecordProcessor_init | /**
* Initializes the descriptor parameters.
*
* @param serviceRecord the service record.
*/
@Override protected void init(ServiceRecord serviceRecord)
throws Exception {
super.init(serviceRecord);
if (getTarget() == null) {
return;
}
try {
this.setTarget(getIpv6Address(getTarget()));
} catch ... | 3.68 |
AreaShop_BuyRegion_buy | /**
* Buy a region.
* @param offlinePlayer The player that wants to buy the region
* @return true if it succeeded and false if not
*/
@SuppressWarnings("deprecation")
public boolean buy(OfflinePlayer offlinePlayer) {
// Check if the player has permission
if(!plugin.hasPermission(offlinePlayer, "areashop.buy")) {
... | 3.68 |
flink_MetricQueryService_replaceInvalidChars | /**
* Lightweight method to replace unsupported characters. If the string does not contain any
* unsupported characters, this method creates no new string (and in fact no new objects at
* all).
*
* <p>Replacements:
*
* <ul>
* <li>{@code space : . ,} are replaced by {@code _} (underscore)
* </ul>
*/
private ... | 3.68 |
framework_ConnectorTracker_markConnectorsDirtyRecursively | /**
* Marks all visible connectors dirty, starting from the given connector and
* going downwards in the hierarchy.
*
* @param c
* The component to start iterating downwards from
*/
private void markConnectorsDirtyRecursively(ClientConnector c) {
if (c instanceof Component && !((Component) c).isVis... | 3.68 |
dubbo_URLParam_removeParameters | /**
* remove specified parameters in URLParam
*
* @param keys keys to being removed
* @return A new URLParam
*/
public URLParam removeParameters(String... keys) {
if (keys == null || keys.length == 0) {
return this;
}
// lazy init, null if no modify
BitSet newKey = null;
int[] newValueA... | 3.68 |
framework_BasicEventMoveHandler_setDates | /**
* Set the start and end dates for the event.
*
* @param event
* The event that the start and end dates should be set
* @param start
* The start date
* @param end
* The end date
*/
protected void setDates(EditableCalendarEvent event, Date start, Date end) {
event.setStar... | 3.68 |
shardingsphere-elasticjob_FailoverService_getFailoveringItems | /**
* Get failovering items.
*
* @param jobInstanceId job instance ID
* @return failovering items
*/
public List<Integer> getFailoveringItems(final String jobInstanceId) {
List<String> items = jobNodeStorage.getJobNodeChildrenKeys(ShardingNode.ROOT);
List<Integer> result = new ArrayList<>(items.size());
... | 3.68 |
flink_FlinkRelMetadataQuery_getUpsertKeys | /**
* Determines the set of upsert minimal keys for this expression. A key is represented as an
* {@link org.apache.calcite.util.ImmutableBitSet}, where each bit position represents a 0-based
* output column ordinal.
*
* <p>Different from the unique keys: In distributed streaming computing, one record may be
* di... | 3.68 |
framework_AbstractComponentConnector_updateWidgetStyleNames | /**
* Updates the user defined, read-only and error style names for the widget
* based the shared state. User defined style names are prefixed with the
* primary style name of the widget returned by {@link #getWidget()}
* <p>
* This method can be overridden to provide additional style names for the
* component, f... | 3.68 |
graphhopper_RouterConfig_setTimeoutMillis | /**
* Limits the runtime of routing requests to the given amount of milliseconds. This only works up to a certain
* precision, but should be sufficient to cancel long-running requests in most cases. The exact implementation of
* the timeout depends on the routing algorithm.
*/
public void setTimeoutMillis(long time... | 3.68 |
hadoop_FsServerDefaults_getKeyProviderUri | /* null means old style namenode.
* "" (empty string) means namenode is upgraded but EZ is not supported.
* some string means that value is the key provider.
*/
public String getKeyProviderUri() {
return keyProviderUri;
} | 3.68 |
hadoop_OperationDuration_time | /**
* Evaluate the system time.
* @return the current clock time.
*/
protected long time() {
return System.currentTimeMillis();
} | 3.68 |
framework_Form_isEmpty | /**
* {@inheritDoc}
* <p>
* A Form is empty if all of its fields are empty.
*
*/
@Override
public boolean isEmpty() {
for (Field<?> f : fields.values()) {
if (f instanceof AbstractField) {
if (!((AbstractField<?>) f).isEmpty()) {
return false;
}
}
}
... | 3.68 |
hadoop_BlockManagerParameters_getBufferPoolSize | /**
* @return The size of the in-memory cache.
*/
public int getBufferPoolSize() {
return bufferPoolSize;
} | 3.68 |
flink_Router_notFound | /**
* Sets the fallback target for use when there's no match at {@link #route(HttpMethod, String)}.
*/
public Router<T> notFound(T target) {
this.notFound = target;
return this;
} | 3.68 |
flink_DefaultBlocklistTracker_tryAddOrMerge | /**
* Try to add a new blocked node record. If the node (identified by node id) already exists, the
* newly added one will be merged with the existing one.
*
* @param newNode the new blocked node record
* @return the add status
*/
private AddStatus tryAddOrMerge(BlockedNode newNode) {
checkNotNull(newNode);
... | 3.68 |
hadoop_LocalSASKeyGeneratorImpl_getAccountNameWithoutDomain | /**
* Helper method that returns the Storage account name without
* the domain name suffix.
* @param fullAccountName Storage account name with domain name suffix
* @return String
*/
private String getAccountNameWithoutDomain(String fullAccountName) {
StringTokenizer tokenizer = new StringTokenizer(fullAccountNam... | 3.68 |
flink_ResourceGuard_acquireResource | /**
* Acquired access from one new client for the guarded resource.
*
* @throws IOException when the resource guard is already closed.
*/
public Lease acquireResource() throws IOException {
synchronized (lock) {
if (closed) {
throw new IOException("Resource guard was already closed.");
... | 3.68 |
flink_ZooKeeperStateHandleStore_getRootLockPath | /**
* Returns the sub-path for lock nodes of the corresponding node (referred to through the passed
* {@code rooPath}. The returned sub-path collects the lock nodes for the {@code rootPath}'s
* node. The {@code rootPath} is marked for deletion if the sub-path for lock nodes is deleted.
*/
@VisibleForTesting
static ... | 3.68 |
framework_Result_of | /**
* Returns a Result representing the result of invoking the given supplier.
* If the supplier returns a value, returns a {@code Result.ok} of the
* value; if an exception is thrown, returns the message in a
* {@code Result.error}.
*
* @param <R>
* the result value type
* @param supplier
* ... | 3.68 |
hadoop_ConnectionPool_getConnection | /**
* Return the next connection round-robin.
*
* @return Connection context.
*/
protected ConnectionContext getConnection() {
this.lastActiveTime = Time.now();
List<ConnectionContext> tmpConnections = this.connections;
for (ConnectionContext tmpConnection : tmpConnections) {
if (tmpConnection != null && ... | 3.68 |
hbase_RegionMover_createConf | /**
* Creates a new configuration and sets region mover specific overrides
*/
private static Configuration createConf() {
Configuration conf = HBaseConfiguration.create();
conf.setInt("hbase.client.prefetch.limit", 1);
conf.setInt("hbase.client.pause", 500);
conf.setInt("hbase.client.retries.number", 100);
... | 3.68 |
flink_StreamExecutionEnvironment_getJobListeners | /** Gets the config JobListeners. */
@PublicEvolving
public List<JobListener> getJobListeners() {
return jobListeners;
} | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.