name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
morf_SchemaModificationAdapter_dropExistingIndexesIfNecessary
/** * Drop all existing indexes of the table that's about to be deployed. */ private void dropExistingIndexesIfNecessary(Table tableToDeploy) { tableToDeploy.indexes().forEach(index -> { Table existingTableWithSameIndex = existingIndexNamesAndTables.remove(index.getName().toUpperCase()); if (existingTableWi...
3.68
hadoop_FederationStateStoreUtils_logAndThrowRetriableException
/** * Throws an <code>FederationStateStoreRetriableException</code> due to an * error in <code>FederationStateStore</code>. * * @param log the logger interface. * @param errMsgFormat the error message format string. * @param args referenced by the format specifiers in the format string. * @throws YarnException o...
3.68
framework_ConnectorFocusAndBlurHandler_removeHandlers
/** * Remove all handlers from the widget and the connector. */ public void removeHandlers() { if (focusRegistration != null) { focusRegistration.removeHandler(); } if (blurRegistration != null) { blurRegistration.removeHandler(); } if (stateChangeRegistration != null) { st...
3.68
hbase_SingleColumnValueFilter_setFilterIfMissing
/** * Set whether entire row should be filtered if column is not found. * <p> * If true, the entire row will be skipped if the column is not found. * <p> * If false, the row will pass if the column is not found. This is default. * @param filterIfMissing flag */ public void setFilterIfMissing(boolean filterIfMiss...
3.68
hadoop_CloseableReferenceCount_unreference
/** * Decrement the reference count. * * @return True if the object is closed and has no outstanding * references. */ public boolean unreference() { int newVal = status.decrementAndGet(); Preconditions.checkState(newVal != 0xffffffff, "called unreference when the reference count...
3.68
hadoop_RegisterApplicationMasterRequest_setPlacementConstraints
/** * Set Placement Constraints applicable to the * {@link org.apache.hadoop.yarn.api.records.SchedulingRequest}s * of this application. * The mapping is from a set of allocation tags to a * <code>PlacementConstraint</code> associated with the tags. * For example: * Map &lt; * &lt;hb_regionserver&gt; -&gt; n...
3.68
framework_ServerRpcQueue_getAll
/** * Returns a collection of all queued method invocations. * <p> * The returned collection must not be modified in any way * * @return a collection of all queued method invocations */ public Collection<MethodInvocation> getAll() { return pendingInvocations.values(); }
3.68
hadoop_NativeAzureFileSystemHelper_logAllLiveStackTraces
/* * Helper method that logs stack traces from all live threads. */ public static void logAllLiveStackTraces() { for (Map.Entry<Thread, StackTraceElement[]> entry : Thread.getAllStackTraces().entrySet()) { LOG.debug("Thread " + entry.getKey().getName()); StackTraceElement[] trace = entry.getValue(); fo...
3.68
framework_VaadinFinderLocatorStrategy_validatePath
/* * (non-Javadoc) * * @see * com.vaadin.client.componentlocator.LocatorStrategy#validatePath(java. * lang.String) */ @Override public boolean validatePath(String path) { // This syntax is so difficult to regexp properly, that we'll just try // to find something with it regardless of the correctness of th...
3.68
hadoop_LoggingAuditor_addAttribute
/** * Add an attribute. * @param key key * @param value value */ public final void addAttribute(String key, String value) { attributes.put(key, value); }
3.68
hbase_MetricsStochasticBalancerSourceImpl_calcMruCap
/** * Calculates the mru cache capacity from the metrics size */ private static int calcMruCap(int metricsSize) { return (int) Math.ceil(metricsSize / MRU_LOAD_FACTOR) + 1; }
3.68
flink_Ordering_appendOrdering
/** * Extends this ordering by appending an additional order requirement. If the index has been * previously appended then the unmodified Ordering is returned. * * @param index Field index of the appended order requirement. * @param type Type of the appended order requirement. * @param order Order of the appended...
3.68
framework_Slider_setValue
/** * Sets the value of this object. If the new value is not equal to * {@code getValue()}, fires a {@link ValueChangeEvent}. Throws * {@code NullPointerException} if the value is null. * * @param value * the new value, not {@code null} * @throws NullPointerException * if {@code value} is...
3.68
framework_ListSet_contains
// Delegate contains operations to the set @Override public boolean contains(Object o) { return itemSet.contains(o); }
3.68
flink_BinaryHashBucketArea_appendRecordAndInsert
/** Append record and insert to bucket. */ boolean appendRecordAndInsert(BinaryRowData record, int hashCode) throws IOException { final int posHashCode = findBucket(hashCode); // get the bucket for the given hash code final int bucketArrayPos = posHashCode >> table.bucketsPerSegmentBits; final int bucke...
3.68
morf_AbstractSqlDialectTest_expectedSelectOrderByNullsFirstDesc
/** * @return Expected SQL for {@link #testSelectOrderByNullsFirstDescendingScript()} */ protected String expectedSelectOrderByNullsFirstDesc() { return "SELECT stringField FROM " + tableName(ALTERNATE_TABLE) + " ORDER BY stringField DESC NULLS FIRST"; }
3.68
morf_Criterion_not
/** * Helper method to create a new "NOT" expression. * * <blockquote><pre> * Criterion.not(Criterion.eq(new Field("agreementnumber"), "A0001"));</pre></blockquote> * * @param criterion the first in the list of criteria * @return a new Criterion object */ public static Criterion not(Criterion criterion) { ...
3.68
MagicPlugin_Mage_getConversations
/** * This isa non-API method that returns the live version of the conversation map */ @Nonnull public Map<Player, MageConversation> getConversations() { return conversations; }
3.68
framework_SelectorPath_generateJavaVariable
/** * Generate Java variable assignment from given selector fragment * * @param pathFragment * Selector fragment * @return piece of java code */ private String generateJavaVariable(String pathFragment) { // Get element type and predicates from fragment List<SelectorPredicate> predicates = Selec...
3.68
flink_ExecutionConfig_isForceAvroEnabled
/** Returns whether the Apache Avro is the default serializer for POJOs. */ public boolean isForceAvroEnabled() { return configuration.get(PipelineOptions.FORCE_AVRO); }
3.68
streampipes_Aggregation_process
// Gets called every time a new event is fired, i.e. when an aggregation has to be calculated protected void process(Iterable<Event> input, Collector<Event> out) { List<Double> values = new ArrayList<>(); Event lastEvent = new Event(); // Adds the values of all recent events in input to aggregate them later //...
3.68
flink_Time_days
/** Creates a new {@link Time} that represents the given number of days. */ public static Time days(long days) { return of(days, TimeUnit.DAYS); }
3.68
graphhopper_AngleCalc_convertAzimuth2xaxisAngle
/** * convert north based clockwise azimuth (0, 360) into x-axis/east based angle (-Pi, Pi) */ public double convertAzimuth2xaxisAngle(double azimuth) { if (Double.compare(azimuth, 360) > 0 || Double.compare(azimuth, 0) < 0) { throw new IllegalArgumentException("Azimuth " + azimuth + " must be in (0, 360)...
3.68
flink_DateTimeUtils_timestampMillisToTime
/** * Get time from a timestamp. * * @param ts the timestamp in milliseconds. * @return the time in milliseconds. */ public static int timestampMillisToTime(long ts) { return (int) (ts % MILLIS_PER_DAY); }
3.68
hbase_Encryption_decrypt
/** * Decrypt a stream of ciphertext given a context and IV */ public static void decrypt(OutputStream out, InputStream in, int outLen, Context context, byte[] iv) throws IOException { Decryptor d = context.getCipher().getDecryptor(); d.setKey(context.getKey()); d.setIv(iv); // can be null decrypt(out, in, ...
3.68
flink_MemorySize_parseBytes
/** * Parses the given string as bytes. The supported expressions are listed under {@link * MemorySize}. * * @param text The string to parse * @return The parsed size, in bytes. * @throws IllegalArgumentException Thrown, if the expression cannot be parsed. */ public static long parseBytes(String text) throws Ill...
3.68
framework_StringToShortConverter_getModelType
/* * (non-Javadoc) * * @see com.vaadin.data.util.converter.Converter#getModelType() */ @Override public Class<Short> getModelType() { return Short.class; }
3.68
hadoop_PublishedConfiguration_asJson
/** * Return the values as json string * @return the JSON representation * @throws IOException marshalling failure */ public String asJson() throws IOException { ObjectMapper mapper = new ObjectMapper(); mapper.configure(SerializationFeature.INDENT_OUTPUT, true); String json = mapper.writeValueAsString(entrie...
3.68
graphhopper_Entity_writeTimeField
/** * Take a time expressed in seconds since noon - 12h (midnight, usually) and write it in HH:MM:SS format. */ protected void writeTimeField (int secsSinceMidnight) throws IOException { if (secsSinceMidnight == INT_MISSING) { writeStringField(""); return; } writeStringField(convertTo...
3.68
hadoop_ManifestCommitter_getTaskAttemptDir
/** * Get the task attempt dir. * May be null. * @return a path or null. */ private Path getTaskAttemptDir() { return taskAttemptDir; }
3.68
hadoop_TaskInfo_getResourceUsageMetrics
/** * @return Resource usage metrics */ public ResourceUsageMetrics getResourceUsageMetrics() { return metrics; }
3.68
hmily_HmilyRepositoryFacade_writeHmilyLocks
/** * Write hmily locks. * * @param locks locks */ public void writeHmilyLocks(final Collection<HmilyLock> locks) { int count = hmilyRepository.writeHmilyLocks(locks); if (count != locks.size()) { HmilyLock lock = locks.iterator().next(); throw new HmilyLockConflictException(String.format("c...
3.68
streampipes_SpGeometryBuilder_isInWGSCoordinateRange
/** * Is in wgs coordinate range boolean. * * @param valueToCheck Any Value * @param min Min value to check * @param max max value to check * @return true if value is in min max range */ private static boolean isInWGSCoordinateRange(double valueToCheck, double min, double max) { return valueT...
3.68
hadoop_CredentialProviderListFactory_loadAWSProviderClasses
/** * Load list of AWS credential provider/credential provider factory classes. * @param conf configuration * @param key key * @param defaultValue list of default values * @return the list of classes, empty if the default list is empty and * there was no match for the key in the configuration. * @throws IOExcept...
3.68
hbase_StorageClusterStatusModel_setRootIndexSizeKB
/** * @param rootIndexSizeKB The current total size of root-level indexes for the region, in KB */ public void setRootIndexSizeKB(int rootIndexSizeKB) { this.rootIndexSizeKB = rootIndexSizeKB; }
3.68
framework_AbstractComponent_getState
/** * Returns the shared state bean with information to be sent from the server * to the client. * * Subclasses should override this method and set any relevant fields of the * state returned by super.getState(). * * @since 7.0 * * @return updated component shared state */ @Override protected AbstractComponen...
3.68
flink_BinarySegmentUtils_find
/** * Find equal segments2 in segments1. * * @param segments1 segs to find. * @param segments2 sub segs. * @return Return the found offset, return -1 if not find. */ public static int find( MemorySegment[] segments1, int offset1, int numBytes1, MemorySegment[] segments2, in...
3.68
flink_TaskSlot_markActive
/** * Mark this slot as active. A slot can only be marked active if it's in state allocated. * * <p>The method returns true if the slot was set to active. Otherwise it returns false. * * @return True if the new state of the slot is active; otherwise false */ public boolean markActive() { if (TaskSlotState.ALL...
3.68
framework_AbstractRemoteDataSource_insertRowData
/** * Informs this data source that new data has been inserted from the server. * * @param firstRowIndex * the destination index of the new row data * @param count * the number of rows inserted */ protected void insertRowData(int firstRowIndex, int count) { Profiler.enter("AbstractRemot...
3.68
flink_TypeTransformations_toNullable
/** * Returns a type transformation that transforms data type to nullable data type but keeps other * information unchanged. */ public static TypeTransformation toNullable() { return DataType::nullable; }
3.68
querydsl_AbstractCollQuery_from
/** * Add a query source * * @param <A> type of expression * @param entity Path for the source * @param col content of the source * @return current object */ public <A> Q from(Path<A> entity, Iterable<? extends A> col) { iterables.put(entity, col); getMetadata().addJoin(JoinType.DEFAULT, entity); ret...
3.68
hadoop_RouterFedBalance_continueJob
/** * Recover and continue the unfinished jobs. */ private int continueJob() throws InterruptedException { BalanceProcedureScheduler scheduler = new BalanceProcedureScheduler(getConf()); try { scheduler.init(true); while (true) { Collection<BalanceJob> jobs = scheduler.getAllJobs(); int ...
3.68
pulsar_MultiRolesTokenAuthorizationProvider_canProduceAsync
/** * Check if the specified role has permission to send messages to the specified fully qualified topic name. * * @param topicName the fully qualified topic name associated with the topic. * @param role the app id used to send messages to the topic. */ @Override public CompletableFuture<Boolean> canProduceAs...
3.68
hudi_HoodieTable_getIndexingMetadataWriter
/** * Gets the metadata writer for async indexer. * * @param triggeringInstantTimestamp The instant that is triggering this metadata write. * @return An instance of {@link HoodieTableMetadataWriter}. */ public Option<HoodieTableMetadataWriter> getIndexingMetadataWriter(String triggeringInstantTimestamp) { return...
3.68
hadoop_S3LogParser_e
/** * Simple entry using the {@link #SIMPLE} pattern. * @param name name of the element (for code clarity only) * @return the pattern for the regexp */ private static String e(String name) { return e(name, SIMPLE); }
3.68
hbase_ServerName_getVersionedBytes
/** * Return {@link #getServerName()} as bytes with a short-sized prefix with the {@link #VERSION} of * this class. */ public synchronized byte[] getVersionedBytes() { if (this.bytes == null) { this.bytes = Bytes.add(VERSION_BYTES, Bytes.toBytes(getServerName())); } return this.bytes; }
3.68
framework_Table_addGeneratedColumn
/** * Adds a generated column to the Table. * <p> * A generated column is a column that exists only in the Table, not as a * property in the underlying Container. It shows up just as a regular * column. * </p> * <p> * A generated column will override a property with the same id, so that the * generated column ...
3.68
flink_FieldReferenceLookup_getInputFields
/** * Gives matching fields of underlying inputs in order of those inputs and order of fields * within input. * * @return concatenated list of matching fields of all inputs. */ public List<FieldReferenceExpression> getInputFields( List<ColumnExpansionStrategy> expansionStrategies) { return fieldReferen...
3.68
shardingsphere-elasticjob_GuaranteeService_clearAllStartedInfo
/** * Clear all started job's info. */ public void clearAllStartedInfo() { jobNodeStorage.removeJobNodeIfExisted(GuaranteeNode.STARTED_ROOT); }
3.68
pulsar_Record_getRecordSequence
/** * Retrieves the sequence of the record from a source partition. * * @return Sequence Id associated with the record */ default Optional<Long> getRecordSequence() { return Optional.empty(); }
3.68
hbase_Chunk_getData
/** Returns This chunk's backing data. */ ByteBuffer getData() { return this.data; }
3.68
morf_SchemaHomology_checkTable
/** * @param table1 First table to compare. * @param table2 Second table to compare. */ private void checkTable(Table table1, Table table2) { matches("Table name", table1.getName().toUpperCase(), table2.getName().toUpperCase()); checkColumns(table1.getName(), table1.columns(), table2.columns()); checkIndexes...
3.68
flink_FlinkDatabaseMetaData_getSchemas
// TODO Flink will support SHOW DATABASES LIKE statement in FLIP-297, this method will be // supported after that issue. @Override public ResultSet getSchemas(String catalog, String schemaPattern) throws SQLException { throw new UnsupportedOperationException(); }
3.68
open-banking-gateway_FintechConsentAccessImpl_createAnonymousConsentNotPersist
/** * Creates consent template, but does not persist it */ public ProtocolFacingConsent createAnonymousConsentNotPersist() { return anonymousPsuConsentAccess.createDoNotPersist(); }
3.68
framework_FieldGroup_discard
/** * Discards all changes done to the bound fields. * <p> * Only has effect if buffered mode is used. * */ public void discard() { for (Field<?> f : fieldToPropertyId.keySet()) { try { f.discard(); } catch (Exception e) { // TODO: handle exception // What ca...
3.68
flink_ExecutionJobVertex_cancelWithFuture
/** * Cancels all currently running vertex executions. * * @return A future that is complete once all tasks have canceled. */ public CompletableFuture<Void> cancelWithFuture() { return FutureUtils.waitForAll(mapExecutionVertices(ExecutionVertex::cancel)); }
3.68
pulsar_AuthenticationDataProvider_hasDataFromCommand
/** * Check if data from Pulsar protocol are available. * * @return true if this authentication data contain data from Pulsar protocol */ default boolean hasDataFromCommand() { return false; }
3.68
hbase_StorageClusterStatusModel_setTotalCompactingKVs
/** * @param totalCompactingKVs The total compacting key values in currently running compaction */ public void setTotalCompactingKVs(long totalCompactingKVs) { this.totalCompactingKVs = totalCompactingKVs; }
3.68
hadoop_TimelineReaderWebServicesUtils_parseStr
/** * Trims the passed string if its not null. * @param str Passed string. * @return trimmed string if string is not null, null otherwise. */ static String parseStr(String str) { return StringUtils.trimToNull(str); }
3.68
hadoop_TFile_end
/** * Get the end location of the TFile. * * @return The location right after the last key-value pair in TFile. */ Location end() { return end; }
3.68
cron-utils_FieldConstraintsBuilder_withValidRange
/** * Allows to set a range of valid values for field. * * @param startRange - start range value * @param endRange - end range value * @return same FieldConstraintsBuilder instance */ public FieldConstraintsBuilder withValidRange(final int startRange, final int endRange) { this.startRange = startRange; ...
3.68
hbase_DefaultMemStore_getNextRow
/** * @param cell Find the row that comes after this one. If null, we return the first. * @return Next row or null if none found. */ Cell getNextRow(final Cell cell) { return getLowest(getNextRow(cell, this.getActive().getCellSet()), getNextRow(cell, this.snapshot.getCellSet())); }
3.68
hbase_BitComparator_getOperator
/** Returns the bitwise operator */ public BitwiseOp getOperator() { return bitOperator; }
3.68
framework_AbstractSelect_isEmpty
/** * For multi-selectable fields, also an empty collection of values is * considered to be an empty field. * * @see LegacyAbstractField#isEmpty(). */ @Override public boolean isEmpty() { if (!multiSelect) { return super.isEmpty(); } else { Object value = getValue(); return super.is...
3.68
hbase_TableName_isLegalFullyQualifiedTableName
/** * Check passed byte array, "tableName", is legal user-space table name. * @return Returns passed <code>tableName</code> param * @throws IllegalArgumentException if passed a tableName is null or is made of other than 'word' * characters or underscores: i.e. * ...
3.68
hbase_VersionModel_getOSVersion
/** Returns the OS name, version, and hardware architecture */ @XmlAttribute(name = "OS") public String getOSVersion() { return osVersion; }
3.68
hudi_PartitionAwareClusteringPlanStrategy_buildClusteringGroupsForPartition
/** * Create Clustering group based on files eligible for clustering in the partition. */ protected Stream<HoodieClusteringGroup> buildClusteringGroupsForPartition(String partitionPath, List<FileSlice> fileSlices) { HoodieWriteConfig writeConfig = getWriteConfig(); List<Pair<List<FileSlice>, Integer>> fileSliceG...
3.68
hadoop_TimelineEntityGroupId_getTimelineEntityGroupId
/** * Get the <code>timelineEntityGroupId</code>. * * @return <code>timelineEntityGroupId</code> */ public String getTimelineEntityGroupId() { return this.id; }
3.68
hbase_FileIOEngine_read
/** * Transfers data from file to the given byte buffer * @param be an {@link BucketEntry} which maintains an (offset, len, refCnt) * @return the {@link Cacheable} with block data inside. * @throws IOException if any IO error happen. */ @Override public Cacheable read(BucketEntry be) throws IOException { long of...
3.68
pulsar_PulsarClientImplementationBindingImpl_convertKeyValueSchemaInfoDataToString
/** * Convert the key/value schema data. * * @param kvSchemaInfo the key/value schema info * @return the convert key/value schema data string */ public String convertKeyValueSchemaInfoDataToString(KeyValue<SchemaInfo, SchemaInfo> kvSchemaInfo) throws IOException { return SchemaUtils.convertKeyValueSche...
3.68
framework_VaadinService_setDefaultClassLoader
/** * Tries to acquire default class loader and sets it as a class loader for * this {@link VaadinService} if found. If current security policy disallows * acquiring class loader instance it will log a message and re-throw * {@link SecurityException} * * @throws SecurityException * If current securit...
3.68
framework_DesignContext_getPackagePrefixes
/** * Gets all registered package prefixes. * * * @since 7.5.0 * @see #getPackage(String) * @return a collection of package prefixes */ public Collection<String> getPackagePrefixes() { return Collections.unmodifiableCollection(prefixToPackage.keySet()); }
3.68
framework_VCalendarPanel_getSubmitListener
/** * Returns the submit listener that listens to selection made from the * panel. * * @return The listener or NULL if no listener has been set */ public SubmitListener getSubmitListener() { return submitListener; }
3.68
dubbo_ServiceModel_getServiceConfig
/** * ServiceModel should be decoupled from AbstractInterfaceConfig and removed in a future version * @return */ @Deprecated public ServiceConfigBase<?> getServiceConfig() { if (config == null) { return null; } if (config instanceof ServiceConfigBase) { return (ServiceConfigBase<?>) confi...
3.68
flink_SinkFunction_invoke
/** * Writes the given value to the sink. This function is called for every record. * * <p>You have to override this method when implementing a {@code SinkFunction}, this is a * {@code default} method for backward compatibility with the old-style method only. * * @param value The input record. * @param context A...
3.68
hudi_AbstractStreamWriteFunction_invalidInstant
/** * Returns whether the pending instant is invalid to write with. */ private boolean invalidInstant(String instant, boolean hasData) { return instant.equals(this.currentInstant) && hasData; }
3.68
hbase_HFileBlockDefaultEncodingContext_close
/** * Releases the compressor this writer uses to compress blocks into the compressor pool. */ @Override public void close() { if (compressor != null) { this.fileContext.getCompression().returnCompressor(compressor); compressor = null; } }
3.68
hadoop_ApplicationServiceRecordProcessor_init
/** * Initializes the descriptor parameters. * * @param serviceRecord the service record. */ @Override protected void init(ServiceRecord serviceRecord) throws Exception { super.init(serviceRecord); if (getTarget() == null) { return; } try { this.setTarget(getIpv6Address(getTarget())); } catch ...
3.68
AreaShop_BuyRegion_buy
/** * Buy a region. * @param offlinePlayer The player that wants to buy the region * @return true if it succeeded and false if not */ @SuppressWarnings("deprecation") public boolean buy(OfflinePlayer offlinePlayer) { // Check if the player has permission if(!plugin.hasPermission(offlinePlayer, "areashop.buy")) { ...
3.68
flink_MetricQueryService_replaceInvalidChars
/** * Lightweight method to replace unsupported characters. If the string does not contain any * unsupported characters, this method creates no new string (and in fact no new objects at * all). * * <p>Replacements: * * <ul> * <li>{@code space : . ,} are replaced by {@code _} (underscore) * </ul> */ private ...
3.68
framework_ConnectorTracker_markConnectorsDirtyRecursively
/** * Marks all visible connectors dirty, starting from the given connector and * going downwards in the hierarchy. * * @param c * The component to start iterating downwards from */ private void markConnectorsDirtyRecursively(ClientConnector c) { if (c instanceof Component && !((Component) c).isVis...
3.68
dubbo_URLParam_removeParameters
/** * remove specified parameters in URLParam * * @param keys keys to being removed * @return A new URLParam */ public URLParam removeParameters(String... keys) { if (keys == null || keys.length == 0) { return this; } // lazy init, null if no modify BitSet newKey = null; int[] newValueA...
3.68
framework_BasicEventMoveHandler_setDates
/** * Set the start and end dates for the event. * * @param event * The event that the start and end dates should be set * @param start * The start date * @param end * The end date */ protected void setDates(EditableCalendarEvent event, Date start, Date end) { event.setStar...
3.68
shardingsphere-elasticjob_FailoverService_getFailoveringItems
/** * Get failovering items. * * @param jobInstanceId job instance ID * @return failovering items */ public List<Integer> getFailoveringItems(final String jobInstanceId) { List<String> items = jobNodeStorage.getJobNodeChildrenKeys(ShardingNode.ROOT); List<Integer> result = new ArrayList<>(items.size()); ...
3.68
flink_FlinkRelMetadataQuery_getUpsertKeys
/** * Determines the set of upsert minimal keys for this expression. A key is represented as an * {@link org.apache.calcite.util.ImmutableBitSet}, where each bit position represents a 0-based * output column ordinal. * * <p>Different from the unique keys: In distributed streaming computing, one record may be * di...
3.68
framework_AbstractComponentConnector_updateWidgetStyleNames
/** * Updates the user defined, read-only and error style names for the widget * based the shared state. User defined style names are prefixed with the * primary style name of the widget returned by {@link #getWidget()} * <p> * This method can be overridden to provide additional style names for the * component, f...
3.68
graphhopper_RouterConfig_setTimeoutMillis
/** * Limits the runtime of routing requests to the given amount of milliseconds. This only works up to a certain * precision, but should be sufficient to cancel long-running requests in most cases. The exact implementation of * the timeout depends on the routing algorithm. */ public void setTimeoutMillis(long time...
3.68
hadoop_FsServerDefaults_getKeyProviderUri
/* null means old style namenode. * "" (empty string) means namenode is upgraded but EZ is not supported. * some string means that value is the key provider. */ public String getKeyProviderUri() { return keyProviderUri; }
3.68
hadoop_OperationDuration_time
/** * Evaluate the system time. * @return the current clock time. */ protected long time() { return System.currentTimeMillis(); }
3.68
framework_Form_isEmpty
/** * {@inheritDoc} * <p> * A Form is empty if all of its fields are empty. * */ @Override public boolean isEmpty() { for (Field<?> f : fields.values()) { if (f instanceof AbstractField) { if (!((AbstractField<?>) f).isEmpty()) { return false; } } } ...
3.68
hadoop_BlockManagerParameters_getBufferPoolSize
/** * @return The size of the in-memory cache. */ public int getBufferPoolSize() { return bufferPoolSize; }
3.68
flink_Router_notFound
/** * Sets the fallback target for use when there's no match at {@link #route(HttpMethod, String)}. */ public Router<T> notFound(T target) { this.notFound = target; return this; }
3.68
flink_DefaultBlocklistTracker_tryAddOrMerge
/** * Try to add a new blocked node record. If the node (identified by node id) already exists, the * newly added one will be merged with the existing one. * * @param newNode the new blocked node record * @return the add status */ private AddStatus tryAddOrMerge(BlockedNode newNode) { checkNotNull(newNode); ...
3.68
hadoop_LocalSASKeyGeneratorImpl_getAccountNameWithoutDomain
/** * Helper method that returns the Storage account name without * the domain name suffix. * @param fullAccountName Storage account name with domain name suffix * @return String */ private String getAccountNameWithoutDomain(String fullAccountName) { StringTokenizer tokenizer = new StringTokenizer(fullAccountNam...
3.68
flink_ResourceGuard_acquireResource
/** * Acquired access from one new client for the guarded resource. * * @throws IOException when the resource guard is already closed. */ public Lease acquireResource() throws IOException { synchronized (lock) { if (closed) { throw new IOException("Resource guard was already closed."); ...
3.68
flink_ZooKeeperStateHandleStore_getRootLockPath
/** * Returns the sub-path for lock nodes of the corresponding node (referred to through the passed * {@code rooPath}. The returned sub-path collects the lock nodes for the {@code rootPath}'s * node. The {@code rootPath} is marked for deletion if the sub-path for lock nodes is deleted. */ @VisibleForTesting static ...
3.68
framework_Result_of
/** * Returns a Result representing the result of invoking the given supplier. * If the supplier returns a value, returns a {@code Result.ok} of the * value; if an exception is thrown, returns the message in a * {@code Result.error}. * * @param <R> * the result value type * @param supplier * ...
3.68
hadoop_ConnectionPool_getConnection
/** * Return the next connection round-robin. * * @return Connection context. */ protected ConnectionContext getConnection() { this.lastActiveTime = Time.now(); List<ConnectionContext> tmpConnections = this.connections; for (ConnectionContext tmpConnection : tmpConnections) { if (tmpConnection != null && ...
3.68
hbase_RegionMover_createConf
/** * Creates a new configuration and sets region mover specific overrides */ private static Configuration createConf() { Configuration conf = HBaseConfiguration.create(); conf.setInt("hbase.client.prefetch.limit", 1); conf.setInt("hbase.client.pause", 500); conf.setInt("hbase.client.retries.number", 100); ...
3.68
flink_StreamExecutionEnvironment_getJobListeners
/** Gets the config JobListeners. */ @PublicEvolving public List<JobListener> getJobListeners() { return jobListeners; }
3.68