name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
hudi_HoodieFlinkWriteClient_cleanHandles | /**
* Clean the write handles within a checkpoint interval.
* All the handles should have been closed already.
*/
public void cleanHandles() {
this.bucketToHandles.clear();
} | 3.68 |
flink_ResultPartitionFactory_isOverdraftBufferNeeded | /** Return whether this result partition need overdraft buffer. */
private static boolean isOverdraftBufferNeeded(ResultPartitionType resultPartitionType) {
// Only pipelined / pipelined-bounded partition needs overdraft buffer. More
// specifically, there is no reason to request more buffers for non-pipelined ... | 3.68 |
hbase_ExcludeDatanodeManager_tryAddExcludeDN | /**
* Try to add a datanode to the regionserver excluding cache
* @param datanodeInfo the datanode to be added to the excluded cache
* @param cause the cause that the datanode is hope to be excluded
* @return True if the datanode is added to the regionserver excluding cache, false otherwise
*/
public boolea... | 3.68 |
hadoop_AbfsManifestStoreOperations_bindToFileSystem | /**
* Bind to the store.
*
* @param filesystem FS.
* @param path path to work under
* @throws IOException binding problems.
*/
@Override
public void bindToFileSystem(FileSystem filesystem, Path path) throws IOException {
if (!(filesystem instanceof AzureBlobFileSystem)) {
throw new PathIOException(path.toSt... | 3.68 |
dubbo_ServiceConfig_exportLocal | /**
* always export injvm
*/
private void exportLocal(URL url) {
URL local = URLBuilder.from(url)
.setProtocol(LOCAL_PROTOCOL)
.setHost(LOCALHOST_VALUE)
.setPort(0)
.build();
local = local.setScopeModel(getScopeModel()).setServiceModel(providerModel);
local ... | 3.68 |
morf_RemoveIndex_apply | /**
* {@inheritDoc}
*
* @see org.alfasoftware.morf.upgrade.SchemaChange#apply(org.alfasoftware.morf.metadata.Schema)
*/
@Override
public Schema apply(Schema schema) {
Table original = schema.getTable(tableName);
boolean foundIndex = false;
List<String> indexes = new ArrayList<>();
for (Index index : origin... | 3.68 |
druid_MySqlStatementParser_parseWhile | /**
* parse while statement with label
*
* @return MySqlWhileStatement
*/
public SQLWhileStatement parseWhile(String label) {
accept(Token.WHILE);
SQLWhileStatement stmt = new SQLWhileStatement();
stmt.setLabelName(label);
stmt.setCondition(this.exprParser.expr());
accept(Token.DO);
thi... | 3.68 |
hadoop_ActiveAuditManagerS3A_modifyResponse | /**
* Forward to the inner span.
* {@inheritDoc}
*/
@Override
public SdkResponse modifyResponse(Context.ModifyResponse context,
ExecutionAttributes executionAttributes) {
return span.modifyResponse(context, executionAttributes);
} | 3.68 |
hbase_OrderedBytes_length | /**
* Return the number of encoded entries remaining in {@code buff}. The state of {@code buff} is
* not modified through use of this method.
*/
public static int length(PositionedByteRange buff) {
PositionedByteRange b =
new SimplePositionedMutableByteRange(buff.getBytes(), buff.getOffset(), buff.getLength())... | 3.68 |
hadoop_AzureFileSystemInstrumentation_directoryDeleted | /**
* Indicate that we just deleted a directory through WASB.
*/
public void directoryDeleted() {
numberOfDirectoriesDeleted.incr();
} | 3.68 |
hbase_ClaimReplicationQueueRemoteProcedure_shouldSkip | // check whether ReplicationSyncUp has already done the work for us, if so, we should skip
// claiming the replication queues and deleting them instead.
private boolean shouldSkip(MasterProcedureEnv env) throws IOException {
MasterFileSystem mfs = env.getMasterFileSystem();
Path syncUpDir = new Path(mfs.getRootDir(... | 3.68 |
hadoop_MagicCommitTracker_outputImmediatelyVisible | /**
* Flag to indicate that output is not visible after the stream
* is closed.
* @return true
*/
@Override
public boolean outputImmediatelyVisible() {
return false;
} | 3.68 |
framework_Potus_getParty | /**
* @return the party
*/
public String getParty() {
return party;
} | 3.68 |
hadoop_ServletUtil_htmlFooter | /**
* HTML footer to be added in the jsps.
* @return the HTML footer.
*/
public static String htmlFooter() {
return HTML_TAIL;
} | 3.68 |
zxing_GenericGF_exp | /**
* @return 2 to the power of a in GF(size)
*/
int exp(int a) {
return expTable[a];
} | 3.68 |
hbase_RegionMover_stripServer | /**
* Remove the servername whose hostname and port portion matches from the passed array of servers.
* Returns as side-effect the servername removed.
* @return server removed from list of Region Servers
*/
private ServerName stripServer(List<ServerName> regionServers, String hostname, int port) {
for (Iterator<S... | 3.68 |
flink_TextElement_wrap | /** Wraps a list of {@link InlineElement}s into a single {@link TextElement}. */
public static InlineElement wrap(InlineElement... elements) {
return text(Strings.repeat("%s", elements.length), elements);
} | 3.68 |
hudi_BaseHoodieQueueBasedExecutor_startConsumingAsync | /**
* Start consumer
*/
private CompletableFuture<Void> startConsumingAsync() {
return consumer.map(consumer ->
CompletableFuture.supplyAsync(() -> {
doConsume(queue, consumer);
return (Void) null;
}, consumerExecutorService)
)
.orElse(CompletableFuture.comple... | 3.68 |
hbase_MasterObserver_postRevoke | /**
* Called after revoking user permissions.
* @param ctx the coprocessor instance's environment
* @param userPermission the user and permissions
*/
default void postRevoke(ObserverContext<MasterCoprocessorEnvironment> ctx,
UserPermission userPermission) throws IOException {
} | 3.68 |
framework_SerializerHelper_readClassArray | /**
* Deserializes a class references serialized by
* {@link #writeClassArray(ObjectOutputStream, Class[])}. Supports null
* class arrays.
*
* @param in
* {@link ObjectInputStream} to read from.
* @return Class array with the class references or null.
* @throws ClassNotFoundException
* I... | 3.68 |
framework_AbstractInMemoryContainer_removeAllFilters | /**
* Remove all container filters for all properties and re-filter the view.
*
* This can be used to implement
* {@link Filterable#removeAllContainerFilters()}.
*/
protected void removeAllFilters() {
if (getFilters().isEmpty()) {
return;
}
getFilters().clear();
filterAll();
} | 3.68 |
hudi_Option_orElse | /**
* Identical to {@code Optional.orElse}
*/
public T orElse(T other) {
return val != null ? val : other;
} | 3.68 |
pulsar_AuthorizationService_allowTopicPolicyOperation | /**
* @deprecated - will be removed after 2.12. Use async variant.
*/
@Deprecated
public Boolean allowTopicPolicyOperation(TopicName topicName,
PolicyName policy,
PolicyOperation operation,
Strin... | 3.68 |
pulsar_PortManager_releaseLockedPort | /**
* Returns whether the port was released successfully.
*
* @return whether the release is successful.
*/
public static synchronized boolean releaseLockedPort(int lockedPort) {
return PORTS.remove(lockedPort);
} | 3.68 |
druid_IPRange_computeNetworkPrefixFromMask | /**
* Compute the extended network prefix from the IP subnet mask.
*
* @param mask Reference to the subnet mask IP number.
* @return Return the extended network prefix. Return -1 if the specified mask cannot be converted into a extended
* prefix network.
*/
private int computeNetworkPrefixFromMask(IPAddress mask)... | 3.68 |
flink_PythonOperatorUtils_setCurrentKeyForStreaming | /** Set the current key for streaming operator. */
public static <K> void setCurrentKeyForStreaming(
KeyedStateBackend<K> stateBackend, K currentKey) {
if (!inBatchExecutionMode(stateBackend)) {
stateBackend.setCurrentKey(currentKey);
}
} | 3.68 |
flink_DataSet_count | /**
* Convenience method to get the count (number of elements) of a DataSet.
*
* @return A long integer that represents the number of elements in the data set.
*/
public long count() throws Exception {
final String id = new AbstractID().toString();
output(new Utils.CountHelper<T>(id)).name("count()");
... | 3.68 |
morf_SqlUtils_truncate | /**
* Constructs a Truncate Statement.
*
* <p>Usage is discouraged; this method will be deprecated at some point. Use
* {@link TruncateStatement#truncate(TableReference)} for preference.</p>
*
* @param table The table to truncate.
* @return The statement.
*/
public static TruncateStatement truncate(TableReferen... | 3.68 |
flink_MemorySegment_putFloatLittleEndian | /**
* Writes the given single-precision float value (32bit, 4 bytes) to the given position in
* little endian byte order. This method's speed depends on the system's native byte order, and
* it is possibly slower than {@link #putFloat(int, float)}. For most cases (such as transient
* storage in memory or serializat... | 3.68 |
hadoop_LocalResolver_chooseFirstNamespace | /**
* Get the local name space. This relies on the RPC Server to get the address
* from the client.
*
* TODO we only support DN and NN locations, we need to add others like
* Resource Managers.
*
* @param path Path ignored by this policy.
* @param loc Federated location with multiple destinations.
* @return Lo... | 3.68 |
rocketmq-connect_Deserializer_deserialize | /**
* Deserialize a record value from a byte array into a value or object.
*/
default T deserialize(String topic, KeyValue extensions, byte[] data) {
return deserialize(topic, data);
} | 3.68 |
dubbo_FileCacheStore_getCacheFilePath | /**
* for unit test only
*/
@Deprecated
protected String getCacheFilePath() {
return cacheFilePath;
} | 3.68 |
flink_StreamProjection_projectTuple18 | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <T0, T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17>
SingleOutputStreamOperator<
Tuple18<
... | 3.68 |
flink_NettyMessage_allocateBuffer | /**
* Allocates a new buffer and adds some header information for the frame decoder.
*
* <p>If the <tt>contentLength</tt> is unknown, you must write the actual length after adding
* the contents as an integer to position <tt>0</tt>!
*
* @param allocator byte buffer allocator to use
* @param id {@link NettyMessag... | 3.68 |
flink_SqlTimestampParser_parseField | /**
* Static utility to parse a field of type Timestamp from a byte sequence that represents text
* characters (such as when read from a file stream).
*
* @param bytes The bytes containing the text data that should be parsed.
* @param startPos The offset to start the parsing.
* @param length The length of the byt... | 3.68 |
hudi_FlinkInMemoryStateIndex_isImplicitWithStorage | /**
* Index needs to be explicitly updated after storage write.
*/
@Override
public boolean isImplicitWithStorage() {
return true;
} | 3.68 |
hbase_MasterObserver_postHasUserPermissions | /**
* Called after checking if user has permissions.
* @param ctx the coprocessor instance's environment
* @param userName the user name
* @param permissions the permission list
*/
default void postHasUserPermissions(ObserverContext<MasterCoprocessorEnvironment> ctx,
String userName, List<Permission> ... | 3.68 |
framework_Page_getCurrent | /**
* Gets the Page to which the current uI belongs. This is automatically
* defined when processing requests to the server. In other cases, (e.g.
* from background threads), the current uI is not automatically defined.
*
* @see UI#getCurrent()
*
* @return the current page instance if available, otherwise
* ... | 3.68 |
framework_Tree_accept | /*
* Uses enhanced server side check
*/
@Override
public boolean accept(DragAndDropEvent dragEvent) {
try {
// must be over tree node and in the middle of it (not top or
// bottom
// part)
TreeTargetDetails eventDetails = (TreeTargetDetails) dragEvent
.getTargetDeta... | 3.68 |
hadoop_SystemErasureCodingPolicies_getReplicationPolicy | /**
* Get the special REPLICATION policy.
*/
public static ErasureCodingPolicy getReplicationPolicy() {
return REPLICATION_POLICY;
} | 3.68 |
flink_CompositeTypeSerializerSnapshot_writeOuterSnapshot | /**
* Writes the outer snapshot, i.e. any information beyond the nested serializers of the outer
* serializer.
*
* <p>The base implementation of this methods writes nothing, i.e. it assumes that the outer
* serializer only has nested serializers and no extra information. Otherwise, if the outer
* serializer conta... | 3.68 |
hbase_HFileReaderImpl_shouldUseHeap | /**
* Whether we use heap or not depends on our intent to cache the block. We want to avoid
* allocating to off-heap if we intend to cache into the on-heap L1 cache. Otherwise, it's more
* efficient to allocate to off-heap since we can control GC ourselves for those. So our decision
* here breaks down as follows: <... | 3.68 |
morf_OracleDialect_buildSQLToStopTracing | /**
* @see org.alfasoftware.morf.jdbc.SqlDialect#buildSQLToStopTracing()
*/
@Override
public List<String> buildSQLToStopTracing() {
return Arrays.asList("ALTER SESSION SET EVENTS '10046 TRACE NAME CONTEXT OFF'");
} | 3.68 |
framework_DefaultFieldGroupFieldFactory_anySelect | /**
* @since 7.4
* @param fieldType
* the type of the field
* @return true if any AbstractSelect can be assigned to the field
*/
@SuppressWarnings("rawtypes")
protected boolean anySelect(Class<? extends Field> fieldType) {
return anyField(fieldType) || fieldType == AbstractSelect.class;
} | 3.68 |
hudi_FlinkConsistentBucketUpdateStrategy_patchFileIdToRecords | /**
* Rewrite the first record with given fileID
*/
private void patchFileIdToRecords(List<HoodieRecord> records, String fileId) {
HoodieRecord first = records.get(0);
HoodieRecord record = new HoodieAvroRecord<>(first.getKey(), (HoodieRecordPayload) first.getData(), first.getOperation());
HoodieRecordLocation ... | 3.68 |
hbase_RestoreTool_checkAndCreateTable | /**
* Prepare the table for bulkload, most codes copied from {@code createTable} method in
* {@code BulkLoadHFilesTool}.
* @param conn connection
* @param targetTableName target table name
* @param regionDirList region directory list
* @param htd table descriptor
* @param truncateIfE... | 3.68 |
hadoop_HttpReferrerAuditHeader_build | /**
* Build.
* @return an HttpReferrerAuditHeader
*/
public HttpReferrerAuditHeader build() {
return new HttpReferrerAuditHeader(this);
} | 3.68 |
framework_ConnectorBundleLoader_notice | // Not using Vaadin notifications (#14597)
private void notice(String productName) {
if (notice == null) {
notice = new HTML();
notice.addClickHandler(event -> notice.removeFromParent());
notice.addTouchStartHandler(event -> notice.removeFromParent());
}
String msg = notice.getText()... | 3.68 |
AreaShop_GithubUpdateCheck_getLatestVersion | /**
* Get the latest version.
* @return Latest version of the plugin (if checking is complete)
*/
public String getLatestVersion() {
return latestVersion;
} | 3.68 |
framework_ConnectorTracker_setWritingResponse | /**
* Sets the current response write status. Connectors can not be marked as
* dirty when the response is written.
* <p>
* This method has a side-effect of incrementing the sync id by one (see
* {@link #getCurrentSyncId()}), if {@link #isWritingResponse()} returns
* <code>true</code> and <code>writingResponse</c... | 3.68 |
hadoop_Configured_getConf | // inherit javadoc
@Override
public Configuration getConf() {
return conf;
} | 3.68 |
hadoop_ConfigurationWithLogging_getBoolean | /**
* See {@link Configuration#getBoolean(String, boolean)}.
*/
@Override
public boolean getBoolean(String name, boolean defaultValue) {
boolean value = super.getBoolean(name, defaultValue);
log.info("Got {} = '{}' (default '{}')", name, value, defaultValue);
return value;
} | 3.68 |
flink_MultiShotLatch_trigger | /** Fires the latch. Code that is blocked on {@link #await()} will now return. */
public void trigger() {
synchronized (lock) {
triggered = true;
lock.notifyAll();
}
} | 3.68 |
flink_OneInputTransformation_getStateKeySelector | /**
* Returns the {@code KeySelector} that must be used for partitioning keyed state in this
* Operation.
*
* @see #setStateKeySelector
*/
public KeySelector<IN, ?> getStateKeySelector() {
return stateKeySelector;
} | 3.68 |
shardingsphere-elasticjob_JobConfiguration_timeZone | /**
* time zone.
*
* @param timeZone the time zone
* @return job configuration builder
*/
public Builder timeZone(final String timeZone) {
if (null != timeZone) {
this.timeZone = timeZone;
}
return this;
} | 3.68 |
pulsar_TopicEventsDispatcher_removeTopicEventListener | /**
* Removes listeners.
* @param listeners
*/
public void removeTopicEventListener(TopicEventsListener... listeners) {
Objects.requireNonNull(listeners);
Arrays.stream(listeners)
.filter(x -> x != null)
.forEach(topicEventListeners::remove);
} | 3.68 |
shardingsphere-elasticjob_GuaranteeService_executeInLeaderForLastStarted | /**
* Invoke doBeforeJobExecutedAtLastStarted method once after last started.
*
* @param listener AbstractDistributeOnceElasticJobListener instance
* @param shardingContexts sharding contexts
*/
public void executeInLeaderForLastStarted(final AbstractDistributeOnceElasticJobListener listener,
... | 3.68 |
hadoop_BondedS3AStatisticsContext_getInstanceStatistics | /**
* The filesystem statistics: know this is thread-local.
* @return FS statistics.
*/
private FileSystem.Statistics getInstanceStatistics() {
return statisticsSource.getInstanceStatistics();
} | 3.68 |
flink_HiveParserUtils_projectNonColumnEquiConditions | /**
* Push any equi join conditions that are not column references as Projections on top of the
* children.
*/
public static RexNode projectNonColumnEquiConditions(
RelFactories.ProjectFactory factory,
RelNode[] inputRels,
List<RexNode> leftJoinKeys,
List<RexNode> rightJoinKeys,
... | 3.68 |
dubbo_ReferenceBean_getObject | /**
* Create bean instance.
*
* <p></p>
* Why we need a lazy proxy?
*
* <p/>
* When Spring searches beans by type, if Spring cannot determine the type of a factory bean, it may try to initialize it.
* The ReferenceBean is also a FactoryBean.
* <br/>
* (This has already been resolved by decorating the BeanDefi... | 3.68 |
hbase_TaskMonitor_get | /**
* Get singleton instance. TODO this would be better off scoped to a single daemon
*/
public static synchronized TaskMonitor get() {
if (instance == null) {
instance = new TaskMonitor(HBaseConfiguration.create());
}
return instance;
} | 3.68 |
hbase_BucketAllocator_freeBlock | /**
* Free a block with the offset
* @param offset block's offset
* @return size freed
*/
public synchronized int freeBlock(long offset, int length) {
int bucketNo = (int) (offset / bucketCapacity);
assert bucketNo >= 0 && bucketNo < buckets.length;
Bucket targetBucket = buckets[bucketNo];
bucketSizeInfos[t... | 3.68 |
flink_CheckpointConfig_setCheckpointIdOfIgnoredInFlightData | /**
* Setup the checkpoint id for which the in-flight data will be ignored for all operators in
* case of the recovery from this checkpoint.
*
* @param checkpointIdOfIgnoredInFlightData Checkpoint id for which in-flight data should be
* ignored.
* @see #setCheckpointIdOfIgnoredInFlightData
*/
@PublicEvolving... | 3.68 |
flink_TimeUtils_parseDuration | /**
* Parse the given string to a java {@link Duration}. The string is in format "{length
* value}{time unit label}", e.g. "123ms", "321 s". If no time unit label is specified, it will
* be considered as milliseconds.
*
* <p>Supported time unit labels are:
*
* <ul>
* <li>DAYS: "d", "day"
* <li>HOURS: "h", ... | 3.68 |
flink_HiveParserCalcitePlanner_genJoinLogicalPlan | // Generate Join Logical Plan Relnode by walking through the join AST.
private RelNode genJoinLogicalPlan(
HiveParserASTNode joinParseTree, Map<String, RelNode> aliasToRel)
throws SemanticException {
RelNode leftRel = null;
RelNode rightRel = null;
JoinType hiveJoinType;
if (joinParseTr... | 3.68 |
morf_ColumnTypeBean_getWidth | /**
* @return the width
*/
@Override
public int getWidth() {
return width;
} | 3.68 |
hbase_HBaseFsckRepair_closeRegionSilentlyAndWait | /**
* Contacts a region server and waits up to hbase.hbck.close.timeout ms (default 120s) to close
* the region. This bypasses the active hmaster.
*/
public static void closeRegionSilentlyAndWait(Connection connection, ServerName server,
RegionInfo region) throws IOException, InterruptedException {
long timeout ... | 3.68 |
druid_DruidPooledConnection_getVariables | /**
* @since 1.0.28
*/
public Map<String, Object> getVariables() {
return this.holder.variables;
} | 3.68 |
hbase_SpaceLimitSettings_validateSizeLimit | // Helper function to validate sizeLimit
private void validateSizeLimit(long sizeLimit) {
if (sizeLimit < 0L) {
throw new IllegalArgumentException("Size limit must be a non-negative value.");
}
} | 3.68 |
framework_GenericFontIcon_getMIMEType | /*
* (non-Javadoc)
*
* @see com.vaadin.server.Resource#getMIMEType()
*/
@Override
public String getMIMEType() {
throw new UnsupportedOperationException(FontIcon.class.getSimpleName()
+ " should not be used where a MIME type is needed.");
} | 3.68 |
flink_BlockInfo_getRecordCount | /**
* Returns the recordCount.
*
* @return the recordCount
*/
public long getRecordCount() {
return this.recordCount;
} | 3.68 |
framework_VCalendar_isEventMoveAllowed | /**
* Is moving an event allowed.
*/
public boolean isEventMoveAllowed() {
return eventMoveAllowed;
} | 3.68 |
hudi_HoodieTableFactory_setupSortOptions | /**
* Sets up the table exec sort options.
*/
private void setupSortOptions(Configuration conf, ReadableConfig contextConfig) {
if (contextConfig.getOptional(TABLE_EXEC_SORT_MAX_NUM_FILE_HANDLES).isPresent()) {
conf.set(TABLE_EXEC_SORT_MAX_NUM_FILE_HANDLES,
contextConfig.get(TABLE_EXEC_SORT_MAX_NUM_FILE... | 3.68 |
flink_FlinkContainersSettings_fullConfiguration | /**
* Sets the {@code flinkConfiguration} value to {@code config} and returns a reference to
* this Builder enabling method chaining.
*
* @param <T> the type parameter
* @param config The {@code config} to set.
* @return A reference to this Builder.
*/
public <T> Builder fullConfiguration(Configuration config) {... | 3.68 |
hadoop_ContentCounts_getSnapshotableDirectoryCount | // Get the number of snapshottable directories.
public long getSnapshotableDirectoryCount() {
return contents.get(Content.SNAPSHOTTABLE_DIRECTORY);
} | 3.68 |
framework_VaadinPortletResponse_getPortletResponse | /**
* Gets the original, unwrapped portlet response.
*
* @return the unwrapped portlet response
*/
public PortletResponse getPortletResponse() {
return response;
} | 3.68 |
querydsl_DateTimeExpression_dayOfMonth | /**
* Create a day of month expression (range 1-31)
*
* @return day of month
*/
public NumberExpression<Integer> dayOfMonth() {
if (dayOfMonth == null) {
dayOfMonth = Expressions.numberOperation(Integer.class, Ops.DateTimeOps.DAY_OF_MONTH, mixin);
}
return dayOfMonth;
} | 3.68 |
flink_DynamicSourceUtils_convertDataStreamToRel | /**
* Converts a given {@link DataStream} to a {@link RelNode}. It adds helper projections if
* necessary.
*/
public static RelNode convertDataStreamToRel(
boolean isBatchMode,
ReadableConfig config,
FlinkRelBuilder relBuilder,
ContextResolvedTable contextResolvedTable,
DataSt... | 3.68 |
flink_TypeSerializerSnapshotSerializationUtil_write | /**
* Binary format layout of a written serializer snapshot is as follows:
*
* <ul>
* <li>1. Format version of this util.
* <li>2. Name of the TypeSerializerSnapshot class.
* <li>3. The version of the TypeSerializerSnapshot's binary format.
* <li>4. The actual serializer snapshot data.
* </ul>
*/
@Supp... | 3.68 |
hudi_TimelineUtils_handleHollowCommitIfNeeded | /**
* Handles hollow commit as per {@link HoodieCommonConfig#INCREMENTAL_READ_HANDLE_HOLLOW_COMMIT}
* and return filtered or non-filtered timeline for incremental query to run against.
*/
public static HoodieTimeline handleHollowCommitIfNeeded(HoodieTimeline completedCommitTimeline,
HoodieTableMetaClient metaCli... | 3.68 |
graphhopper_AngleCalc_calcOrientation | /**
* Return orientation of line relative to east.
* <p>
*
* @param exact If false the atan gets calculated faster, but it might contain small errors
* @return Orientation in interval -pi to +pi where 0 is east
*/
public double calcOrientation(double lat1, double lon1, double lat2, double lon2, boolean exact) {
... | 3.68 |
framework_DesignContext_setCustomAttribute | /**
* Sets a custom attribute not handled by the component. These attributes
* are directly written to the component tag.
*
* @since 7.7
* @param component
* the component to set the attribute for
* @param attribute
* the attribute to set
* @param value
* the value of the attr... | 3.68 |
hadoop_AbfsOutputStreamStatisticsImpl_blockAllocated | /**
* Increment the counter to indicate a block has been allocated.
*/
@Override
public void blockAllocated() {
blocksAllocated.incrementAndGet();
} | 3.68 |
hbase_RSGroupAdminClient_removeRSGroup | /**
* Removes RegionServer group associated with the given name.
*/
public void removeRSGroup(String name) throws IOException {
RemoveRSGroupRequest request = RemoveRSGroupRequest.newBuilder().setRSGroupName(name).build();
try {
stub.removeRSGroup(null, request);
} catch (ServiceException e) {
throw Pro... | 3.68 |
framework_DesignContext_mapCaption | /**
* Creates a mapping between the given caption and the component. Returns
* true if caption was already mapped to some component.
*
* Note that unlike mapGlobalId, if some component already has the given
* caption, the caption is not cleared from the component. This allows
* non-unique captions. However, only ... | 3.68 |
flink_FlinkContainersSettings_numSlotsPerTaskManager | /**
* Sets the {@code numSlotsPerTaskManager} and returns a reference to this Builder enabling
* method chaining. It also adds this property into the {@code flinkConfiguration} field.
*
* @param numSlotsPerTaskManager The {@code numSlotsPerTaskManager} to set.
* @return A reference to this Builder.
*/
public Buil... | 3.68 |
framework_DragHandle_addTo | /**
* Adds this drag handle to an HTML element.
*
* @param elem
* an element
*/
public void addTo(Element elem) {
removeFromParent();
parent = elem;
parent.appendChild(element);
} | 3.68 |
flink_FlinkContainersSettings_getCheckpointPath | /**
* Gets checkpoint path.
*
* @return The checkpoint path.
*/
public String getCheckpointPath() {
return checkpointPath;
} | 3.68 |
graphhopper_Path_getDistance | /**
* @return distance in meter
*/
public double getDistance() {
return distance;
} | 3.68 |
framework_Table_removeColumnReorderListener | /**
* Removes a column reorder listener from the Table.
*
* @param listener
* The listener to remove
*/
public void removeColumnReorderListener(ColumnReorderListener listener) {
removeListener(TableConstants.COLUMN_REORDER_EVENT_ID,
ColumnReorderEvent.class, listener);
} | 3.68 |
flink_JoinHintsResolver_matchIdentifier | /**
* Check whether the given hint option matches the table qualified names. For convenience, we
* follow a simple rule: the matching is successful if the option is the suffix of the table
* qualified names.
*/
private boolean matchIdentifier(String option, String tableIdentifier) {
String[] optionNames = optio... | 3.68 |
zxing_URIParsedResult_isPossiblyMaliciousURI | /**
* @return true if the URI contains suspicious patterns that may suggest it intends to
* mislead the user about its true nature
* @deprecated see {@link URIResultParser#isPossiblyMaliciousURI(String)}
*/
@Deprecated
public boolean isPossiblyMaliciousURI() {
return URIResultParser.isPossiblyMaliciousURI(uri);
... | 3.68 |
hmily_Binder_of | /**
* Of binder.
*
* @param source the source
* @return the binder
*/
public static Binder of(final ConfigPropertySource source) {
return new Binder(source);
} | 3.68 |
flink_HadoopConfigLoader_getOrLoadHadoopConfig | /** get the loaded Hadoop config (or fall back to one loaded from the classpath). */
public org.apache.hadoop.conf.Configuration getOrLoadHadoopConfig() {
org.apache.hadoop.conf.Configuration hadoopConfig = this.hadoopConfig;
if (hadoopConfig == null) {
if (flinkConfig != null) {
hadoopConfi... | 3.68 |
flink_Router_removePathPattern | /** Removes the route specified by the path pattern. */
public void removePathPattern(String pathPattern) {
for (MethodlessRouter<T> router : routers.values()) {
router.removePathPattern(pathPattern);
}
anyMethodRouter.removePathPattern(pathPattern);
} | 3.68 |
hadoop_LocalityMulticastAMRMProxyPolicy_routeNodeRequestIfNeeded | /**
* When certain subcluster is too loaded, reroute Node requests going there.
*
* @param targetId current subClusterId where request is sent
* @param maxThreshold threshold for Pending count
* @param activeAndEnabledSCs list of active sc
* @return subClusterId target sc id
*/
protected SubClusterId routeNodeRe... | 3.68 |
hadoop_ServerWebApp_isSslEnabled | /**
*
*/
public boolean isSslEnabled() {
return Boolean.parseBoolean(
System.getProperty(getName() + SSL_ENABLED, "false"));
} | 3.68 |
hbase_ZkSplitLogWorkerCoordination_init | /**
* Override setter from {@link SplitLogWorkerCoordination}
*/
@Override
public void init(RegionServerServices server, Configuration conf, TaskExecutor splitExecutor,
SplitLogWorker worker) {
this.server = server;
this.worker = worker;
this.splitTaskExecutor = splitExecutor;
maxConcurrentTasks =
conf.... | 3.68 |
framework_CheckBoxGroup_setItemDescriptionGenerator | /**
* Sets the description generator that is used for generating descriptions
* for items. Description is shown as a tooltip when hovering on
* corresponding element. If the generator returns {@code null}, no tooltip
* is shown.
*
*
* @param descriptionGenerator
* the item description generator to se... | 3.68 |
hudi_RDDConsistentBucketBulkInsertPartitioner_generateFileIdPfx | /**
* Initialize fileIdPfx for each data partition. Specifically, the following fields is constructed:
* - fileIdPfxList: the Nth element corresponds to the Nth data partition, indicating its fileIdPfx
* - partitionToFileIdPfxIdxMap (return value): (table partition) -> (fileIdPfx -> idx) mapping
* - doAppend: repre... | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.