name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
hadoop_WrappedFailoverProxyProvider_close | /**
* Close the proxy,
*/
@Override
public synchronized void close() throws IOException {
proxyProvider.close();
} | 3.68 |
flink_FromClasspathEntryClassInformationProvider_create | /**
* Creates a {@code FromClasspathEntryClassInformationProvider} based on the passed job class
* and classpath.
*
* @param jobClassName The job's class name.
* @param classpath The classpath the job class should be part of.
* @return The {@code FromClasspathEntryClassInformationProvider} instances collecting th... | 3.68 |
hbase_Scan_isRaw | /** Returns True if this Scan is in "raw" mode. */
public boolean isRaw() {
byte[] attr = getAttribute(RAW_ATTR);
return attr == null ? false : Bytes.toBoolean(attr);
} | 3.68 |
framework_AbstractEmbedded_setAlternateText | /**
* Sets this component's alternate text that can be presented instead of the
* component's normal content for accessibility purposes.
*
* @param altText
* A short, human-readable description of this component's
* content.
*/
public void setAlternateText(String altText) {
getState().a... | 3.68 |
hudi_EmbeddedTimelineService_getRemoteFileSystemViewConfig | /**
* Retrieves proper view storage configs for remote clients to access this service.
*/
public FileSystemViewStorageConfig getRemoteFileSystemViewConfig() {
FileSystemViewStorageType viewStorageType = writeConfig.getClientSpecifiedViewStorageConfig()
.shouldEnableBackupForRemoteFileSystemView()
? File... | 3.68 |
hadoop_CDFPiecewiseLinearRandomGenerator_valueAt | /**
* TODO This code assumes that the empirical minimum resp. maximum is the
* epistomological minimum resp. maximum. This is probably okay for the
* minimum, because that likely represents a task where everything went well,
* but for the maximum we may want to develop a way of extrapolating past the
* maximum.
*... | 3.68 |
hadoop_DatanodeAdminProperties_getHostName | /**
* Return the host name of the datanode.
* @return the host name of the datanode.
*/
public String getHostName() {
return hostName;
} | 3.68 |
morf_DatabaseUpgradeTableContribution_upgradeAuditTable | /**
* @return The Table descriptor of UpgradeAudit
*/
public static Table upgradeAuditTable() {
return table(UPGRADE_AUDIT_NAME)
.columns(
column("upgradeUUID", DataType.STRING, 100).primaryKey(),
column("description", DataType.STRING, 200).nullable(),
column("appliedTime", DataType.DE... | 3.68 |
hudi_HoodieCatalogUtil_createHiveConf | /**
* Returns a new {@code HiveConf}.
*
* @param hiveConfDir Hive conf directory path.
* @return A HiveConf instance.
*/
public static HiveConf createHiveConf(@Nullable String hiveConfDir, org.apache.flink.configuration.Configuration flinkConf) {
// create HiveConf from hadoop configuration with hadoop conf dire... | 3.68 |
rocketmq-connect_CountDownLatch2_await | /**
* Causes the current thread to wait until the latch has counted down to zero, unless the thread is {@linkplain
* Thread#interrupt interrupted}, or the specified waiting time elapses.
*
* <p>If the current count is zero then this method returns immediately
* with the value {@code true}.
*
* <p>If the current ... | 3.68 |
zxing_FinderPatternFinder_selectBestPatterns | /**
* @return the 3 best {@link FinderPattern}s from our list of candidates. The "best" are
* those have similar module size and form a shape closer to a isosceles right triangle.
* @throws NotFoundException if 3 such finder patterns do not exist
*/
private FinderPattern[] selectBestPatterns() throws NotFou... | 3.68 |
open-banking-gateway_Xs2aConsentInfo_isOauth2AuthenticationPreStep | /**
* Is the current consent in OAUTH-Pre-step (authentication) mode.
*/
public boolean isOauth2AuthenticationPreStep(Xs2aContext ctx) {
return ctx.isOauth2PreStepNeeded() || ctx.isEmbeddedPreAuthNeeded();
} | 3.68 |
rocketmq-connect_PluginUtils_isAliasUnique | /**
* Verify whether a given plugin's alias matches another alias in a collection of plugins.
*
* @param alias the plugin descriptor to test for alias matching.
* @param plugins the collection of plugins to test against.
* @param <U> the plugin type.
* @return false if a match was found in the collection, o... | 3.68 |
hudi_SparkInternalSchemaConverter_convertAndPruneStructTypeToInternalSchema | /**
* Convert Spark schema to Hudi internal schema, and prune fields.
* Fields without IDs are kept and assigned fallback IDs.
*
* @param sparkSchema a pruned spark schema
* @param originSchema a internal schema for hoodie table
* @return a pruned internal schema for the provided spark schema
*/
public static In... | 3.68 |
hadoop_FSDirSatisfyStoragePolicyOp_satisfyStoragePolicy | /**
* Satisfy storage policy function which will add the entry to SPS call queue
* and will perform satisfaction async way.
*
* @param fsd
* fs directory
* @param bm
* block manager
* @param src
* source path
* @param logRetryCache
* whether to record RPC ids in editlog fo... | 3.68 |
hbase_ServerManager_countOfRegionServers | /** Returns the count of active regionservers */
public int countOfRegionServers() {
// Presumes onlineServers is a concurrent map
return this.onlineServers.size();
} | 3.68 |
flink_LocatableInputSplitAssigner_addInputSplit | /**
* Adds a single input split
*
* @param split The input split to add
*/
public void addInputSplit(LocatableInputSplitWithCount split) {
int localCount = split.getLocalCount();
if (minLocalCount == -1) {
// first split to add
this.minLocalCount = localCount;
this.elementCycleCount... | 3.68 |
framework_DefaultConnectionStateHandler_showDialog | /**
* Called when the reconnect dialog should be shown. This is typically when
* N seconds has passed since a problem with the connection has been
* detected
*/
protected void showDialog() {
reconnectDialog.setReconnecting(true);
reconnectDialog.show(connection);
// We never want to show loading indica... | 3.68 |
framework_VaadinService_setSessionLock | /**
* Associates the given lock with this service and the given wrapped
* session. This method should not be called more than once when the lock is
* initialized for the session.
*
* @see #getSessionLock(WrappedSession)
* @param wrappedSession
* The wrapped session the lock is associated with
* @para... | 3.68 |
framework_VCalendarPanel_onKeyDown | /*
* (non-Javadoc)
*
* @see
* com.google.gwt.event.dom.client.KeyDownHandler#onKeyDown(com.google.gwt
* .event.dom.client.KeyDownEvent)
*/
@Override
public void onKeyDown(KeyDownEvent event) {
handleKeyPress(event);
} | 3.68 |
hadoop_ResourceRequest_relaxLocality | /**
* Set the <code>relaxLocality</code> of the request.
* @see ResourceRequest#setRelaxLocality(boolean)
* @param relaxLocality <code>relaxLocality</code> of the request
* @return {@link ResourceRequestBuilder}
*/
@Public
@Stable
public ResourceRequestBuilder relaxLocality(boolean relaxLocality) {
resourceReque... | 3.68 |
flink_LongHashPartition_finalizeBuildPhase | /**
* After build phase.
*
* @return build spill return buffer, if have spilled, it returns the current write buffer,
* because it was used all the time in build phase, so it can only be returned at this time.
*/
int finalizeBuildPhase(IOManager ioAccess, FileIOChannel.Enumerator probeChannelEnumerator)
... | 3.68 |
hadoop_DiskBalancerWorkStatus_getResult | /**
* Returns result.
*
* @return long
*/
public Result getResult() {
return result;
} | 3.68 |
hbase_ParseFilter_reduce | /**
* This function is called while parsing the filterString and an operator is parsed
* <p>
* @param operatorStack the stack containing the operators and parenthesis
* @param filterStack the stack containing the filters
* @param operator the operator found while parsing the filterString
*/
public void red... | 3.68 |
Activiti_DelegateInvocation_getInvocationParameters | /**
* @return an array of invocation parameters (null if the invocation takes no parameters)
*/
public Object[] getInvocationParameters() {
return invocationParameters;
} | 3.68 |
pulsar_ConsumerImpl_notifyPendingReceivedCallback | /**
* Notify waiting asyncReceive request with the received message.
*
* @param message
*/
void notifyPendingReceivedCallback(final Message<T> message, Exception exception) {
if (pendingReceives.isEmpty()) {
return;
}
// fetch receivedCallback from queue
final CompletableFuture<Message<T>> ... | 3.68 |
hbase_ScannerContext_mayHaveMoreCellsInRow | /**
* @return true when we have more cells for the current row. This usually because we have reached
* a limit in the middle of a row
*/
boolean mayHaveMoreCellsInRow() {
return scannerState == NextState.SIZE_LIMIT_REACHED_MID_ROW
|| scannerState == NextState.TIME_LIMIT_REACHED_MID_ROW
|| scannerSt... | 3.68 |
pulsar_MetadataStore_getDefaultMetadataCacheConfig | /**
* Returns the default metadata cache config.
*
* @return default metadata cache config
*/
default MetadataCacheConfig getDefaultMetadataCacheConfig() {
return MetadataCacheConfig.builder().build();
} | 3.68 |
flink_LogicalTypeMerging_findAvgAggType | /** Finds the result type of a decimal average aggregation. */
public static LogicalType findAvgAggType(LogicalType argType) {
final LogicalType resultType;
if (argType.is(DECIMAL)) {
// a hack to make legacy types possible until we drop them
if (argType instanceof LegacyTypeInformationType) {
... | 3.68 |
AreaShop_GithubUpdateCheck_isChecking | /**
* Check if an update check is running.
* @return true if an update check is running
*/
public boolean isChecking() {
return checking;
} | 3.68 |
flink_MutableHashTable_buildBloomFilterForBucket | /**
* Set all the bucket memory except bucket header as the bit set of bloom filter, and use hash
* code of build records to build bloom filter.
*/
final void buildBloomFilterForBucket(
int bucketInSegmentPos, MemorySegment bucket, HashPartition<BT, PT> p) {
final int count = bucket.getShort(bucketInSegm... | 3.68 |
dubbo_ConfigurableMetadataServiceExporter_setMetadataService | // for unit test
public void setMetadataService(MetadataServiceDelegation metadataService) {
this.metadataService = metadataService;
} | 3.68 |
framework_AbstractComponent_setCaptionAsHtml | /**
* Sets whether the caption is rendered as HTML.
* <p>
* If set to true, the captions are rendered in the browser as HTML and the
* developer is responsible for ensuring no harmful HTML is used. If set to
* false, the caption is rendered in the browser as plain text.
* <p>
* The default is false, i.e. to rend... | 3.68 |
hbase_AvlUtil_get | /**
* Return the node that matches the specified key or null in case of node not found.
* @param root the current root of the tree
* @param key the key for the node we are trying to find
* @param keyComparator the comparator to use to match node and key
* @return the node that matches the specif... | 3.68 |
AreaShop_RentRegion_isRenter | /**
* Check if a player is the renter of this region.
* @param player Player to check
* @return true if this player rents this region, otherwise false
*/
public boolean isRenter(Player player) {
return player != null && isRenter(player.getUniqueId());
} | 3.68 |
streampipes_SpOpcUaClient_connect | /***
* Establishes appropriate connection to OPC UA endpoint depending on the {@link SpOpcUaClient} instance
*
* @throws UaException An exception occurring during OPC connection
*/
public void connect()
throws UaException, ExecutionException, InterruptedException, SpConfigurationException, URISyntaxException {
... | 3.68 |
hadoop_HsController_singleTaskCounter | /*
* (non-Javadoc)
* @see org.apache.hadoop.mapreduce.v2.app.webapp.AppController#singleTaskCounter()
*/
@Override
public void singleTaskCounter() throws IOException{
super.singleTaskCounter();
} | 3.68 |
hudi_HoodieMergeHandle_initializeIncomingRecordsMap | /**
* Initialize a spillable map for incoming records.
*/
protected void initializeIncomingRecordsMap() {
try {
// Load the new records in a map
long memoryForMerge = IOUtils.getMaxMemoryPerPartitionMerge(taskContextSupplier, config);
LOG.info("MaxMemoryPerPartitionMerge => " + memoryForMerge);
this... | 3.68 |
framework_VaadinService_writeUncachedStringResponse | /**
* Writes the given string as a response with headers to prevent caching and
* using the given content type.
*
* @param response
* The response reference
* @param contentType
* The content type of the response
* @param responseString
* The actual response
* @throws IOExcept... | 3.68 |
hadoop_LongLong_set | /** Set the values. */
LongLong set(long d0, long d1) {
this.d0 = d0;
this.d1 = d1;
return this;
} | 3.68 |
framework_AbstractOrderedLayout_getCustomAttributes | /*
* (non-Javadoc)
*
* @see com.vaadin.ui.AbstractComponent#getCustomAttributes()
*/
@Override
protected Collection<String> getCustomAttributes() {
Collection<String> customAttributes = super.getCustomAttributes();
customAttributes.add("margin");
customAttributes.add("margin-left");
customAttributes... | 3.68 |
hadoop_DiffList_unmodifiableList | /**
* Returns an unmodifiable diffList.
* @param diffs DiffList
* @param <T> Type of the object in the the diffList
* @return Unmodifiable diffList
*/
static <T extends Comparable<Integer>> DiffList<T> unmodifiableList(
DiffList<T> diffs) {
return new DiffList<T>() {
@Override
public T get(int i) {
... | 3.68 |
flink_LargeRecordHandler_close | /**
* Closes all structures and deletes all temporary files. Even in the presence of failures, this
* method will try and continue closing files and deleting temporary files.
*
* @throws IOException Thrown if an error occurred while closing/deleting the files.
*/
public void close() throws IOException {
// we... | 3.68 |
hadoop_ContainerStatus_getCapability | /**
* Get the <code>Resource</code> allocated to the container.
* @return <code>Resource</code> allocated to the container
*/
@Public
@Unstable
public Resource getCapability() {
throw new UnsupportedOperationException(
"subclass must implement this method");
} | 3.68 |
morf_DeleteStatement_copyOnWriteOrMutate | /**
* Either shallow copies and mutates the result, returning it,
* or mutates the statement directly, depending on
* {@link AliasedField#immutableDslEnabled()}.
*
* TODO for removal along with mutable behaviour.
*
* @param transform A transform which modifies the shallow copy builder.
* @param mutator Code whi... | 3.68 |
hbase_HRegion_getReadPoint | /** Returns readpoint considering given IsolationLevel. Pass {@code null} for default */
public long getReadPoint(IsolationLevel isolationLevel) {
if (isolationLevel != null && isolationLevel == IsolationLevel.READ_UNCOMMITTED) {
// This scan can read even uncommitted transactions
return Long.MAX_VALUE;
}
... | 3.68 |
flink_PatternStream_inEventTime | /** Sets the time characteristic to event time. */
public PatternStream<T> inEventTime() {
return new PatternStream<>(builder.inEventTime());
} | 3.68 |
hadoop_CoderUtil_getNullIndexes | /**
* Get indexes array for items marked as null, either erased or
* not to read.
* @return indexes array
*/
static <T> int[] getNullIndexes(T[] inputs) {
int[] nullIndexes = new int[inputs.length];
int idx = 0;
for (int i = 0; i < inputs.length; i++) {
if (inputs[i] == null) {
nullIndexes[idx++] = ... | 3.68 |
framework_GridDropEvent_getDropLocation | /**
* Get the location of the drop within the row.
* <p>
* <em>NOTE: the location will be {@link DropLocation#EMPTY} if:
* <ul>
* <li>dropped on an empty grid</li>
* <li>dropping on rows was not possible because of
* {@link DropMode#ON_GRID } was used</li>
* <li>{@link DropMode#ON_TOP} is used and the drop happ... | 3.68 |
hbase_CellUtil_matchingRowColumn | /** Compares the row and column of two keyvalues for equality */
public static boolean matchingRowColumn(final Cell left, final Cell right) {
short lrowlength = left.getRowLength();
short rrowlength = right.getRowLength();
// match length
if (lrowlength != rrowlength) {
return false;
}
byte lfamlength ... | 3.68 |
framework_VaadinService_getSystemMessages | /**
* Gets the system message to use for a specific locale. This method may
* also be implemented to use information from current instances of various
* objects, which means that this method might return different values for
* the same locale under different circumstances.
*
* @param locale
* the desi... | 3.68 |
flink_CheckpointStatsCounts_incrementRestoredCheckpoints | /** Increments the number of restored checkpoints. */
void incrementRestoredCheckpoints() {
numRestoredCheckpoints++;
} | 3.68 |
flink_CheckpointProperties_forceCheckpoint | /**
* Returns whether the checkpoint should be forced.
*
* <p>Forced checkpoints ignore the configured maximum number of concurrent checkpoints and
* minimum time between checkpoints. Furthermore, they are not subsumed by more recent
* checkpoints as long as they are pending.
*
* @return <code>true</code> if the... | 3.68 |
flink_CommonExecTableSourceScan_createSourceFunctionTransformation | /**
* Adopted from {@link StreamExecutionEnvironment#addSource(SourceFunction, String,
* TypeInformation)} but with custom {@link Boundedness}.
*
* @deprecated This method relies on the {@link
* org.apache.flink.streaming.api.functions.source.SourceFunction} API, which is due to be
* removed.
*/
@Depreca... | 3.68 |
hudi_HoodieRowDataCreateHandle_canWrite | /**
* Returns {@code true} if this handle can take in more writes. else {@code false}.
*/
public boolean canWrite() {
return fileWriter.canWrite();
} | 3.68 |
graphhopper_BitUtil_toUnsignedLong | /**
* This method handles the specified (potentially negative) int as unsigned bit representation
* and returns the positive converted long.
*/
public static long toUnsignedLong(int x) {
return ((long) x) & 0xFFFF_FFFFL;
} | 3.68 |
hadoop_LocalJobOutputFiles_getSpillIndexFileForWrite | /**
* Create a local map spill index file name.
*
* @param spillNumber the number
* @param size the size of the file
*/
public Path getSpillIndexFileForWrite(int spillNumber, long size) throws IOException {
String path = String
.format(SPILL_INDEX_FILE_FORMAT_STRING, TASKTRACKER_OUTPUT, spillNum... | 3.68 |
morf_OracleMetaDataProvider_getTable | /**
* {@inheritDoc}
*
* <p>The {@link Table} implementation returned may contain {@link Column} implementations
* which evaluate the metadata elements ({@link Column#getType()}, {@link Column#getWidth()}
* etc.) lazily. If the database column type is not supported, this may throw an
* {@link UnexpectedDataTypeEx... | 3.68 |
pulsar_FunctionMetaDataManager_getAllFunctionMetaData | /**
* Get a list of all the meta for every function.
* @return list of function metadata
*/
public synchronized List<FunctionMetaData> getAllFunctionMetaData() {
List<FunctionMetaData> ret = new LinkedList<>();
for (Map<String, Map<String, FunctionMetaData>> i : this.functionMetaDataMap.values()) {
f... | 3.68 |
streampipes_StreamPipesClient_create | /**
* Create a new StreamPipes API client with custom port and HTTPS settings
*
* @param streamPipesHost The hostname of the StreamPipes instance without scheme
* @param streamPipesPort The port of the StreamPipes instance
* @param credentials The credentials object
* @param httpsDisabled Set true if the in... | 3.68 |
zilla_DefaultBufferPool_release | /**
* Releases a slot so it may be used by other streams
* @param slot - Id of a previously acquired slot
*/
@Override
public void release(
int slot)
{
assert used.get(slot);
used.clear(slot);
availableSlots.value++;
poolBuffer.putLongOrdered(usedIndex + (slot << 3), 0L);
} | 3.68 |
flink_LogicalFile_advanceLastCheckpointId | /**
* A logical file may share across checkpoints (especially for shared state). When this logical
* file is used/reused by a checkpoint, update the last checkpoint id that uses this logical
* file.
*
* @param checkpointId the checkpoint that uses this logical file.
*/
public void advanceLastCheckpointId(long che... | 3.68 |
hbase_Addressing_createInetSocketAddressFromHostAndPortStr | /**
* Create a socket address
* @param hostAndPort Formatted as <code><hostname> ':' <port></code>
* @return An InetSocketInstance
*/
public static InetSocketAddress
createInetSocketAddressFromHostAndPortStr(final String hostAndPort) {
return new InetSocketAddress(parseHostname(hostAndPort), parsePo... | 3.68 |
flink_SubtaskStateStats_getPersistedData | /** @return the total number of persisted bytes during the checkpoint. */
public long getPersistedData() {
return persistedData;
} | 3.68 |
hadoop_WritableFactories_getFactory | /**
* Define a factory for a class.
* @param c input c.
* @return a factory for a class.
*/
public static WritableFactory getFactory(Class c) {
return CLASS_TO_FACTORY.get(c);
} | 3.68 |
hadoop_ExternalCall_run | // invoked by ipc handler
@Override
public final Void run() throws IOException {
try {
result = action.run();
sendResponse();
} catch (Throwable t) {
abortResponse(t);
}
return null;
} | 3.68 |
hadoop_CommonAuditContext_createInstance | /**
* Demand invoked to create the instance for this thread.
* @return an instance.
*/
private static CommonAuditContext createInstance() {
CommonAuditContext context = new CommonAuditContext();
context.init();
return context;
} | 3.68 |
hbase_HRegion_flush | /**
* Flush the cache.
* <p>
* When this method is called the cache will be flushed unless:
* <ol>
* <li>the cache is empty</li>
* <li>the region is closed.</li>
* <li>a flush is already in progress</li>
* <li>writes are disabled</li>
* </ol>
* <p>
* This method may block for some time, so it should not be c... | 3.68 |
hadoop_SliderFileSystem_deleteComponentDir | /**
* Deletes the component directory.
*
* @param serviceVersion
* @param compName
* @throws IOException
*/
public void deleteComponentDir(String serviceVersion, String compName)
throws IOException {
Path path = getComponentDir(serviceVersion, compName);
if (fileSystem.exists(path)) {
fileSystem.delet... | 3.68 |
zxing_GridSampler_setGridSampler | /**
* Sets the implementation of GridSampler used by the library. One global
* instance is stored, which may sound problematic. But, the implementation provided
* ought to be appropriate for the entire platform, and all uses of this library
* in the whole lifetime of the JVM. For instance, an Android activity can s... | 3.68 |
flink_PekkoUtils_createDefaultActorSystem | /**
* Creates an actor system with the default config and listening on a random port of the
* localhost.
*
* @return default actor system listening on a random port of the localhost
*/
@VisibleForTesting
public static ActorSystem createDefaultActorSystem() {
return createActorSystem(getDefaultConfig());
} | 3.68 |
flink_BlobServer_getServerSocket | /** Access to the server socket, for testing. */
ServerSocket getServerSocket() {
return this.serverSocket;
} | 3.68 |
hmily_ThreadLocalHmilyContext_remove | /**
* clean threadLocal for gc.
*/
public void remove() {
CURRENT_LOCAL.remove();
} | 3.68 |
hbase_GroupingTableMapper_map | /**
* Extract the grouping columns from value to construct a new key. Pass the new key and value to
* reduce. If any of the grouping columns are not found in the value, the record is skipped.
* @param key The current key.
* @param value The current value.
* @param context The current context.
* @throws IOEx... | 3.68 |
shardingsphere-elasticjob_SensitiveInfoUtils_filterSensitiveIps | /**
* Filter sensitive IP addresses.
*
* @param target IP addresses to be filtered
* @return filtered IP addresses
*/
public static List<String> filterSensitiveIps(final List<String> target) {
final Map<String, String> fakeIpMap = new HashMap<>();
final AtomicInteger step = new AtomicInteger();
return... | 3.68 |
flink_FileCatalogStore_storeCatalog | /**
* Stores the specified catalog in the catalog store.
*
* @param catalogName the name of the catalog
* @param catalog the catalog descriptor to store
* @throws CatalogException if the catalog store is not open or if there is an error storing the
* catalog
*/
@Override
public void storeCatalog(String catal... | 3.68 |
dubbo_AbstractServiceRestMetadataResolver_findRestCapableMethod | /**
* Find the method with the capable for REST from the specified service method and its override method
*
* @param processingEnv {@link ProcessingEnvironment}
* @param serviceType
* @param serviceInterfaceType
* @param serviceMethod
* @return <code>null</code> if can't be found
*/
private ExecutableEle... | 3.68 |
dubbo_BasicJsonWriter_print | /**
* Write the specified text.
*
* @param string the content to write
*/
public IndentingWriter print(String string) {
write(string.toCharArray(), 0, string.length());
return this;
} | 3.68 |
hbase_Bytes_toByteArrays | /**
* Create a byte[][] where first and only entry is <code>column</code>
* @param column operand
* @return A byte array of a byte array where first and only entry is <code>column</code>
*/
public static byte[][] toByteArrays(final byte[] column) {
byte[][] result = new byte[1][];
result[0] = column;
return r... | 3.68 |
flink_DelegationTokenProvider_serviceConfigPrefix | /** Config prefix of the service. */
default String serviceConfigPrefix() {
return String.format("%s.%s", CONFIG_PREFIX, serviceName());
} | 3.68 |
hadoop_HAState_prepareToEnterState | /**
* Method to be overridden by subclasses to prepare to enter a state.
* This method is called <em>without</em> the context being locked,
* and after {@link #prepareToExitState(HAContext)} has been called
* for the previous state, but before {@link #exitState(HAContext)}
* has been called for the previous state.... | 3.68 |
framework_StringToIntegerConverter_convertToModel | /*
* (non-Javadoc)
*
* @see
* com.vaadin.data.util.converter.Converter#convertToModel(java.lang.Object,
* java.lang.Class, java.util.Locale)
*/
@Override
public Integer convertToModel(String value,
Class<? extends Integer> targetType, Locale locale)
throws ConversionException {
Number n = conv... | 3.68 |
AreaShop_GeneralRegion_setup | /**
* Shared setup of all constructors.
*/
public void setup() {
features = new HashMap<>();
} | 3.68 |
hbase_HMaster_shutdown | /**
* Shutdown the cluster. Master runs a coordinated stop of all RegionServers and then itself.
*/
public void shutdown() throws IOException {
TraceUtil.trace(() -> {
if (cpHost != null) {
cpHost.preShutdown();
}
// Tell the servermanager cluster shutdown has been called. This makes it so when M... | 3.68 |
hbase_AsyncAdmin_restoreSnapshot | /**
* Restore the specified snapshot on the original table. (The table must be disabled) If
* 'takeFailSafeSnapshot' is set to true, a snapshot of the current table is taken before
* executing the restore operation. In case of restore failure, the failsafe snapshot will be
* restored. If the restore completes witho... | 3.68 |
hadoop_AbstractConfigurableFederationPolicy_validate | /**
* Overridable validation step for the policy configuration.
*
* @param newPolicyInfo the configuration to test.
*
* @throws FederationPolicyInitializationException if the configuration is not
* valid.
*/
public void validate(WeightedPolicyInfo newPolicyInfo)
throws FederationPolicyInitializatio... | 3.68 |
flink_Schema_primaryKey | /**
* Declares a primary key constraint for a set of given columns. Primary key uniquely
* identify a row in a table. Neither of columns in a primary can be nullable. The primary
* key is informational only. It will not be enforced. It can be used for optimizations. It
* is the data owner's responsibility to ensure... | 3.68 |
flink_RowData_createFieldGetter | /**
* Creates an accessor for getting elements in an internal row data structure at the given
* position.
*
* @param fieldType the element type of the row
* @param fieldPos the element position of the row
*/
static FieldGetter createFieldGetter(LogicalType fieldType, int fieldPos) {
final FieldGetter fieldGet... | 3.68 |
hadoop_TFile_getLocationNear | /**
* Get the location pointing to the beginning of the first key-value pair in
* a compressed block whose byte offset in the TFile is greater than or
* equal to the specified offset.
*
* @param offset
* the user supplied offset.
* @return the location to the corresponding entry; or end() if no such
*... | 3.68 |
flink_MemorySegmentFactory_allocateUnpooledOffHeapMemory | /**
* Allocates some unpooled off-heap memory and creates a new memory segment that represents that
* memory.
*
* @param size The size of the off-heap memory segment to allocate.
* @param owner The owner to associate with the off-heap memory segment.
* @return A new memory segment, backed by unpooled off-heap mem... | 3.68 |
hadoop_AbfsThrottlingInterceptFactory_getInstance | /**
* Returns an instance of AbfsThrottlingIntercept.
*
* @param accountName The account for which we need instance of throttling intercept.
@param abfsConfiguration The object of abfsconfiguration class.
* @return Instance of AbfsThrottlingIntercept.
*/
static synchronized AbfsThrottlingIntercept getInstance(... | 3.68 |
framework_ColorPickerPopup_setSwatchesTabVisible | /**
* Sets the visibility of the Swatches tab.
*
* @param visible
* The visibility of the Swatches tab
*/
public void setSwatchesTabVisible(boolean visible) {
if (visible && !isTabVisible(swatchesTab)) {
tabs.addTab(swatchesTab, "Swatches", null);
checkIfTabsNeeded();
} else if (... | 3.68 |
morf_MySqlDialect_dropPrimaryKey | /**
* ALTER TABLE `XYZ` DROP PRIMARY KEY
*/
private String dropPrimaryKey(String tableName) {
return "ALTER TABLE `" + tableName + "` DROP PRIMARY KEY";
} | 3.68 |
incubator-hugegraph-toolchain_FileMappingController_loadParameter | /**
* TODO: All file mapping share one load paramter now, should be separated
* in actually
*/
@PostMapping("load-parameter")
public void loadParameter(@RequestBody LoadParameter newEntity) {
this.checkLoadParameter(newEntity);
List<FileMapping> mappings = this.service.listAll();
for (FileMapping mappin... | 3.68 |
flink_GenericInMemoryCatalog_isFullPartitionSpec | /** Check if the given partitionSpec is full partition spec for the given table. */
private boolean isFullPartitionSpec(ObjectPath tablePath, CatalogPartitionSpec partitionSpec)
throws TableNotExistException {
CatalogBaseTable baseTable = getTable(tablePath);
if (!(baseTable instanceof CatalogTable)) {... | 3.68 |
hadoop_FilePool_refresh | /**
* (Re)generate cache of input FileStatus objects.
*/
public void refresh() throws IOException {
updateLock.writeLock().lock();
try {
root = new InnerDesc(fs, fs.getFileStatus(path),
new MinFileFilter(conf.getLong(GRIDMIX_MIN_FILE, 128 * 1024 * 1024),
conf.getLong(GRIDMIX_MAX_... | 3.68 |
hbase_CatalogFamilyFormat_getStartCodeColumn | /**
* Returns the column qualifier for server start code column for replicaId
* @param replicaId the replicaId of the region
* @return a byte[] for server start code column qualifier
*/
public static byte[] getStartCodeColumn(int replicaId) {
return replicaId == 0
? HConstants.STARTCODE_QUALIFIER
: Bytes.... | 3.68 |
hadoop_NativeS3FileSystem_initialize | /**
* Always fail to initialize.
* @throws IOException always.
*/
@Override
public void initialize(URI uri, Configuration conf) throws IOException {
super.initialize(uri, conf);
throw new IOException(UNSUPPORTED);
} | 3.68 |
flink_QueryableStateConfiguration_getProxyPortRange | /**
* Returns the port range where the queryable state client proxy can listen. See {@link
* org.apache.flink.configuration.QueryableStateOptions#PROXY_PORT_RANGE
* QueryableStateOptions.PROXY_PORT_RANGE}.
*/
public Iterator<Integer> getProxyPortRange() {
return proxyPortRange;
} | 3.68 |
framework_Payload_getValue | /**
* Gets the value of this payload.
*
* @return value of this payload
*/
public String getValue() {
return value;
} | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.