name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
hadoop_SelectTool_run | /**
* Execute the select operation.
* @param args argument list
* @param out output stream
* @return an exit code
* @throws IOException IO failure
* @throws ExitUtil.ExitException managed failure
*/
public int run(String[] args, PrintStream out)
throws IOException, ExitUtil.ExitException {
final List<Strin... | 3.68 |
querydsl_AbstractGroupExpression_as | /**
* Create an alias for the expression
*
* @return alias expression
*/
public DslExpression<R> as(String alias) {
return as(ExpressionUtils.path(getType(), alias));
} | 3.68 |
framework_HasStyleNames_removeStyleNames | /**
* Removes one or more style names from component. Multiple styles can be
* specified by using multiple parameters.
*
* <p>
* The parameter must be a valid CSS style name. Only user-defined style
* names added with {@link #addStyleName(String) addStyleName()} or
* {@link #setStyleName(String) setStyleName()} ... | 3.68 |
flink_ExceptionHistoryEntry_createGlobal | /** Creates an {@code ExceptionHistoryEntry} that is not based on an {@code Execution}. */
public static ExceptionHistoryEntry createGlobal(
Throwable cause, CompletableFuture<Map<String, String>> failureLabels) {
return new ExceptionHistoryEntry(
cause,
System.currentTimeMillis(),
... | 3.68 |
hbase_StoreScanner_parallelSeek | /**
* Seek storefiles in parallel to optimize IO latency as much as possible
* @param scanners the list {@link KeyValueScanner}s to be read from
* @param kv the KeyValue on which the operation is being requested
*/
private void parallelSeek(final List<? extends KeyValueScanner> scanners, final Cell kv)
thro... | 3.68 |
framework_VCaption_updateCaption | /**
* Updates the caption from UIDL.
*
* This method may only be called when the caption has an owner - otherwise,
* use {@link #updateCaptionWithoutOwner(UIDL, String, boolean, boolean)}.
*
* @return true if the position where the caption should be placed has
* changed
*/
public boolean updateCaption()... | 3.68 |
hbase_TableDescriptorBuilder_isNormalizationEnabled | /**
* Check if normalization enable flag of the table is true. If flag is false then no region
* normalizer won't attempt to normalize this table.
* @return true if region normalization is enabled for this table
**/
@Override
public boolean isNormalizationEnabled() {
return getOrDefault(NORMALIZATION_ENABLED_KEY,... | 3.68 |
hbase_TableDescriptorBuilder_isReadOnly | /**
* Check if the readOnly flag of the table is set. If the readOnly flag is set then the contents
* of the table can only be read from but not modified.
* @return true if all columns in the table should be read only
*/
@Override
public boolean isReadOnly() {
return getOrDefault(READONLY_KEY, Boolean::valueOf, D... | 3.68 |
hadoop_PlacementConstraints_nodePartition | /**
* Constructs a target expression on a node partition. It is satisfied if
* the specified node partition has one of the specified nodePartitions.
*
* @param nodePartitions the set of values that the attribute should take
* values from
* @return the resulting expression on the node attribute
*/
public... | 3.68 |
pulsar_AuthenticationDataHttps_hasDataFromTls | /*
* TLS
*/
@Override
public boolean hasDataFromTls() {
return (certificates != null);
} | 3.68 |
hadoop_DiskBalancerDataNode_getDataNodeIP | /**
* Returns the IP address of this Node.
*
* @return IP Address string
*/
public String getDataNodeIP() {
return dataNodeIP;
} | 3.68 |
hadoop_AbfsHttpOperation_getConnUrl | /**
* Gets the connection url.
* @return url.
*/
URL getConnUrl() {
return connection.getURL();
} | 3.68 |
hudi_IncrSourceHelper_generateQueryInfo | /**
* Find begin and end instants to be set for the next fetch.
*
* @param jssc Java Spark Context
* @param srcBasePath Base path of Hudi source table
* @param numInstantsPerFetch Max Instants per fetch
* @param beginInstant Last Checkpoint String
* @param mi... | 3.68 |
hbase_SnapshotInfo_getSnapshotDescription | /** Returns the snapshot descriptor */
public SnapshotDescription getSnapshotDescription() {
return ProtobufUtil.createSnapshotDesc(this.snapshot);
} | 3.68 |
hadoop_StartupProgress_getCounter | /**
* Returns a counter associated with the specified phase and step. Typical
* usage is to increment a counter within a tight loop. Callers may use this
* method to obtain a counter once and then increment that instance repeatedly
* within a loop. This prevents redundant lookup operations and object
* creation... | 3.68 |
dubbo_Invoker_invoke | // This method will never be called for a legacy invoker.
@Override
default org.apache.dubbo.rpc.Result invoke(org.apache.dubbo.rpc.Invocation invocation)
throws org.apache.dubbo.rpc.RpcException {
return null;
} | 3.68 |
dubbo_Server_start | /**
* start server, bind port
*/
public void start() throws Throwable {
if (!started.compareAndSet(false, true)) {
return;
}
boss = new NioEventLoopGroup(1, new DefaultThreadFactory("qos-boss", true));
worker = new NioEventLoopGroup(0, new DefaultThreadFactory("qos-worker", true));
ServerB... | 3.68 |
framework_TableQuery_writeObject | /**
* Custom writeObject to call rollback() if object is serialized.
*/
private void writeObject(java.io.ObjectOutputStream out)
throws IOException {
try {
rollback();
} catch (SQLException ignored) {
}
out.defaultWriteObject();
} | 3.68 |
flink_EmbeddedRocksDBStateBackend_getDbStoragePaths | /**
* Gets the configured local DB storage paths, or null, if none were configured.
*
* <p>Under these directories on the TaskManager, RocksDB stores its SST files and metadata
* files. These directories do not need to be persistent, they can be ephermeral, meaning that
* they are lost on a machine failure, becaus... | 3.68 |
flink_SolutionSetNode_getOperator | /**
* Gets the contract object for this data source node.
*
* @return The contract.
*/
@Override
public SolutionSetPlaceHolder<?> getOperator() {
return (SolutionSetPlaceHolder<?>) super.getOperator();
} | 3.68 |
flink_ResultPartitionType_isHybridResultPartition | /**
* {@link #isHybridResultPartition()} is used to judge whether it is the specified {@link
* #HYBRID_FULL} or {@link #HYBRID_SELECTIVE} resultPartitionType.
*
* <p>this method suitable for judgment conditions related to the specific implementation of
* {@link ResultPartitionType}.
*
* <p>this method not relate... | 3.68 |
flink_HiveParserTypeCheckProcFactory_getIntervalExprProcessor | /** Factory method to get IntervalExprProcessor. */
public HiveParserTypeCheckProcFactory.IntervalExprProcessor getIntervalExprProcessor() {
return new HiveParserTypeCheckProcFactory.IntervalExprProcessor();
} | 3.68 |
morf_SqlServerDialect_dropPrimaryKey | /**
* Drops the primary key from a {@link Table}.
*
* @param table The table to drop the primary key from
*/
private String dropPrimaryKey(Table table) {
StringBuilder dropPkStatement = new StringBuilder();
dropPkStatement.append("ALTER TABLE ").append(schemaNamePrefix()).append(table.getName()).append(" DROP "... | 3.68 |
flink_SourceTestSuiteBase_testTaskManagerFailure | /**
* Test connector source with task manager failover.
*
* <p>This test will create 1 split in the external system, write test record set A into the
* split, restart task manager to trigger job failover, write test record set B into the split,
* and terminate the Flink job finally.
*
* <p>The number and order o... | 3.68 |
hadoop_TypedBytesOutput_writeFloat | /**
* Writes a float as a typed bytes sequence.
*
* @param f the float to be written
* @throws IOException
*/
public void writeFloat(float f) throws IOException {
out.write(Type.FLOAT.code);
out.writeFloat(f);
} | 3.68 |
pulsar_AuthenticationProviderToken_getTokenAudienceClaim | // get Token Audience Claim from configuration, if not configured return null.
private String getTokenAudienceClaim(ServiceConfiguration conf) throws IllegalArgumentException {
String tokenAudienceClaim = (String) conf.getProperty(confTokenAudienceClaimSettingName);
if (StringUtils.isNotBlank(tokenAudienceClaim... | 3.68 |
hbase_KeyValueHeap_requestSeek | /**
* {@inheritDoc}
*/
@Override
public boolean requestSeek(Cell key, boolean forward, boolean useBloom) throws IOException {
return generalizedSeek(true, key, forward, useBloom);
} | 3.68 |
hudi_TableHeader_addTableHeaderField | /**
* Add a field (column) to table.
*
* @param fieldName field Name
*/
public TableHeader addTableHeaderField(String fieldName) {
fieldNames.add(fieldName);
return this;
} | 3.68 |
flink_PluginLoader_load | /**
* Returns in iterator over all available implementations of the given service interface (SPI)
* for the plugin.
*
* @param service the service interface (SPI) for which implementations are requested.
* @param <P> Type of the requested plugin service.
* @return An iterator of all implementations of the given s... | 3.68 |
flink_SegmentsUtil_setFloat | /**
* set float from segments.
*
* @param segments target segments.
* @param offset value offset.
*/
public static void setFloat(MemorySegment[] segments, int offset, float value) {
if (inFirstSegment(segments, offset, 4)) {
segments[0].putFloat(offset, value);
} else {
setFloatMultiSegment... | 3.68 |
framework_Escalator_scrollToRow | /**
* Scrolls the body vertically so that the row at the given index is visible
* and there is at least {@literal padding} pixels to the given scroll
* destination.
*
* @param rowIndex
* the index of the logical row to scroll to
* @param destination
* where the row should be aligned visual... | 3.68 |
morf_ResultSetMismatch_getKey | /**
* @return key identifying the mismatch.
*/
public String[] getKey() {
return Arrays.copyOf(key, key.length);
} | 3.68 |
hadoop_RegistryTypeUtils_urlEndpoint | /**
* Create a URL endpoint from a list of URIs
* @param api implemented API
* @param protocolType protocol type
* @param uris URIs
* @return a new endpoint
*/
public static Endpoint urlEndpoint(String api,
String protocolType,
URI... uris) {
return new Endpoint(api, protocolType, uris);
} | 3.68 |
hadoop_IOStatisticsLogging_ioStatisticsToPrettyString | /**
* Convert IOStatistics to a string form, with all the metrics sorted
* and empty value stripped.
* This is more expensive than the simple conversion, so should only
* be used for logging/output where it's known/highly likely that the
* caller wants to see the values. Not for debug logging.
* @param statistics... | 3.68 |
zxing_BitMatrix_getRow | /**
* A fast method to retrieve one row of data from the matrix as a BitArray.
*
* @param y The row to retrieve
* @param row An optional caller-allocated BitArray, will be allocated if null or too small
* @return The resulting BitArray - this reference should always be used even when passing
* your own ro... | 3.68 |
dubbo_ScriptStateRouter_getEngine | /**
* create ScriptEngine instance by type from url parameters, then cache it
*/
private ScriptEngine getEngine(URL url) {
String type = url.getParameter(TYPE_KEY, DEFAULT_SCRIPT_TYPE_KEY);
return ConcurrentHashMapUtils.computeIfAbsent(ENGINES, type, t -> {
ScriptEngine scriptEngine = new ScriptEngin... | 3.68 |
flink_SharedObjectsExtension_create | /**
* Creates a new instance. Usually that should be done inside a JUnit test class as an
* instance-field annotated with {@link org.junit.Rule}.
*/
public static SharedObjectsExtension create() {
return new SharedObjectsExtension(LAST_ID.getAndIncrement());
} | 3.68 |
hbase_ReflectionUtils_invokeMethod | /**
* Get and invoke the target method from the given object with given parameters
* @param obj the object to get and invoke method from
* @param methodName the name of the method to invoke
* @param params the parameters for the method to invoke
* @return the return value of the method invocation
*/
@N... | 3.68 |
framework_VTooltip_connectHandlersToWidget | /**
* Connects DOM handlers to widget that are needed for tooltip presentation.
*
* @param widget
* Widget which DOM handlers are connected
*/
public void connectHandlersToWidget(Widget widget) {
Profiler.enter("VTooltip.connectHandlersToWidget");
widget.addDomHandler(tooltipEventHandler, MouseO... | 3.68 |
hadoop_JavaCommandLineBuilder_addConfOptionToCLI | /**
* Ass a configuration option to the command line of the application
* @param conf configuration
* @param key key
* @param defVal default value
* @return the resolved configuration option
* @throws IllegalArgumentException if key is null or the looked up value
* is null (that is: the argument is missing and ... | 3.68 |
open-banking-gateway_AuthorizationPossibleErrorHandler_handlePossibleAuthorizationError | /**
* Swallows retryable (like wrong password) authorization exceptions.
* @param tryAuthorize Authorization function to call
* @param onFail Fallback function to call if retryable exception occurred.
*/
public void handlePossibleAuthorizationError(Runnable tryAuthorize, Consumer<ErrorResponseException> onFail) {
... | 3.68 |
hbase_PermissionStorage_isAclTable | /**
* Returns {@code true} if the given table is {@code _acl_} metadata table.
*/
static boolean isAclTable(TableDescriptor desc) {
return ACL_TABLE_NAME.equals(desc.getTableName());
} | 3.68 |
morf_DataSourceAdapter_unwrap | /**
* @see java.sql.Wrapper#unwrap(java.lang.Class)
*/
@Override
public <T> T unwrap(Class<T> iface) throws SQLException {
throw new UnsupportedOperationException("Wrappers not supported");
} | 3.68 |
hudi_BaseHoodieTableServiceClient_scheduleCompactionAtInstant | /**
* Schedules a new compaction instant with passed-in instant time.
*
* @param instantTime Compaction Instant Time
* @param extraMetadata Extra Metadata to be stored
*/
public boolean scheduleCompactionAtInstant(String instantTime, Option<Map<String, String>> extraMetadata) throws HoodieIOException {
return ... | 3.68 |
hibernate-validator_SizeValidatorForArraysOfDouble_isValid | /**
* Checks the number of entries in an array.
*
* @param array The array to validate.
* @param constraintValidatorContext context in which the constraint is evaluated.
*
* @return Returns {@code true} if the array is {@code null} or the number of entries in
* {@code array} is between the specified {@co... | 3.68 |
flink_TableConfig_getPlannerConfig | /** Returns the current configuration of Planner for Table API and SQL queries. */
public PlannerConfig getPlannerConfig() {
return plannerConfig;
} | 3.68 |
hbase_MasterObserver_postTableFlush | /**
* Called after the table memstore is flushed to disk.
* @param ctx the environment to interact with the framework and master
* @param tableName the name of the table
*/
default void postTableFlush(final ObserverContext<MasterCoprocessorEnvironment> ctx,
final TableName tableName) throws IOException {
} | 3.68 |
hbase_CompactingMemStore_setCompositeSnapshot | // the following three methods allow to manipulate the settings of composite snapshot
public void setCompositeSnapshot(boolean useCompositeSnapshot) {
this.compositeSnapshot = useCompositeSnapshot;
} | 3.68 |
flink_OperatingSystem_readOSFromSystemProperties | /**
* Parses the operating system that the JVM runs on from the java system properties. If the
* operating system was not successfully determined, this method returns {@code UNKNOWN}.
*
* @return The enum constant for the operating system, or {@code UNKNOWN}, if it was not
* possible to determine.
*/
private ... | 3.68 |
framework_VScrollTable_resizeCaptionContainer | /**
* Makes room for the sorting indicator in case the column that the
* header cell belongs to is sorted. This is done by resizing the width
* of the caption container element by the correct amount
*/
public void resizeCaptionContainer(int rightSpacing) {
int captionContainerWidth = width - colResizeWidget.get... | 3.68 |
flink_VertexInputInfoComputationUtils_computeVertexInputInfoForAllToAll | /**
* Compute the {@link JobVertexInputInfo} for a {@link DistributionPattern#ALL_TO_ALL} edge.
* This computation algorithm will evenly distribute subpartitions to downstream subtasks
* according to the number of subpartitions. Different downstream subtasks consume roughly the
* same number of subpartitions.
*
*... | 3.68 |
hadoop_TFile_getValueLength | /**
* Get the length of the value. isValueLengthKnown() must be tested
* true.
*
* @return the length of the value.
*/
public int getValueLength() {
if (vlen >= 0) {
return vlen;
}
throw new RuntimeException("Value length unknown.");
} | 3.68 |
framework_GridLayout_isHideEmptyRowsAndColumns | /**
* Checks whether whether empty rows and columns should be considered as
* non-existent when rendering or not.
*
* @see #setHideEmptyRowsAndColumns(boolean)
* @since 7.3
* @return true if empty rows and columns are hidden, false otherwise
*/
public boolean isHideEmptyRowsAndColumns() {
return getState(fal... | 3.68 |
hadoop_SolverPreprocessor_getDiscreteSkyline | /**
* Discretize job's lifespan into intervals, and return the number of
* containers used by the job within each interval.
* <p> Note that here we assume all containers allocated to the job have the
* same {@link Resource}. This is due to the limit of
* {@link RLESparseResourceAllocation}.
*
* @param skyList ... | 3.68 |
framework_AbstractExtension_extend | /**
* Add this extension to the target connector. This method is protected to
* allow subclasses to require a more specific type of target.
*
* @param target
* the connector to attach this extension to
*/
protected void extend(AbstractClientConnector target) {
target.addExtension(this);
} | 3.68 |
hudi_BootstrapExecutorUtils_execute | /**
* Executes Bootstrap.
*/
public void execute() throws IOException {
initializeTable();
try (SparkRDDWriteClient bootstrapClient = new SparkRDDWriteClient(new HoodieSparkEngineContext(jssc), bootstrapConfig)) {
HashMap<String, String> checkpointCommitMetadata = new HashMap<>();
checkpointCommitMetadat... | 3.68 |
hbase_SnapshotManifest_open | /**
* Return a SnapshotManifest instance with the information already loaded in-memory.
* SnapshotManifest manifest = SnapshotManifest.open(...) TableDescriptor htd =
* manifest.getDescriptor() for (SnapshotRegionManifest regionManifest:
* manifest.getRegionManifests()) hri = regionManifest.getRegionInfo() for
* (... | 3.68 |
framework_Escalator_isScrollLocked | /**
* Checks whether or not an direction is locked for scrolling.
*
* @param direction
* the direction of the scroll of which to check the lock status
* @return <code>true</code> if the direction is locked
*/
public boolean isScrollLocked(ScrollbarBundle.Direction direction) {
switch (direction) {
... | 3.68 |
hadoop_MappingRuleActionBase_setFallbackSkip | /**
* Sets the fallback method to skip, if the action cannot be executed
* We move onto the next rule, ignoring this one.
* @return MappingRuleAction The same object for method chaining.
*/
public MappingRuleAction setFallbackSkip() {
fallback = MappingRuleResult.createSkipResult();
return this;
} | 3.68 |
flink_SkipListUtils_getKeyLen | /**
* Returns the length of the key.
*
* @param memorySegment memory segment for key space.
* @param offset offset of key space in the memory segment.
*/
public static int getKeyLen(MemorySegment memorySegment, int offset) {
return memorySegment.getInt(offset + KEY_LEN_OFFSET);
} | 3.68 |
zxing_BinaryBitmap_crop | /**
* Returns a new object with cropped image data. Implementations may keep a reference to the
* original data rather than a copy. Only callable if isCropSupported() is true.
*
* @param left The left coordinate, which must be in [0,getWidth())
* @param top The top coordinate, which must be in [0,getHeight())
* @... | 3.68 |
framework_ComboBox_setTextInputAllowed | /**
* Sets whether it is possible to input text into the field or whether the
* field area of the component is just used to show what is selected. By
* disabling text input, the comboBox will work in the same way as a
* {@link NativeSelect}
*
* @see #isTextInputAllowed()
*
* @param textInputAllowed
* ... | 3.68 |
flink_FlinkContainersSettings_getHaStoragePath | /**
* Gets HA storage path.
*
* @return The ha storage path.
*/
public String getHaStoragePath() {
return haStoragePath;
} | 3.68 |
flink_StreamExecutionEnvironment_fromParallelCollection | // private helper for passing different names
private <OUT> DataStreamSource<OUT> fromParallelCollection(
SplittableIterator<OUT> iterator, TypeInformation<OUT> typeInfo, String operatorName) {
return addSource(
new FromSplittableIteratorFunction<>(iterator),
operatorName,
... | 3.68 |
graphhopper_Entity_checkRangeInclusive | /** @return whether the number actual is in the range [min, max] */
protected boolean checkRangeInclusive(double min, double max, double actual) {
if (actual < min || actual > max) {
feed.errors.add(new RangeError(tableName, row, null, min, max, actual)); // TODO set column name in loader so it's available ... | 3.68 |
flink_KeyMap_getCurrentTableCapacity | /**
* Gets the current table capacity, i.e., the number of slots in the hash table, without and
* overflow chaining.
*
* @return The number of slots in the hash table.
*/
public int getCurrentTableCapacity() {
return table.length;
} | 3.68 |
hudi_HoodieFlinkTableServiceClient_initMetadataWriter | /**
* Initialize the table metadata writer, for e.g, bootstrap the metadata table
* from the filesystem if it does not exist.
*/
private HoodieBackedTableMetadataWriter initMetadataWriter(Option<String> latestPendingInstant) {
return (HoodieBackedTableMetadataWriter) FlinkHoodieBackedTableMetadataWriter.create(
... | 3.68 |
flink_Channel_setSerializer | /**
* Sets the serializer for this Channel.
*
* @param serializer The serializer to set.
*/
public void setSerializer(TypeSerializerFactory<?> serializer) {
this.serializer = serializer;
} | 3.68 |
pulsar_ClientCnx_addPendingLookupRequests | // caller of this method needs to be protected under pendingLookupRequestSemaphore
private void addPendingLookupRequests(long requestId, TimedCompletableFuture<LookupDataResult> future) {
pendingRequests.put(requestId, future);
requestTimeoutQueue.add(new RequestTime(requestId, RequestType.Lookup));
} | 3.68 |
dubbo_MetadataService_getServiceDefinition | /**
* Interface definition.
*
* @return
*/
default String getServiceDefinition(String interfaceName, String version, String group) {
return getServiceDefinition(buildKey(interfaceName, group, version));
} | 3.68 |
framework_VTabsheet_sendTabClosedEvent | /**
* Informs the server that closing of a tab has been requested.
*
* @param tabIndex
* the index of the closed to close
*/
void sendTabClosedEvent(int tabIndex) {
getRpcProxy().closeTab(tabKeys.get(tabIndex));
} | 3.68 |
hbase_SimpleRpcServer_bind | /**
* A convenience method to bind to a given address and report better exceptions if the address is
* not a valid host.
* @param socket the socket to bind
* @param address the address to bind to
* @param backlog the number of connections allowed in the queue
* @throws BindException if the address can't b... | 3.68 |
hbase_StructBuilder_toStruct | /**
* Retrieve the {@link Struct} represented by {@code this}.
*/
public Struct toStruct() {
return new Struct(fields.toArray(new DataType<?>[fields.size()]));
} | 3.68 |
hbase_MasterProcedureUtil_submitProcedure | /**
* Helper used to deal with submitting procs with nonce. Internally the
* NonceProcedureRunnable.run() will be called only if no one else registered the nonce. any
* Exception thrown by the run() method will be collected/handled and rethrown. <code>
* long procId = MasterProcedureUtil.submitProcedure(
* ne... | 3.68 |
hadoop_S3ClientFactory_withRequesterPays | /**
* Set requester pays option.
* @param value new value
* @return the builder
*/
public S3ClientCreationParameters withRequesterPays(
final boolean value) {
requesterPays = value;
return this;
} | 3.68 |
hbase_MemStoreLABImpl_getOrMakeChunk | /**
* Get the current chunk, or, if there is no current chunk, allocate a new one from the JVM.
*/
private Chunk getOrMakeChunk() {
// Try to get the chunk
Chunk c = currChunk.get();
if (c != null) {
return c;
}
// No current chunk, so we want to allocate one. We race
// against other allocators to CA... | 3.68 |
hadoop_AbfsClientThrottlingIntercept_getReadThrottler | /**
* Returns the analyzer for read operations.
* @return AbfsClientThrottlingAnalyzer for read.
*/
AbfsClientThrottlingAnalyzer getReadThrottler() {
return readThrottler;
} | 3.68 |
hbase_MultithreadedTableMapper_run | /**
* Run the application's maps using a thread pool.
*/
@Override
public void run(Context context) throws IOException, InterruptedException {
outer = context;
int numberOfThreads = getNumberOfThreads(context);
mapClass = getMapperClass(context);
if (LOG.isDebugEnabled()) {
LOG.debug("Configuring multithr... | 3.68 |
hadoop_FilterFileSystem_getDefaultBlockSize | // path variants delegate to underlying filesystem
@Override
public long getDefaultBlockSize(Path f) {
return fs.getDefaultBlockSize(f);
} | 3.68 |
hbase_TableMapReduceUtil_getConfiguredInputFormat | /**
* @return {@link TableInputFormat} .class unless Configuration has something else at
* {@link #TABLE_INPUT_CLASS_KEY}.
*/
private static Class<? extends InputFormat> getConfiguredInputFormat(Job job) {
return (Class<? extends InputFormat>) job.getConfiguration().getClass(TABLE_INPUT_CLASS_KEY,
Tabl... | 3.68 |
flink_ExecutionEnvironment_resetContextEnvironment | /**
* Un-sets the context environment factory. After this method is called, the call to {@link
* #getExecutionEnvironment()} will again return a default local execution environment, and it
* is possible to explicitly instantiate the LocalEnvironment and the RemoteEnvironment.
*/
protected static void resetContextEn... | 3.68 |
flink_ChangelogTruncateHelper_checkpoint | /**
* Set the highest {@link SequenceNumber} of changelog used by the given checkpoint.
*
* @param lastUploadedTo exclusive
*/
public void checkpoint(long checkpointId, SequenceNumber lastUploadedTo) {
checkpointedUpTo.put(checkpointId, lastUploadedTo);
} | 3.68 |
morf_OracleDialect_getSqlForAddMonths | /**
* @see org.alfasoftware.morf.jdbc.SqlDialect#getSqlForAddMonths(org.alfasoftware.morf.sql.element.Function)
*/
@Override
protected String getSqlForAddMonths(Function function) {
return "ADD_MONTHS(" +
getSqlFrom(function.getArguments().get(0)) + ", " +
getSqlFrom(function.getArguments().get(1)) + ")";
} | 3.68 |
morf_RemoveIndex_getIndexToBeRemoved | /**
* @return The index to be removed.
*/
public Index getIndexToBeRemoved() {
return indexToBeRemoved;
} | 3.68 |
hadoop_BalanceJob_removeAfterDone | /**
* Automatically remove this job from the scheduler cache when the job is
* done.
*/
public Builder removeAfterDone(boolean remove) {
removeAfterDone = remove;
return this;
} | 3.68 |
hbase_HRegionServer_getFavoredNodesForRegion | /**
* Return the favored nodes for a region given its encoded name. Look at the comment around
* {@link #regionFavoredNodesMap} on why we convert to InetSocketAddress[] here.
* @param encodedRegionName the encoded region name.
* @return array of favored locations
*/
@Override
public InetSocketAddress[] getFavoredN... | 3.68 |
graphhopper_Path_getWeight | /**
* This weight will be updated during the algorithm. The initial value is maximum double.
*/
public double getWeight() {
return weight;
} | 3.68 |
hadoop_JavaCommandLineBuilder_defineIfSet | /**
* Add a <code>-D key=val</code> command to the CLI if <code>val</code>
* is not null
* @param key key
* @param val value
*/
public boolean defineIfSet(String key, String val) {
Preconditions.checkArgument(key != null, "null key");
if (val != null) {
define(key, val);
return true;
} else {
ret... | 3.68 |
hadoop_Nfs3Constant_getValue | /** @return the int value representing the procedure. */
public int getValue() {
return ordinal();
} | 3.68 |
hadoop_CachingBlockManager_numReadErrors | /**
* Number of errors encountered when reading.
*
* @return the number of errors encountered when reading.
*/
public int numReadErrors() {
return numReadErrors.get();
} | 3.68 |
hbase_ConfigurationUtil_setKeyValues | /**
* Store a collection of Map.Entry's in conf, with each entry separated by ',' and key values
* delimited by delimiter.
* @param conf configuration to store the collection in
* @param key overall key to store keyValues under
* @param keyValues kvps to be stored under key in conf
* @param delimiter c... | 3.68 |
flink_OutputFormatProvider_of | /** Helper method for creating a static provider with a provided sink parallelism. */
static OutputFormatProvider of(OutputFormat<RowData> outputFormat, Integer sinkParallelism) {
return new OutputFormatProvider() {
@Override
public OutputFormat<RowData> createOutputFormat() {
return out... | 3.68 |
framework_SerializerHelper_readClass | /**
* Deserializes a class reference serialized by
* {@link #writeClass(ObjectOutputStream, Class)}. Supports null class
* references.
*
* @param in
* {@code ObjectInputStream} to read from.
* @return Class reference to the resolved class
* @throws ClassNotFoundException
* If the class c... | 3.68 |
flink_DamBehavior_isMaterializing | /**
* Checks whether this enumeration represents some form of materialization, either with a full
* dam or without.
*
* @return True, if this enumeration constant represents a materializing behavior, false
* otherwise.
*/
public boolean isMaterializing() {
return this != PIPELINED;
} | 3.68 |
hbase_Response_setCode | /**
* @param code the HTTP response code
*/
public void setCode(int code) {
this.code = code;
} | 3.68 |
flink_HybridShuffleConfiguration_getBufferPoolSizeCheckIntervalMs | /** Check interval of buffer pool's size. */
public long getBufferPoolSizeCheckIntervalMs() {
return bufferPoolSizeCheckIntervalMs;
} | 3.68 |
dubbo_StringUtils_join | /**
* join string like javascript.
*
* @param array String array.
* @param split split
* @return String.
*/
public static String join(String[] array, String split) {
if (ArrayUtils.isEmpty(array)) {
return EMPTY_STRING;
}
StringBuilder sb = new StringBuilder();
for (int i = 0; i < array.le... | 3.68 |
flink_MutableHashTable_getNewInMemoryPartition | /**
* Returns a new inMemoryPartition object. This is required as a plug for
* ReOpenableMutableHashTable.
*/
protected HashPartition<BT, PT> getNewInMemoryPartition(int number, int recursionLevel) {
return new HashPartition<BT, PT>(
this.buildSideSerializer,
this.probeSideSerializer,
... | 3.68 |
morf_SqlUtils_clobLiteral | /**
* Constructs a new ClobFieldLiteral from a given string.
* @param value - the literal value to use.
* @return ClobFieldLiteral
*/
public static ClobFieldLiteral clobLiteral(String value) {
return new ClobFieldLiteral(value);
} | 3.68 |
flink_ResultPartitionType_mustBePipelinedConsumed | /** return if this partition's upstream and downstream must be scheduled in the same time. */
public boolean mustBePipelinedConsumed() {
return consumingConstraint == ConsumingConstraint.MUST_BE_PIPELINED;
} | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.