name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_AllWindowedStream_allowedLateness | /**
* Sets the time by which elements are allowed to be late. Elements that arrive behind the
* watermark by more than the specified time will be dropped. By default, the allowed lateness
* is {@code 0L}.
*
* <p>Setting an allowed lateness is only valid for event-time windows.
*/
@PublicEvolving
public AllWindowe... | 3.68 |
hadoop_ServiceLauncher_getClassLoader | /**
* Override point: get the classloader to use.
* @return the classloader for loading a service class.
*/
protected ClassLoader getClassLoader() {
return this.getClass().getClassLoader();
} | 3.68 |
flink_FlinkSemiAntiJoinFilterTransposeRule_onMatch | // implement RelOptRule
public void onMatch(RelOptRuleCall call) {
LogicalJoin join = call.rel(0);
LogicalFilter filter = call.rel(1);
RelNode newJoin =
LogicalJoin.create(
filter.getInput(),
join.getRight(),
join.getHints(),
... | 3.68 |
framework_LoginForm_setUsernameCaption | /**
* Sets the caption of the user name field. Note that the caption can only
* be set with this method before the login form has been initialized
* (attached).
* <p>
* As an alternative to calling this method, the method
* {@link #createUsernameField()} can be overridden.
*
* @param usernameCaption
* ... | 3.68 |
hbase_LocalHBaseCluster_getLiveMasters | /**
* @return List of running master servers (Some servers may have been killed or aborted during
* lifetime of cluster; these servers are not included in this list).
*/
public List<JVMClusterUtil.MasterThread> getLiveMasters() {
List<JVMClusterUtil.MasterThread> liveServers = new ArrayList<>();
List<JVM... | 3.68 |
pulsar_AuthorizationProvider_allowTopicOperation | /**
* @deprecated - will be removed after 2.12. Use async variant.
*/
@Deprecated
default Boolean allowTopicOperation(TopicName topicName,
String role,
TopicOperation operation,
AuthenticationDataSource authDat... | 3.68 |
zxing_AztecReader_decode | /**
* Locates and decodes a Data Matrix code in an image.
*
* @return a String representing the content encoded by the Data Matrix code
* @throws NotFoundException if a Data Matrix code cannot be found
* @throws FormatException if a Data Matrix code cannot be decoded
*/
@Override
public Result decode(BinaryBitmap... | 3.68 |
framework_StaticSection_removeRow | /**
* Removes the given row from this section.
*
* @param row
* the row to remove, not null
* @throws IllegalArgumentException
* if this section does not contain the row
*/
public void removeRow(Object row) {
Objects.requireNonNull(row, "row cannot be null");
int index = rows.index... | 3.68 |
flink_Pool_add | /**
* Adds an entry to the pool with an optional payload. This method fails if called more often
* than the pool capacity specified during construction.
*/
public synchronized void add(T object) {
if (poolSize >= poolCapacity) {
throw new IllegalStateException("No space left in pool");
}
poolSize... | 3.68 |
morf_AbstractSqlDialectTest_testSpecifiedValueInsertWithTableInDifferentSchema | /**
* Tests an insert statement where the value for each column (except the id) has been explicitly specified,
*/
@Test
public void testSpecifiedValueInsertWithTableInDifferentSchema() {
InsertStatement stmt = new InsertStatement().into(new TableReference("MYSCHEMA", TEST_TABLE)).values(
new FieldLiteral("Escap... | 3.68 |
hibernate-validator_ReflectionHelper_isCollection | /**
* Indicates whether the given type represents a collection of elements or not (i.e. whether it is an
* {@code Iterable}, {@code Map} or array type).
*/
public static boolean isCollection(Type type) {
return isIterable( type ) ||
isMap( type ) ||
TypeHelper.isArray( type );
} | 3.68 |
morf_SqlServerDialect_renameIndexStatements | /**
* @see org.alfasoftware.morf.jdbc.SqlDialect#renameIndexStatements(org.alfasoftware.morf.metadata.Table, java.lang.String, java.lang.String)
*/
@Override
public Collection<String> renameIndexStatements(Table table, String fromIndexName, String toIndexName) {
return ImmutableList.of(String.format("sp_rename N'%s... | 3.68 |
flink_StreamExecutionEnvironment_createRemoteEnvironment | /**
* Creates a {@link RemoteStreamEnvironment}. The remote environment sends (parts of) the
* program to a cluster for execution. Note that all file paths used in the program must be
* accessible from the cluster. The execution will use the specified parallelism.
*
* @param host The host name or address of the ma... | 3.68 |
flink_Tuple9_of | /**
* Creates a new tuple and assigns the given values to the tuple's fields. This is more
* convenient than using the constructor, because the compiler can infer the generic type
* arguments implicitly. For example: {@code Tuple3.of(n, x, s)} instead of {@code new
* Tuple3<Integer, Double, String>(n, x, s)}
*/
pu... | 3.68 |
flink_PartitionTempFileManager_listTaskTemporaryPaths | /** Returns task temporary paths in this checkpoint. */
public static List<Path> listTaskTemporaryPaths(
FileSystem fs, Path basePath, BiPredicate<Integer, Integer> taskAttemptFilter)
throws Exception {
List<Path> taskTmpPaths = new ArrayList<>();
if (fs.exists(basePath)) {
for (FileSta... | 3.68 |
hbase_AuthManager_refreshTableCacheFromWritable | /**
* Update acl info for table.
* @param table name of table
* @param data updated acl data
* @throws IOException exception when deserialize data
*/
public void refreshTableCacheFromWritable(TableName table, byte[] data) throws IOException {
if (data != null && data.length > 0) {
try {
ListMultimap<S... | 3.68 |
hadoop_ErasureCodingPolicyState_write | /** Write to out. */
public void write(DataOutput out) throws IOException {
out.writeByte(ordinal());
} | 3.68 |
flink_MailboxProcessor_sendPoisonMail | /** Send mail in first priority for internal needs. */
private void sendPoisonMail(RunnableWithException mail) {
mailbox.runExclusively(
() -> {
// keep state check and poison mail enqueuing atomic, such that no intermediate
// #close may cause a
// Mailbo... | 3.68 |
framework_AbsoluteLayout_readDesign | /*
* (non-Javadoc)
*
* @see com.vaadin.ui.AbstractComponent#readDesign(org.jsoup.nodes .Node,
* com.vaadin.ui.declarative.DesignContext)
*/
@Override
public void readDesign(Element design, DesignContext designContext) {
// process default attributes
super.readDesign(design, designContext);
// handle ch... | 3.68 |
framework_CalendarComponentEvents_getWeek | /**
* Get week as a integer. See {@link java.util.Calendar} for the allowed
* values.
*
* @return Week as a integer.
*/
public int getWeek() {
return week;
} | 3.68 |
flink_TypeExtractor_getUnaryOperatorReturnType | /**
* Returns the unary operator's return type.
*
* <p>This method can extract a type in 4 different ways:
*
* <p>1. By using the generics of the base class like MyFunction<X, Y, Z, IN, OUT>. This is what
* outputTypeArgumentIndex (in this example "4") is good for.
*
* <p>2. By using input type inference SubMyF... | 3.68 |
framework_VDateField_getResolutionVariable | /**
* Returns a resolution variable name for the given {@code resolution}.
*
* @param resolution
* the given resolution
* @return the resolution variable name
*/
public String getResolutionVariable(R resolution) {
return resolution.name().toLowerCase(Locale.ROOT);
} | 3.68 |
hbase_RateLimiter_update | /**
* Sets the current instance of RateLimiter to a new values. if current limit is smaller than the
* new limit, bump up the available resources. Otherwise allow clients to use up the previously
* available resources.
*/
public synchronized void update(final RateLimiter other) {
this.tunit = other.tunit;
if (t... | 3.68 |
hadoop_NodePlan_getNodeName | /**
* Returns the DataNode URI.
*
* @return URI
*/
public String getNodeName() {
return nodeName;
} | 3.68 |
framework_ProgressIndicator_getPollingInterval | /**
* Gets the interval that component checks for progress.
*
* @return the interval in milliseconds.
*/
public int getPollingInterval() {
return getState(false).pollingInterval;
} | 3.68 |
hadoop_StoragePolicySatisfyManager_addPathId | /**
* Adds the sps path to SPSPathIds list.
* @param id
*/
public void addPathId(long id) {
synchronized (pathsToBeTraversed) {
pathsToBeTraversed.add(id);
}
} | 3.68 |
hbase_ScannerContext_incrementSizeProgress | /**
* Progress towards the size limit has been made. Increment internal tracking of size progress
*/
void incrementSizeProgress(long dataSize, long heapSize) {
if (skippingRow) {
return;
}
long curDataSize = progress.getDataSize();
progress.setDataSize(curDataSize + dataSize);
long curHeapSize = progres... | 3.68 |
hbase_TableSplit_getStartRow | /**
* Returns the start row.
* @return The start row.
*/
public byte[] getStartRow() {
return startRow;
} | 3.68 |
hbase_User_create | /**
* Wraps an underlying {@code UserGroupInformation} instance.
* @param ugi The base Hadoop user
*/
public static User create(UserGroupInformation ugi) {
if (ugi == null) {
return null;
}
return new SecureHadoopUser(ugi);
} | 3.68 |
pulsar_ManagedLedgerConfig_setRetentionTime | /**
* Set the retention time for the ManagedLedger.
* <p>
* Retention time and retention size ({@link #setRetentionSizeInMB(long)}) are together used to retain the
* ledger data when there are no cursors or when all the cursors have marked the data for deletion.
* Data will be deleted in this case when both retent... | 3.68 |
druid_Lexer_token | /**
* Return the current token, set by nextToken().
*/
public final Token token() {
return token;
} | 3.68 |
morf_AbstractSqlDialectTest_testSelectSimpleJoinScript | /**
* Tests a select with a simple join.
*/
@Test
public void testSelectSimpleJoinScript() {
SelectStatement stmt = new SelectStatement().from(new TableReference(TEST_TABLE))
.innerJoin(new TableReference(ALTERNATE_TABLE),
eq(new FieldReference(new TableReference(TEST_TABLE), STRING_FIELD),
... | 3.68 |
flink_CatalogManager_resolveCatalogBaseTable | /** Resolves a {@link CatalogBaseTable} to a validated {@link ResolvedCatalogBaseTable}. */
public ResolvedCatalogBaseTable<?> resolveCatalogBaseTable(CatalogBaseTable baseTable) {
Preconditions.checkNotNull(schemaResolver, "Schema resolver is not initialized.");
if (baseTable instanceof CatalogTable) {
... | 3.68 |
flink_PythonConfigUtil_alignTransformation | /**
* Configure the {@link AbstractExternalOneInputPythonFunctionOperator} to be chained with the
* upstream/downstream operator by setting their parallelism, slot sharing group, co-location
* group to be the same, and applying a {@link ForwardPartitioner}. 1. operator with name
* "_keyed_stream_values_operator" sh... | 3.68 |
pulsar_ModularLoadManagerImpl_updateLoadBalancingBundlesMetrics | /**
* As any broker, update its bundle metrics.
*
* @param bundlesData
*/
private void updateLoadBalancingBundlesMetrics(Map<String, NamespaceBundleStats> bundlesData) {
List<Metrics> metrics = new ArrayList<>();
for (Map.Entry<String, NamespaceBundleStats> entry: bundlesData.entrySet()) {
final Str... | 3.68 |
AreaShop_SignsFeature_getSignsRef | /**
* Get the signs of this region.
* @return Map with signs: locationString -> RegionSign
*/
Map<String, RegionSign> getSignsRef() {
return signs;
} | 3.68 |
hbase_ScannerModel_getJasonProvider | /**
* Get the <code>JacksonJaxbJsonProvider</code> instance;
* @return A <code>JacksonJaxbJsonProvider</code>.
*/
private static JacksonJaxbJsonProvider getJasonProvider() {
return JaxbJsonProviderHolder.INSTANCE;
} | 3.68 |
hbase_CleanerChore_deleteAction | /**
* Perform a delete on a specified type.
* @param deletion a delete
* @param type possible values are 'files', 'subdirs', 'dirs'
* @return true if it deleted successfully, false otherwise
*/
private boolean deleteAction(Action<Boolean> deletion, String type, Path dir) {
boolean deleted;
try {
LOG.tr... | 3.68 |
framework_DragSourceExtension_clearDataTransferData | /**
* Clears all data for this drag source element.
*/
public void clearDataTransferData() {
getState().types.clear();
getState().data.clear();
} | 3.68 |
dubbo_AbstractJSONImpl_getObject | /**
* Gets an object from an object for the given key. If the key is not present, this returns null.
* If the value is not a Map, throws an exception.
*/
@SuppressWarnings("unchecked")
@Override
public Map<String, ?> getObject(Map<String, ?> obj, String key) {
assert obj != null;
assert key != null;
if ... | 3.68 |
hbase_NamedQueueRecorder_persistAll | /**
* Add all in memory queue records to system table. The implementors can use system table or
* direct HDFS file or ZK as persistence system.
*/
public void persistAll(NamedQueuePayload.NamedQueueEvent namedQueueEvent, Connection connection) {
if (this.logEventHandler != null) {
this.logEventHandler.persistA... | 3.68 |
flink_UploadThrottle_hasCapacity | /** Test whether some capacity is available. */
public boolean hasCapacity() {
return inFlightBytesCounter < maxBytesInFlight;
} | 3.68 |
flink_BeamPythonFunctionRunner_notifyNoMoreResults | /** Interrupts the progress of takeResult. */
public void notifyNoMoreResults() {
resultBuffer.add(Tuple2.of(null, new byte[0]));
} | 3.68 |
morf_SchemaChangeSequence_addTableFrom | /**
* @see org.alfasoftware.morf.upgrade.SchemaEditor#addTableFrom(org.alfasoftware.morf.metadata.Table, org.alfasoftware.morf.sql.SelectStatement)
*/
@Override
public void addTableFrom(Table table, SelectStatement select) {
// track added tables...
tableAdditions.add(table.getName());
AddTable addTable = new ... | 3.68 |
hbase_MetricsUserAggregateImpl_getActiveUser | /**
* Returns the active user to which authorization checks should be applied. If we are in the
* context of an RPC call, the remote user is used, otherwise the currently logged in user is
* used.
*/
private String getActiveUser() {
Optional<User> user = RpcServer.getRequestUser();
if (!user.isPresent()) {
... | 3.68 |
hbase_KeyValue_getDelimiter | /**
* Find index of passed delimiter walking from start of buffer forwards.
* @param b the kv serialized byte[] to process
* @param delimiter input delimeter to fetch index from start
* @return Index of delimiter having started from start of <code>b</code> moving rightward.
*/
public static int getDelimite... | 3.68 |
framework_Tree_addListener | /**
* @deprecated As of 7.0, replaced by
* {@link #addItemClickListener(ItemClickListener)}
*/
@Override
@Deprecated
public void addListener(ItemClickListener listener) {
addItemClickListener(listener);
} | 3.68 |
hadoop_IdentifierResolver_getOutputReaderClass | /**
* Returns the resolved {@link OutputReader} class.
*/
public Class<? extends OutputReader> getOutputReaderClass() {
return outputReaderClass;
} | 3.68 |
druid_CharsetConvert_encode | /**
* 字符串编码
*
* @param s String
* @return String
* @throws UnsupportedEncodingException
*/
public String encode(String s) throws UnsupportedEncodingException {
if (enable && !isEmpty(s)) {
s = new String(s.getBytes(clientEncoding), serverEncoding);
}
return s;
} | 3.68 |
hudi_HoodieLogBlock_getLogMetadata | /**
* Convert bytes to LogMetadata, follow the same order as {@link HoodieLogBlock#getLogMetadataBytes}.
*/
public static Map<HeaderMetadataType, String> getLogMetadata(DataInputStream dis) throws IOException {
Map<HeaderMetadataType, String> metadata = new HashMap<>();
// 1. Read the metadata written out
int ... | 3.68 |
pulsar_IScheduler_rebalance | /**
* Rebalances function instances scheduled to workers.
*
* @param currentAssignments
* current assignments
* @param workers
* current list of active workers
* @return
* A list of new assignments
*/
default List<Function.Assignment> rebalance(List<Function.Assignment> current... | 3.68 |
framework_DragEndEvent_isCanceled | /**
* Returns whether the drag event was cancelled. This is a shorthand for
* {@code dropEffect == NONE}.
*
* @return {@code true} if the drop event was cancelled, {@code false}
* otherwise.
*/
public boolean isCanceled() {
return getDropEffect() == DropEffect.NONE;
} | 3.68 |
hbase_Get_getTimeRange | /**
* Method for retrieving the get's TimeRange
*/
public TimeRange getTimeRange() {
return this.tr;
} | 3.68 |
morf_Function_randomString | /**
* Helper method to create a function for generating random strings via SQL.
*
* @param length The length of the generated string
* @return an instance of the randomString function.
*/
public static Function randomString(AliasedField length) {
return new Function(FunctionType.RANDOM_STRING, length);
} | 3.68 |
framework_VScrollTable_addAndRemoveRows | /**
* Inserts rows in the table body or removes them from the table body based
* on the commands in the UIDL.
* <p>
* For internal use only. May be removed or replaced in the future.
*
* @param partialRowAdditions
* the UIDL containing row updates.
*/
public void addAndRemoveRows(UIDL partialRowAddit... | 3.68 |
hbase_FavoredNodeAssignmentHelper_placePrimaryRSAsRoundRobin | // Place the regions round-robin across the racks picking one server from each
// rack at a time. Start with a random rack, and a random server from every rack.
// If a rack doesn't have enough servers it will go to the next rack and so on.
// for choosing a primary.
// For example, if 4 racks (r1 .. r4) with 8 servers... | 3.68 |
hbase_ReplicationPeerConfigUtil_convert2Map | /**
* Convert tableCFs Object to Map.
*/
public static Map<TableName, List<String>> convert2Map(ReplicationProtos.TableCF[] tableCFs) {
if (tableCFs == null || tableCFs.length == 0) {
return null;
}
Map<TableName, List<String>> tableCFsMap = new HashMap<>();
for (int i = 0, n = tableCFs.length; i < n; i++... | 3.68 |
flink_DynamicTableFactory_getPrimaryKeyIndexes | /**
* Returns the primary key indexes, if any, otherwise returns an empty array. A factory can
* use it to compute the schema projection of the key fields with {@code
* Projection.of(ctx.getPrimaryKeyIndexes()).project(dataType)}.
*
* <p>Shortcut for {@code getCatalogTable().getResolvedSchema().getPrimaryKeyIndexe... | 3.68 |
hbase_MasterObserver_postCompletedMergeRegionsAction | /**
* called after the regions merge.
* @param ctx the environment to interact with the framework and master
*/
default void postCompletedMergeRegionsAction(
final ObserverContext<MasterCoprocessorEnvironment> ctx, final RegionInfo[] regionsToMerge,
final RegionInfo mergedRegion) throws IOException {
} | 3.68 |
querydsl_GeometryExpression_difference | /**
* Returns a geometric object that represents the Point
* set difference of this geometric object with anotherGeometry.
*
* @param geometry other geometry
* @return difference between this and the other geometry
*/
public GeometryExpression<Geometry> difference(Expression<? extends Geometry> geometry) {
re... | 3.68 |
hadoop_HadoopLogsAnalyzer_main | /**
* @param args
*
* Last arg is the input file. That file can be a directory, in which
* case you get all the files in sorted order. We will decompress
* files whose nmes end in .gz .
*
* switches: -c collect line types.
*
* -d debug mode
*
* -delay... | 3.68 |
flink_TestcontainersSettings_baseImage | /**
* Sets the {@code baseImage} and returns a reference to this Builder enabling method
* chaining.
*
* @param baseImage The {@code baseImage} to set.
* @return A reference to this Builder.
*/
public Builder baseImage(String baseImage) {
this.baseImage = baseImage;
return this;
} | 3.68 |
framework_BootstrapHandler_getApplicationParameters | /**
* Gets the application parameters specified by the BootstrapHandler.
*
* @return the application parameters which will be written on the page
*/
public JsonObject getApplicationParameters() {
if (applicationParameters == null) {
applicationParameters = BootstrapHandler.this
.getAppli... | 3.68 |
flink_Execution_sendOperatorEvent | /**
* Sends the operator event to the Task on the Task Executor.
*
* @return True, of the message was sent, false is the task is currently not running.
*/
public CompletableFuture<Acknowledge> sendOperatorEvent(
OperatorID operatorId, SerializedValue<OperatorEvent> event) {
assertRunningInJobMasterMain... | 3.68 |
hudi_CleanPlanner_getFilesToCleanKeepingLatestHours | /**
* This method finds the files to be cleaned based on the number of hours. If {@code config.getCleanerHoursRetained()} is set to 5,
* all the files with commit time earlier than 5 hours will be removed. Also the latest file for any file group is retained.
* This policy gives much more flexibility to users for ret... | 3.68 |
hudi_DagScheduler_schedule | /**
* Method to start executing workflow DAGs.
*
* @throws Exception Thrown if schedule failed.
*/
public void schedule() throws Exception {
ExecutorService service = Executors.newFixedThreadPool(2);
try {
execute(service, workflowDag);
service.shutdown();
} finally {
if (!service.isShutdown()) {
... | 3.68 |
dubbo_DubboBootstrap_isStarting | /**
* @return true if the dubbo application is starting.
* @see #isStarted()
*/
public boolean isStarting() {
return applicationDeployer.isStarting();
} | 3.68 |
hadoop_SinglePendingCommit_getSaved | /**
* When was the upload saved?
* @return timestamp
*/
public long getSaved() {
return saved;
} | 3.68 |
framework_PureGWTTestApplication_addChildMenu | /**
* Adds a child menu entry to this menu. The title for this entry is
* taken from the Menu object argument.
*
* @param m
* another Menu object
*/
public void addChildMenu(Menu m) {
menubar.addItem(m.title, m.menubar);
children.add(m);
} | 3.68 |
hbase_Query_setACL | /**
* Set the ACL for the operation.
* @param perms A map of permissions for a user or users
*/
public Query setACL(Map<String, Permission> perms) {
ListMultimap<String, Permission> permMap = ArrayListMultimap.create();
for (Map.Entry<String, Permission> entry : perms.entrySet()) {
permMap.put(entry.getKey()... | 3.68 |
framework_FilesystemContainer_addItemIds | /**
* Internal recursive method to add the files under the specified directory
* to the collection.
*
* @param col
* the collection where the found items are added
* @param f
* the root file where to start adding files
*/
private void addItemIds(Collection<File> col, File f) {
File[] l... | 3.68 |
dubbo_StringUtils_splitToList | /**
* Splits String around matches of the given character.
* <p>
* Note: Compare with {@link StringUtils#split(String, char)}, this method reduce memory copy.
*/
public static List<String> splitToList(String str, char ch) {
if (isEmpty(str)) {
return Collections.emptyList();
}
return splitToList... | 3.68 |
MagicPlugin_MapController_getMapItem | /**
* A helper function to get an ItemStack from a MapView.
*
* @param name The display name to give the new item. Optional.
*/
public ItemStack getMapItem(String name, int mapId) {
ItemStack newMapItem = createMap(mapId);
if (name != null) {
ItemMeta meta = newMapItem.getItemMeta();
meta.se... | 3.68 |
hbase_ClientUtil_calculateTheClosestNextRowKeyForPrefix | /**
* <p>
* When scanning for a prefix the scan should stop immediately after the the last row that has the
* specified prefix. This method calculates the closest next rowKey immediately following the
* given rowKeyPrefix.
* </p>
* <p>
* <b>IMPORTANT: This converts a rowKey<u>Prefix</u> into a rowKey</b>.
* </p... | 3.68 |
framework_ConnectorTracker_markDirty | /**
* Mark the connector as dirty and notifies any marked as dirty listeners.
* This should not be done while the response is being written.
*
* @see #getDirtyConnectors()
* @see #isWritingResponse()
*
* @param connector
* The connector that should be marked clean.
*/
public void markDirty(ClientCon... | 3.68 |
hbase_KeyValueUtil_keyLength | /**
* Returns number of bytes this cell's key part would have been used if serialized as in
* {@link KeyValue}. Key includes rowkey, family, qualifier, timestamp and type.
* @return the key length
*/
public static int keyLength(final Cell cell) {
return keyLength(cell.getRowLength(), cell.getFamilyLength(), cell.... | 3.68 |
hadoop_FileSystemReleaseFilter_setFileSystem | /**
* Static method that sets the <code>FileSystem</code> to release back to
* the {@link FileSystemAccess} service on servlet request completion.
*
* @param fs a filesystem instance.
*/
public static void setFileSystem(FileSystem fs) {
FILE_SYSTEM_TL.set(fs);
} | 3.68 |
hudi_AvroSchemaCompatibility_getWriterFragment | /**
* Returns the fragment of the writer schema that failed compatibility check.
*
* @return a Schema instance (fragment of the writer schema).
*/
public Schema getWriterFragment() {
return mWriterFragment;
} | 3.68 |
hadoop_MultipleOutputFormat_generateActualValue | /**
* Generate the actual value from the given key and value. The default behavior is that
* the actual value is equal to the given value
*
* @param key
* the key of the output data
* @param value
* the value of the output data
* @return the actual value derived from the given key/value
*/
p... | 3.68 |
hbase_ServerManager_getAverageLoad | /**
* Compute the average load across all region servers. Currently, this uses a very naive
* computation - just uses the number of regions being served, ignoring stats about number of
* requests.
* @return the average load
*/
public double getAverageLoad() {
int totalLoad = 0;
int numServers = 0;
for (Serve... | 3.68 |
zxing_ITFReader_decodeStart | /**
* Identify where the start of the middle / payload section starts.
*
* @param row row of black/white values to search
* @return Array, containing index of start of 'start block' and end of
* 'start block'
*/
private int[] decodeStart(BitArray row) throws NotFoundException {
int endStart = skipWhiteS... | 3.68 |
hbase_ReportMakingVisitor_checkServer | /**
* Run through referenced servers and save off unknown and the dead.
*/
private void checkServer(RegionLocations locations) {
if (this.services == null) {
// Can't do this test if no services.
return;
}
if (locations == null) {
return;
}
if (locations.getRegionLocations() == null) {
retur... | 3.68 |
hadoop_ReconfigurationTaskStatus_stopped | /**
* Return true if the latest reconfiguration task has finished and there is
* no another active task running.
* @return true if endTime > 0; false if not.
*/
public boolean stopped() {
return endTime > 0;
} | 3.68 |
graphhopper_Downloader_fetch | /**
* This method initiates a connect call of the provided connection and returns the response
* stream. It only returns the error stream if it is available and readErrorStreamNoException is
* true otherwise it throws an IOException if an error happens. Furthermore it wraps the stream
* to decompress it if the conn... | 3.68 |
framework_Table_handleClickEvent | /**
* Handles click event
*
* @param variables
*/
private void handleClickEvent(Map<String, Object> variables) {
// Item click event
if (variables.containsKey("clickEvent")) {
String key = (String) variables.get("clickedKey");
Object itemId = itemIdMapper.get(key);
Object propertyId... | 3.68 |
framework_Escalator_setScrollLeft | /**
* Sets the logical horizontal scroll offset. Note that will not necessarily
* become the same as the {@code scrollLeft} attribute in the DOM.
*
* @param scrollLeft
* the number of pixels to scroll horizontally
*/
public void setScrollLeft(final double scrollLeft) {
horizontalScrollbar.setScroll... | 3.68 |
morf_AbstractSqlDialectTest_expectedHints3 | /**
* @return The expected SQL for the {@link UpdateStatement#useParallelDml()} directive.
*/
protected String expectedHints3() {
return "UPDATE " + tableName("Foo") + " SET a = b";
} | 3.68 |
dubbo_DubboBootstrap_isStarted | /**
* @return true if the dubbo application has been started.
* @see #start()
* @see #isStarting()
*/
public boolean isStarted() {
return applicationDeployer.isStarted();
} | 3.68 |
hmily_TransactionManagerImpl_initialized | /**
* Initialized.
*/
public void initialized() {
//initialize_it @see HmilyXaTransactionManager;
hmilyXaTransactionManager = HmilyXaTransactionManager.initialized();
} | 3.68 |
hbase_RawBytesTerminated_decode | /**
* Read a {@code byte[]} from the buffer {@code src}.
*/
public byte[] decode(PositionedByteRange src, int length) {
return ((RawBytes) wrapped).decode(src, length);
} | 3.68 |
hadoop_Trash_moveToAppropriateTrash | /**
* In case of the symlinks or mount points, one has to move the appropriate
* trashbin in the actual volume of the path p being deleted.
*
* Hence we get the file system of the fully-qualified resolved-path and
* then move the path p to the trashbin in that volume,
* @param fs - the filesystem of path p
* @pa... | 3.68 |
morf_HumanReadableStatementHelper_generateDeleteStatementString | /**
* Generates a human-readable description of a data delete operation.
*
* @param statement the data upgrade statement to describe.
* @return a string containing the human-readable description of the operation.
*/
private static String generateDeleteStatementString(final DeleteStatement statement) {
if (statem... | 3.68 |
framework_VCalendarPanel_buildTime | /**
* Constructs the ListBoxes and updates their value
*
* @param redraw
* Should new instances of the listboxes be created
*/
private void buildTime() {
clear();
hours = createListBox();
if (getDateTimeService().isTwelveHourClock()) {
hours.addItem("12");
for (int i = 1; i ... | 3.68 |
hudi_HoodieTableFactory_checkPreCombineKey | /**
* Validate pre_combine key.
*/
private void checkPreCombineKey(Configuration conf, ResolvedSchema schema) {
List<String> fields = schema.getColumnNames();
String preCombineField = conf.get(FlinkOptions.PRECOMBINE_FIELD);
if (!fields.contains(preCombineField)) {
if (OptionsResolver.isDefaultHoodieRecordP... | 3.68 |
flink_BinaryStringData_startsWith | /**
* Tests if this BinaryStringData starts with the specified prefix.
*
* @param prefix the prefix.
* @return {@code true} if the bytes represented by the argument is a prefix of the bytes
* represented by this string; {@code false} otherwise. Note also that {@code true} will be
* returned if the argumen... | 3.68 |
hbase_DeadServer_cleanPreviousInstance | /**
* Handles restart of a server. The new server instance has a different start code. The new start
* code should be greater than the old one. We don't check that here. Removes the old server from
* deadserver list.
* @param newServerName Servername as either <code>host:port</code> or
* <code... | 3.68 |
hudi_AbstractColumnReader_supportLazyDecode | /**
* Support lazy dictionary ids decode. See more in {@link ParquetDictionary}.
* If return false, we will decode all the data first.
*/
protected boolean supportLazyDecode() {
return true;
} | 3.68 |
framework_AbstractRemoteDataSource_getRequestedAvailability | /**
* Gets the row index range that was requested by the previous call to
* {@link #ensureAvailability(int, int)}.
*
* @return the requested availability range
*/
public Range getRequestedAvailability() {
return requestedAvailability;
} | 3.68 |
hadoop_AbstractStoreOperation_getAuditSpan | /**
* Get the audit span this object was created with.
* @return the current span or null
*/
public AuditSpan getAuditSpan() {
return auditSpan;
} | 3.68 |
flink_WebMonitorUtils_resolveFileLocation | /**
* Verify log file location.
*
* @param logFilePath Path to log file
* @return File or null if not a valid log file
*/
private static File resolveFileLocation(String logFilePath) {
File logFile = new File(logFilePath);
return (logFile.exists() && logFile.canRead()) ? logFile : null;
} | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.