name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
framework_GridDragSource_addGridDragEndListener | /**
* Attaches dragend listener for the current drag source grid.
*
* @param listener
* Listener to handle the dragend event.
* @return Handle to be used to remove this listener.
* @see GridDragEndEvent
*/
public Registration addGridDragEndListener(
GridDragEndListener<T> listener) {
retur... | 3.68 |
flink_EnvironmentInformation_getSizeOfFreeHeapMemoryWithDefrag | /**
* Gets an estimate of the size of the free heap memory.
*
* <p>NOTE: This method is heavy-weight. It triggers a garbage collection to reduce
* fragmentation and get a better estimate at the size of free memory. It is typically more
* accurate than the plain version {@link #getSizeOfFreeHeapMemory()}.
*
* @re... | 3.68 |
MagicPlugin_CompatibilityUtilsBase_getOrCreatePotionEntity | /**
* Lazily creates potion entities that can be used when damaging players.
*
* @param location The location the potion should be placed at.
* @return A potion entity placed ad the given location.
*/
protected ThrownPotion getOrCreatePotionEntity(Location location) {
World world = location.getWorld();
//... | 3.68 |
flink_HsFileDataManager_release | /** Releases this file data manager and delete shuffle data after all readers is removed. */
public void release() {
synchronized (lock) {
if (isReleased) {
return;
}
isReleased = true;
List<HsSubpartitionFileReader> pendingReaders = new ArrayList<>(allReaders);
... | 3.68 |
hadoop_SaveSuccessFileStage_getStageName | /**
* Stage name is always job commit.
* @param arguments args to the invocation.
* @return stage name
*/
@Override
protected String getStageName(ManifestSuccessData arguments) {
// set it to the job commit stage, always.
return OP_STAGE_JOB_COMMIT;
} | 3.68 |
flink_TableFunction_getParameterTypes | /**
* Returns {@link TypeInformation} about the operands of the evaluation method with a given
* signature.
*
* @deprecated This method uses the old type system and is based on the old reflective
* extraction logic. The method will be removed in future versions and is only called when
* using the deprecat... | 3.68 |
flink_BroadcastConnectedStream_getSecondInput | /**
* Returns the {@link BroadcastStream}.
*
* @return The stream which, by convention, is the broadcast one.
*/
public BroadcastStream<IN2> getSecondInput() {
return broadcastStream;
} | 3.68 |
hbase_MasterObserver_preMergeRegionsCommitAction | /**
* This will be called before update META step as part of regions merge transaction.
* @param ctx the environment to interact with the framework and master
* @param metaEntries mutations to execute on hbase:meta atomically with regions merge updates.
* Any puts or deletes to execute on... | 3.68 |
hbase_Import_instantiateFilter | /**
* Create a {@link Filter} to apply to all incoming keys ({@link KeyValue KeyValues}) to
* optionally not include in the job output
* @param conf {@link Configuration} from which to load the filter
* @return the filter to use for the task, or <tt>null</tt> if no filter to should be used
* @throws IllegalArgumen... | 3.68 |
flink_StreamExecutionEnvironment_getDefaultLocalParallelism | /**
* Gets the default parallelism that will be used for the local execution environment created by
* {@link #createLocalEnvironment()}.
*
* @return The default local parallelism
*/
@PublicEvolving
public static int getDefaultLocalParallelism() {
return defaultLocalParallelism;
} | 3.68 |
flink_BeamPythonFunctionRunner_create | // the input value type is always byte array
@SuppressWarnings("unchecked")
@Override
public FnDataReceiver<WindowedValue<byte[]>> create(String pCollectionId) {
return input -> {
resultBuffer.add(Tuple2.of(pCollectionId, input.getValue()));
};
} | 3.68 |
flink_JobGraph_getJobID | /**
* Returns the ID of the job.
*
* @return the ID of the job
*/
public JobID getJobID() {
return this.jobID;
} | 3.68 |
graphhopper_RAMIntDataAccess_setStore | /**
* @param store true if in-memory data should be saved when calling flush
*/
public RAMIntDataAccess setStore(boolean store) {
this.store = store;
return this;
} | 3.68 |
pulsar_BytesSchemaVersion_hashCode | /**
* The hashcode is cached except for the case where it is computed as 0, in which
* case we compute the hashcode on every call.
*
* @return the hashcode
*/
@Override
public int hashCode() {
if (hashCode == 0) {
hashCode = Arrays.hashCode(bytes);
}
return hashCode;
} | 3.68 |
hudi_HoodieSyncClient_getPartitionValuesToPathMapping | /**
* Gets the partition values to the absolute path mapping based on the
* partition information from the metastore.
*
* @param partitionsInMetastore Partitions in the metastore.
* @return The partition values to the absolute path mapping.
*/
private Map<String, String> getPartitionValuesToPathMapping(List<Parti... | 3.68 |
hadoop_QuotaUsage_getHeader | /** Return the header of the output.
* @return the header of the output
*/
public static String getHeader() {
return QUOTA_HEADER;
} | 3.68 |
hadoop_StageConfig_getEnterStageEventHandler | /**
* Handler for stage entry events.
* @return the handler.
*/
public StageEventCallbacks getEnterStageEventHandler() {
return enterStageEventHandler;
} | 3.68 |
morf_WindowFunction_over | /**
* Starts a new window function Builder.
* @param function the function to construct the window function over.
* @return the window function builder
*/
public static Builder over(Function function) {
return new BuilderImpl(function);
} | 3.68 |
hudi_HoodieTableMetadataUtil_tryUpcastDecimal | /**
* Does an upcast for {@link BigDecimal} instance to align it with scale/precision expected by
* the {@link org.apache.avro.LogicalTypes.Decimal} Avro logical type
*/
public static BigDecimal tryUpcastDecimal(BigDecimal value, final LogicalTypes.Decimal decimal) {
final int scale = decimal.getScale();
final i... | 3.68 |
framework_GridConnector_handleServerInitiated | /**
* Used to handle the case where the editor calls us because it was
* invoked by the server via RPC and not by the client. In that case,
* the request can be simply synchronously completed.
*
* @param request
* the request object
* @return true if the request was originally triggered by the server,... | 3.68 |
framework_ComponentSizeValidator_isForm | /**
* Comparability form component which is defined in the different jar.
*
* TODO : Normally this logic shouldn't be here. But it means that the whole
* this class has wrong design and implementation and should be refactored.
*/
private static boolean isForm(Component component) {
if (!(component instanceof H... | 3.68 |
hudi_StreamSync_close | /**
* Close all resources.
*/
public void close() {
if (writeClient != null) {
writeClient.close();
writeClient = null;
}
if (formatAdapter != null) {
formatAdapter.close();
}
LOG.info("Shutting down embedded timeline server");
if (embeddedTimelineService.isPresent()) {
embeddedTimelineS... | 3.68 |
framework_ConnectorTracker_getSeckey | /**
* Returns the security key associated with the given StreamVariable.
*
* @param variable
* @return matching security key if one exists, null otherwise
*/
public String getSeckey(StreamVariable variable) {
if (streamVariableToSeckey == null) {
return null;
}
return streamVariableToSeckey.get... | 3.68 |
hibernate-validator_MessagerAdapter_reportError | /**
* Reports the given error. Message parameters will be put into the template
* retrieved from the resource bundle if applicable.
*
* @param error The error to report.
*/
private void reportError(ConstraintCheckIssue error) {
report( error, diagnosticKind );
} | 3.68 |
hmily_SingletonHolder_register | /**
* register.
*
* @param <T> the type parameter
* @param clazz the clazz
* @param o the o
*/
public <T> void register(final Class<T> clazz, final Object o) {
SINGLES.put(clazz.getName(), o);
} | 3.68 |
hbase_QuotaSettingsFactory_unthrottleRegionServer | /**
* Remove the throttling for the specified region server.
* @param regionServer the region Server
* @return the quota settings
*/
public static QuotaSettings unthrottleRegionServer(final String regionServer) {
return throttle(null, null, null, regionServer, null, 0, null, QuotaScope.MACHINE);
} | 3.68 |
querydsl_JTSGeometryCollectionExpression_numGeometries | /**
* Returns the number of geometries in this GeometryCollection.
*
* @return numbers of geometries
*/
public NumberExpression<Integer> numGeometries() {
if (numGeometries == null) {
numGeometries = Expressions.numberOperation(Integer.class, SpatialOps.NUM_GEOMETRIES, mixin);
}
return numGeomet... | 3.68 |
hbase_TableDescriptorBuilder_setValue | /**
* Setter for storing metadata as a (key, value) pair in {@link #values} map
* @param key The key.
* @param value The value. If null, removes the setting.
*/
public ModifyableTableDescriptor setValue(final Bytes key, final Bytes value) {
if (value == null || value.getLength() == 0) {
values.remove(key);
... | 3.68 |
pulsar_ModularLoadManagerImpl_writeBrokerDataOnZooKeeper | /**
* As any broker, write the local broker data to metadata store.
*/
@Override
public void writeBrokerDataOnZooKeeper() {
writeBrokerDataOnZooKeeper(false);
} | 3.68 |
framework_VUpload_rebuildPanel | /**
* Re-creates file input field and populates panel. This is needed as we
* want to clear existing values from our current file input field.
*/
private void rebuildPanel() {
panel.remove(submitButton);
panel.remove(fu);
fu = new VFileUpload();
fu.setName(paintableId + "_file");
fu.getElement().... | 3.68 |
flink_JoinedStreams_trigger | /** Sets the {@code Trigger} that should be used to trigger window emission. */
@PublicEvolving
public WithWindow<T1, T2, KEY, W> trigger(
Trigger<? super TaggedUnion<T1, T2>, ? super W> newTrigger) {
return new WithWindow<>(
input1,
input2,
keySelector1,
keyS... | 3.68 |
shardingsphere-elasticjob_FailoverService_getLocalFailoverItems | /**
* Get failover items which execute on localhost.
*
* @return failover items which execute on localhost
*/
public List<Integer> getLocalFailoverItems() {
if (JobRegistry.getInstance().isShutdown(jobName)) {
return Collections.emptyList();
}
return getFailoverItems(JobRegistry.getInstance().g... | 3.68 |
flink_ParquetRowDataWriter_write | /**
* It writes a record to Parquet.
*
* @param record Contains the record that is going to be written.
*/
public void write(final RowData record) {
recordConsumer.startMessage();
rowWriter.write(record);
recordConsumer.endMessage();
} | 3.68 |
flink_PekkoUtils_getRpcURL | /**
* Returns the given {@link ActorRef}'s path string representation with host and port of the
* {@link ActorSystem} in which the actor is running.
*
* @param system {@link ActorSystem} in which the given {@link ActorRef} is running
* @param actor {@link ActorRef} of the actor for which the URL has to be generate... | 3.68 |
pulsar_KubernetesServiceAccountTokenAuthProvider_cleanUpAuthData | /**
* No need to clean up anything. Kubernetes cleans up the secret when the pod is deleted.
*/
@Override
public void cleanUpAuthData(Function.FunctionDetails funcDetails, Optional<FunctionAuthData> functionAuthData)
throws Exception {
} | 3.68 |
hudi_HoodieOperation_isUpdateAfter | /**
* Returns whether the operation is UPDATE_AFTER.
*/
public static boolean isUpdateAfter(HoodieOperation operation) {
return operation == UPDATE_AFTER;
} | 3.68 |
hbase_ZKUtil_getChildDataAndWatchForNewChildren | /**
* Returns the date of child znodes of the specified znode. Also sets a watch on the specified
* znode which will capture a NodeDeleted event on the specified znode as well as
* NodeChildrenChanged if any children of the specified znode are created or deleted. Returns null
* if the specified node does not exist.... | 3.68 |
open-banking-gateway_EncryptionKeySerde_writeKey | /**
* Write public-private key pair into OutputStream
* @param publicKey Public key of pair
* @param privKey Private key of pair
* @param os Output stream to write to
*/
@SneakyThrows
public void writeKey(PublicKey publicKey, PrivateKey privKey, OutputStream os) {
// Mapper may choose to close the stream if us... | 3.68 |
hudi_CleanPlanner_getFilesToCleanKeepingLatestVersions | /**
* Selects the older versions of files for cleaning, such that it bounds the number of versions of each file. This
* policy is useful, if you are simply interested in querying the table, and you don't want too many versions for a
* single file (i.e., run it with versionsRetained = 1)
*/
private Pair<Boolean, Lis... | 3.68 |
flink_SqlConstraintValidator_getFullConstraints | /** Returns the column constraints plus the table constraints. */
public static List<SqlTableConstraint> getFullConstraints(
List<SqlTableConstraint> tableConstraints, SqlNodeList columnList) {
List<SqlTableConstraint> ret = new ArrayList<>();
columnList.forEach(
column -> {
... | 3.68 |
flink_BeamOperatorStateStore_getListState | /** Currently list state and union-list state is not supported. */
@Override
public ListState<byte[]> getListState(BeamFnApi.StateRequest request) throws Exception {
throw new RuntimeException("Operator list state is still not supported");
} | 3.68 |
morf_RemoveTable_isApplied | /**
* {@inheritDoc}
*
* @see org.alfasoftware.morf.upgrade.SchemaChange#isApplied(Schema, ConnectionResources)
*/
@Override
public boolean isApplied(Schema schema, ConnectionResources database) {
for (String tableName : schema.tableNames()) {
if (tableName.equalsIgnoreCase(tableToBeRemoved.getName())) {
... | 3.68 |
framework_Table_setMultiSelectMode | /**
* Sets the behavior of how the multi-select mode should behave when the
* table is both selectable and in multi-select mode.
* <p>
* Note, that on some clients the mode may not be respected. E.g. on touch
* based devices CTRL/SHIFT base selection method is invalid, so touch based
* browsers always use the {@l... | 3.68 |
framework_CustomLayoutDemo_init | /**
* Initialize Application. Demo components are added to main window.
*/
@Override
public void init() {
final LegacyWindow mainWindow = new LegacyWindow("CustomLayout demo");
setMainWindow(mainWindow);
// set the application to use example -theme
setTheme("tests-components");
// Create custom ... | 3.68 |
flink_OptionalFailure_createFrom | /**
* @return wrapped {@link OptionalFailure} returned by {@code valueSupplier} or wrapped failure
* if {@code valueSupplier} has thrown an {@link Exception}.
*/
public static <T> OptionalFailure<T> createFrom(CheckedSupplier<T> valueSupplier) {
try {
return of(valueSupplier.get());
} catch (Exce... | 3.68 |
querydsl_GroupBy_sortedSet | /**
* Create a new aggregating set expression using a backing TreeSet using the given comparator
*
* @param groupExpression values for this expression will be accumulated into a set
* @param comparator comparator of the created TreeSet instance
* @return wrapper expression
*/
public static <E, F> GroupExpression<... | 3.68 |
hudi_WriteMarkers_stripMarkerSuffix | /**
* Strips the marker file suffix from the input path, i.e., ".marker.[IO_type]".
*
* @param path file path
* @return Stripped path
*/
public static String stripMarkerSuffix(String path) {
return path.substring(0, path.indexOf(HoodieTableMetaClient.MARKER_EXTN));
} | 3.68 |
hadoop_OBSObjectBucketUtils_copyFile | /**
* Copy a single object in the bucket via a COPY operation.
*
* @param owner OBS File System instance
* @param srcKey source object path
* @param dstKey destination object path
* @param size object size
* @throws InterruptedIOException the operation was interrupted
* @throws IOException Other I... | 3.68 |
flink_KerberosLoginProvider_doLogin | /**
* Does kerberos login and sets current user. Must be called when isLoginPossible returns true.
*/
public void doLogin(boolean supportProxyUser) throws IOException {
if (principal != null) {
LOG.info(
"Attempting to login to KDC using principal: {} keytab: {}", principal, keytab);
... | 3.68 |
streampipes_Operations_validatePipeline | /**
* @param pipeline the pipeline to validate
* @return PipelineModificationMessage a message containing desired pipeline modifications
*/
public static PipelineModificationMessage validatePipeline(Pipeline pipeline) throws Exception {
return new PipelineVerificationHandlerV2(pipeline).verifyPipeline();
} | 3.68 |
hadoop_StagingCommitter_deleteStagingUploadsParentDirectory | /**
* Delete the multipart upload staging directory.
* @param context job context
* @throws IOException IO failure
*/
protected void deleteStagingUploadsParentDirectory(JobContext context)
throws IOException {
Path stagingUploadsPath = Paths.getStagingUploadsParentDirectory(
context.getConfigura... | 3.68 |
framework_BaseLayoutTestUI_getTestDescription | /*
* (non-Javadoc)
*
* @see com.vaadin.tests.components.AbstractTestUI#getTestDescription()
*/
@Override
protected String getTestDescription() {
return null;
} | 3.68 |
hudi_SparkPreCommitValidator_validate | /**
* Verify the data written as part of specified instant.
* Throw HoodieValidationException if any unexpected data is written (Example: data files are not readable for some reason).
*/
public void validate(String instantTime, HoodieWriteMetadata<O> writeResult, Dataset<Row> before, Dataset<Row> after) throws Hood... | 3.68 |
hbase_ResponseConverter_isClosed | /**
* Check if the region is closed from a CloseRegionResponse
* @param proto the CloseRegionResponse
* @return the region close state
*/
public static boolean isClosed(final CloseRegionResponse proto) {
if (proto == null || !proto.hasClosed()) return false;
return proto.getClosed();
} | 3.68 |
hbase_StorageClusterStatusModel_getDeadNodes | /** Returns the list of dead nodes */
@XmlElement(name = "Node")
@XmlElementWrapper(name = "DeadNodes")
// workaround https://github.com/FasterXML/jackson-dataformat-xml/issues/192
@JsonProperty("DeadNodes")
public List<String> getDeadNodes() {
return deadNodes;
} | 3.68 |
flink_RestfulGateway_deliverCoordinationRequestToCoordinator | /**
* Deliver a coordination request to a specified coordinator and return the response.
*
* @param jobId identifying the job which the coordinator belongs to
* @param operatorId identifying the coordinator to receive the request
* @param serializedRequest serialized request to deliver
* @param timeout RPC timeou... | 3.68 |
hbase_SpaceLimitSettings_buildProtoRemoveQuota | /**
* Builds a {@link SpaceQuota} protobuf object to remove a quota.
* @return The protobuf SpaceQuota representation.
*/
private SpaceLimitRequest buildProtoRemoveQuota() {
return SpaceLimitRequest.newBuilder().setQuota(SpaceQuota.newBuilder().setRemove(true).build())
.build();
} | 3.68 |
hbase_RegionReplicaCandidateGenerator_selectCoHostedRegionPerGroup | /**
* Randomly select one regionIndex out of all region replicas co-hosted in the same group (a group
* is a server, host or rack)
* @param colocatedReplicaCountsPerGroup either Cluster.colocatedReplicaCountsPerServer,
* colocatedReplicaCountsPerHost or
* ... | 3.68 |
zilla_HpackIntegerFW_integer | /*
* Encodes integer in HPACK representation
*
* if I < 2^N - 1, encode I on N bits
* else
* encode (2^N - 1) on N bits
* I = I - (2^N - 1)
* while I >= 128
* encode (I % 128 + 128) on 8 bits
* I = I / 128
* encode I on 8 bits
*
* @param offset offset for current octe... | 3.68 |
hadoop_QuotaUsage_isTypeConsumedAvailable | /**
* Return true if any storage type consumption information is available.
*
* @return if any storage type consumption information
* is available, not false.
*/
public boolean isTypeConsumedAvailable() {
if (typeConsumed != null) {
for (StorageType t : StorageType.getTypesSupportingQuota()) {
if (type... | 3.68 |
flink_TaskExecutorManager_allocateWorkers | /**
* Allocate a number of workers based on the input param.
*
* @param workerNum the number of workers to allocate
* @return the number of successfully allocated workers
*/
private int allocateWorkers(int workerNum) {
int allocatedWorkerNum = 0;
for (int i = 0; i < workerNum; ++i) {
if (allocateWo... | 3.68 |
flink_RocksDBNativeMetricOptions_enableSizeAllMemTables | /**
* Returns approximate size of active, unflushed immutable, and pinned immutable memtables
* (bytes).
*/
public void enableSizeAllMemTables() {
this.properties.add(RocksDBProperty.SizeAllMemTables.getRocksDBProperty());
} | 3.68 |
framework_UIProvider_getPushMode | /**
* Finds the {@link PushMode} to use for a specific UI. If no specific push
* mode is required, <code>null</code> is returned.
* <p>
* The default implementation uses the @{@link Push} annotation if it's
* defined for the UI class.
*
* @param event
* the UI create event with information about the ... | 3.68 |
hadoop_OBSLoginHelper_extractLoginDetails | /**
* Extract the login details from a URI.
*
* @param name URI of the filesystem
* @return a login tuple, possibly empty.
*/
public static Login extractLoginDetails(final URI name) {
try {
String authority = name.getAuthority();
if (authority == null) {
return Login.EMPTY;
}
int loginIndex... | 3.68 |
hadoop_LengthInputStream_getLength | /** @return the length. */
public long getLength() {
return length;
} | 3.68 |
hbase_HFileBlock_createFromBuff | /**
* Creates a block from an existing buffer starting with a header. Rewinds and takes ownership of
* the buffer. By definition of rewind, ignores the buffer position, but if you slice the buffer
* beforehand, it will rewind to that point.
* @param buf Has header, content, and trailing checksums if present.
*/
st... | 3.68 |
framework_Table_getCurrentWidth | /**
* Get the width in pixels of the column after the resize event.
*
* @return Width in pixels
*/
public int getCurrentWidth() {
return currentWidth;
} | 3.68 |
flink_TemporalTableJoinUtil_isRowTimeTemporalTableJoinCondition | /** Check if the given rexCall is a rewrote join condition on event time. */
public static boolean isRowTimeTemporalTableJoinCondition(RexCall call) {
// (LEFT_TIME_ATTRIBUTE, RIGHT_TIME_ATTRIBUTE, LEFT_KEY, RIGHT_KEY, PRIMARY_KEY)
return call.getOperator() == TemporalJoinUtil.TEMPORAL_JOIN_CONDITION()
... | 3.68 |
hudi_HFileBootstrapIndex_getUserKeyFromCellKey | /**
* HFile stores cell key in the format example : "2020/03/18//LATEST_TIMESTAMP/Put/vlen=3692/seqid=0".
* This API returns only the user key part from it.
* @param cellKey HFIle Cell Key
* @return
*/
private static String getUserKeyFromCellKey(String cellKey) {
int hfileSuffixBeginIndex = cellKey.lastIndexOf(H... | 3.68 |
framework_DragSourceExtensionConnector_isNativeDragEvent | /**
* Returns whether the given event is a native (android) drag start/end
* event, and not produced by the drag-drop-polyfill.
*
* @param nativeEvent
* the event to test
* @return {@code true} if native event, {@code false} if not (polyfill
* event)
*/
protected boolean isNativeDragEvent(Nat... | 3.68 |
hadoop_DiskBalancerDataNode_compareTo | /**
* Compares this object with the specified object for order. Returns a
* negative integer, zero, or a positive integer as this object is less than,
* equal to, or greater than the specified object.
*
* @param that the object to be compared.
* @return a negative integer, zero, or a positive integer as this obj... | 3.68 |
hadoop_HsLogsPage_preHead | /*
* (non-Javadoc)
* @see org.apache.hadoop.mapreduce.v2.hs.webapp.HsView#preHead(org.apache.hadoop.yarn.webapp.hamlet.Hamlet.HTML)
*/
@Override protected void preHead(Page.HTML<__> html) {
commonPreHead(html);
setActiveNavColumnForTask();
} | 3.68 |
hudi_HoodieDataSourceHelpers_allCompletedCommitsCompactions | /**
* Obtain all the commits, compactions that have occurred on the timeline, whose instant times could be fed into the
* datasource options.
*/
@PublicAPIMethod(maturity = ApiMaturityLevel.STABLE)
public static HoodieTimeline allCompletedCommitsCompactions(FileSystem fs, String basePath) {
HoodieTableMetaClient m... | 3.68 |
framework_HierarchicalContainer_hasChildren | /*
* Is the Item corresponding to the given ID a leaf node? Don't add a
* JavaDoc comment here, we use the default documentation from implemented
* interface.
*/
@Override
public boolean hasChildren(Object itemId) {
if (filteredChildren != null) {
return filteredChildren.containsKey(itemId);
} else ... | 3.68 |
morf_AnalyseTable_getTableName | /**
* Method to get the table that is being analysed.
*
* @return {@link Statement} to be executed
*/
public String getTableName() {
return tableName;
} | 3.68 |
hbase_Client_delete | /**
* Send a DELETE request
* @param cluster the cluster definition
* @param path the path or URI
* @return a Response object with response detail
* @throws IOException for error
*/
public Response delete(Cluster cluster, String path, Header extraHdr) throws IOException {
HttpDelete method = new HttpDelete(p... | 3.68 |
hadoop_AltKerberosAuthenticationHandler_authenticate | /**
* It enforces the the Kerberos SPNEGO authentication sequence returning an
* {@link AuthenticationToken} only after the Kerberos SPNEGO sequence has
* completed successfully (in the case of Java access) and only after the
* custom authentication implemented by the subclass in alternateAuthenticate
* has comple... | 3.68 |
hadoop_DockerKillCommand_setSignal | /**
* Set the signal for the {@link DockerKillCommand}.
*
* @param signal the signal to send to the container.
* @return the {@link DockerKillCommand} with the signal set.
*/
public DockerKillCommand setSignal(String signal) {
super.addCommandArguments("signal", signal);
return this;
} | 3.68 |
framework_TextArea_setRows | /**
* Sets the number of rows in the text area.
*
* @param rows
* the number of rows for this text area.
*/
public void setRows(int rows) {
if (rows < 0) {
rows = 0;
}
getState().rows = rows;
} | 3.68 |
flink_KeyMap_get | /**
* Looks up the value mapped under the given key. Returns null if no value is mapped under this
* key.
*
* @param key The key to look up.
* @return The value associated with the key, or null, if no value is found for the key.
* @throws java.lang.NullPointerException Thrown, if the key is null.
*/
public V get... | 3.68 |
hbase_TableHFileArchiveTracker_keepHFiles | /**
* Determine if the given table should or should not allow its hfiles to be deleted
* @param tableName name of the table to check
* @return <tt>true</tt> if its store files should be retained, <tt>false</tt> otherwise
*/
public boolean keepHFiles(String tableName) {
return getMonitor().shouldArchiveTable(table... | 3.68 |
framework_InfoSection_uidl | /*
* (non-Javadoc)
*
* @see com.vaadin.client.debug.internal.Section#uidl(com.vaadin.client.
* ApplicationConnection, com.vaadin.client.ValueMap)
*/
@Override
public void uidl(ApplicationConnection ac, ValueMap uidl) {
} | 3.68 |
dubbo_StringUtils_toString | /**
* @param msg
* @param e
* @return string
*/
public static String toString(String msg, Throwable e) {
UnsafeStringWriter w = new UnsafeStringWriter();
w.write(msg + "\n");
PrintWriter p = new PrintWriter(w);
try {
e.printStackTrace(p);
return w.toString();
} finally {
... | 3.68 |
flink_BlobCacheSizeTracker_untrack | /** Remove the BLOB from the tracker. */
private void untrack(JobID jobId, BlobKey blobKey) {
checkNotNull(jobId);
checkNotNull(blobKey);
untrack(Tuple2.of(jobId, blobKey));
} | 3.68 |
framework_DateField_setLenient | /**
* Specifies whether or not date/time interpretation in component is to be
* lenient.
*
* @see Calendar#setLenient(boolean)
* @see #isLenient()
*
* @param lenient
* true if the lenient mode is to be turned on; false if it is to
* be turned off.
*/
public void setLenient(boolean lenien... | 3.68 |
hudi_ExpressionPredicates_bindFieldReference | /**
* Binds field reference to create a column predicate.
*
* @param fieldReference The field reference to negate.
* @return A column predicate.
*/
public ColumnPredicate bindFieldReference(FieldReferenceExpression fieldReference) {
this.literalType = fieldReference.getOutputDataType().getLogicalType();
this.c... | 3.68 |
hadoop_OBSFileSystem_initialize | /**
* Initialize a FileSystem. Called after a new FileSystem instance is
* constructed.
*
* @param name a URI whose authority section names the host, port,
* etc. for this FileSystem
* @param originalConf the configuration to use for the FS. The
* bucket-specific o... | 3.68 |
flink_ExtractionUtils_createMethodSignatureString | /** Creates a method signature string like {@code int eval(Integer, String)}. */
public static String createMethodSignatureString(
String methodName, Class<?>[] parameters, @Nullable Class<?> returnType) {
final StringBuilder builder = new StringBuilder();
if (returnType != null) {
builder.appen... | 3.68 |
morf_UpgradeTestHelper_validateUpgradeStepProperties | /**
* Validate that each upgrade step meets the basic requirements.
* For example UUID, Sequence, JIRA ID and Description are all populated.
*/
public void validateUpgradeStepProperties(Iterable<Class<? extends UpgradeStep>> upgradeSteps) {
instantiateAndValidateUpgradeSteps(upgradeSteps);
} | 3.68 |
pulsar_WRRPlacementStrategy_findBrokerForPlacement | /**
* Function : getByWeightedRoundRobin returns ResourceUnit selected by WRR algorithm
* based on available resource on RU.
* <code>
* ^
* |
* |
* |
* | | | | |
* | | | ... | 3.68 |
framework_IndexedContainer_equals | /**
* Tests if the given object is the same as the this object. Two
* Properties got from an Item with the same ID are equal.
*
* @param obj
* an object to compare with this object
* @return <code>true</code> if the given object is the same as this
* object, <code>false</code> if not
*/
@Over... | 3.68 |
hudi_SparkInternalSchemaConverter_convertIntLongType | /**
* Convert Int/long type to other Type.
* Now only support int/long -> long/float/double/string/Decimal
* TODO: support more types
*/
private static boolean convertIntLongType(WritableColumnVector oldV, WritableColumnVector newV, DataType newType, int len) {
boolean isInt = oldV.dataType() instanceof IntegerTy... | 3.68 |
framework_ShowLastItem_setup | /*
* (non-Javadoc)
*
* @see com.vaadin.tests.components.AbstractTestUI#setup(com.vaadin.server.
* VaadinRequest)
*/
@Override
protected void setup(VaadinRequest request) {
final Table table = new Table();
table.setHeight("210px");
table.addContainerProperty("Col", String.class, "");
for (int i = ... | 3.68 |
flink_ScriptProcessBuilder_getAbsolutePath | /** Returns the full path name of this file if it is listed in the path. */
public File getAbsolutePath(String filename) {
if (pathenv == null || pathSep == null || fileSep == null) {
return null;
}
int val;
String classvalue = pathenv + pathSep;
while (((val = classvalue.indexOf(pathSep))... | 3.68 |
framework_VDebugWindow_readPositionAndSize | /**
* Reads position and size from the DOM to local variables (which in turn
* can be stored to localStorage)
*/
private void readPositionAndSize() {
int x = getPopupLeft();
int fromRight = Window.getClientWidth() - x - getOffsetWidth();
if (fromRight < x) {
x -= Window.getClientWidth();
}
... | 3.68 |
hadoop_NMClient_createNMClient | /**
* Create a new instance of NMClient.
*/
@Public
public static NMClient createNMClient(String name) {
NMClient client = new NMClientImpl(name);
return client;
} | 3.68 |
flink_ExpressionResolver_resolveExpanding | /**
* Resolves given expressions with configured set of rules. All expressions of an operation
* should be given at once as some rules might assume the order of expressions.
*
* <p>After this method is applied the returned expressions might contain unresolved expression
* that can be used for further API transform... | 3.68 |
graphhopper_WaySegmentParser_setWayPreprocessor | /**
* @param wayPreprocessor callback function that is called for each accepted OSM way during the second pass
*/
public Builder setWayPreprocessor(WayPreprocessor wayPreprocessor) {
waySegmentParser.wayPreprocessor = wayPreprocessor;
return this;
} | 3.68 |
hbase_ClientExceptionsUtil_findException | /**
* Look for an exception we know in the remote exception: - hadoop.ipc wrapped exceptions - nested
* exceptions Looks for: RegionMovedException / RegionOpeningException / RegionTooBusyException /
* RpcThrottlingException
* @return null if we didn't find the exception, the exception otherwise.
*/
public static T... | 3.68 |
hadoop_FlowActivityRowKey_getRowKey | /**
* Constructs a row key for the flow activity table as follows:
* {@code clusterId!dayTimestamp!user!flowName}.
*
* @return byte array for the row key
*/
public byte[] getRowKey() {
return flowActivityRowKeyConverter.encode(this);
} | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.