name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
framework_Table_setColumnHeaders | /**
* Sets the headers of the columns.
*
* <p>
* The headers match the property id:s given by the set visible column
* headers. The table must be set in either
* {@link #COLUMN_HEADER_MODE_EXPLICIT} or
* {@link #COLUMN_HEADER_MODE_EXPLICIT_DEFAULTS_ID} mode to show the
* headers. In the defaults mode any nulls ... | 3.68 |
druid_DruidPooledConnection_getPhysicalConnectNanoSpan | /**
* @since 1.0.17
*/
public long getPhysicalConnectNanoSpan() {
return this.holder.getCreateNanoSpan();
} | 3.68 |
framework_LayoutManager_addElementResizeListener | /**
* Adds a listener that will be notified whenever the size of a specific
* element changes. Adding a listener to an element also ensures that all
* sizes for that element will be available starting from the next layout
* phase.
*
* @param element
* the element that should be checked for size change... | 3.68 |
hbase_SnapshotInfo_isMissing | /** Returns true if the file is missing */
public boolean isMissing() {
return this.size < 0;
} | 3.68 |
hmily_MapBinder_bindEntries | /**
* Bind entries.
*
* @param source the source
* @param map the map
*/
void bindEntries(final ConfigPropertySource source, final Map<Object, Object> map) {
source.stream().forEach(name -> {
boolean ancestorOf = root.isAncestorOf(name);
if (ancestorOf) {
BindData<?> valueBindDat... | 3.68 |
flink_RemoteInputChannel_requestBuffer | /**
* Requests buffer from input channel directly for receiving network data. It should always
* return an available buffer in credit-based mode unless the channel has been released.
*
* @return The available buffer.
*/
@Nullable
public Buffer requestBuffer() {
return bufferManager.requestBuffer();
} | 3.68 |
hudi_HoodieExampleDataGenerator_generateInsertsOnPartition | /**
* Generates new inserts, across a single partition path. It also updates the list of existing keys.
*/
public List<HoodieRecord<T>> generateInsertsOnPartition(String commitTime, Integer n, String partitionPath) {
return generateInsertsStreamOnPartition(commitTime, n, partitionPath).collect(Collectors.toList());... | 3.68 |
flink_EntropyInjector_addEntropy | /**
* Handles entropy injection across regular and entropy-aware file systems.
*
* <p>If the given file system is entropy-aware (a implements {@link
* EntropyInjectingFileSystem}), then this method replaces the entropy marker in the path with
* random characters. The entropy marker is defined by {@link
* EntropyI... | 3.68 |
hbase_ServerRegionReplicaUtil_isRegionReplicaStoreFileRefreshEnabled | /** Returns True if we are to refresh user-space hfiles in Region Read Replicas. */
public static boolean isRegionReplicaStoreFileRefreshEnabled(Configuration conf) {
return conf.getBoolean(REGION_REPLICA_STORE_FILE_REFRESH,
DEFAULT_REGION_REPLICA_STORE_FILE_REFRESH);
} | 3.68 |
morf_SqlDialect_escapeSql | /**
* Turn a string value into an SQL string literal which has that value.
* This escapes single quotes as double-single quotes.
* @param literalValue the value to escape
* @return escaped value
*/
protected String escapeSql(String literalValue) {
if (literalValue == null) {
return null;
}
return StringUtil... | 3.68 |
hadoop_FutureDataInputStreamBuilderImpl_bufferSize | /**
* Set the size of the buffer to be used.
*
* @param bufSize buffer size.
* @return FutureDataInputStreamBuilder.
*/
public FutureDataInputStreamBuilder bufferSize(int bufSize) {
bufferSize = bufSize;
return getThisBuilder();
} | 3.68 |
hbase_MetricsConnection_getScanTracker | /** scanTracker metric */
public CallTracker getScanTracker() {
return scanTracker;
} | 3.68 |
hbase_SaslClientAuthenticationProviders_addProviderIfNotExists | /**
* Adds the given {@code provider} to the set, only if an equivalent provider does not already
* exist in the set.
*/
static void addProviderIfNotExists(SaslClientAuthenticationProvider provider,
HashMap<Byte, SaslClientAuthenticationProvider> providers) {
Byte code = provider.getSaslAuthMethod().getCode();
... | 3.68 |
flink_CrossOperator_projectSecond | /**
* Continues a ProjectCross transformation and adds fields of the second cross input.
*
* <p>If the second cross input is a {@link Tuple} {@link DataSet}, fields can be selected
* by their index. If the second cross input is not a Tuple DataSet, no parameters should be
* passed.
*
* <p>Fields of the first and... | 3.68 |
querydsl_MultiSurfaceExpression_centroid | /**
* The mathematical centroid for this MultiSurface. The result is not guaranteed to be on
* this MultiSurface.
*
* @return centroid
*/
public PointExpression<Point> centroid() {
if (centroid == null) {
centroid = GeometryExpressions.pointOperation(SpatialOps.CENTROID, mixin);
}
return centro... | 3.68 |
morf_AbstractSqlDialectTest_shouldGenerateCorrectSqlForMathOperations7 | /**
* Test for proper SQL mathematics operation generation from DSL expressions
* that use brackets.
* <p>
* Subexpression "a+b" is put to bracket explicitly, and
* the subexpression "c-d" should be put to the bracket implicitly, even without explicit
* {@link org.alfasoftware.morf.sql.SqlUtils#bracket(MathsField... | 3.68 |
hbase_RegionInfo_isEncodedRegionName | /**
* Figure if the passed bytes represent an encoded region name or not.
* @param regionName A Region name either encoded or not.
* @return True if <code>regionName</code> represents an encoded name.
*/
@InterfaceAudience.Private // For use by internals only.
public static boolean isEncodedRegionName(byte[] region... | 3.68 |
hadoop_OBSBlockOutputStream_complete | /**
* This completes a multipart upload. Sometimes it fails; here retries are
* handled to avoid losing all data on a transient failure.
*
* @param partETags list of partial uploads
* @return result for completing multipart upload
* @throws IOException on any problem
*/
private CompleteMultipartUploadResult comp... | 3.68 |
framework_Payload_setKey | /**
* Sets the key of this payload.
*
* @param key
* key that identifies the payload
*/
public void setKey(String key) {
this.key = key;
} | 3.68 |
framework_VUpload_disableTitle | /**
* For internal use only. May be removed or replaced in the future.
*
* @param disable
* {@code true} if the built-in browser-dependent tooltip should
* be hidden in favor of a Vaadin tooltip, {@code false}
* otherwise
*/
public void disableTitle(boolean disable) {
if (dis... | 3.68 |
hbase_ParseFilter_removeQuotesFromByteArray | /**
* Takes a quoted byte array and converts it into an unquoted byte array For example: given a byte
* array representing 'abc', it returns a byte array representing abc
* <p>
* @param quotedByteArray the quoted byte array
* @return Unquoted byte array
*/
public static byte[] removeQuotesFromByteArray(byte[] quo... | 3.68 |
hbase_SnapshotInfo_getStoreFilesSize | /** Returns the total size of the store files referenced by the snapshot */
public long getStoreFilesSize() {
return hfilesSize.get() + hfilesArchiveSize.get() + hfilesMobSize.get();
} | 3.68 |
hbase_MasterObserver_preDeleteSnapshot | /**
* Called before a snapshot is deleted. Called as part of deleteSnapshot RPC call.
* @param ctx the environment to interact with the framework and master
* @param snapshot the SnapshotDescriptor of the snapshot to delete
*/
default void preDeleteSnapshot(final ObserverContext<MasterCoprocessorEnvironment> c... | 3.68 |
flink_FlinkJoinToMultiJoinRule_combineOuterJoins | /**
* Combines the outer join conditions and join types from the left and right join inputs. If the
* join itself is either a left or right outer join, then the join condition corresponding to
* the join is also set in the position corresponding to the null-generating input into the
* join. The join type is also se... | 3.68 |
hbase_Scan_includeStartRow | /** Returns if we should include start row when scan */
public boolean includeStartRow() {
return includeStartRow;
} | 3.68 |
graphhopper_VectorTile_getFeaturesList | /**
* <pre>
* The actual features in this tile.
* </pre>
*
* <code>repeated .vector_tile.Tile.Feature features = 2;</code>
*/
public java.util.List<vector_tile.VectorTile.Tile.Feature> getFeaturesList() {
if (featuresBuilder_ == null) {
return java.util.Collections.unmodifiableList(features_);
} else {
... | 3.68 |
hudi_HoodieTableMetaClient_getBasePathV2 | /**
* Returns base path of the table
*/
public Path getBasePathV2() {
return basePath.get();
} | 3.68 |
shardingsphere-elasticjob_JobNodeStorage_getJobNodeData | /**
* Get job node data.
*
* @param node node
* @return data of job node
*/
public String getJobNodeData(final String node) {
return regCenter.get(jobNodePath.getFullPath(node));
} | 3.68 |
hadoop_BytesWritable_setCapacity | /**
* Change the capacity of the backing storage. The data is preserved.
*
* @param capacity The new capacity in bytes.
*/
public void setCapacity(final int capacity) {
if (capacity != getCapacity()) {
this.size = Math.min(size, capacity);
this.bytes = Arrays.copyOf(this.bytes, capacity);
}
} | 3.68 |
morf_AbstractSqlDialectTest_testInsertFromSelectWithSourceInDifferentSchema | /**
* Tests that an insert from a select works when the source table is in a different schema.
*/
@Test
public void testInsertFromSelectWithSourceInDifferentSchema() {
SelectStatement sourceStmt = new SelectStatement(new FieldReference("id"),
new FieldReference("version"),
new FieldReference(STRING_FIELD),
... | 3.68 |
hudi_RequestHandler_registerDataFilesAPI | /**
* Register Data-Files API calls.
*/
private void registerDataFilesAPI() {
app.get(RemoteHoodieTableFileSystemView.LATEST_PARTITION_DATA_FILES_URL, new ViewHandler(ctx -> {
metricsRegistry.add("LATEST_PARTITION_DATA_FILES", 1);
List<BaseFileDTO> dtos = dataFileHandler.getLatestDataFiles(
ctx.quer... | 3.68 |
hbase_ReplicationSourceManager_getOldLogDir | /**
* Get the directory where wals are archived
* @return the directory where wals are archived
*/
public Path getOldLogDir() {
return this.oldLogDir;
} | 3.68 |
hudi_BufferedRandomAccessFile_init | /**
*
* @param size - capacity of the buffer
*/
private void init(int size) {
this.capacity = Math.max(DEFAULT_BUFFER_SIZE, size);
this.dataBuffer = ByteBuffer.wrap(new byte[this.capacity]);
} | 3.68 |
framework_ContainerHierarchicalWrapper_removeAllItems | /**
* Removes all items from the underlying container and from the hierarcy.
*
* @return <code>true</code> if the operation succeeded, <code>false</code>
* if not
* @throws UnsupportedOperationException
* if the removeAllItems is not supported.
*/
@Override
public boolean removeAllItems() thr... | 3.68 |
morf_SqlScriptExecutor_executeStatementBatch | /**
* Runs the specified SQL statement (which should contain parameters), repeatedly for
* each record, mapping the contents of the records into the statement parameters in
* their defined order. Use to insert, merge or update a large batch of records
* efficiently.
*
* @param sqlStatement the SQL statement.
* ... | 3.68 |
hbase_FirstKeyValueMatchingQualifiersFilter_parseFrom | /**
* Parses a serialized representation of {@link FirstKeyValueMatchingQualifiersFilter}
* @param pbBytes A pb serialized {@link FirstKeyValueMatchingQualifiersFilter} instance
* @return An instance of {@link FirstKeyValueMatchingQualifiersFilter} made from
* <code>bytes</code>
* @throws DeserializationEx... | 3.68 |
querydsl_PathBuilder_getComparable | /**
* Create a new Comparable typed path
*
* @param <A>
* @param property property name
* @param type property type
* @return property path
*/
@SuppressWarnings("unchecked")
public <A extends Comparable<?>> ComparablePath<A> getComparable(String property, Class<A> type) {
Class<? extends A> vtype = validate(... | 3.68 |
flink_HiveParserTypeCheckCtx_getUnparseTranslator | /** @return the unparseTranslator */
public HiveParserUnparseTranslator getUnparseTranslator() {
return unparseTranslator;
} | 3.68 |
hadoop_AbfsDtFetcher_getServiceName | /**
* Returns the service name for the scheme..
*/
public Text getServiceName() {
return new Text(getScheme());
} | 3.68 |
flink_PackagedProgram_invokeInteractiveModeForExecution | /**
* This method assumes that the context environment is prepared, or the execution will be a
* local execution by default.
*/
public void invokeInteractiveModeForExecution() throws ProgramInvocationException {
FlinkSecurityManager.monitorUserSystemExitForCurrentThread();
try {
callMainMethod(mainCl... | 3.68 |
dubbo_ScopeClusterInvoker_init | /**
* Initializes the ScopeClusterInvoker instance.
*/
private void init() {
Boolean peer = (Boolean) getUrl().getAttribute(PEER_KEY);
String isInjvm = getUrl().getParameter(LOCAL_PROTOCOL);
// When the point-to-point direct connection is directly connected,
// the initialization is directly ended
... | 3.68 |
hbase_SnapshotOfRegionAssignmentFromMeta_getExistingAssignmentPlan | /**
* Get the favored nodes plan
* @return the existing favored nodes plan
*/
public FavoredNodesPlan getExistingAssignmentPlan() {
return this.existingAssignmentPlan;
} | 3.68 |
hbase_UnsafeAccess_copy | /**
* Copies specified number of bytes from given offset of {@code src} buffer into the {@code dest}
* buffer.
* @param src source buffer
* @param srcOffset offset into source buffer
* @param dest destination buffer
* @param destOffset offset into destination buffer
* @param length length of da... | 3.68 |
framework_AbstractProperty_removeListener | /**
* @deprecated As of 7.0, replaced by
* {@link #removeValueChangeListener(Property.ValueChangeListener)}
*/
@Override
@Deprecated
public void removeListener(ValueChangeListener listener) {
removeValueChangeListener(listener);
} | 3.68 |
framework_AbsoluteLayout_setRight | /**
* Sets the 'right' attribute; distance from the right of the component
* to the right edge of the layout.
*
* @param rightValue
* The value of the 'right' attribute
* @param rightUnits
* The unit of the 'right' attribute. See UNIT_SYMBOLS for a
* description of the available... | 3.68 |
flink_ProcessFunction_onTimer | /**
* Called when a timer set using {@link TimerService} fires.
*
* @param timestamp The timestamp of the firing timer.
* @param ctx An {@link OnTimerContext} that allows querying the timestamp of the firing timer,
* querying the {@link TimeDomain} of the firing timer and getting a {@link TimerService}
* ... | 3.68 |
framework_ColorPickerGrid_getPosition | /**
* Gets the position.
*
* @return the position
*/
public int[] getPosition() {
return new int[] { x, y };
} | 3.68 |
incubator-hugegraph-toolchain_FileLineFetcher_checkMatchHeader | /**
* Just match header for second or subsequent file first line
*/
private boolean checkMatchHeader(String line) {
if (!this.source().format().needHeader() ||
this.offset() != FIRST_LINE_OFFSET) {
return false;
}
assert this.source().header() != null;
String[] columns = this.parser.s... | 3.68 |
framework_JavaScriptConnectorHelper_removeResizeListener | // Called from JSNI to remove a listener
private void removeResizeListener(Element element,
JavaScriptObject callbackFunction) {
Map<JavaScriptObject, ElementResizeListener> listenerMap = resizeListeners
.get(element);
if (listenerMap == null) {
return;
}
ElementResizeListen... | 3.68 |
flink_StreamExecutionEnvironment_addSource | /**
* Ads a data source with a custom type information thus opening a {@link DataStream}. Only in
* very special cases does the user need to support type information. Otherwise use {@link
* #addSource(org.apache.flink.streaming.api.functions.source.SourceFunction)}
*
* @param function the user defined function
* ... | 3.68 |
zxing_BitArray_get | /**
* @param i bit to get
* @return true iff bit i is set
*/
public boolean get(int i) {
return (bits[i / 32] & (1 << (i & 0x1F))) != 0;
} | 3.68 |
dubbo_TTable_getBorder | /**
* get border
*
* @return table border
*/
public Border getBorder() {
return border;
} | 3.68 |
hbase_WALFactory_getInstance | // Public only for FSHLog
public static WALFactory getInstance(Configuration configuration) {
WALFactory factory = singleton.get();
if (null == factory) {
WALFactory temp = new WALFactory(configuration);
if (singleton.compareAndSet(null, temp)) {
factory = temp;
} else {
// someone else beat... | 3.68 |
flink_ListView_setList | /** Replaces the entire view's content with the content of the given {@link List}. */
public void setList(List<T> list) {
this.list = list;
} | 3.68 |
framework_AbstractClientConnector_detach | /**
* {@inheritDoc}
*
* <p>
* The {@link #getSession()} and {@link #getUI()} methods might return
* <code>null</code> after this method is called.
* </p>
*/
@Override
public void detach() {
for (ClientConnector connector : getAllChildrenIterable(this)) {
connector.detach();
}
fireEvent(new D... | 3.68 |
hbase_TableMapReduceUtil_convertScanToString | /**
* Writes the given scan into a Base64 encoded string.
* @param scan The scan to write out.
* @return The scan saved in a Base64 encoded string.
* @throws IOException When writing the scan fails.
*/
public static String convertScanToString(Scan scan) throws IOException {
ClientProtos.Scan proto = ProtobufUtil... | 3.68 |
hadoop_ReencryptionHandler_throttle | /**
* Throttles the ReencryptionHandler in 3 aspects:
* 1. Prevents generating more Callables than the CPU could possibly
* handle.
* 2. Prevents generating more Callables than the ReencryptionUpdater
* can handle, under its own throttling.
* 3. Prevents contending FSN/FSD read locks. This is done based
* on the... | 3.68 |
hbase_AuthManager_authorizeUserNamespace | /**
* Check if user has given action privilige in namespace scope.
* @param user user name
* @param namespace namespace
* @param action one of action in [Read, Write, Create, Exec, Admin]
* @return true if user has, false otherwise
*/
public boolean authorizeUserNamespace(User user, String namespace, Perm... | 3.68 |
morf_ViewURLAsFile_createTempFile | /**
* Wrapper for {@link java.io.File#createTempFile(String, String, File)} that
* wraps any exceptions in a {@link RuntimeException} and propagates it.
*/
private File createTempFile(String prefix, String suffix, File file) {
try {
return File.createTempFile(prefix, suffix, file);
} catch (IOException e) {
... | 3.68 |
hadoop_DistributedCache_setFileTimestamps | /**
* This is to check the timestamp of the files to be localized.
* Used by internal MapReduce code.
* @param conf Configuration which stores the timestamp's
* @param timestamps comma separated list of timestamps of files.
* The order should be the same as the order in which the files are added.
*/
@Deprecated
p... | 3.68 |
framework_ConnectorBundleLoader_ensureDeferredBundleLoaded | /**
* Starts loading the deferred bundle if it hasn't already been started.
*
* @since 8.0.3
*/
public void ensureDeferredBundleLoaded() {
if (!isBundleLoaded(DEFERRED_BUNDLE_NAME)) {
loadBundle(DEFERRED_BUNDLE_NAME, new BundleLoadCallback() {
@Override
public void loaded() {
... | 3.68 |
hadoop_JobBase_getLongValue | /**
*
* @param name
* the counter name
* @return return the value of the given counter.
*/
protected Long getLongValue(Object name) {
return this.longCounters.get(name);
} | 3.68 |
hadoop_OBSBlockOutputStream_clearHFlushOrSync | /**
* Clear for hflush or hsync.
*/
private synchronized void clearHFlushOrSync() {
appendAble.set(true);
multiPartUpload = null;
} | 3.68 |
framework_NullValidator_isNullAllowed | /**
* Returns <code>true</code> if nulls are allowed otherwise
* <code>false</code>.
*/
public final boolean isNullAllowed() {
return onlyNullAllowed;
} | 3.68 |
framework_VAbstractSplitPanel_setSecondWidget | /**
* For internal use only. May be removed or replaced in the future.
*
* @param w
* the widget to set to the second region or {@code null} to
* remove previously set widget
*/
public void setSecondWidget(Widget w) {
if (secondChild == w) {
return;
}
if (secondChild != n... | 3.68 |
hadoop_ArrayFile_get | /**
* Return the <code>n</code>th value in the file.
* @param n n key.
* @param value value.
* @throws IOException raised on errors performing I/O.
* @return writable.
*/
public synchronized Writable get(long n, Writable value)
throws IOException {
key.set(n);
return get(key, value);
} | 3.68 |
MagicPlugin_CompatibilityUtilsBase_loadChunk | /**
* This will load chunks asynchronously if possible.
*
* <p>But note that it will never be truly asynchronous, it is important not to call this in a tight retry loop,
* the main server thread needs to free up to actually process the async chunk loads.
*/
@Override
public void loadChunk(World world, int x, int z... | 3.68 |
hudi_HoodieAvroUtils_bytesToAvro | /**
* Convert serialized bytes back into avro record.
*/
public static GenericRecord bytesToAvro(byte[] bytes, Schema writerSchema, Schema readerSchema) throws IOException {
BinaryDecoder decoder = DecoderFactory.get().binaryDecoder(bytes, BINARY_DECODER.get());
BINARY_DECODER.set(decoder);
GenericDatumReader<G... | 3.68 |
hbase_ServerNonceManager_startOperation | /**
* Starts the operation if operation with such nonce has not already succeeded. If the operation
* is in progress, waits for it to end and checks whether it has succeeded.
* @param group Nonce group.
* @param nonce Nonce.
* @param stoppable Stoppable that terminates waiting (if any) when the server is s... | 3.68 |
hadoop_NodeIDsInfo_add | /**
* This method will generate a new NodeIDsInfo object based on the two NodeIDsInfo objects.
* The information to be combined includes the node list (removed duplicate node)
* and partitionInfo object.
*
* @param left left NodeIDsInfo Object.
* @param right right NodeIDsInfo Object.
* @return new NodeIDsInfo O... | 3.68 |
framework_HasStyleNames_addStyleNames | /**
* Adds one or more style names to this component by using one or multiple
* parameters.
*
* @since 8.7
* @param styles
* the style name or style names to be added to the component
* @see #addStyleName(String)
* @see #setStyleName(String)
* @see #removeStyleName(String)
*/
default void addStyleN... | 3.68 |
morf_SqlServerDialect_internalTableDeploymentStatements | /**
* @see org.alfasoftware.morf.jdbc.SqlDialect#tableDeploymentStatements(org.alfasoftware.morf.metadata.Table)
*/
@Override
public Collection<String> internalTableDeploymentStatements(Table table) {
List<String> statements = new ArrayList<>();
// Create the table deployment statement
StringBuilder createTabl... | 3.68 |
hudi_HoodieKeyLookupHandle_addKey | /**
* Adds the key for look up.
*/
public void addKey(String recordKey) {
// check record key against bloom filter of current file & add to possible keys if needed
if (bloomFilter.mightContain(recordKey)) {
if (LOG.isDebugEnabled()) {
LOG.debug("Record key " + recordKey + " matches bloom filter in " + ... | 3.68 |
hudi_HoodieIngestionService_requestShutdownIfNeeded | /**
* To determine if shutdown should be requested to allow gracefully terminate the ingestion in continuous mode.
* <p>
* Subclasses should implement the logic to make the decision. If the shutdown condition is met, the implementation
* should call {@link #shutdown(boolean)} to indicate the request.
*
* @see Pos... | 3.68 |
hadoop_MawoConfiguration_readConfigFile | /**
* Find, read, and parse the configuration file.
*
* @return the properties that were found or empty if no file was found
*/
private static Properties readConfigFile() {
Properties properties = new Properties();
// Get property file stream from classpath
LOG.info("Configuration file being loaded: " + CONF... | 3.68 |
pulsar_RangeCache_getNumberOfEntries | /**
* Just for testing. Getting the number of entries is very expensive on the conncurrent map
*/
protected long getNumberOfEntries() {
return entries.size();
} | 3.68 |
flink_MimeTypes_getMimeTypeForFileName | /**
* Gets the MIME type for the file with the given name, by extension. This method tries to
* extract the file extension and then use the {@link #getMimeTypeForExtension(String)} to
* determine the MIME type. If the extension cannot be determined, or the extension is
* unrecognized, this method return {@code null... | 3.68 |
framework_VLoadingIndicator_isVisible | /**
* Returns whether or not the loading indicator is showing.
*
* @return true if the loading indicator is visible, false otherwise
*/
public boolean isVisible() {
if (getElement().getStyle().getDisplay()
.equals(Display.NONE.getCssName())) {
return false;
}
return true;
} | 3.68 |
framework_VaadinSession_getCurrent | /**
* Gets the currently used session. The current session is automatically
* defined when processing requests related to the session (see
* {@link ThreadLocal}) and in {@link VaadinSession#access(Runnable)} and
* {@link UI#access(Runnable)}. In other cases, (e.g. from background
* threads, the current session is ... | 3.68 |
flink_MapView_putAll | /**
* Inserts all mappings from the specified map to this map view.
*
* @param map The map whose entries are inserted into this map view.
* @throws Exception Thrown if the system cannot access the map.
*/
public void putAll(Map<K, V> map) throws Exception {
this.map.putAll(map);
} | 3.68 |
flink_BlobServer_getPort | /**
* Returns the port on which the server is listening.
*
* @return port on which the server is listening
*/
@Override
public int getPort() {
return this.serverSocket.getLocalPort();
} | 3.68 |
zxing_MinimalECIInput_isFNC1 | /**
* Determines if a value is the FNC1 character
*
* @param index the index of the value
*
* @return true if the value at position {@code index} is the FNC1 character
*
* @throws IndexOutOfBoundsException
* if the {@code index} argument is negative or not less than
* {@code length()}
*/... | 3.68 |
flink_InputGate_getChannelInfos | /** Returns the channel infos of this gate. */
public List<InputChannelInfo> getChannelInfos() {
return IntStream.range(0, getNumberOfInputChannels())
.mapToObj(index -> getChannel(index).getChannelInfo())
.collect(Collectors.toList());
} | 3.68 |
framework_AbsoluteLayout_toString | /*
* (non-Javadoc)
*
* @see java.lang.Object#toString()
*/
@Override
public String toString() {
return getCSSString();
} | 3.68 |
hudi_AWSDmsAvroPayload_handleDeleteOperation | /**
*
* Handle a possible delete - check for "D" in Op column and return empty row if found.
* @param insertValue The new row that is being "inserted".
*/
private Option<IndexedRecord> handleDeleteOperation(IndexedRecord insertValue) throws IOException {
boolean delete = false;
if (insertValue instanceof Generi... | 3.68 |
flink_StreamProjection_projectTuple24 | /**
* Projects a {@link Tuple} {@link DataStream} to the previously selected fields.
*
* @return The projected DataStream.
* @see Tuple
* @see DataStream
*/
public <
T0,
T1,
T2,
T3,
T4,
T5,
T6,
... | 3.68 |
hadoop_FileIoProvider_read | /**
* {@inheritDoc}.
*/
@Override
public int read(@Nonnull byte[] b, int off, int len) throws IOException {
final long begin = profilingEventHook.beforeFileIo(volume, READ, len);
try {
faultInjectorEventHook.beforeFileIo(volume, READ, len);
int numBytesRead = super.read(b, off, len);
profilingEventHoo... | 3.68 |
flink_NullableSerializer_wrap | /**
* This method wraps the {@code originalSerializer} with the {@code NullableSerializer} if not
* already wrapped.
*
* @param originalSerializer serializer to wrap and add {@code null} support
* @param padNullValueIfFixedLen pad null value to preserve the fixed length of original
* serializer
* @return wra... | 3.68 |
hadoop_AbfsOutputStream_writeAppendBlobCurrentBufferToService | /**
* Appending the current active data block to service. Clearing the active
* data block and releasing all buffered data.
* @throws IOException if there is any failure while starting an upload for
* the dataBlock or while closing the BlockUploadData.
*/
private void writeAppendBlobCurrentBuff... | 3.68 |
hadoop_Base64_decodeAsByteObjectArray | /**
* Decodes a given Base64 string into its corresponding byte array.
*
* @param data
* the Base64 string, as a <code>String</code> object, to decode
*
* @return the corresponding decoded byte array
* @throws IllegalArgumentException
* If the string is not a valid base64 encoded string
... | 3.68 |
framework_ScrollbarBundle_isScrollerVisible | /**
* Checks whether the scroll handle is currently visible or not.
*
* @return <code>true</code> if the scroll handle is currently visible.
* <code>false</code> if not.
*/
public boolean isScrollerVisible() {
return isScrollerVisible;
} | 3.68 |
framework_Button_getRelativeY | /**
* Returns the relative mouse position (y coordinate) when the click
* took place. The position is relative to the clicked component.
*
* @return The mouse cursor y position relative to the clicked layout
* component or -1 if no y coordinate available
*/
public int getRelativeY() {
if (null != deta... | 3.68 |
hibernate-validator_NotEmptyValidatorForArraysOfFloat_isValid | /**
* Checks the array is not {@code null} and not empty.
*
* @param array the array to validate
* @param constraintValidatorContext context in which the constraint is evaluated
* @return returns {@code true} if the array is not {@code null} and the array is not empty
*/
@Override
public boolean isValid(float[] a... | 3.68 |
querydsl_GenericExporter_setTargetFolder | /**
* Set the target folder for generated sources
*
* @param targetFolder
*/
public void setTargetFolder(File targetFolder) {
this.targetFolder = targetFolder;
} | 3.68 |
hudi_FileSystemBasedLockProvider_getLockConfig | /**
* Returns a filesystem based lock config with given table path.
*/
public static TypedProperties getLockConfig(String tablePath) {
TypedProperties props = new TypedProperties();
props.put(HoodieLockConfig.LOCK_PROVIDER_CLASS_NAME.key(), FileSystemBasedLockProvider.class.getName());
props.put(HoodieLockConfi... | 3.68 |
hadoop_RequestFactoryImpl_withMultipartPartCountLimit | /**
* Multipart limit.
* @param value new value
* @return the builder
*/
public RequestFactoryBuilder withMultipartPartCountLimit(
final long value) {
multipartPartCountLimit = value;
return this;
} | 3.68 |
hbase_PermissionStorage_removeNamespacePermissions | /**
* Remove specified namespace from the acl table.
*/
static void removeNamespacePermissions(Configuration conf, String namespace, Table t)
throws IOException {
Delete d = new Delete(Bytes.toBytes(toNamespaceEntry(namespace)));
d.addFamily(ACL_LIST_FAMILY);
if (LOG.isDebugEnabled()) {
LOG.debug("Removin... | 3.68 |
framework_Table_getItemDescriptionGenerator | /**
* Get the item description generator which generates tooltips for cells and
* rows in the Table.
*/
public ItemDescriptionGenerator getItemDescriptionGenerator() {
return itemDescriptionGenerator;
} | 3.68 |
hbase_StoreFileScanner_seekToPreviousRowStateless | /**
* This variant of the {@link StoreFileScanner#seekToPreviousRow(Cell)} method requires two seeks.
* It should be used if the cost for seeking is lower i.e. when using a fast seeking data block
* encoding like RIV1.
*/
private boolean seekToPreviousRowStateless(Cell originalKey) throws IOException {
Cell key =... | 3.68 |
flink_PekkoRpcActor_handleRpcInvocation | /**
* Handle rpc invocations by looking up the rpc method on the rpc endpoint and calling this
* method with the provided method arguments. If the method has a return value, it is returned
* to the sender of the call.
*
* @param rpcInvocation Rpc invocation message
*/
private void handleRpcInvocation(RpcInvocatio... | 3.68 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.