name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
framework_Table_setColumnHeaders
/** * Sets the headers of the columns. * * <p> * The headers match the property id:s given by the set visible column * headers. The table must be set in either * {@link #COLUMN_HEADER_MODE_EXPLICIT} or * {@link #COLUMN_HEADER_MODE_EXPLICIT_DEFAULTS_ID} mode to show the * headers. In the defaults mode any nulls ...
3.68
druid_DruidPooledConnection_getPhysicalConnectNanoSpan
/** * @since 1.0.17 */ public long getPhysicalConnectNanoSpan() { return this.holder.getCreateNanoSpan(); }
3.68
framework_LayoutManager_addElementResizeListener
/** * Adds a listener that will be notified whenever the size of a specific * element changes. Adding a listener to an element also ensures that all * sizes for that element will be available starting from the next layout * phase. * * @param element * the element that should be checked for size change...
3.68
hbase_SnapshotInfo_isMissing
/** Returns true if the file is missing */ public boolean isMissing() { return this.size < 0; }
3.68
hmily_MapBinder_bindEntries
/** * Bind entries. * * @param source the source * @param map the map */ void bindEntries(final ConfigPropertySource source, final Map<Object, Object> map) { source.stream().forEach(name -> { boolean ancestorOf = root.isAncestorOf(name); if (ancestorOf) { BindData<?> valueBindDat...
3.68
flink_RemoteInputChannel_requestBuffer
/** * Requests buffer from input channel directly for receiving network data. It should always * return an available buffer in credit-based mode unless the channel has been released. * * @return The available buffer. */ @Nullable public Buffer requestBuffer() { return bufferManager.requestBuffer(); }
3.68
hudi_HoodieExampleDataGenerator_generateInsertsOnPartition
/** * Generates new inserts, across a single partition path. It also updates the list of existing keys. */ public List<HoodieRecord<T>> generateInsertsOnPartition(String commitTime, Integer n, String partitionPath) { return generateInsertsStreamOnPartition(commitTime, n, partitionPath).collect(Collectors.toList());...
3.68
flink_EntropyInjector_addEntropy
/** * Handles entropy injection across regular and entropy-aware file systems. * * <p>If the given file system is entropy-aware (a implements {@link * EntropyInjectingFileSystem}), then this method replaces the entropy marker in the path with * random characters. The entropy marker is defined by {@link * EntropyI...
3.68
hbase_ServerRegionReplicaUtil_isRegionReplicaStoreFileRefreshEnabled
/** Returns True if we are to refresh user-space hfiles in Region Read Replicas. */ public static boolean isRegionReplicaStoreFileRefreshEnabled(Configuration conf) { return conf.getBoolean(REGION_REPLICA_STORE_FILE_REFRESH, DEFAULT_REGION_REPLICA_STORE_FILE_REFRESH); }
3.68
morf_SqlDialect_escapeSql
/** * Turn a string value into an SQL string literal which has that value. * This escapes single quotes as double-single quotes. * @param literalValue the value to escape * @return escaped value */ protected String escapeSql(String literalValue) { if (literalValue == null) { return null; } return StringUtil...
3.68
hadoop_FutureDataInputStreamBuilderImpl_bufferSize
/** * Set the size of the buffer to be used. * * @param bufSize buffer size. * @return FutureDataInputStreamBuilder. */ public FutureDataInputStreamBuilder bufferSize(int bufSize) { bufferSize = bufSize; return getThisBuilder(); }
3.68
hbase_MetricsConnection_getScanTracker
/** scanTracker metric */ public CallTracker getScanTracker() { return scanTracker; }
3.68
hbase_SaslClientAuthenticationProviders_addProviderIfNotExists
/** * Adds the given {@code provider} to the set, only if an equivalent provider does not already * exist in the set. */ static void addProviderIfNotExists(SaslClientAuthenticationProvider provider, HashMap<Byte, SaslClientAuthenticationProvider> providers) { Byte code = provider.getSaslAuthMethod().getCode(); ...
3.68
flink_CrossOperator_projectSecond
/** * Continues a ProjectCross transformation and adds fields of the second cross input. * * <p>If the second cross input is a {@link Tuple} {@link DataSet}, fields can be selected * by their index. If the second cross input is not a Tuple DataSet, no parameters should be * passed. * * <p>Fields of the first and...
3.68
querydsl_MultiSurfaceExpression_centroid
/** * The mathematical centroid for this MultiSurface. The result is not guaranteed to be on * this MultiSurface. * * @return centroid */ public PointExpression<Point> centroid() { if (centroid == null) { centroid = GeometryExpressions.pointOperation(SpatialOps.CENTROID, mixin); } return centro...
3.68
morf_AbstractSqlDialectTest_shouldGenerateCorrectSqlForMathOperations7
/** * Test for proper SQL mathematics operation generation from DSL expressions * that use brackets. * <p> * Subexpression "a+b" is put to bracket explicitly, and * the subexpression "c-d" should be put to the bracket implicitly, even without explicit * {@link org.alfasoftware.morf.sql.SqlUtils#bracket(MathsField...
3.68
hbase_RegionInfo_isEncodedRegionName
/** * Figure if the passed bytes represent an encoded region name or not. * @param regionName A Region name either encoded or not. * @return True if <code>regionName</code> represents an encoded name. */ @InterfaceAudience.Private // For use by internals only. public static boolean isEncodedRegionName(byte[] region...
3.68
hadoop_OBSBlockOutputStream_complete
/** * This completes a multipart upload. Sometimes it fails; here retries are * handled to avoid losing all data on a transient failure. * * @param partETags list of partial uploads * @return result for completing multipart upload * @throws IOException on any problem */ private CompleteMultipartUploadResult comp...
3.68
framework_Payload_setKey
/** * Sets the key of this payload. * * @param key * key that identifies the payload */ public void setKey(String key) { this.key = key; }
3.68
framework_VUpload_disableTitle
/** * For internal use only. May be removed or replaced in the future. * * @param disable * {@code true} if the built-in browser-dependent tooltip should * be hidden in favor of a Vaadin tooltip, {@code false} * otherwise */ public void disableTitle(boolean disable) { if (dis...
3.68
hbase_ParseFilter_removeQuotesFromByteArray
/** * Takes a quoted byte array and converts it into an unquoted byte array For example: given a byte * array representing 'abc', it returns a byte array representing abc * <p> * @param quotedByteArray the quoted byte array * @return Unquoted byte array */ public static byte[] removeQuotesFromByteArray(byte[] quo...
3.68
hbase_SnapshotInfo_getStoreFilesSize
/** Returns the total size of the store files referenced by the snapshot */ public long getStoreFilesSize() { return hfilesSize.get() + hfilesArchiveSize.get() + hfilesMobSize.get(); }
3.68
hbase_MasterObserver_preDeleteSnapshot
/** * Called before a snapshot is deleted. Called as part of deleteSnapshot RPC call. * @param ctx the environment to interact with the framework and master * @param snapshot the SnapshotDescriptor of the snapshot to delete */ default void preDeleteSnapshot(final ObserverContext<MasterCoprocessorEnvironment> c...
3.68
flink_FlinkJoinToMultiJoinRule_combineOuterJoins
/** * Combines the outer join conditions and join types from the left and right join inputs. If the * join itself is either a left or right outer join, then the join condition corresponding to * the join is also set in the position corresponding to the null-generating input into the * join. The join type is also se...
3.68
hbase_Scan_includeStartRow
/** Returns if we should include start row when scan */ public boolean includeStartRow() { return includeStartRow; }
3.68
graphhopper_VectorTile_getFeaturesList
/** * <pre> * The actual features in this tile. * </pre> * * <code>repeated .vector_tile.Tile.Feature features = 2;</code> */ public java.util.List<vector_tile.VectorTile.Tile.Feature> getFeaturesList() { if (featuresBuilder_ == null) { return java.util.Collections.unmodifiableList(features_); } else { ...
3.68
hudi_HoodieTableMetaClient_getBasePathV2
/** * Returns base path of the table */ public Path getBasePathV2() { return basePath.get(); }
3.68
shardingsphere-elasticjob_JobNodeStorage_getJobNodeData
/** * Get job node data. * * @param node node * @return data of job node */ public String getJobNodeData(final String node) { return regCenter.get(jobNodePath.getFullPath(node)); }
3.68
hadoop_BytesWritable_setCapacity
/** * Change the capacity of the backing storage. The data is preserved. * * @param capacity The new capacity in bytes. */ public void setCapacity(final int capacity) { if (capacity != getCapacity()) { this.size = Math.min(size, capacity); this.bytes = Arrays.copyOf(this.bytes, capacity); } }
3.68
morf_AbstractSqlDialectTest_testInsertFromSelectWithSourceInDifferentSchema
/** * Tests that an insert from a select works when the source table is in a different schema. */ @Test public void testInsertFromSelectWithSourceInDifferentSchema() { SelectStatement sourceStmt = new SelectStatement(new FieldReference("id"), new FieldReference("version"), new FieldReference(STRING_FIELD), ...
3.68
hudi_RequestHandler_registerDataFilesAPI
/** * Register Data-Files API calls. */ private void registerDataFilesAPI() { app.get(RemoteHoodieTableFileSystemView.LATEST_PARTITION_DATA_FILES_URL, new ViewHandler(ctx -> { metricsRegistry.add("LATEST_PARTITION_DATA_FILES", 1); List<BaseFileDTO> dtos = dataFileHandler.getLatestDataFiles( ctx.quer...
3.68
hbase_ReplicationSourceManager_getOldLogDir
/** * Get the directory where wals are archived * @return the directory where wals are archived */ public Path getOldLogDir() { return this.oldLogDir; }
3.68
hudi_BufferedRandomAccessFile_init
/** * * @param size - capacity of the buffer */ private void init(int size) { this.capacity = Math.max(DEFAULT_BUFFER_SIZE, size); this.dataBuffer = ByteBuffer.wrap(new byte[this.capacity]); }
3.68
framework_ContainerHierarchicalWrapper_removeAllItems
/** * Removes all items from the underlying container and from the hierarcy. * * @return <code>true</code> if the operation succeeded, <code>false</code> * if not * @throws UnsupportedOperationException * if the removeAllItems is not supported. */ @Override public boolean removeAllItems() thr...
3.68
morf_SqlScriptExecutor_executeStatementBatch
/** * Runs the specified SQL statement (which should contain parameters), repeatedly for * each record, mapping the contents of the records into the statement parameters in * their defined order. Use to insert, merge or update a large batch of records * efficiently. * * @param sqlStatement the SQL statement. * ...
3.68
hbase_FirstKeyValueMatchingQualifiersFilter_parseFrom
/** * Parses a serialized representation of {@link FirstKeyValueMatchingQualifiersFilter} * @param pbBytes A pb serialized {@link FirstKeyValueMatchingQualifiersFilter} instance * @return An instance of {@link FirstKeyValueMatchingQualifiersFilter} made from * <code>bytes</code> * @throws DeserializationEx...
3.68
querydsl_PathBuilder_getComparable
/** * Create a new Comparable typed path * * @param <A> * @param property property name * @param type property type * @return property path */ @SuppressWarnings("unchecked") public <A extends Comparable<?>> ComparablePath<A> getComparable(String property, Class<A> type) { Class<? extends A> vtype = validate(...
3.68
flink_HiveParserTypeCheckCtx_getUnparseTranslator
/** @return the unparseTranslator */ public HiveParserUnparseTranslator getUnparseTranslator() { return unparseTranslator; }
3.68
hadoop_AbfsDtFetcher_getServiceName
/** * Returns the service name for the scheme.. */ public Text getServiceName() { return new Text(getScheme()); }
3.68
flink_PackagedProgram_invokeInteractiveModeForExecution
/** * This method assumes that the context environment is prepared, or the execution will be a * local execution by default. */ public void invokeInteractiveModeForExecution() throws ProgramInvocationException { FlinkSecurityManager.monitorUserSystemExitForCurrentThread(); try { callMainMethod(mainCl...
3.68
dubbo_ScopeClusterInvoker_init
/** * Initializes the ScopeClusterInvoker instance. */ private void init() { Boolean peer = (Boolean) getUrl().getAttribute(PEER_KEY); String isInjvm = getUrl().getParameter(LOCAL_PROTOCOL); // When the point-to-point direct connection is directly connected, // the initialization is directly ended ...
3.68
hbase_SnapshotOfRegionAssignmentFromMeta_getExistingAssignmentPlan
/** * Get the favored nodes plan * @return the existing favored nodes plan */ public FavoredNodesPlan getExistingAssignmentPlan() { return this.existingAssignmentPlan; }
3.68
hbase_UnsafeAccess_copy
/** * Copies specified number of bytes from given offset of {@code src} buffer into the {@code dest} * buffer. * @param src source buffer * @param srcOffset offset into source buffer * @param dest destination buffer * @param destOffset offset into destination buffer * @param length length of da...
3.68
framework_AbstractProperty_removeListener
/** * @deprecated As of 7.0, replaced by * {@link #removeValueChangeListener(Property.ValueChangeListener)} */ @Override @Deprecated public void removeListener(ValueChangeListener listener) { removeValueChangeListener(listener); }
3.68
framework_AbsoluteLayout_setRight
/** * Sets the 'right' attribute; distance from the right of the component * to the right edge of the layout. * * @param rightValue * The value of the 'right' attribute * @param rightUnits * The unit of the 'right' attribute. See UNIT_SYMBOLS for a * description of the available...
3.68
flink_ProcessFunction_onTimer
/** * Called when a timer set using {@link TimerService} fires. * * @param timestamp The timestamp of the firing timer. * @param ctx An {@link OnTimerContext} that allows querying the timestamp of the firing timer, * querying the {@link TimeDomain} of the firing timer and getting a {@link TimerService} * ...
3.68
framework_ColorPickerGrid_getPosition
/** * Gets the position. * * @return the position */ public int[] getPosition() { return new int[] { x, y }; }
3.68
incubator-hugegraph-toolchain_FileLineFetcher_checkMatchHeader
/** * Just match header for second or subsequent file first line */ private boolean checkMatchHeader(String line) { if (!this.source().format().needHeader() || this.offset() != FIRST_LINE_OFFSET) { return false; } assert this.source().header() != null; String[] columns = this.parser.s...
3.68
framework_JavaScriptConnectorHelper_removeResizeListener
// Called from JSNI to remove a listener private void removeResizeListener(Element element, JavaScriptObject callbackFunction) { Map<JavaScriptObject, ElementResizeListener> listenerMap = resizeListeners .get(element); if (listenerMap == null) { return; } ElementResizeListen...
3.68
flink_StreamExecutionEnvironment_addSource
/** * Ads a data source with a custom type information thus opening a {@link DataStream}. Only in * very special cases does the user need to support type information. Otherwise use {@link * #addSource(org.apache.flink.streaming.api.functions.source.SourceFunction)} * * @param function the user defined function * ...
3.68
zxing_BitArray_get
/** * @param i bit to get * @return true iff bit i is set */ public boolean get(int i) { return (bits[i / 32] & (1 << (i & 0x1F))) != 0; }
3.68
dubbo_TTable_getBorder
/** * get border * * @return table border */ public Border getBorder() { return border; }
3.68
hbase_WALFactory_getInstance
// Public only for FSHLog public static WALFactory getInstance(Configuration configuration) { WALFactory factory = singleton.get(); if (null == factory) { WALFactory temp = new WALFactory(configuration); if (singleton.compareAndSet(null, temp)) { factory = temp; } else { // someone else beat...
3.68
flink_ListView_setList
/** Replaces the entire view's content with the content of the given {@link List}. */ public void setList(List<T> list) { this.list = list; }
3.68
framework_AbstractClientConnector_detach
/** * {@inheritDoc} * * <p> * The {@link #getSession()} and {@link #getUI()} methods might return * <code>null</code> after this method is called. * </p> */ @Override public void detach() { for (ClientConnector connector : getAllChildrenIterable(this)) { connector.detach(); } fireEvent(new D...
3.68
hbase_TableMapReduceUtil_convertScanToString
/** * Writes the given scan into a Base64 encoded string. * @param scan The scan to write out. * @return The scan saved in a Base64 encoded string. * @throws IOException When writing the scan fails. */ public static String convertScanToString(Scan scan) throws IOException { ClientProtos.Scan proto = ProtobufUtil...
3.68
hadoop_ReencryptionHandler_throttle
/** * Throttles the ReencryptionHandler in 3 aspects: * 1. Prevents generating more Callables than the CPU could possibly * handle. * 2. Prevents generating more Callables than the ReencryptionUpdater * can handle, under its own throttling. * 3. Prevents contending FSN/FSD read locks. This is done based * on the...
3.68
hbase_AuthManager_authorizeUserNamespace
/** * Check if user has given action privilige in namespace scope. * @param user user name * @param namespace namespace * @param action one of action in [Read, Write, Create, Exec, Admin] * @return true if user has, false otherwise */ public boolean authorizeUserNamespace(User user, String namespace, Perm...
3.68
morf_ViewURLAsFile_createTempFile
/** * Wrapper for {@link java.io.File#createTempFile(String, String, File)} that * wraps any exceptions in a {@link RuntimeException} and propagates it. */ private File createTempFile(String prefix, String suffix, File file) { try { return File.createTempFile(prefix, suffix, file); } catch (IOException e) { ...
3.68
hadoop_DistributedCache_setFileTimestamps
/** * This is to check the timestamp of the files to be localized. * Used by internal MapReduce code. * @param conf Configuration which stores the timestamp's * @param timestamps comma separated list of timestamps of files. * The order should be the same as the order in which the files are added. */ @Deprecated p...
3.68
framework_ConnectorBundleLoader_ensureDeferredBundleLoaded
/** * Starts loading the deferred bundle if it hasn't already been started. * * @since 8.0.3 */ public void ensureDeferredBundleLoaded() { if (!isBundleLoaded(DEFERRED_BUNDLE_NAME)) { loadBundle(DEFERRED_BUNDLE_NAME, new BundleLoadCallback() { @Override public void loaded() { ...
3.68
hadoop_JobBase_getLongValue
/** * * @param name * the counter name * @return return the value of the given counter. */ protected Long getLongValue(Object name) { return this.longCounters.get(name); }
3.68
hadoop_OBSBlockOutputStream_clearHFlushOrSync
/** * Clear for hflush or hsync. */ private synchronized void clearHFlushOrSync() { appendAble.set(true); multiPartUpload = null; }
3.68
framework_NullValidator_isNullAllowed
/** * Returns <code>true</code> if nulls are allowed otherwise * <code>false</code>. */ public final boolean isNullAllowed() { return onlyNullAllowed; }
3.68
framework_VAbstractSplitPanel_setSecondWidget
/** * For internal use only. May be removed or replaced in the future. * * @param w * the widget to set to the second region or {@code null} to * remove previously set widget */ public void setSecondWidget(Widget w) { if (secondChild == w) { return; } if (secondChild != n...
3.68
hadoop_ArrayFile_get
/** * Return the <code>n</code>th value in the file. * @param n n key. * @param value value. * @throws IOException raised on errors performing I/O. * @return writable. */ public synchronized Writable get(long n, Writable value) throws IOException { key.set(n); return get(key, value); }
3.68
MagicPlugin_CompatibilityUtilsBase_loadChunk
/** * This will load chunks asynchronously if possible. * * <p>But note that it will never be truly asynchronous, it is important not to call this in a tight retry loop, * the main server thread needs to free up to actually process the async chunk loads. */ @Override public void loadChunk(World world, int x, int z...
3.68
hudi_HoodieAvroUtils_bytesToAvro
/** * Convert serialized bytes back into avro record. */ public static GenericRecord bytesToAvro(byte[] bytes, Schema writerSchema, Schema readerSchema) throws IOException { BinaryDecoder decoder = DecoderFactory.get().binaryDecoder(bytes, BINARY_DECODER.get()); BINARY_DECODER.set(decoder); GenericDatumReader<G...
3.68
hbase_ServerNonceManager_startOperation
/** * Starts the operation if operation with such nonce has not already succeeded. If the operation * is in progress, waits for it to end and checks whether it has succeeded. * @param group Nonce group. * @param nonce Nonce. * @param stoppable Stoppable that terminates waiting (if any) when the server is s...
3.68
hadoop_NodeIDsInfo_add
/** * This method will generate a new NodeIDsInfo object based on the two NodeIDsInfo objects. * The information to be combined includes the node list (removed duplicate node) * and partitionInfo object. * * @param left left NodeIDsInfo Object. * @param right right NodeIDsInfo Object. * @return new NodeIDsInfo O...
3.68
framework_HasStyleNames_addStyleNames
/** * Adds one or more style names to this component by using one or multiple * parameters. * * @since 8.7 * @param styles * the style name or style names to be added to the component * @see #addStyleName(String) * @see #setStyleName(String) * @see #removeStyleName(String) */ default void addStyleN...
3.68
morf_SqlServerDialect_internalTableDeploymentStatements
/** * @see org.alfasoftware.morf.jdbc.SqlDialect#tableDeploymentStatements(org.alfasoftware.morf.metadata.Table) */ @Override public Collection<String> internalTableDeploymentStatements(Table table) { List<String> statements = new ArrayList<>(); // Create the table deployment statement StringBuilder createTabl...
3.68
hudi_HoodieKeyLookupHandle_addKey
/** * Adds the key for look up. */ public void addKey(String recordKey) { // check record key against bloom filter of current file & add to possible keys if needed if (bloomFilter.mightContain(recordKey)) { if (LOG.isDebugEnabled()) { LOG.debug("Record key " + recordKey + " matches bloom filter in " + ...
3.68
hudi_HoodieIngestionService_requestShutdownIfNeeded
/** * To determine if shutdown should be requested to allow gracefully terminate the ingestion in continuous mode. * <p> * Subclasses should implement the logic to make the decision. If the shutdown condition is met, the implementation * should call {@link #shutdown(boolean)} to indicate the request. * * @see Pos...
3.68
hadoop_MawoConfiguration_readConfigFile
/** * Find, read, and parse the configuration file. * * @return the properties that were found or empty if no file was found */ private static Properties readConfigFile() { Properties properties = new Properties(); // Get property file stream from classpath LOG.info("Configuration file being loaded: " + CONF...
3.68
pulsar_RangeCache_getNumberOfEntries
/** * Just for testing. Getting the number of entries is very expensive on the conncurrent map */ protected long getNumberOfEntries() { return entries.size(); }
3.68
flink_MimeTypes_getMimeTypeForFileName
/** * Gets the MIME type for the file with the given name, by extension. This method tries to * extract the file extension and then use the {@link #getMimeTypeForExtension(String)} to * determine the MIME type. If the extension cannot be determined, or the extension is * unrecognized, this method return {@code null...
3.68
framework_VLoadingIndicator_isVisible
/** * Returns whether or not the loading indicator is showing. * * @return true if the loading indicator is visible, false otherwise */ public boolean isVisible() { if (getElement().getStyle().getDisplay() .equals(Display.NONE.getCssName())) { return false; } return true; }
3.68
framework_VaadinSession_getCurrent
/** * Gets the currently used session. The current session is automatically * defined when processing requests related to the session (see * {@link ThreadLocal}) and in {@link VaadinSession#access(Runnable)} and * {@link UI#access(Runnable)}. In other cases, (e.g. from background * threads, the current session is ...
3.68
flink_MapView_putAll
/** * Inserts all mappings from the specified map to this map view. * * @param map The map whose entries are inserted into this map view. * @throws Exception Thrown if the system cannot access the map. */ public void putAll(Map<K, V> map) throws Exception { this.map.putAll(map); }
3.68
flink_BlobServer_getPort
/** * Returns the port on which the server is listening. * * @return port on which the server is listening */ @Override public int getPort() { return this.serverSocket.getLocalPort(); }
3.68
zxing_MinimalECIInput_isFNC1
/** * Determines if a value is the FNC1 character * * @param index the index of the value * * @return true if the value at position {@code index} is the FNC1 character * * @throws IndexOutOfBoundsException * if the {@code index} argument is negative or not less than * {@code length()} */...
3.68
flink_InputGate_getChannelInfos
/** Returns the channel infos of this gate. */ public List<InputChannelInfo> getChannelInfos() { return IntStream.range(0, getNumberOfInputChannels()) .mapToObj(index -> getChannel(index).getChannelInfo()) .collect(Collectors.toList()); }
3.68
framework_AbsoluteLayout_toString
/* * (non-Javadoc) * * @see java.lang.Object#toString() */ @Override public String toString() { return getCSSString(); }
3.68
hudi_AWSDmsAvroPayload_handleDeleteOperation
/** * * Handle a possible delete - check for "D" in Op column and return empty row if found. * @param insertValue The new row that is being "inserted". */ private Option<IndexedRecord> handleDeleteOperation(IndexedRecord insertValue) throws IOException { boolean delete = false; if (insertValue instanceof Generi...
3.68
flink_StreamProjection_projectTuple24
/** * Projects a {@link Tuple} {@link DataStream} to the previously selected fields. * * @return The projected DataStream. * @see Tuple * @see DataStream */ public < T0, T1, T2, T3, T4, T5, T6, ...
3.68
hadoop_FileIoProvider_read
/** * {@inheritDoc}. */ @Override public int read(@Nonnull byte[] b, int off, int len) throws IOException { final long begin = profilingEventHook.beforeFileIo(volume, READ, len); try { faultInjectorEventHook.beforeFileIo(volume, READ, len); int numBytesRead = super.read(b, off, len); profilingEventHoo...
3.68
flink_NullableSerializer_wrap
/** * This method wraps the {@code originalSerializer} with the {@code NullableSerializer} if not * already wrapped. * * @param originalSerializer serializer to wrap and add {@code null} support * @param padNullValueIfFixedLen pad null value to preserve the fixed length of original * serializer * @return wra...
3.68
hadoop_AbfsOutputStream_writeAppendBlobCurrentBufferToService
/** * Appending the current active data block to service. Clearing the active * data block and releasing all buffered data. * @throws IOException if there is any failure while starting an upload for * the dataBlock or while closing the BlockUploadData. */ private void writeAppendBlobCurrentBuff...
3.68
hadoop_Base64_decodeAsByteObjectArray
/** * Decodes a given Base64 string into its corresponding byte array. * * @param data * the Base64 string, as a <code>String</code> object, to decode * * @return the corresponding decoded byte array * @throws IllegalArgumentException * If the string is not a valid base64 encoded string ...
3.68
framework_ScrollbarBundle_isScrollerVisible
/** * Checks whether the scroll handle is currently visible or not. * * @return <code>true</code> if the scroll handle is currently visible. * <code>false</code> if not. */ public boolean isScrollerVisible() { return isScrollerVisible; }
3.68
framework_Button_getRelativeY
/** * Returns the relative mouse position (y coordinate) when the click * took place. The position is relative to the clicked component. * * @return The mouse cursor y position relative to the clicked layout * component or -1 if no y coordinate available */ public int getRelativeY() { if (null != deta...
3.68
hibernate-validator_NotEmptyValidatorForArraysOfFloat_isValid
/** * Checks the array is not {@code null} and not empty. * * @param array the array to validate * @param constraintValidatorContext context in which the constraint is evaluated * @return returns {@code true} if the array is not {@code null} and the array is not empty */ @Override public boolean isValid(float[] a...
3.68
querydsl_GenericExporter_setTargetFolder
/** * Set the target folder for generated sources * * @param targetFolder */ public void setTargetFolder(File targetFolder) { this.targetFolder = targetFolder; }
3.68
hudi_FileSystemBasedLockProvider_getLockConfig
/** * Returns a filesystem based lock config with given table path. */ public static TypedProperties getLockConfig(String tablePath) { TypedProperties props = new TypedProperties(); props.put(HoodieLockConfig.LOCK_PROVIDER_CLASS_NAME.key(), FileSystemBasedLockProvider.class.getName()); props.put(HoodieLockConfi...
3.68
hadoop_RequestFactoryImpl_withMultipartPartCountLimit
/** * Multipart limit. * @param value new value * @return the builder */ public RequestFactoryBuilder withMultipartPartCountLimit( final long value) { multipartPartCountLimit = value; return this; }
3.68
hbase_PermissionStorage_removeNamespacePermissions
/** * Remove specified namespace from the acl table. */ static void removeNamespacePermissions(Configuration conf, String namespace, Table t) throws IOException { Delete d = new Delete(Bytes.toBytes(toNamespaceEntry(namespace))); d.addFamily(ACL_LIST_FAMILY); if (LOG.isDebugEnabled()) { LOG.debug("Removin...
3.68
framework_Table_getItemDescriptionGenerator
/** * Get the item description generator which generates tooltips for cells and * rows in the Table. */ public ItemDescriptionGenerator getItemDescriptionGenerator() { return itemDescriptionGenerator; }
3.68
hbase_StoreFileScanner_seekToPreviousRowStateless
/** * This variant of the {@link StoreFileScanner#seekToPreviousRow(Cell)} method requires two seeks. * It should be used if the cost for seeking is lower i.e. when using a fast seeking data block * encoding like RIV1. */ private boolean seekToPreviousRowStateless(Cell originalKey) throws IOException { Cell key =...
3.68
flink_PekkoRpcActor_handleRpcInvocation
/** * Handle rpc invocations by looking up the rpc method on the rpc endpoint and calling this * method with the provided method arguments. If the method has a return value, it is returned * to the sender of the call. * * @param rpcInvocation Rpc invocation message */ private void handleRpcInvocation(RpcInvocatio...
3.68