name stringlengths 12 178 | code_snippet stringlengths 8 36.5k | score float64 3.26 3.68 |
|---|---|---|
flink_KubernetesUtils_getDeploymentName_rdh | /**
* Generate name of the Deployment.
*/public static String getDeploymentName(String clusterId) {
return clusterId;
} | 3.26 |
flink_KubernetesUtils_parsePort_rdh | /**
* Parse a valid port for the config option. A fixed port is expected, and do not support a
* range of ports.
*
* @param flinkConfig
* flink config
* @param port
* port config option
* @return valid port
*/
public static Integer parsePort(Configuration flinkConfig, ConfigOption<String> port) {
check... | 3.26 |
flink_KubernetesUtils_createCompletedCheckpointStore_rdh | /**
* Create a {@link DefaultCompletedCheckpointStore} with {@link KubernetesStateHandleStore}.
*
* @param configuration
* configuration to build a RetrievableStateStorageHelper
* @param kubeClient
* flink kubernetes client
* @param configMapName
* ConfigMap name
* @param executor
* executor to run bl... | 3.26 |
flink_KubernetesUtils_getServiceAccount_rdh | /**
* Get the service account from the input pod first, if not specified, the service account name
* will be used.
*
* @param flinkPod
* the Flink pod to parse the service account
* @return the parsed service account
*/
@Nullable
public static String getServiceAccount(FlinkPod flinkPod) {
final String serv... | 3.26 |
flink_KubernetesUtils_getCommonLabels_rdh | /**
* Get the common labels for Flink native clusters. All the Kubernetes resources will be set
* with these labels.
*
* @param clusterId
* cluster id
* @return Return common labels map
*/
public static Map<String, String> getCommonLabels(String clusterId) {
final Map<String, String> commonLabels = new Has... | 3.26 |
flink_KubernetesUtils_tryToGetPrettyPrintYaml_rdh | /**
* Try to get the pretty print yaml for Kubernetes resource.
*
* @param kubernetesResource
* kubernetes resource
* @return the pretty print yaml, or the {@link KubernetesResource#toString()} if parse failed.
*/
public static String tryToGetPrettyPrintYaml(KubernetesResource kubernetesResource) {
try {
... | 3.26 |
flink_KubernetesUtils_checkAndUpdatePortConfigOption_rdh | /**
* Check whether the port config option is a fixed port. If not, the fallback port will be set
* to configuration.
*
* @param flinkConfig
* flink configuration
* @param port
* config option need to be checked
* @param fallbackPort
* the fallback port that will be set to the configuration
*/
public st... | 3.26 |
flink_KubernetesUtils_resolveDNSPolicy_rdh | /**
* Resolve the DNS policy defined value. Return DNS_POLICY_HOSTNETWORK if host network enabled.
* If not, check whether there is a DNS policy overridden in pod template.
*
* @param dnsPolicy
* DNS policy defined in pod template spec
* @param hostNetworkEnabled
* Host network enabled or not
* @return the ... | 3.26 |
flink_KubernetesUtils_resolveUserDefinedValue_rdh | /**
* Resolve the user defined value with the precedence. First an explicit config option value is
* taken, then the value in pod template and at last the default value of a config option if
* nothing is specified.
*
* @param flinkConfig
* flink configuration
* @param configOption
* the config option to def... | 3.26 |
flink_KubernetesUtils_createJobGraphStateHandleStore_rdh | /**
* Create a {@link KubernetesStateHandleStore} which storing {@link JobGraph}.
*
* @param configuration
* configuration to build a RetrievableStateStorageHelper
* @param flinkKubeClient
* flink kubernetes client
* @param configMapName
* ConfigMap name
* @param lockIdentity
* lock identity to check ... | 3.26 |
flink_KubernetesUtils_getResourceRequirements_rdh | /**
* Get resource requirements from memory and cpu.
*
* @param resourceRequirements
* resource requirements in pod template
* @param mem
* Memory in mb.
* @param memoryLimitFactor
* limit factor for the memory, used to set the limit resources.
* @param cpu
* cpu.
* @param cpuLimitFactor
* limit f... | 3.26 |
flink_KubernetesUtils_getOnlyConfigMap_rdh | /**
* Check the ConfigMap list should only contain the expected one.
*
* @param configMaps
* ConfigMap list to check
* @param expectedConfigMapName
* expected ConfigMap Name
* @return Return the expected ConfigMap
*/
public static KubernetesConfigMap getOnlyConfigMap(List<KubernetesConfigMap> configMaps, St... | 3.26 |
flink_KubernetesUtils_getNamespacedServiceName_rdh | /**
* Generate namespaced name of the service.
*/
public static String getNamespacedServiceName(Service service) {
return (service.getMetadata().getName() + ".") + service.getMetadata().getNamespace();
} | 3.26 |
flink_KubernetesUtils_getConfigMapLabels_rdh | /**
* Get ConfigMap labels for the current Flink cluster. They could be used to filter and clean-up
* the resources.
*
* @param clusterId
* cluster id
* @param type
* the config map use case. It could only be {@link Constants#LABEL_CONFIGMAP_TYPE_HIGH_AVAILABILITY} now.
* @return Return ConfigMap labels.
*... | 3.26 |
flink_KubernetesUtils_isHostNetwork_rdh | /**
* Checks if hostNetwork is enabled.
*/
public static boolean isHostNetwork(Configuration configuration) {
return configuration.getBoolean(KubernetesConfigOptions.KUBERNETES_HOSTNETWORK_ENABLED);
} | 3.26 |
flink_RawFormatFactory_validateAndExtractSingleField_rdh | /**
* Validates and extract the single field type from the given physical row schema.
*/
private static LogicalType validateAndExtractSingleField(RowType physicalRowType) {
if (physicalRowType.getFieldCount() != 1) {
String schemaString = physicalRowType.getFields().stream().map(RowType.Ro... | 3.26 |
flink_RawFormatFactory_checkFieldType_rdh | /**
* Checks the given field type is supported.
*/
private static void checkFieldType(LogicalType fieldType) {
if (!supportedTypes.contains(fieldType.getTypeRoot())) {
throw new ValidationException(String.format("The 'raw' format doesn't supports '%s' as column type.", fieldType.asSummaryString()));
... | 3.26 |
flink_TableFunctionCollector_reset_rdh | /**
* Resets the flag to indicate whether [[collect(T)]] has been called.
*/
public void reset() {
this.collected = false;
} | 3.26 |
flink_TableFunctionCollector_setInput_rdh | /**
* Sets the input row from left table, which will be used to cross join with the result of table
* function.
*/
public void setInput(Object input) {
this.input = input;
} | 3.26 |
flink_TableFunctionCollector_getInput_rdh | /**
* Gets the input value from left table, which will be used to cross join with the result of
* table function.
*/public Object getInput() {
return input;
} | 3.26 |
flink_TableFunctionCollector_setCollector_rdh | /**
* Sets the current collector, which used to emit the final row.
*/
public void setCollector(Collector<?> collector) {
this.collector = collector;
} | 3.26 |
flink_TableFunctionCollector_isCollected_rdh | /**
* Whether {@link #collect(Object)} has been called.
*
* @return True if {@link #collect(Object)} has been called.
*/
public boolean isCollected() {return collected;
} | 3.26 |
flink_SavepointMetadata_getNewOperators_rdh | /**
*
* @return List of new operator states for the savepoint, represented by their target {@link OperatorID} and {@link BootstrapTransformation}.
*/
public List<BootstrapTransformationWithID<?>> getNewOperators() {
return operatorStateIndex.values().stream().filter(OperatorStateSpec::isNewStateTransformation).m... | 3.26 |
flink_SavepointMetadata_getExistingOperators_rdh | /**
*
* @return List of {@link OperatorState} that already exists within the savepoint.
*/
public List<OperatorState> getExistingOperators() {
return operatorStateIndex.values().stream().filter(OperatorStateSpec::isExistingState).map(OperatorStateSpec::asExistingState).collect(Collectors.toList());
} | 3.26 |
flink_SavepointMetadata_getOperatorState_rdh | /**
*
* @return Operator state for the given UID.
* @throws IOException
* If the savepoint does not contain operator state with the given uid.
*/
public OperatorState getOperatorState(String uid) throws IOException {
OperatorID
operatorID = OperatorIDGenerator.fromUid(uid);
OperatorStateSpec operator... | 3.26 |
flink_PythonFunctionFactory_getPythonFunction_rdh | /**
* Returns PythonFunction according to the fully qualified name of the Python UDF i.e
* ${moduleName}.${functionName} or ${moduleName}.${className}.
*
* @param fullyQualifiedName
* The fully qualified name of the Python UDF.
* @param config
* The configuration of python dependencies.
* @param classLoader... | 3.26 |
flink_BroadcastConnectedStream_getSecondInput_rdh | /**
* Returns the {@link BroadcastStream}.
*
* @return The stream which, by convention, is the broadcast one.
*/
public BroadcastStream<IN2> getSecondInput() {
return broadcastStream;
} | 3.26 |
flink_BroadcastConnectedStream_process_rdh | /**
* Assumes as inputs a {@link BroadcastStream} and a non-keyed {@link DataStream} and applies
* the given {@link BroadcastProcessFunction} on them, thereby creating a transformed output
* stream.
*
* @param function
* The {@link BroadcastProcessFunction} that is called for each element in the
* stream.
*... | 3.26 |
flink_SecurityFactoryServiceLoader_findModuleFactory_rdh | /**
* Find a suitable {@link SecurityModuleFactory} based on canonical name.
*/
public static SecurityModuleFactory findModuleFactory(String securityModuleFactoryClass) throws NoMatchSecurityFactoryException {
return
findFactoryInternal(securityModuleFactoryClass, SecurityModuleFactory.class, SecurityModuleF... | 3.26 |
flink_SecurityFactoryServiceLoader_findContextFactory_rdh | /**
* Find a suitable {@link SecurityContextFactory} based on canonical name.
*/
public static SecurityContextFactory findContextFactory(String securityContextFactoryClass) throws NoMatchSecurityFactoryException {
return findFactoryInternal(securityContextFactoryClass, SecurityContextFactory.class, SecurityContex... | 3.26 |
flink_HeapListState_get_rdh | // ------------------------------------------------------------------------
// state access
// ------------------------------------------------------------------------
@Override
public Iterable<V> get() {
return getInternal();} | 3.26 |
flink_HeapListState_mergeState_rdh | // ------------------------------------------------------------------------
// state merging
// ------------------------------------------------------------------------
@Override
protected List<V> mergeState(List<V> a, List<V> b)
{
a.addAll(b);
return a;
} | 3.26 |
flink_HistoryServer_createOrGetFile_rdh | // ------------------------------------------------------------------------
// File generation
// ------------------------------------------------------------------------
static FileWriter createOrGetFile(File folder, String name) throws IOException {
File file = new File(folder, name + ".json");
if (!file.ex... | 3.26 |
flink_HistoryServer_start_rdh | // ------------------------------------------------------------------------
// Life-cycle
// ------------------------------------------------------------------------
void start() throws IOException, InterruptedException {
synchronized(startupShutdownLock) {
LOG.info("Starting history server.");
Fil... | 3.26 |
flink_MaxWithRetractAggFunction_getArgumentDataTypes_rdh | // --------------------------------------------------------------------------------------------
// Planning
// --------------------------------------------------------------------------------------------
@Override
public List<DataType> getArgumentDataTypes() {
return Collections.singletonList(valueDataType);
} | 3.26 |
flink_AggregatorWithName_getName_rdh | /**
* Gets the name that the aggregator is registered under.
*
* @return The name that the aggregator is registered under.
*/
public String getName() {
return
name;
} | 3.26 |
flink_AggregatorWithName_getAggregator_rdh | /**
* Gets the aggregator.
*
* @return The aggregator.
*/
public Aggregator<T> getAggregator() {
return aggregator;
} | 3.26 |
flink_BigDecComparator_putNormalizedKey_rdh | /**
* Adds a normalized key containing a normalized order of magnitude of the given record. 2 bits
* determine the sign (negative, zero, positive), 33 bits determine the magnitude. This method
* adds at most 5 bytes that contain information.
*/
@Override
public void putNormalizedKey(BigDecimal record, MemorySegment... | 3.26 |
flink_EnrichedRowData_from_rdh | /**
* Creates a new {@link EnrichedRowData} with the provided {@code fixedRow} as the immutable
* static row, and uses the {@code producedRowFields}, {@code fixedRowFields} and {@code mutableRowFields} arguments to compute the indexes mapping.
*
* <p>The {@code producedRowFields} should include the name of fields o... | 3.26 |
flink_EnrichedRowData_computeIndexMapping_rdh | /**
* This method computes the index mapping for {@link EnrichedRowData}.
*
* @see EnrichedRowData#from(RowData, List, List, List)
*/
public static int[] computeIndexMapping(List<String> producedRowFields, List<String> mutableRowFields, List<String> fixedRowFields) {
... | 3.26 |
flink_EnrichedRowData_getArity_rdh | // ---------------------------------------------------------------------------------------------
@Override
public int getArity() {
return indexMapping.length;} | 3.26 |
flink_BroadcastVariableManager_getNumberOfVariablesWithReferences_rdh | // --------------------------------------------------------------------------------------------
public int getNumberOfVariablesWithReferences() {
return this.variables.size();
} | 3.26 |
flink_BroadcastVariableManager_materializeBroadcastVariable_rdh | // --------------------------------------------------------------------------------------------
/**
* Materializes the broadcast variable for the given name, scoped to the given task and its
* iteration superstep. An existing materialization created by another parallel subtask may be
* returned, ... | 3.26 |
flink_LocalProperties_addUniqueFields_rdh | /**
* Adds a combination of fields that are unique in these data properties.
*
* @param uniqueFields
* The fields that are unique in these data properties.
*/
public LocalProperties addUniqueFields(FieldSet uniqueFields) {
LocalProperties copy = clone();
if (copy.uniqu... | 3.26 |
flink_LocalProperties_getOrdering_rdh | // --------------------------------------------------------------------------------------------
/**
* Gets the key order.
*
* @return The key order, or <code>null</code> if nothing is ordered.
*/
public Ordering getOrdering() {
return ordering;
} | 3.26 |
flink_LocalProperties_areFieldsUnique_rdh | /**
* Checks whether the given set of fields is unique, as specified in these local properties.
*
* @param set
* The set to check.
* @return True, if the given column combination is unique, false if not.
*/
public boolean
areFieldsUnique(FieldSet set) {
return (this.uniqueFields != null) && this.uniqueField... | 3.26 |
flink_LocalProperties_combine_rdh | // --------------------------------------------------------------------------------------------
public static LocalProperties combine(LocalProperties lp1, LocalProperties lp2) {
if (lp1.ordering != null) {
return lp1;
} else if (lp2.ordering != null) {
return lp2;
} else if (lp1.groupedFields != null) {
ret... | 3.26 |
flink_LocalProperties_filterBySemanticProperties_rdh | // --------------------------------------------------------------------------------------------
/**
* Filters these LocalProperties by the fields that are forwarded to the output as described by
* the SemanticProperties.
*
* @param props
* The semantic properties holding information about forwarded fields.
* @p... | 3.26 |
flink_LocalProperties_forOrdering_rdh | // --------------------------------------------------------------------------------------------
public static LocalProperties forOrdering(Ordering o) {
LocalProperties props = new LocalProperties();
props.ordering = o;
props.groupedFields = o.getInvolvedIndexes();
return props;
} | 3.26 |
flink_LocalProperties_isTrivial_rdh | /**
* Checks, if the properties in this object are trivial, i.e. only standard values.
*/
public boolean isTrivial() {
return ((ordering == null) && (this.groupedFields == null)) && (this.uniqueFields == null);
} | 3.26 |
flink_FlinkPipelineTranslationUtil_getJobGraph_rdh | /**
* Transmogrifies the given {@link Pipeline} to a {@link JobGraph}.
*/
public static JobGraph getJobGraph(ClassLoader userClassloader, Pipeline pipeline, Configuration optimizerConfiguration, int defaultParallelism) {
FlinkPipelineTranslator pipelineTranslator = getPipelineTranslator(userClassloader, pipeline)... | 3.26 |
flink_FlinkPipelineTranslationUtil_getJobGraphUnderUserClassLoader_rdh | /**
* Transmogrifies the given {@link Pipeline} under the userClassloader to a {@link JobGraph}.
*/
public static JobGraph getJobGraphUnderUserClassLoader(final ClassLoader userClassloader, final Pipeline pipeline, final Configuration configuration, final int defaultParallelism) {
final Class... | 3.26 |
flink_FlinkPipelineTranslationUtil_translateToJSONExecutionPlan_rdh | /**
* Extracts the execution plan (as JSON) from the given {@link Pipeline}.
*/
public static String
translateToJSONExecutionPlan(ClassLoader userClassloader, Pipeline pipeline) {
FlinkPipelineTranslator pipelineTranslator = getPipelineTranslator(userClassloader, pipeline);
return pipelineTranslator.translateToJSONEx... | 3.26 |
flink_NettyMessageClientDecoderDelegate_channelInactive_rdh | /**
* Releases resources when the channel is closed. When exceptions are thrown during processing
* received netty buffers, {@link CreditBasedPartitionRequestClientHandler} is expected to catch
* the exception and close the channel and trigger this notification.
*
* @param ctx
* The context of the channel close... | 3.26 |
flink_HighAvailabilityServices_getWebMonitorLeaderElection_rdh | /**
* Gets the {@link LeaderElection} for the cluster's rest endpoint.
*
* @deprecated Use {@link #getClusterRestEndpointLeaderElection()} instead.
*/@Deprecateddefault LeaderElection getWebMonitorLeaderElection() {
throw new UnsupportedOperationException(((("getWebMonitorLeaderElectionService should no longer ... | 3.26 |
flink_HighAvailabilityServices_getWebMonitorLeaderRetriever_rdh | /**
* This retriever should no longer be used on the cluster side. The web monitor retriever is
* only required on the client-side and we have a dedicated high-availability services for the
* client, named {@link ClientHighAvailabilityServices}. See also FLINK-13750.
*
* @return the leader retriever for web monito... | 3.26 |
flink_HighAvailabilityServices_closeWithOptionalClean_rdh | /**
* Calls {@link #cleanupAllData()} (if {@code true} is passed as a parameter) before calling
* {@link #close()} on this instance. Any error that appeared during the cleanup will be
* propagated after calling {@code close()}.
*/
default void closeWithOptionalClean(boolean cleanupData) throws Exception {
Throw... | 3.26 |
flink_HighAvailabilityServices_getClusterRestEndpointLeaderElection_rdh | /**
* Gets the {@link LeaderElection} for the cluster's rest endpoint.
*/
default LeaderElection getClusterRestEndpointLeaderElection() {
// for backwards compatibility we delegate to getWebMonitorLeaderElectionService
// all implementations of this interface should override
// getClusterRestEndpointLeade... | 3.26 |
flink_ObjectIdentifier_toList_rdh | /**
* List of the component names of this object identifier.
*/
public List<String> toList() {
if (catalogName == null) {
return Collections.singletonList(getObjectName());
}
return Arrays.asList(getCatalogName(), m0(), getObjectName());
} | 3.26 |
flink_ObjectIdentifier_toObjectPath_rdh | /**
* Convert this {@link ObjectIdentifier} to {@link ObjectPath}.
*
* @throws TableException
* if the identifier cannot be converted
*/
public ObjectPath toObjectPath() throws TableException {if (catalogName == null) {
throw new TableException("This ObjectIdentifier instance refers to an anonymous obje... | 3.26 |
flink_ObjectIdentifier_asSerializableString_rdh | /**
* Returns a string that fully serializes this instance. The serialized string can be used for
* transmitting or persisting an object identifier.
*
* @throws TableException
* if the identifier cannot be serialized
*/
public String asSerializableString() throws TableException {
if (catalogName == null)
... | 3.26 |
flink_ObjectIdentifier_m1_rdh | /**
* Returns a string that summarizes this instance for printing to a console or log.
*/
public String m1() {
if (catalogName == null) {
return objectName;
}
return String.join(".", catalogName, databaseName, objectName);
} | 3.26 |
flink_ObjectIdentifier_ofAnonymous_rdh | /**
* This method allows to create an {@link ObjectIdentifier} without catalog and database name,
* in order to propagate anonymous objects with unique identifiers throughout the stack.
*
* <p>This method for no reason should be exposed to users, as this should be used only when
* creating anonymous tables with un... | 3.26 |
flink_NettyShuffleEnvironmentConfiguration_numNetworkBuffers_rdh | // ------------------------------------------------------------------------
public int numNetworkBuffers() {
return numNetworkBuffers;
} | 3.26 |
flink_NettyShuffleEnvironmentConfiguration_m1_rdh | // ------------------------------------------------------------------------
@Override
public int m1() {
int result = 1;
result = (31
* result) + numNetworkBuffers;
result = (31 * result) + networkBufferSize; result = (31 * result) + partitionRequestInitialBackoff;
result = (31 * result) + partition... | 3.26 |
flink_NettyShuffleEnvironmentConfiguration_calculateNumberOfNetworkBuffers_rdh | /**
* Calculates the number of network buffers based on configuration and jvm heap size.
*
* @param configuration
* configuration object
* @param networkMemorySize
* the size of memory reserved for shuffle environment
... | 3.26 |
flink_NettyShuffleEnvironmentConfiguration_createNettyConfig_rdh | /**
* Generates {@link NettyConfig} from Flink {@link Configuration}.
*
* @param configuration
* configuration object
* @param localTaskManagerCommunication
* true, to skip initializing the network stack
* @param taskManagerAddress
* identifying the IP address under which the TaskManager will be
* acce... | 3.26 |
flink_NettyShuffleEnvironmentConfiguration_fromConfiguration_rdh | // ------------------------------------------------------------------------
/**
* Utility method to extract network related parameters from the configuration and to sanity
* check them.
*
* @param configuration
* configuration object
* @param networkMemorySize
* the size of memory reserved for shuffle enviro... | 3.26 |
flink_NettyShuffleEnvironmentConfiguration_getDataBindPortRange_rdh | /**
* Parses the hosts / ports for communication and data exchange from configuration.
*
* @param configuration
* configuration object
* @return the data port
*/
private static PortRange getDataBindPortRange(Configuration configuration) {
if (configuration.contains(NettyShuffleEnvironmentOptions.DATA_BIND_... | 3.26 |
flink_SingleInputSemanticProperties_m0_rdh | /**
* Adds, to the existing information, a field that is forwarded directly from the source
* record(s) to the destination record(s).
*
* @param sourceField
* the position in the source record(s)
* @param targetField
* the position in the destination record(s)
*/
public void m0(int sourceField, int targetFi... | 3.26 |
flink_SingleInputSemanticProperties_addReadFields_rdh | /**
* Adds, to the existing information, field(s) that are read in the source record(s).
*
* @param readFields
* the position(s) in the source record(s)
*/
public void addReadFields(FieldSet readFields) {
if (this.readFields == null) {
this.readFields = readFields;
} else {
this.readField... | 3.26 |
flink_RegisteredRpcConnection_createNewRegistration_rdh | // ------------------------------------------------------------------------
// Internal methods
// ------------------------------------------------------------------------
private RetryingRegistration<F, G, S, R> createNewRegistration() {
RetryingRegistration<F, G, S, R> newRegistration = checkNotNull(generateReg... | 3.26 |
flink_RegisteredRpcConnection_start_rdh | // ------------------------------------------------------------------------
// Life cycle
// ------------------------------------------------------------------------
public void start() {
checkState(!closed, "The RPC connection is already closed");
checkState((!isConnected()) && (pendingRegistration == null), "... | 3.26 |
flink_RegisteredRpcConnection_getTargetLeaderId_rdh | // ------------------------------------------------------------------------
// Properties
// ------------------------------------------------------------------------
public F getTargetLeaderId() {
return fencingToken;
} | 3.26 |
flink_RegisteredRpcConnection_toString_rdh | // ------------------------------------------------------------------------
@Override
public String toString() {
String connectionInfo = ((("(ADDRESS: " + targetAddress) + " FENCINGTOKEN: ") + fencingToken) + ")";
if (isConnected()) {
connectionInfo = (("RPC connection to " + targetGateway.getClass... | 3.26 |
flink_RegisteredRpcConnection_close_rdh | /**
* Close connection.
*/
public void close() {
closed = true;
// make sure we do not keep re-trying forever
if (pendingRegistration != null) {
pendingRegistration.cancel();
}
} | 3.26 |
flink_RegisteredRpcConnection_getTargetGateway_rdh | /**
* Gets the RegisteredGateway. This returns null until the registration is completed.
*/
public G getTargetGateway() {
return targetGateway;
} | 3.26 |
flink_ChangelogCollectResult_processRecord_rdh | // --------------------------------------------------------------------------------------------
@Override
protected void processRecord(RowData row) {synchronized(resultLock) {
// wait if the buffer is full
if (changeRecordBuffer.size() >= CHANGE_RECORD_BUFFER_SIZE) {
try {
resultLock.wait();... | 3.26 |
flink_RowKind_fromByteValue_rdh | /**
* Creates a {@link RowKind} from the given byte value. Each {@link RowKind} has a byte value
* representation.
*
* @see #toByteValue() for mapping of byte value and {@link RowKind}.
*/
public static RowKind fromByteValue(byte value) {
switch (value) {
case 0 :
return f0;
case ... | 3.26 |
flink_AndCondition_getLeft_rdh | /**
*
* @return One of the {@link IterativeCondition conditions} combined in this condition.
*/
public IterativeCondition<T> getLeft() {
return f0;
} | 3.26 |
flink_AndCondition_getRight_rdh | /**
*
* @return One of the {@link IterativeCondition conditions} combined in this condition.
*/
public IterativeCondition<T> getRight() {
return right;
} | 3.26 |
flink_BooleanParser_byteArrayEquals_rdh | /**
* Checks if a part of a byte array matches another byte array with chars (case-insensitive).
*
* @param source
* The source byte array.
* @param start
* The offset into the source byte array.
* @param length
* The length of the match.
* @param other
* The byte array which is fully compared to the ... | 3.26 |
flink_DataStatistics_cacheBaseStatistics_rdh | /**
* Caches the given statistics. They are later retrievable under the given identifier.
*
* @param statistics
* The statistics to cache.
* @param identifier
* The identifier which may be later used to retrieve the statistics.
*/
public void cacheBaseStatistics(BaseStatistics statistics, String identifier) ... | 3.26 |
flink_DataStatistics_m0_rdh | // --------------------------------------------------------------------------------------------
/**
* Gets the base statistics for the input identified by the given identifier.
*
* @param inputIdentifier
* The identifier for the input.
* @return The statistics that were cached for this input.
*/
public BaseStat... | 3.26 |
flink_LinkedListSerializer_isImmutableType_rdh | // ------------------------------------------------------------------------
// Type Serializer implementation
// ------------------------------------------------------------------------
@Override
public boolean isImmutableType() {
return false;
} | 3.26 |
flink_LinkedListSerializer_equals_rdh | // --------------------------------------------------------------------
@Overridepublic boolean equals(Object obj) {
return (obj == this) || (((obj != null) && (obj.getClass() == getClass())) && elementSerializer.equals(((LinkedListSerializer<?>) (obj)).elementSerializer));
} | 3.26 |
flink_LinkedListSerializer_getElementSerializer_rdh | // ------------------------------------------------------------------------
// LinkedListSerializer specific properties
// ------------------------------------------------------------------------
/**
* Gets the serializer for the elements of the list.
*
* @return The serializer for the el... | 3.26 |
flink_LinkedListSerializer_snapshotConfiguration_rdh | // --------------------------------------------------------------------------------------------
// Serializer configuration snapshot & compatibility
// --------------------------------------------------------------------------------------------
@Override
public TypeSerializerSnapshot<LinkedList<T>> snapshotConfigurat... | 3.26 |
flink_DispatcherId_fromUuid_rdh | /**
* Creates a new DispatcherId that corresponds to the UUID.
*/
public static DispatcherId fromUuid(UUID uuid) {
return new DispatcherId(uuid);
} | 3.26 |
flink_DispatcherId_generate_rdh | /**
* Generates a new random DispatcherId.
*/
public static DispatcherId generate() {
return new DispatcherId();
} | 3.26 |
flink_AbstractReader_handleEvent_rdh | /**
* Handles the event and returns whether the reader reached an end-of-stream event (either the
* end of the whole stream or the end of an superstep).
*/
protected boolean handleEvent(AbstractEvent event) throws IOException {
final Class<?> eventType = event.getClass();
try {
// -----------------... | 3.26 |
flink_AbstractReader_registerTaskEventListener_rdh | // ------------------------------------------------------------------------
// Events
// ------------------------------------------------------------------------
@Override
public void registerTaskEventListener(EventListener<TaskEvent> listener, Class<? extends TaskEvent> eventType) {
taskEventHandler.subscribe(list... | 3.26 |
flink_AbstractReader_setIterativeReader_rdh | // ------------------------------------------------------------------------
// Iterations
// ------------------------------------------------------------------------
@Override
public void setIterativeReader() {
isIterative = true;
} | 3.26 |
flink_RateLimiterStrategy_perCheckpoint_rdh | /**
* Creates a {@code RateLimiterStrategy} that is limiting the number of records per checkpoint.
*
* @param recordsPerCheckpoint
* The number of records produced per checkpoint. This value has to
* be greater or equal to parallelism. The actual number of produced records is subject to
* rounding due to di... | 3.26 |
flink_RateLimiterStrategy_noOp_rdh | /**
* Creates a convenience {@code RateLimiterStrategy} that is not limiting the records rate.
*/
static RateLimiterStrategy noOp() {
return parallelism -> new NoOpRateLimiter();
} | 3.26 |
flink_RateLimiterStrategy_perSecond_rdh | /**
* Creates a {@code RateLimiterStrategy} that is limiting the number of records per second.
*
* @param recordsPerSecond
* The number of records produced per second. The actual number of
* produced records is subject to rounding due to dividing the number of produced records
* among the parallel instances... | 3.26 |
flink_ExecNodePlanDumper_treeToString_rdh | /**
* Converts an {@link ExecNode} tree to a string as a tree style.
*
* <p>The following tree of {@link ExecNode}
*
* <pre>{@code Sink
* |
* Join
* / \
* Filter1 Filter2
* \ /
* Project
* |
* Scan}</pre>
*
* <p>would be converted to the tree style as following:
... | 3.26 |
flink_ExecNodePlanDumper_getReuseId_rdh | /**
* Returns reuse id if the given node is a reuse node (that means it has multiple outputs),
* else -1.
*/
public Integer getReuseId(ExecNode<?> node) {
return mapReuseNodeToReuseId.getOrDefault(node, -1);
} | 3.26 |
flink_ExecNodePlanDumper_m0_rdh | /**
* Converts an {@link ExecNode} tree to a string as a tree style.
*
* @param node
* the ExecNode to convert
* @param borders
* node sets that stop visit when meet them
* @param includingBorders
* Whether print the border nodes
* @return the plan of ExecNode
*/
public static String m0(ExecNode<?> node... | 3.26 |
flink_ExecNodePlanDumper_addVisitedTimes_rdh | /**
* Updates visited times for given node, return the new times.
*/
int addVisitedTimes(ExecNode<?> node) {
return mapNodeToVisitedTimes.compute(node, (k, v) -> v == null ? 1 : v + 1);
} | 3.26 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.