name
stringlengths
12
178
code_snippet
stringlengths
8
36.5k
score
float64
3.26
3.68
flink_KubernetesUtils_getDeploymentName_rdh
/** * Generate name of the Deployment. */public static String getDeploymentName(String clusterId) { return clusterId; }
3.26
flink_KubernetesUtils_parsePort_rdh
/** * Parse a valid port for the config option. A fixed port is expected, and do not support a * range of ports. * * @param flinkConfig * flink config * @param port * port config option * @return valid port */ public static Integer parsePort(Configuration flinkConfig, ConfigOption<String> port) { check...
3.26
flink_KubernetesUtils_createCompletedCheckpointStore_rdh
/** * Create a {@link DefaultCompletedCheckpointStore} with {@link KubernetesStateHandleStore}. * * @param configuration * configuration to build a RetrievableStateStorageHelper * @param kubeClient * flink kubernetes client * @param configMapName * ConfigMap name * @param executor * executor to run bl...
3.26
flink_KubernetesUtils_getServiceAccount_rdh
/** * Get the service account from the input pod first, if not specified, the service account name * will be used. * * @param flinkPod * the Flink pod to parse the service account * @return the parsed service account */ @Nullable public static String getServiceAccount(FlinkPod flinkPod) { final String serv...
3.26
flink_KubernetesUtils_getCommonLabels_rdh
/** * Get the common labels for Flink native clusters. All the Kubernetes resources will be set * with these labels. * * @param clusterId * cluster id * @return Return common labels map */ public static Map<String, String> getCommonLabels(String clusterId) { final Map<String, String> commonLabels = new Has...
3.26
flink_KubernetesUtils_tryToGetPrettyPrintYaml_rdh
/** * Try to get the pretty print yaml for Kubernetes resource. * * @param kubernetesResource * kubernetes resource * @return the pretty print yaml, or the {@link KubernetesResource#toString()} if parse failed. */ public static String tryToGetPrettyPrintYaml(KubernetesResource kubernetesResource) { try { ...
3.26
flink_KubernetesUtils_checkAndUpdatePortConfigOption_rdh
/** * Check whether the port config option is a fixed port. If not, the fallback port will be set * to configuration. * * @param flinkConfig * flink configuration * @param port * config option need to be checked * @param fallbackPort * the fallback port that will be set to the configuration */ public st...
3.26
flink_KubernetesUtils_resolveDNSPolicy_rdh
/** * Resolve the DNS policy defined value. Return DNS_POLICY_HOSTNETWORK if host network enabled. * If not, check whether there is a DNS policy overridden in pod template. * * @param dnsPolicy * DNS policy defined in pod template spec * @param hostNetworkEnabled * Host network enabled or not * @return the ...
3.26
flink_KubernetesUtils_resolveUserDefinedValue_rdh
/** * Resolve the user defined value with the precedence. First an explicit config option value is * taken, then the value in pod template and at last the default value of a config option if * nothing is specified. * * @param flinkConfig * flink configuration * @param configOption * the config option to def...
3.26
flink_KubernetesUtils_createJobGraphStateHandleStore_rdh
/** * Create a {@link KubernetesStateHandleStore} which storing {@link JobGraph}. * * @param configuration * configuration to build a RetrievableStateStorageHelper * @param flinkKubeClient * flink kubernetes client * @param configMapName * ConfigMap name * @param lockIdentity * lock identity to check ...
3.26
flink_KubernetesUtils_getResourceRequirements_rdh
/** * Get resource requirements from memory and cpu. * * @param resourceRequirements * resource requirements in pod template * @param mem * Memory in mb. * @param memoryLimitFactor * limit factor for the memory, used to set the limit resources. * @param cpu * cpu. * @param cpuLimitFactor * limit f...
3.26
flink_KubernetesUtils_getOnlyConfigMap_rdh
/** * Check the ConfigMap list should only contain the expected one. * * @param configMaps * ConfigMap list to check * @param expectedConfigMapName * expected ConfigMap Name * @return Return the expected ConfigMap */ public static KubernetesConfigMap getOnlyConfigMap(List<KubernetesConfigMap> configMaps, St...
3.26
flink_KubernetesUtils_getNamespacedServiceName_rdh
/** * Generate namespaced name of the service. */ public static String getNamespacedServiceName(Service service) { return (service.getMetadata().getName() + ".") + service.getMetadata().getNamespace(); }
3.26
flink_KubernetesUtils_getConfigMapLabels_rdh
/** * Get ConfigMap labels for the current Flink cluster. They could be used to filter and clean-up * the resources. * * @param clusterId * cluster id * @param type * the config map use case. It could only be {@link Constants#LABEL_CONFIGMAP_TYPE_HIGH_AVAILABILITY} now. * @return Return ConfigMap labels. *...
3.26
flink_KubernetesUtils_isHostNetwork_rdh
/** * Checks if hostNetwork is enabled. */ public static boolean isHostNetwork(Configuration configuration) { return configuration.getBoolean(KubernetesConfigOptions.KUBERNETES_HOSTNETWORK_ENABLED); }
3.26
flink_RawFormatFactory_validateAndExtractSingleField_rdh
/** * Validates and extract the single field type from the given physical row schema. */ private static LogicalType validateAndExtractSingleField(RowType physicalRowType) { if (physicalRowType.getFieldCount() != 1) { String schemaString = physicalRowType.getFields().stream().map(RowType.Ro...
3.26
flink_RawFormatFactory_checkFieldType_rdh
/** * Checks the given field type is supported. */ private static void checkFieldType(LogicalType fieldType) { if (!supportedTypes.contains(fieldType.getTypeRoot())) { throw new ValidationException(String.format("The 'raw' format doesn't supports '%s' as column type.", fieldType.asSummaryString())); ...
3.26
flink_TableFunctionCollector_reset_rdh
/** * Resets the flag to indicate whether [[collect(T)]] has been called. */ public void reset() { this.collected = false; }
3.26
flink_TableFunctionCollector_setInput_rdh
/** * Sets the input row from left table, which will be used to cross join with the result of table * function. */ public void setInput(Object input) { this.input = input; }
3.26
flink_TableFunctionCollector_getInput_rdh
/** * Gets the input value from left table, which will be used to cross join with the result of * table function. */public Object getInput() { return input; }
3.26
flink_TableFunctionCollector_setCollector_rdh
/** * Sets the current collector, which used to emit the final row. */ public void setCollector(Collector<?> collector) { this.collector = collector; }
3.26
flink_TableFunctionCollector_isCollected_rdh
/** * Whether {@link #collect(Object)} has been called. * * @return True if {@link #collect(Object)} has been called. */ public boolean isCollected() {return collected; }
3.26
flink_SavepointMetadata_getNewOperators_rdh
/** * * @return List of new operator states for the savepoint, represented by their target {@link OperatorID} and {@link BootstrapTransformation}. */ public List<BootstrapTransformationWithID<?>> getNewOperators() { return operatorStateIndex.values().stream().filter(OperatorStateSpec::isNewStateTransformation).m...
3.26
flink_SavepointMetadata_getExistingOperators_rdh
/** * * @return List of {@link OperatorState} that already exists within the savepoint. */ public List<OperatorState> getExistingOperators() { return operatorStateIndex.values().stream().filter(OperatorStateSpec::isExistingState).map(OperatorStateSpec::asExistingState).collect(Collectors.toList()); }
3.26
flink_SavepointMetadata_getOperatorState_rdh
/** * * @return Operator state for the given UID. * @throws IOException * If the savepoint does not contain operator state with the given uid. */ public OperatorState getOperatorState(String uid) throws IOException { OperatorID operatorID = OperatorIDGenerator.fromUid(uid); OperatorStateSpec operator...
3.26
flink_PythonFunctionFactory_getPythonFunction_rdh
/** * Returns PythonFunction according to the fully qualified name of the Python UDF i.e * ${moduleName}.${functionName} or ${moduleName}.${className}. * * @param fullyQualifiedName * The fully qualified name of the Python UDF. * @param config * The configuration of python dependencies. * @param classLoader...
3.26
flink_BroadcastConnectedStream_getSecondInput_rdh
/** * Returns the {@link BroadcastStream}. * * @return The stream which, by convention, is the broadcast one. */ public BroadcastStream<IN2> getSecondInput() { return broadcastStream; }
3.26
flink_BroadcastConnectedStream_process_rdh
/** * Assumes as inputs a {@link BroadcastStream} and a non-keyed {@link DataStream} and applies * the given {@link BroadcastProcessFunction} on them, thereby creating a transformed output * stream. * * @param function * The {@link BroadcastProcessFunction} that is called for each element in the * stream. *...
3.26
flink_SecurityFactoryServiceLoader_findModuleFactory_rdh
/** * Find a suitable {@link SecurityModuleFactory} based on canonical name. */ public static SecurityModuleFactory findModuleFactory(String securityModuleFactoryClass) throws NoMatchSecurityFactoryException { return findFactoryInternal(securityModuleFactoryClass, SecurityModuleFactory.class, SecurityModuleF...
3.26
flink_SecurityFactoryServiceLoader_findContextFactory_rdh
/** * Find a suitable {@link SecurityContextFactory} based on canonical name. */ public static SecurityContextFactory findContextFactory(String securityContextFactoryClass) throws NoMatchSecurityFactoryException { return findFactoryInternal(securityContextFactoryClass, SecurityContextFactory.class, SecurityContex...
3.26
flink_HeapListState_get_rdh
// ------------------------------------------------------------------------ // state access // ------------------------------------------------------------------------ @Override public Iterable<V> get() { return getInternal();}
3.26
flink_HeapListState_mergeState_rdh
// ------------------------------------------------------------------------ // state merging // ------------------------------------------------------------------------ @Override protected List<V> mergeState(List<V> a, List<V> b) { a.addAll(b); return a; }
3.26
flink_HistoryServer_createOrGetFile_rdh
// ------------------------------------------------------------------------ // File generation // ------------------------------------------------------------------------ static FileWriter createOrGetFile(File folder, String name) throws IOException { File file = new File(folder, name + ".json"); if (!file.ex...
3.26
flink_HistoryServer_start_rdh
// ------------------------------------------------------------------------ // Life-cycle // ------------------------------------------------------------------------ void start() throws IOException, InterruptedException { synchronized(startupShutdownLock) { LOG.info("Starting history server."); Fil...
3.26
flink_MaxWithRetractAggFunction_getArgumentDataTypes_rdh
// -------------------------------------------------------------------------------------------- // Planning // -------------------------------------------------------------------------------------------- @Override public List<DataType> getArgumentDataTypes() { return Collections.singletonList(valueDataType); }
3.26
flink_AggregatorWithName_getName_rdh
/** * Gets the name that the aggregator is registered under. * * @return The name that the aggregator is registered under. */ public String getName() { return name; }
3.26
flink_AggregatorWithName_getAggregator_rdh
/** * Gets the aggregator. * * @return The aggregator. */ public Aggregator<T> getAggregator() { return aggregator; }
3.26
flink_BigDecComparator_putNormalizedKey_rdh
/** * Adds a normalized key containing a normalized order of magnitude of the given record. 2 bits * determine the sign (negative, zero, positive), 33 bits determine the magnitude. This method * adds at most 5 bytes that contain information. */ @Override public void putNormalizedKey(BigDecimal record, MemorySegment...
3.26
flink_EnrichedRowData_from_rdh
/** * Creates a new {@link EnrichedRowData} with the provided {@code fixedRow} as the immutable * static row, and uses the {@code producedRowFields}, {@code fixedRowFields} and {@code mutableRowFields} arguments to compute the indexes mapping. * * <p>The {@code producedRowFields} should include the name of fields o...
3.26
flink_EnrichedRowData_computeIndexMapping_rdh
/** * This method computes the index mapping for {@link EnrichedRowData}. * * @see EnrichedRowData#from(RowData, List, List, List) */ public static int[] computeIndexMapping(List<String> producedRowFields, List<String> mutableRowFields, List<String> fixedRowFields) { ...
3.26
flink_EnrichedRowData_getArity_rdh
// --------------------------------------------------------------------------------------------- @Override public int getArity() { return indexMapping.length;}
3.26
flink_BroadcastVariableManager_getNumberOfVariablesWithReferences_rdh
// -------------------------------------------------------------------------------------------- public int getNumberOfVariablesWithReferences() { return this.variables.size(); }
3.26
flink_BroadcastVariableManager_materializeBroadcastVariable_rdh
// -------------------------------------------------------------------------------------------- /** * Materializes the broadcast variable for the given name, scoped to the given task and its * iteration superstep. An existing materialization created by another parallel subtask may be * returned, ...
3.26
flink_LocalProperties_addUniqueFields_rdh
/** * Adds a combination of fields that are unique in these data properties. * * @param uniqueFields * The fields that are unique in these data properties. */ public LocalProperties addUniqueFields(FieldSet uniqueFields) { LocalProperties copy = clone(); if (copy.uniqu...
3.26
flink_LocalProperties_getOrdering_rdh
// -------------------------------------------------------------------------------------------- /** * Gets the key order. * * @return The key order, or <code>null</code> if nothing is ordered. */ public Ordering getOrdering() { return ordering; }
3.26
flink_LocalProperties_areFieldsUnique_rdh
/** * Checks whether the given set of fields is unique, as specified in these local properties. * * @param set * The set to check. * @return True, if the given column combination is unique, false if not. */ public boolean areFieldsUnique(FieldSet set) { return (this.uniqueFields != null) && this.uniqueField...
3.26
flink_LocalProperties_combine_rdh
// -------------------------------------------------------------------------------------------- public static LocalProperties combine(LocalProperties lp1, LocalProperties lp2) { if (lp1.ordering != null) { return lp1; } else if (lp2.ordering != null) { return lp2; } else if (lp1.groupedFields != null) { ret...
3.26
flink_LocalProperties_filterBySemanticProperties_rdh
// -------------------------------------------------------------------------------------------- /** * Filters these LocalProperties by the fields that are forwarded to the output as described by * the SemanticProperties. * * @param props * The semantic properties holding information about forwarded fields. * @p...
3.26
flink_LocalProperties_forOrdering_rdh
// -------------------------------------------------------------------------------------------- public static LocalProperties forOrdering(Ordering o) { LocalProperties props = new LocalProperties(); props.ordering = o; props.groupedFields = o.getInvolvedIndexes(); return props; }
3.26
flink_LocalProperties_isTrivial_rdh
/** * Checks, if the properties in this object are trivial, i.e. only standard values. */ public boolean isTrivial() { return ((ordering == null) && (this.groupedFields == null)) && (this.uniqueFields == null); }
3.26
flink_FlinkPipelineTranslationUtil_getJobGraph_rdh
/** * Transmogrifies the given {@link Pipeline} to a {@link JobGraph}. */ public static JobGraph getJobGraph(ClassLoader userClassloader, Pipeline pipeline, Configuration optimizerConfiguration, int defaultParallelism) { FlinkPipelineTranslator pipelineTranslator = getPipelineTranslator(userClassloader, pipeline)...
3.26
flink_FlinkPipelineTranslationUtil_getJobGraphUnderUserClassLoader_rdh
/** * Transmogrifies the given {@link Pipeline} under the userClassloader to a {@link JobGraph}. */ public static JobGraph getJobGraphUnderUserClassLoader(final ClassLoader userClassloader, final Pipeline pipeline, final Configuration configuration, final int defaultParallelism) { final Class...
3.26
flink_FlinkPipelineTranslationUtil_translateToJSONExecutionPlan_rdh
/** * Extracts the execution plan (as JSON) from the given {@link Pipeline}. */ public static String translateToJSONExecutionPlan(ClassLoader userClassloader, Pipeline pipeline) { FlinkPipelineTranslator pipelineTranslator = getPipelineTranslator(userClassloader, pipeline); return pipelineTranslator.translateToJSONEx...
3.26
flink_NettyMessageClientDecoderDelegate_channelInactive_rdh
/** * Releases resources when the channel is closed. When exceptions are thrown during processing * received netty buffers, {@link CreditBasedPartitionRequestClientHandler} is expected to catch * the exception and close the channel and trigger this notification. * * @param ctx * The context of the channel close...
3.26
flink_HighAvailabilityServices_getWebMonitorLeaderElection_rdh
/** * Gets the {@link LeaderElection} for the cluster's rest endpoint. * * @deprecated Use {@link #getClusterRestEndpointLeaderElection()} instead. */@Deprecateddefault LeaderElection getWebMonitorLeaderElection() { throw new UnsupportedOperationException(((("getWebMonitorLeaderElectionService should no longer ...
3.26
flink_HighAvailabilityServices_getWebMonitorLeaderRetriever_rdh
/** * This retriever should no longer be used on the cluster side. The web monitor retriever is * only required on the client-side and we have a dedicated high-availability services for the * client, named {@link ClientHighAvailabilityServices}. See also FLINK-13750. * * @return the leader retriever for web monito...
3.26
flink_HighAvailabilityServices_closeWithOptionalClean_rdh
/** * Calls {@link #cleanupAllData()} (if {@code true} is passed as a parameter) before calling * {@link #close()} on this instance. Any error that appeared during the cleanup will be * propagated after calling {@code close()}. */ default void closeWithOptionalClean(boolean cleanupData) throws Exception { Throw...
3.26
flink_HighAvailabilityServices_getClusterRestEndpointLeaderElection_rdh
/** * Gets the {@link LeaderElection} for the cluster's rest endpoint. */ default LeaderElection getClusterRestEndpointLeaderElection() { // for backwards compatibility we delegate to getWebMonitorLeaderElectionService // all implementations of this interface should override // getClusterRestEndpointLeade...
3.26
flink_ObjectIdentifier_toList_rdh
/** * List of the component names of this object identifier. */ public List<String> toList() { if (catalogName == null) { return Collections.singletonList(getObjectName()); } return Arrays.asList(getCatalogName(), m0(), getObjectName()); }
3.26
flink_ObjectIdentifier_toObjectPath_rdh
/** * Convert this {@link ObjectIdentifier} to {@link ObjectPath}. * * @throws TableException * if the identifier cannot be converted */ public ObjectPath toObjectPath() throws TableException {if (catalogName == null) { throw new TableException("This ObjectIdentifier instance refers to an anonymous obje...
3.26
flink_ObjectIdentifier_asSerializableString_rdh
/** * Returns a string that fully serializes this instance. The serialized string can be used for * transmitting or persisting an object identifier. * * @throws TableException * if the identifier cannot be serialized */ public String asSerializableString() throws TableException { if (catalogName == null) ...
3.26
flink_ObjectIdentifier_m1_rdh
/** * Returns a string that summarizes this instance for printing to a console or log. */ public String m1() { if (catalogName == null) { return objectName; } return String.join(".", catalogName, databaseName, objectName); }
3.26
flink_ObjectIdentifier_ofAnonymous_rdh
/** * This method allows to create an {@link ObjectIdentifier} without catalog and database name, * in order to propagate anonymous objects with unique identifiers throughout the stack. * * <p>This method for no reason should be exposed to users, as this should be used only when * creating anonymous tables with un...
3.26
flink_NettyShuffleEnvironmentConfiguration_numNetworkBuffers_rdh
// ------------------------------------------------------------------------ public int numNetworkBuffers() { return numNetworkBuffers; }
3.26
flink_NettyShuffleEnvironmentConfiguration_m1_rdh
// ------------------------------------------------------------------------ @Override public int m1() { int result = 1; result = (31 * result) + numNetworkBuffers; result = (31 * result) + networkBufferSize; result = (31 * result) + partitionRequestInitialBackoff; result = (31 * result) + partition...
3.26
flink_NettyShuffleEnvironmentConfiguration_calculateNumberOfNetworkBuffers_rdh
/** * Calculates the number of network buffers based on configuration and jvm heap size. * * @param configuration * configuration object * @param networkMemorySize * the size of memory reserved for shuffle environment ...
3.26
flink_NettyShuffleEnvironmentConfiguration_createNettyConfig_rdh
/** * Generates {@link NettyConfig} from Flink {@link Configuration}. * * @param configuration * configuration object * @param localTaskManagerCommunication * true, to skip initializing the network stack * @param taskManagerAddress * identifying the IP address under which the TaskManager will be * acce...
3.26
flink_NettyShuffleEnvironmentConfiguration_fromConfiguration_rdh
// ------------------------------------------------------------------------ /** * Utility method to extract network related parameters from the configuration and to sanity * check them. * * @param configuration * configuration object * @param networkMemorySize * the size of memory reserved for shuffle enviro...
3.26
flink_NettyShuffleEnvironmentConfiguration_getDataBindPortRange_rdh
/** * Parses the hosts / ports for communication and data exchange from configuration. * * @param configuration * configuration object * @return the data port */ private static PortRange getDataBindPortRange(Configuration configuration) { if (configuration.contains(NettyShuffleEnvironmentOptions.DATA_BIND_...
3.26
flink_SingleInputSemanticProperties_m0_rdh
/** * Adds, to the existing information, a field that is forwarded directly from the source * record(s) to the destination record(s). * * @param sourceField * the position in the source record(s) * @param targetField * the position in the destination record(s) */ public void m0(int sourceField, int targetFi...
3.26
flink_SingleInputSemanticProperties_addReadFields_rdh
/** * Adds, to the existing information, field(s) that are read in the source record(s). * * @param readFields * the position(s) in the source record(s) */ public void addReadFields(FieldSet readFields) { if (this.readFields == null) { this.readFields = readFields; } else { this.readField...
3.26
flink_RegisteredRpcConnection_createNewRegistration_rdh
// ------------------------------------------------------------------------ // Internal methods // ------------------------------------------------------------------------ private RetryingRegistration<F, G, S, R> createNewRegistration() { RetryingRegistration<F, G, S, R> newRegistration = checkNotNull(generateReg...
3.26
flink_RegisteredRpcConnection_start_rdh
// ------------------------------------------------------------------------ // Life cycle // ------------------------------------------------------------------------ public void start() { checkState(!closed, "The RPC connection is already closed"); checkState((!isConnected()) && (pendingRegistration == null), "...
3.26
flink_RegisteredRpcConnection_getTargetLeaderId_rdh
// ------------------------------------------------------------------------ // Properties // ------------------------------------------------------------------------ public F getTargetLeaderId() { return fencingToken; }
3.26
flink_RegisteredRpcConnection_toString_rdh
// ------------------------------------------------------------------------ @Override public String toString() { String connectionInfo = ((("(ADDRESS: " + targetAddress) + " FENCINGTOKEN: ") + fencingToken) + ")"; if (isConnected()) { connectionInfo = (("RPC connection to " + targetGateway.getClass...
3.26
flink_RegisteredRpcConnection_close_rdh
/** * Close connection. */ public void close() { closed = true; // make sure we do not keep re-trying forever if (pendingRegistration != null) { pendingRegistration.cancel(); } }
3.26
flink_RegisteredRpcConnection_getTargetGateway_rdh
/** * Gets the RegisteredGateway. This returns null until the registration is completed. */ public G getTargetGateway() { return targetGateway; }
3.26
flink_ChangelogCollectResult_processRecord_rdh
// -------------------------------------------------------------------------------------------- @Override protected void processRecord(RowData row) {synchronized(resultLock) { // wait if the buffer is full if (changeRecordBuffer.size() >= CHANGE_RECORD_BUFFER_SIZE) { try { resultLock.wait();...
3.26
flink_RowKind_fromByteValue_rdh
/** * Creates a {@link RowKind} from the given byte value. Each {@link RowKind} has a byte value * representation. * * @see #toByteValue() for mapping of byte value and {@link RowKind}. */ public static RowKind fromByteValue(byte value) { switch (value) { case 0 : return f0; case ...
3.26
flink_AndCondition_getLeft_rdh
/** * * @return One of the {@link IterativeCondition conditions} combined in this condition. */ public IterativeCondition<T> getLeft() { return f0; }
3.26
flink_AndCondition_getRight_rdh
/** * * @return One of the {@link IterativeCondition conditions} combined in this condition. */ public IterativeCondition<T> getRight() { return right; }
3.26
flink_BooleanParser_byteArrayEquals_rdh
/** * Checks if a part of a byte array matches another byte array with chars (case-insensitive). * * @param source * The source byte array. * @param start * The offset into the source byte array. * @param length * The length of the match. * @param other * The byte array which is fully compared to the ...
3.26
flink_DataStatistics_cacheBaseStatistics_rdh
/** * Caches the given statistics. They are later retrievable under the given identifier. * * @param statistics * The statistics to cache. * @param identifier * The identifier which may be later used to retrieve the statistics. */ public void cacheBaseStatistics(BaseStatistics statistics, String identifier) ...
3.26
flink_DataStatistics_m0_rdh
// -------------------------------------------------------------------------------------------- /** * Gets the base statistics for the input identified by the given identifier. * * @param inputIdentifier * The identifier for the input. * @return The statistics that were cached for this input. */ public BaseStat...
3.26
flink_LinkedListSerializer_isImmutableType_rdh
// ------------------------------------------------------------------------ // Type Serializer implementation // ------------------------------------------------------------------------ @Override public boolean isImmutableType() { return false; }
3.26
flink_LinkedListSerializer_equals_rdh
// -------------------------------------------------------------------- @Overridepublic boolean equals(Object obj) { return (obj == this) || (((obj != null) && (obj.getClass() == getClass())) && elementSerializer.equals(((LinkedListSerializer<?>) (obj)).elementSerializer)); }
3.26
flink_LinkedListSerializer_getElementSerializer_rdh
// ------------------------------------------------------------------------ // LinkedListSerializer specific properties // ------------------------------------------------------------------------ /** * Gets the serializer for the elements of the list. * * @return The serializer for the el...
3.26
flink_LinkedListSerializer_snapshotConfiguration_rdh
// -------------------------------------------------------------------------------------------- // Serializer configuration snapshot & compatibility // -------------------------------------------------------------------------------------------- @Override public TypeSerializerSnapshot<LinkedList<T>> snapshotConfigurat...
3.26
flink_DispatcherId_fromUuid_rdh
/** * Creates a new DispatcherId that corresponds to the UUID. */ public static DispatcherId fromUuid(UUID uuid) { return new DispatcherId(uuid); }
3.26
flink_DispatcherId_generate_rdh
/** * Generates a new random DispatcherId. */ public static DispatcherId generate() { return new DispatcherId(); }
3.26
flink_AbstractReader_handleEvent_rdh
/** * Handles the event and returns whether the reader reached an end-of-stream event (either the * end of the whole stream or the end of an superstep). */ protected boolean handleEvent(AbstractEvent event) throws IOException { final Class<?> eventType = event.getClass(); try { // -----------------...
3.26
flink_AbstractReader_registerTaskEventListener_rdh
// ------------------------------------------------------------------------ // Events // ------------------------------------------------------------------------ @Override public void registerTaskEventListener(EventListener<TaskEvent> listener, Class<? extends TaskEvent> eventType) { taskEventHandler.subscribe(list...
3.26
flink_AbstractReader_setIterativeReader_rdh
// ------------------------------------------------------------------------ // Iterations // ------------------------------------------------------------------------ @Override public void setIterativeReader() { isIterative = true; }
3.26
flink_RateLimiterStrategy_perCheckpoint_rdh
/** * Creates a {@code RateLimiterStrategy} that is limiting the number of records per checkpoint. * * @param recordsPerCheckpoint * The number of records produced per checkpoint. This value has to * be greater or equal to parallelism. The actual number of produced records is subject to * rounding due to di...
3.26
flink_RateLimiterStrategy_noOp_rdh
/** * Creates a convenience {@code RateLimiterStrategy} that is not limiting the records rate. */ static RateLimiterStrategy noOp() { return parallelism -> new NoOpRateLimiter(); }
3.26
flink_RateLimiterStrategy_perSecond_rdh
/** * Creates a {@code RateLimiterStrategy} that is limiting the number of records per second. * * @param recordsPerSecond * The number of records produced per second. The actual number of * produced records is subject to rounding due to dividing the number of produced records * among the parallel instances...
3.26
flink_ExecNodePlanDumper_treeToString_rdh
/** * Converts an {@link ExecNode} tree to a string as a tree style. * * <p>The following tree of {@link ExecNode} * * <pre>{@code Sink * | * Join * / \ * Filter1 Filter2 * \ / * Project * | * Scan}</pre> * * <p>would be converted to the tree style as following: ...
3.26
flink_ExecNodePlanDumper_getReuseId_rdh
/** * Returns reuse id if the given node is a reuse node (that means it has multiple outputs), * else -1. */ public Integer getReuseId(ExecNode<?> node) { return mapReuseNodeToReuseId.getOrDefault(node, -1); }
3.26
flink_ExecNodePlanDumper_m0_rdh
/** * Converts an {@link ExecNode} tree to a string as a tree style. * * @param node * the ExecNode to convert * @param borders * node sets that stop visit when meet them * @param includingBorders * Whether print the border nodes * @return the plan of ExecNode */ public static String m0(ExecNode<?> node...
3.26
flink_ExecNodePlanDumper_addVisitedTimes_rdh
/** * Updates visited times for given node, return the new times. */ int addVisitedTimes(ExecNode<?> node) { return mapNodeToVisitedTimes.compute(node, (k, v) -> v == null ? 1 : v + 1); }
3.26