There are 163 nodes that can be used as successor for a node with an output port of type Spark Data (legacy).
Executes a Spark SQL query statement
Concatenates Spark DataFrame/RDDs row wise, inputs are optional.
The Spark GroupBy allows to group by the selected columns and output aggregated data to the generated groups.
Splits input data into two partitions.
Pivots and groups the given Spark DataFrame/RDD by the selected columns for pivoting and grouping. Also performs aggregations for each pivot value.
The Spark Row Filter allows rows to be excluded from the input Spark DataFrame/RDD.
Extracts a sample from the input data.
Sorts the rows according to user-defined criteria.
Filters out correlated columns.
Computes the correlation matrix for the selected input columns.
Do you have feedback, questions, comments about NodePit, want to support this platform, or want your own nodes or workflows listed here as well? Do you think, the search results could be improved or something is missing? Then please get in touch! Alternatively, you can send us an email to mail@nodepit.com.
Please note that this is only about NodePit. We do not provide general support for KNIME — please use the KNIME forums instead.