There are 96 nodes that can be used as predessesor for a node with an input port of type Spark Context.
Execute arbitrary Python code in Spark.
Execute arbitrary Python code in Spark.
Execute arbitrary Python code in Spark.
Execute arbitrary Python code in Spark.
Repartitions a Spark DataFrame.
Executes a Spark SQL query statement
Concatenates Spark DataFrame/RDDs row wise, inputs are optional.
The Spark GroupBy allows to group by the selected columns and output aggregated data to the generated groups.
Splits input data into two partitions.
Pivots and groups the given Spark DataFrame/RDD by the selected columns for pivoting and grouping. Also performs aggregations for each pivot value.
Do you have feedback, questions, comments about NodePit, want to support this platform, or want your own nodes or workflows listed here as well? Do you think, the search results could be improved or something is missing? Then please get in touch! Alternatively, you can send us an email to mail@nodepit.com.
Please note that this is only about NodePit. We do not provide general support for KNIME — please use the KNIME forums instead.