This node loads a KNIME data table into Impala. Impala requires imported data to be present in the HDFS file system on the Impala server, therefore this node first copies the data onto the Impala server. You can use the HDFS Connection node to establish a connection to the HDFS file system. The data is then loaded into a Impala table and the uploaded file is deleted.
Additionally the data can be partitioned by selecting one or more compatible columns (e.g. integer or string). The node relies on Impala's dynamic partitioning.
You want to see the source code for this node? Click the following button and we’ll use our super-powers to find it for you.
A zipped version of the software site can be downloaded here.
Deploy, schedule, execute, and monitor your KNIME workflows locally, in the cloud or on-premises – with our brand new NodePit Runner.Try NodePit Runner!
Do you have feedback, questions, comments about NodePit, want to support this platform, or want your own nodes or workflows listed here as well? Do you think, the search results could be improved or something is missing? Then please get in touch! Alternatively, you can send us an email to firstname.lastname@example.org, follow @NodePit on Twitter, or chat on Gitter!
Please note that this is only about NodePit. We do not provide general support for KNIME — please use the KNIME forums instead.