This node is part of the legacy database framework. For more information on how to migrate to the new database framework see the migration section of the database documentation.
This node loads a KNIME data table into Hive. Hive requires imported data to be present on the Hive server, therefore this node first copies the data onto the Hive server. You can use any of the protocols supported by the file handling nodes, e.g. SSH/SCP or FTP. The data is then loaded into a Hive table and the uploaded file is deleted.
Additionally the data can be partitioned by selecting one or more compatible columns (e.g. integer or string). The node relies on Hive's dynamic partitioning.
You want to see the source code for this node? Click the following button and we’ll use our super-powers to find it for you.
To use this node in KNIME, install the extension KNIME Big Data Connectors from the below update site following our NodePit Product and Node Installation Guide:
A zipped version of the software site can be downloaded here.
Deploy, schedule, execute, and monitor your KNIME workflows locally, in the cloud or on-premises – with our brand new NodePit Runner.
Try NodePit Runner!