This workflow demonstrates how to connect to various Azure services such as HDInsight clusters, Azure Blob Storage, and AzureSQL from within KNIME Analytics Platform.
The Azure Blob Store Connection node connects KNIME Analytics Platform with your Azure Blob Storage and allows you to work with your files using the file handling nodes. The Azure Blob Store File Picker node creates a pre-signed URL that can be used in the reader nodes in KNIME to read directly from Azure Blob Storage or that can be shared with other users to access the dedicated files without the need for authentication.
The Create Spark Context (Livy) node can be used to set up a Spark context in your Azure HDInsight cluster. For more details see link to the documentation. Once a context is created you can use all the existing Spark nodes to visually assemble your Spark analysis flow.
Additionally, it is also possible to connect to Hive within your HDInsight cluster via the Hive Connector node.
The Microsoft SQL Server Connector node provides a DB Connection which can be used with the existing DB nodes to visually assemble queries that are executed within your AzureSQL. To upload large amounts of data into AzureSQL use the DB Loader node.
To use this workflow in KNIME, download it from the below URL and open it in KNIME:
Download WorkflowDeploy, schedule, execute, and monitor your KNIME workflows locally, in the cloud or on-premises – with our brand new NodePit Runner.
Try NodePit Runner!Do you have feedback, questions, comments about NodePit, want to support this platform, or want your own nodes or workflows listed here as well? Do you think, the search results could be improved or something is missing? Then please get in touch! Alternatively, you can send us an email to mail@nodepit.com.
Please note that this is only about NodePit. We do not provide general support for KNIME — please use the KNIME forums instead.