HDFS Connector (KNOX)

This node connects to Hadoop Distributed File System (HDFS) using an Apache KNOX gateway. The resulting output port allows downstream nodes to access the files of the remote file system, e.g. to read or write, or to perform other file system operations (browse/list files, copy, move, ...).

Path syntax: Paths for HDFS are specified with a UNIX-like syntax, for example /myfolder/file.csv, which is an absolute path that consists of:

  1. A leading slash (/).
  2. The name of a folder (myfolder), followed by a slash.
  3. Followed by the name of a file (file.csv).

SSL: This node uses the JVM SSL settings.

Options

URL
KNOX URL including cluster name, for example https://<server>:8443/gateway/<default> (replace <server> with the fully qualified hostname and <default> with your cluster name). The protocol should be https and port 8443 on most deployments.
Username & Password
Credentials for username and password authentication.
Working directory
Specify the working directory of the resulting file system connection, using the UNIX-like path syntax. The working directory must be specified as an absolute path (starting with "/"). A working directory allows downstream nodes to access files/folders using relative paths, i.e. paths that do not have a leading slash. The default working directory is the root "/".
Connection timeout (seconds)
Timeout in seconds to establish a connection, or 0 for an infinite timeout.
Read timeout (seconds)
Timeout in seconds to read data from the connection, or 0 for an infinite timeout.

Input Ports

This node has no input ports

Output Ports

Icon
HDFS (via KNOX) File System Connection.

Views

This node has no views

Workflows

Links

Developers

You want to see the source code for this node? Click the following button and we’ll use our super-powers to find it for you.