Keras CuDNN LSTM Layer

This Node Is Deprecated — This version of the node has been replaced with a new and improved version. The old version is kept for backwards-compatibility, but for all new workflows we suggest to use the version linked below.
Go to Suggested ReplacementKeras CuDNN LSTM Layer

Can only be run on GPU, with the TensorFlow back end. Corresponds to the CuDNNLSTM Keras layer .

Options

Name prefix
The name prefix of the layer. The prefix is complemented by an index suffix to obtain a unique layer name. If this option is unchecked, the name prefix is derived from the layer type.
Input tensor
The tensor to use as input for the layer.
First hidden state tensor
The tensor to use as initial state for the first hidden state in case the corresponding port is connected.
Second hidden state tensor
The tensor to use as initial state for the second hidden state in case the corresponding port is connected.
Units
Dimensionality of the output space.
Return sequences
Whether to return the last output in the output sequence or the full output sequence. If selected the output will have shape [time, units] otherwise the output will have shape [units].
Return state
Whether to return the hidden state in addition to the layer output. If selected the layer returns two tensors, the normal output and the hidden state of the layer. If sequences are returned, this also applies to the hidden state.

Initializers

Kernel initializer
Initializer for the weight matrix used for the linear transformations of the input. See initializers for details on the available initializers.
Recurrent initializer
Initializer for the weight matrix used for the linear transformation of the recurrent connection. See initializers for details on the available initializers.
Bias initializer
Initializer for the bias vector (if a bias is used). See initializers for details on the available initializers.

Regularizers

Kernel regularizer
Regularizer function applied to the weight matrix. See regularizers for details on the available regularizers.
Recurrent regularizer
Regularizer function applied to the weight matrix for the recurrent connection. See regularizers for details on the available regularizers.
Bias regularizer
Regularizer function applied to the bias vector. See regularizers for details on the available regularizers.
Activity regularizer
Regularizer function applied to the output of the layer i.e. its activation. See regularizers for details on the available regularizers.

Constraints

Kernel constraint
Constraint on the weight matrix for the input connection. See constraints for details on the available constraints.
Recurrent constraint
Constraint on the weight matrix for the recurrent connection. See constraints for details on the available constraints.
Bias constraint
Constraint on the bias vector. See constraints for details on the available constraints.

Input Ports

Icon
The Keras deep learning network to which to add an CuDNN LSTM layer.
Icon
An optional Keras deep learning network providing the first initial state for this CuDNN LSTM layer. Note that if this port is connected, you also have to connect the second hidden state port. The hidden state must have shape [units], where units must correspond to the number of units this layer uses.
Icon
> An optional Keras deep learning network providing the second initial state for this CuDNN LSTM layer. Note that if this port is connected, you also have to connect the first hidden state port. The hidden state must have shape [units], where units must correspond to the number of units this layer uses.

Output Ports

Icon
The Keras deep learning network with an added CuDNN LSTM layer.

Views

This node has no views

Workflows

  • No workflows found

Links

Developers

You want to see the source code for this node? Click the following button and we’ll use our super-powers to find it for you.