TD_​GLMPerSegment

This function trains data to get generalized linear models. This function can train multiple models simutanously.

Options

Alpha
Specify the Elasticnet parameter for penalty computation. It only becomes effective if RegularizationLambda > 0. The value represents the contribution ratio of L1 in the penalty. A value of 1.0 indicates L1 (LASSO) only, a value of 0 indicates L2 (Ridge) only, and a value in between is a combination of L1 and L2. Default: 0.15 (15% L1, 85% L2). Must be a float value between 0 and 1.
BatchSize
Specify the number of observations ( training samples ) to be parsed in one mini-batch. Must be a non-negative integer value. A value of 0 indicates no mini-batches, the entire dataset is processed in each iteration, and the algorithm becomes Gradient Descent. A value higher than the number of rows on any AMP will also default to Gradient Descent.
ClassWeights
Specify weights associated with classes. Only applicable for Binomial Family. The format is ‘0:weight,1:weight’. For example, ‘0:1.0,1:0.5’ will give twice the weight to each observation in class 0. If the weight of a class is omitted, it is assumed to be 1.0.
DecayRate
Specify the decay rate for the learning rate (invtime and adaptive).
DecaySteps
Specify the decay steps (number of iterations) for the ‘adaptive’ learning rate. The learning rate changes by decay rate after this many number of iterations.
Family
Specify the distribution exponential family. Acceptable values are Gaussian, Binomial.
InitialEta
Specify the initial value of eta for the learning rate. For ‘constant’, this value is the learning rate for all iterations.
InputColumns
Specify the names of the input table columns that need to be used for training the model (predictors, features or independent variables).
Intercept
Specify whether intercept should be estimated or not (based on whether data is already centered or not).
IterationMode
Specify the iteration mode. Acceptable values are Batch, Epoch, where Batch is one iteration per batch and Epoch is one iteration per epoch.
IterNumNoChange
Specify the number of iterations (batches) with no improvement in loss (including the tolerance) to stop training (early stopping). Must be a non-negative integer. A value of 0 indicates no early stopping and the algorithm will continue till MaxIterNum iterations are reached.
LearningRate
Specify the learning rate algorithm for SGD iterations.
MaxIterNum
Specify the maximum number of iterations over the training data batches. In epoch mode, MaxIterNum equals the number of epochs (an epoch is a single pass over entire training data). In batch mode, if the batch size is 0, MaxIterNum equals the number of epochs. If there are 1000 rows in an AMP, and batch size is 10, then 100 iterations will result into one epoch and 500 iterations will result into 5 epochs over this AMP’s data.
Momentum
Specify the value to use for the momentum learning rate optimizer. Must be a non-negative float value between 0 and 1. A larger value indicates a higher momentum contribution. A value of 0 means the momentum optimizer is disabled. For a good momentum contribution, a value between 0.6-0.95 is recommended.
Nesterov
Specify whether Nesterov optimization should be applied to the momentum optimizer or not. Only applicable when momentum > 0.
PartitionColumn
Specify the name of the InputTable columns on which to partition the input. The name should be consistent with the partition_by_column in the ON clause. If the partition_by_column is unicode with foreign language characters, it is necessary to specify PartitionColumn argument. Note: Column range is not supported for PartitionColumn argument.
RegularizationLambda
Specify the amount of regularization to be added. The higher the value, the stronger the regularization. It is also used to compute the learning rate when the learning rate is set to ‘optimal’. Must be a non-negative float value. A value of 0 means no regularization.
ResponseColumn
Specify the name of the column that contains the class label for classification or target value (dependent variable) for regression.
Output Schema
Output Schema, if Volatile is true then use user login as the schema.
Output Table
Output Table
VAL Location
VAL Location
Volatile
Specifies whether the table should be a VOLATILE table. If true, then the table is automatically deleted, otherwise it is users responsibility to remove or clean it up for space.
Tolerance
Specify the stopping criteria in terms of loss function improvement. Only applicable when IterNumNoChange > 0. Training stops when loss > best_loss – tolerance for IterNumNoChange times. Must be a non-negative value.

Input Ports

Icon
Connection to a Teradata Database Instance
Icon
Specifies the table containing the input data.
Icon
Specifies the name of the attribute table.
Icon
Specifies the name of the parameter table.

Output Ports

Icon
output of TD_GLMPerSegment

Nodes

Extensions

Links