Icon

Simple cross validation

<p>This workflow is designed to illustrate how to estimate test error using cross validation.</p><p>The workflow reads in the dataset, splits it into training and test sets multiple times (cross-validation), trains a <strong>polynomial regression model</strong> on each training set, predicts outcomes on the corresponding test set, and then evaluates prediction accuracy. By repeating this process for different model complexities, it helps identify which model best balances <strong>fit</strong> and <strong>generalization</strong>, reducing the risk of overfitting.</p>

This workflow is designed to illustrate how to estimate test error using cross validation.

The workflow reads in the dataset, splits it into training and test sets multiple times (cross-validation), trains a polynomial regression model on each training set, predicts outcomes on the corresponding test set, and then evaluates prediction accuracy. By repeating this process for different model complexities, it helps identify which model best balances fit and generalization, reducing the risk of overfitting.

Cross-Validation for Model Selection

This sequence reads in the dataset, splits it into training and test sets multiple times (cross-validation), trains a polynomial regression model on each training set, predicts outcomes on the corresponding test set, and then evaluates prediction accuracy. By repeating this process for different model complexities, it helps identify which model best balances fit and generalization, reducing the risk of overfitting.

Numeric Scorer
Polynomial Regression Learner
Read data. Target generated withg[x_] := 200 + 1000 x - 80 x^2 + x^3;plus noise
CSV Reader
X-Partitioner
X-Aggregator
Regression Predictor
Column Renamer

Nodes

Extensions

Links