Data Quality Check
Imported airline stock data (AAL, DAL, LUV, UAL), jet fuel prices, and NYSE ARCA Index
Conducted data quality check using Statistics node
Focused on identifying missing values and unique values
Evaluated data completeness and consistency before preprocessing
Data Preprocessing Overview
Stock and ARCA Prices
Cleaned dataset by removing duplicates and converting variables into numeric format
Standardized and cleaned datasets across consistent weekly base intervals
Transformed price data into return-based metrics
Constructed volume-weighted average price and return relative to the market benchmark
Generated lagged variables to capture delayed market effects
Fuel Price
Cleaned dataset by removing duplicates and converting variables into numeric format
Transforming raw fuel price data into structured, lagged return-based features for time-series regression.
Then, do the -1,-2,-3,-4 lag accordingly ⟶ capture delayed market reactions
Regression Overview
Steps
Separated models into Linear and Exponential regression frameworks
Linear model captures direct (additive) relationships
Exponential model captures non-linear (multiplicative) effects
Applied multiple lag structures (Lag 1–4) to test delayed impact
Compared model performance to identify the best predictive specification
Components
Each regression model is applied across three time periods
Full sample: 2019–2025
Sub-period Chow test analysis: Pre-war and Post-war (2022-8 break)
Allows comparison of model behavior under different market conditions
To use this workflow in KNIME, download it from the below URL and open it in KNIME:
Download WorkflowDeploy, schedule, execute, and monitor your KNIME workflows locally, in the cloud or on-premises – with our brand new NodePit Runner.
Try NodePit Runner!