Files
Tension-Board-2-Analysis/data/05_predictive_modelling/model_summary.txt
Pawel Sarkowicz 1530c02961 fixed leakage
2026-03-28 12:19:09 -04:00

36 lines
1.8 KiB
Plaintext
Raw Blame History

This file contains ambiguous Unicode characters
This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.
### Model Performance Summary
| Model | MAE | RMSE | R² | Within ±1 | Within ±2 | Exact V | Within ±1 V |
|-------|-----|------|----|-----------|-----------|---------|-------------|
| Linear Regression | 2.191 | 2.742 | 0.537 | 28.1% | 53.1% | 23.9% | 61.3% |
| Ridge Regression | 2.191 | 2.742 | 0.537 | 28.1% | 53.1% | 23.9% | 61.3% |
| Lasso Regression | 2.192 | 2.741 | 0.538 | 27.9% | 53.1% | 23.8% | 61.3% |
| Random Forest (Tuned) | 1.788 | 2.293 | 0.676 | 36.1% | 64.3% | 30.2% | 70.8% |
### Key Findings
1. **Tree-based models remain strongest on this structured feature set.**
- Random Forest (Tuned) achieves the best overall balance of MAE, RMSE, and grouped V-grade performance.
- Linear models remain useful baselines but leave clear nonlinear signal unexplained.
2. **Fine-grained difficulty prediction is meaningfully harder than grouped grade prediction.**
- On the held-out test set, the best model is within ±1 fine-grained difficulty score 36.1% of the time.
- The same model is within ±1 grouped V-grade 70.8% of the time.
3. **This gap is expected and informative.**
- Small numeric errors often stay inside the same or adjacent V-grade buckets.
- The model captures broad difficulty bands more reliably than exact score distinctions.
4. **The projects main predictive takeaway is practical rather than perfect.**
- The models are not exact grade replicators.
- They are reasonably strong at placing climbs into the correct neighborhood of difficulty.
### Portfolio Interpretation
From a modelling perspective, this project shows:
- feature engineering grounded in domain structure,
- comparison of linear and nonlinear models,
- honest evaluation on a held-out test set,
- and the ability to translate raw regression performance into climbing-relevant grouped V-grade metrics.