Discussion 2 – Benchmarking & Forecasting#
This session pivots from dashboards to the theory and tooling behind the Financial Time-Series Forecasting Repository (FTSFR).
Slide Talk: Why a Financial Forecasting Benchmark?#
We’ll walk through the slides in slides_ftsfr.pdf.
Topics include: the motivation for a dedicated finance benchmark, dataset design, train/test protocol, and how the team evaluates competing models.
Bring questions about asset classes, evaluation metrics, or how the benchmark might align with your own research interests.
Repository Walkthrough#
Tour the FTSFR repo structure (
src/
,_data/
,_output/
,datasets.toml
,subscriptions.toml
).Review environment setup,
.env
credentials, and the keydoit
scripts you’ll use (dodo_01_pull.py
,dodo_02_forecasting.py
).Highlight where forecasting jobs drop their predictions and error metrics so you can wire them into the Streamlit apps.
Hands-On Forecasting#
Kick off at least one forecasting job (or use the provided sample outputs if compute is limited).
Inspect the generated CSV/parquet files and verify the schema matches what your dashboard expects.
Capture any TODOs needed to plug the forecasts into
app_01.py
orapp_04_crsp.py
before Discussion 3.
Suggested Next Steps#
Run a full forecasting batch overnight and stash the outputs for integration.
Start weaving forecasts into your dashboard—new tabs, comparison charts, or narrative callouts.
Note any blockers (data access, compute, modeling questions) so we can troubleshoot them in the next session.