1. Bayesian & Probabilistic Approaches
1.1 Bayesian Neural Networks (BNNs) for Time Series
-
Core Idea: Replace deterministic weight parameters with distributions to capture uncertainty in predictions and model parameters.
-
Interpretability:
-
The distributions over parameters and predictions provide “credible intervals,” offering insight into how certain or uncertain the model is at different time steps or forecast horizons.
-
Techniques like Bayes by Backprop and Variational Inference enable posterior approximation for high-dimensional networks.
-
1.2 Deep State-Space & Structural Time Series Models
-
Examples: DeepAR, DeepState, DeepFactor (from Amazon’s forecasting frameworks), which are often extended with Bayesian components.
-
Interpretability:
-
Breakdowns into latent states or factors can be inspected for domain-specific meaning (e.g., seasonality, trend components).
-
Credible intervals around latent states yield an interpretable decomposition of the forecast.
-
1.3 Bayesian Shap (or Probabilistic Shap)
-
Motivation: Classic SHAP can be extended with Bayesian treatments to capture the uncertainty in feature-attribution scores.
-
Key Advantage:
-
You get a distribution over the Shapley values for each time step or segment, rather than just point estimates.
-
2. Advanced Sequence Attribution Methods
2.1 TimeSHAP
-
Overview: An adaptation of the SHAP framework tailored specifically for sequential data (e.g., time series).
-
How It Works: Considers temporal dependencies by removing or “abating” parts of the series and measuring the change in model output.
-
Advantages:
-
More faithful to time-series structure.
-
Works well for RNNs, LSTMs, Transformers, or any black-box model.
-
2.2 RETAIN (Reverse Time Attention for Interpretation)
-
Context: Originally developed for healthcare time series (Electronic Health Records).
-
Mechanism: Uses a two-level attention mechanism—one for feature embeddings and another for time steps (in reverse).
-
Interpretability:
-
We can quantify how much each feature at each time step contributes to a final prediction.
-
This is especially useful for irregular medical time series but generalizable to other domains.
-
2.3 Temporal Occlusion & Perturbation Methods
-
Concept: Systematically occlude (mask) segments of the time series (e.g., a window of days) and observe how the model’s output changes.
-
Benefits:
-
Offers a global view of which intervals the model relies on the most.
-
Extends typical occlusion-based interpretability from images to time series.
-
3. Architectures With Built-In Interpretability
3.1 N-BEATS (Neural Basis Expansion Analysis for Time Series)
-
Core Idea: A sequence-to-sequence model that decomposes the forecast into additive basis functions (trend, seasonality, etc.).
-
Interpretability:
-
Each block has forward and backward “basis expansions,” which can be directly inspected for interpretable components.
-
Offers a “white-box” approach: you can see how each basis contributes to the final prediction.
-
3.2 Temporal Fusion Transformers (TFT)
-
Proposed By: Google (Brian Lim et al.)
-
What’s New: Combines a gating mechanism, static covariate encoders, and multi-head attention.
-
Interpretability:
-
Variable selection networks highlight which input variables matter most at a given time.
-
Attention can be dissected to see how the model weighs past time steps.
-
Provides quantile forecasts with interpretability into how different time steps and features affect each quantile.
-
3.3 Hybrid CNN–Transformer Models
-
Motivation: Address the computational overhead of full self-attention on long sequences by using CNN-based local encoders combined with global attention modules.
-
Interpretability:
-
Local convolutions can be probed via filter activations to see local pattern significance.
-
Attention can highlight which time steps matter globally.
-
4. Counterfactual & Contrastive Explanations
-
Counterfactual Explanations
-
For time series, a counterfactual might involve “changing” or “perturbing” a subset of time steps to see how the prediction shifts.
-
Example approach: If a model predicts a spike in energy consumption, a counterfactual explanation could show which historical segments or features—if altered—would have reduced that spike.
-
-
Contrastive Explanation Methods
-
Provide reasons why a certain prediction was made instead of a plausible alternative.
-
For time series, this can highlight the key intervals or signals that differentiate one predictive outcome from another.
-
5. Surrogate Models & Global Explainability
5.1 Surrogate Modeling (e.g., LIME for Time Series)
-
Concept: Train a simpler (often linear or rule-based) “explanation model” around local regions of the time series to approximate the behavior of a complex deep network.
-
Time-Series Twist:
-
Must preserve temporal dependency in how samples are perturbed.
-
Segment-based or shapelet-based surrogates can be used to better approximate local behavior.
-
5.2 Global Rule Extraction
-
Rule Distillation from RNNs or Transformers can yield high-level logical statements describing how certain patterns in the input lead to particular predictions.
-
While more nascent in time series, research is ongoing to unify sequential rule-mining with deep networks (some methods rely on symbolic metamodels or tree-based surrogates).
6. Things to Consider When Selecting Models
-
Trade-off: Interpretability vs. Accuracy
-
More transparent methods (e.g., N-BEATS or simpler Bayesian structures) can sometimes be outperformed by large black-box models (e.g., huge Transformers).
-
Knowledge distillation or surrogate modeling is a good strategy to balance performance and interpretability.
-
-
Multi-Scale Interpretations
-
Time series often have multi-scale phenomena (seasonal, weekly, daily cycles).
-
Consider hierarchical or multi-resolution models (like Swin-like Transformers for sequences) that let you interpret contributions at each scale.
-
-
Uncertainty Quantification
-
In time series, you often care about predictive intervals.
-
Bayesian or quantile-based approaches (TFT, quantile regression, etc.) can show how the model’s uncertainty changes over time—this can be more critical than a single deterministic forecast.
-
-
Domain-Specific Visualizations
-
Visualization can make or break interpretability. Methods like saliency maps for temporal data, attention heatmaps, or breakdown plots of additive components (like in N-BEATS) are crucial.
-