Concept Graph & Resume using Claude 3.5 Sonnet | Chat GPT4o | Llama 3:
Resume:
1.- Time series definition: sequence of signal observations measured at uniform intervals, with examples from finance, weather, and medicine.
2.- Autoregressive (AR) model: each observation is a noisy linear combination of previous observations, with Gaussian noise. Focus is on AR(p) model.
3.- Motivation: real-life time series often have missing data due to mistakes or equipment issues. Existing models make strict statistical assumptions.
4.- Limitations of existing approaches: assume data follows AR model with Gaussian noise and missing data is random. Misspecification leads to suboptimal results.
5.- Goal: relax or remove statistical assumptions on time series with missing data.
6.- Setting: adversary chooses signal value arbitrarily at each time point and decides whether to reveal it. Player predicts and suffers loss.
7.- Regret criterion: compares player's cumulative loss to loss of best fixed predictor in hindsight. Standard in online learning.
8.- Challenge: AR prediction not well-defined for missing data since previous missing values are needed for prediction.
9.- Solution: replace missing data with predictions of missing data, using recursive AR predictor up to finite number of steps back (D).
10.- Example: For AR(1) with D=2, prediction uses observed value if available, else uses previous prediction recursively.
11.- Non-linearity issue: recursive prediction is non-linear in AR coefficients, preventing learning of optimal coefficients using standard online convex optimization techniques.
12.- Non-proper learning: generate predictions from a richer model and compete against best AR predictor, without directly learning AR coefficients.
13.- Expanded prediction: use weight vector W of dimension 2^D and feature vector phi, giving a much richer linear-in-W prediction.
14.- Advantage of expanded prediction: problem becomes linear in W, enabling online convex optimization. Prediction is richer than original AR.
15.- Main result: regret bound of O(sqrt(T)) w.r.t. best fixed recursive AR predictor in hindsight, using standard online learning algorithms.
16.- Issues: constant factor exponential in D, and time/space complexity also exponential in D due to maintaining/updating high-dimensional W.
17.- Partial solution: efficient inner product computation between feature vectors phi enables avoiding explicit maintenance of W.
18.- Inner product computation: coefficients depend on number of "common good paths" - paths through missing data points.
19.- Efficient path counting: number of common good paths of length k equals 2^(num. common missing points), enabling efficient computation.
20.- Experimental results: comparisons on synthetic data capture merits of different models. Real data experiments in paper.
21.- Stochastic data: all approaches perform similarly when data and missingness are stochastic. Online methods more robust to AR coefficient changes.
22.- Heteroscedastic noise: baselines not robust to non-Gaussian noise, but online approach is robust. Online algorithms faster and simpler to implement.
23.- Conclusions: new approach for time series prediction without requiring stochastic assumptions. Theoretical guarantees and good practical performance.
24.- Open questions: reducing exponential constant to polynomial in D (some progress in follow-up work), and generalizing to multivariate case.
25.- Key aspects: online learning framework, recursive prediction for missing data, non-proper learning, efficient path counting, robustness to model misspecification.
Knowledge Vault built byDavid Vivancos 2024