Skip to content

Commit c91da2e

Browse files
committed
Create sde.tex
1 parent beec890 commit c91da2e

File tree

1 file changed

+274
-0
lines changed
  • doc/src/Projects/2026/Project1

1 file changed

+274
-0
lines changed
Lines changed: 274 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,274 @@
1+
\documentclass[11pt,a4paper]{article}
2+
\usepackage{amsmath,amssymb,physics}
3+
\usepackage{graphicx}
4+
\usepackage{hyperref}
5+
\usepackage{geometry}
6+
\usepackage{enumitem}
7+
\geometry{margin=2.5cm}
8+
9+
\title{\textbf{Methodological Roadmap for Machine Learning-Based Inference of Lévy Jump-Diffusion Processes}}
10+
\author{}
11+
\date{}
12+
13+
\begin{document}
14+
\maketitle
15+
16+
\section*{1. Scientific Objective}
17+
18+
The objective of this project is to develop and compare machine learning methodologies for estimating the parameters and structure of Lévy jump-diffusion processes underlying financial asset prices.
19+
20+
We consider stochastic processes of the form
21+
\begin{equation}
22+
dS_t = \mu S_t dt + \sigma S_t dW_t + S_{t^-} dJ_t,
23+
\end{equation}
24+
where
25+
\begin{itemize}
26+
\item $\mu$ is the drift,
27+
\item $\sigma^2$ is diffusion variance,
28+
\item $W_t$ is Brownian motion,
29+
\item $J_t$ is a compound Poisson jump process with intensity $\lambda$ and jump distribution $\nu(dx)$.
30+
\end{itemize}
31+
32+
The parameter vector is
33+
\[
34+
\theta = (\mu, \sigma^2, \lambda, \nu).
35+
\]
36+
37+
The task is to infer $\theta$ (or approximate the underlying distribution) from discrete-time observations of $S_t$.
38+
39+
\section*{2. Mathematical Framework}
40+
41+
\subsection*{2.1 Kolmogorov Forward Equation}
42+
43+
The probability density $p(x,t)$ satisfies a partial integro-differential equation (PIDE):
44+
45+
\begin{equation}
46+
\partial_t p = -\mu \partial_x p + \frac{\sigma^2}{2}\partial_x^2 p
47+
+ \lambda \int_{\mathbb{R}} \left[ p(x-y,t) - p(x,t) \right] \nu(dy).
48+
\end{equation}
49+
50+
This provides a deterministic constraint that can be embedded in physics-informed neural networks.
51+
52+
\subsection*{2.2 Inverse Problem}
53+
54+
Given data $\mathcal{D} = \{S_{t_i}\}_{i=1}^N$, estimate:
55+
56+
\[
57+
\theta^* = \arg\max_\theta p(\mathcal{D} \mid \theta),
58+
\]
59+
60+
or approximate the posterior distribution:
61+
62+
\[
63+
p(\theta \mid \mathcal{D}).
64+
\]
65+
66+
\section*{3. Methodological Architecture}
67+
68+
The project is structured in five stages.
69+
70+
\section*{Stage I: Synthetic Data Generation}
71+
72+
\begin{itemize}
73+
\item Implement simulation of Lévy processes.
74+
\item Generate datasets across parameter regimes.
75+
\item Study identifiability and sensitivity.
76+
\item Validate statistical estimators (MLE, method of moments).
77+
\end{itemize}
78+
79+
Deliverable: Baseline statistical inference benchmark.
80+
81+
\section*{Stage II: Supervised Neural Network Estimation}
82+
83+
\subsection*{2A. Direct Parameter Regression}
84+
85+
Train networks:
86+
\[
87+
\text{Time Series} \rightarrow (\mu, \sigma^2, \lambda, \text{jump parameters})
88+
\]
89+
90+
Architectures:
91+
\begin{itemize}
92+
\item MLP (baseline)
93+
\item 1D CNN
94+
\item LSTM / GRU
95+
\item Transformer encoder
96+
\end{itemize}
97+
98+
Loss function:
99+
\[
100+
\mathcal{L} = \sum_i \norm{\hat{\theta}_i - \theta_i}^2.
101+
\]
102+
103+
Evaluation:
104+
\begin{itemize}
105+
\item Parameter estimation error
106+
\item Sensitivity to sampling frequency
107+
\item Robustness under model misspecification
108+
\end{itemize}
109+
110+
\section*{Stage III: Physics-Informed Neural Networks (PINNs)}
111+
112+
Instead of regressing parameters directly, approximate $p(x,t)$ by a neural network $p_\phi(x,t)$.
113+
114+
Loss function:
115+
116+
\begin{equation}
117+
\mathcal{L} = \mathcal{L}_{data} + \alpha \mathcal{L}_{PIDE},
118+
\end{equation}
119+
120+
where
121+
122+
\begin{equation}
123+
\mathcal{L}_{PIDE} = \norm{
124+
\partial_t p_\phi
125+
+ \mu \partial_x p_\phi
126+
- \frac{\sigma^2}{2}\partial_x^2 p_\phi
127+
- \lambda \int (p_\phi(x-y)-p_\phi(x)) \nu(dy)
128+
}^2.
129+
\end{equation}
130+
131+
Goals:
132+
\begin{itemize}
133+
\item Enforce physical consistency
134+
\item Improve generalization
135+
\item Study stability of integro-differential operator learning
136+
\end{itemize}
137+
138+
\section*{Stage IV: Bayesian Machine Learning}
139+
140+
\subsection*{4A. Bayesian Neural Networks}
141+
142+
Place priors on weights and infer posterior:
143+
144+
\[
145+
p(\theta \mid \mathcal{D})
146+
\]
147+
148+
via:
149+
\begin{itemize}
150+
\item Variational inference
151+
\item Monte Carlo dropout
152+
\item Hamiltonian Monte Carlo (if feasible)
153+
\end{itemize}
154+
155+
\subsection*{4B. Gaussian Process Hybrid Models}
156+
157+
Model:
158+
\[
159+
X_t = \text{GP diffusion} + \text{Sparse jump process}
160+
\]
161+
162+
Study:
163+
\begin{itemize}
164+
\item Jump detection
165+
\item Volatility decomposition
166+
\item Uncertainty quantification
167+
\end{itemize}
168+
169+
Deliverable: Credible intervals for parameters.
170+
171+
\section*{Stage V: Neural Stochastic Differential Equations}
172+
173+
Generalize to neural SDE framework:
174+
175+
\begin{equation}
176+
dX_t = f_\theta(X_t,t) dt + g_\theta(X_t,t) dW_t + dJ_t.
177+
\end{equation}
178+
179+
Objectives:
180+
\begin{itemize}
181+
\item Learn drift and diffusion functions non-parametrically.
182+
\item Compare structured Lévy assumption vs neural SDE flexibility.
183+
\item Study overfitting vs interpretability trade-offs.
184+
\end{itemize}
185+
186+
\section*{4. Comparative Evaluation Framework}
187+
188+
All methods will be benchmarked against:
189+
190+
\begin{itemize}
191+
\item Maximum likelihood estimation
192+
\item Expectation–maximization methods
193+
\item Classical jump-diffusion calibration
194+
\end{itemize}
195+
196+
Metrics:
197+
198+
\begin{itemize}
199+
\item Parameter RMSE
200+
\item Log-likelihood on unseen data
201+
\item Predictive density calibration
202+
\item Computational cost
203+
\item Uncertainty quantification quality
204+
\end{itemize}
205+
206+
\section*{5. Application to Real Financial Data}
207+
208+
\begin{itemize}
209+
\item Calibrate on synthetic data.
210+
\item Apply to historical equity and FX data.
211+
\item Compare parameter estimates to standard econometric methods.
212+
\item Study stability across market regimes.
213+
\end{itemize}
214+
215+
\section*{6. Risk Analysis}
216+
217+
\begin{itemize}
218+
\item Identifiability issues for $\lambda$ at low frequency.
219+
\item Jump distribution non-uniqueness.
220+
\item Overfitting of neural SDE models.
221+
\item Numerical instability in PINN integro-differential terms.
222+
\end{itemize}
223+
224+
Mitigation:
225+
\begin{itemize}
226+
\item Regularization.
227+
\item Bayesian priors.
228+
\item Cross-validation.
229+
\item Theoretical error bounds.
230+
\end{itemize}
231+
232+
\section*{7. Expected Scientific Contributions}
233+
234+
\begin{itemize}
235+
\item Systematic comparison of ML paradigms for Lévy inference.
236+
\item Demonstration of PINNs for integro-differential stochastic systems.
237+
\item Uncertainty-aware neural parameter estimation.
238+
\item Insights into interpretability vs flexibility trade-offs.
239+
\end{itemize}
240+
241+
\section*{8. Publication Strategy}
242+
243+
Potential venues:
244+
245+
\begin{itemize}
246+
\item Quantitative Finance
247+
\item Journal of Computational Finance
248+
\item SIAM Journal on Financial Mathematics
249+
\item Machine Learning in Finance
250+
\item NeurIPS / ICML workshops (if neural SDE focus)
251+
\end{itemize}
252+
253+
\section*{9. Timeline (12–18 Months)}
254+
255+
\begin{itemize}
256+
\item Months 1–3: Simulation + classical benchmarks.
257+
\item Months 4–6: Supervised neural models.
258+
\item Months 7–9: PINN development.
259+
\item Months 10–12: Bayesian extensions.
260+
\item Months 13–15: Neural SDE models.
261+
\item Months 16–18: Real data application + publication.
262+
\end{itemize}
263+
264+
\section*{Conclusion}
265+
266+
This roadmap provides a structured, progressively sophisticated
267+
approach to applying machine learning to Lévy jump-diffusion
268+
inference. The project balances statistical rigor, physical
269+
constraints, uncertainty quantification, and modern deep learning
270+
architectures. It offers both methodological innovation and practical
271+
financial relevance.
272+
273+
\end{document}
274+

0 commit comments

Comments
 (0)