Mathematical Finance Seminar – Spring 2023

Schedule for Spring 2023

Seminars are on Thursdays
Time: 4:10pm – 5:25pm
Location: Room 903 SSW

Organizers: Gokce Dayanikli, Ioannis Karatzas, Marcel Nutz, Philip Protter, Johannes Wiesel

1/26/23

 

Emma Hubert (Princeton)

Title: Large-scale principal-agent problems

Abstract: In this talk, we introduce two problems of contract theory, in continuous-time, with a multitude of agents. First, we will study a model of optimal contracting in a hierarchy, which generalises the one-period framework of Sung (2015). The hierarchy is modelled by a series of interlinked principal-agent problems, leading to a sequence of Stackelberg equilibria. More precisely, the principal (she) can contract with a manager (he), to incentivise him to act in her best interest, despite only observing the net benefits of the total hierarchy. The manager in turn subcontracts the agents below him. We will see through a simple example that, while the agents only control the drift of their outcome, the manager controls the volatility of the Agents’ continuation utility. Therefore, even this relatively simple introductory example justifies the use of recent results on optimal contracting for drift and volatility control, and therefore the theory on 2BSDEs.

This will lead us to introduce the second problem, namely optimal contracting for demand- response management, which consists in extending the model by Aid, Possamai, and Touzi (2022)  to a mean-field of consumers. More precisely, the principal (an electricity producer, or provider) contracts with a continuum of agents (the consumers), to incentivise them to decrease the mean and the volatility of their energy consumption during high peak demand. In addition, we introduce a common noise, impacting all consumption processes, to take into account the impact of weather conditions on the agents’ electricity consumption. This mean- field framework with common noise leads us to consider a more extensive class of contracts. In particular, we prove that these results can be improved by indexing the contracts on the consumption of one agent and aggregate consumption statistics from the distribution of the entire population of consumers.

 

2/2/23

 

 

 

Renyuan Xu (USC)

Title: Asymptotic Analysis of Deep Residual Networks and Global Convergence of Gradient Descent Methods

Abstract: Residual networks (ResNets) have displayed impressive results in pattern recognition and, recently, have garnered considerable theoretical interest due to a perceived link with neural ordinary differential equations (neural ODEs). This link relies on the convergence of network weights to a smooth function as the number of layers increases. We investigate the properties of weights trained by stochastic gradient descent and their scaling with network depth through detailed numerical experiments. We observe the existence of scaling regimes markedly different from those assumed in neural ODE literature. Depending on certain features of the network architecture, such as the smoothness of the activation function, we prove the existence of an alternative ODE limit, a stochastic differential equation, or neither of these. For each case, we also derive the limit of the backpropagation dynamics and address its adaptiveness issue. These findings cast doubts on the validity of the neural ODE model as an adequate asymptotic description of deep ResNets and point to an alternative class of differential equations as a better description of the deep network limit.

When the gradient descent method is applied to the training of ResNets, we prove that it converges linearly to a global minimum if the network is sufficiently deep and the initialization is sufficiently small. In addition, the global minimum found by the gradient descent method has finite quadratic variation without using any regularization in the training. This confirms existing empirical results that the gradient descent method enjoys an implicit regularization property and is capable of generalizing to unseen data.

This is based on several papers with Rama Cont (Oxford), Alain Rossier (Oxford), and Alain-Sam Cohen (InstaDeep).

2/9/23

 

Samuel Cohen (Oxford)

Title: Neural Q-learning solutions to elliptic PDEs

Abstract: Solving high-dimensional partial differential equations (PDEs) is a major challenge in scientific computing. We develop a new numerical method for solving elliptic-type PDEs by adapting the Q-learning algorithm in reinforcement learning. Using a neural tangent kernel (NTK) approach, we prove that the neural network approximator for the PDE solution, trained with the Q-PDE algorithm, converges to the trajectory of an infinite-dimensional ordinary differential equation (ODE) as the number of hidden units becomes infinite. For monotone PDE (i.e. those given by monotone operators), despite the lack of a spectral gap in the NTK, we then prove that the limit neural network, which satisfies the infinite-dimensional ODE, converges in $L^2$ to the PDE solution as the training time increases. The numerical performance of the Q-PDE algorithm is studied for several elliptic PDEs. Based on joint work with Deqing Jiang and Justin Sirignano.

2/16/23

 

 

Ulrich Horst (HU Berlin)

Title: Mean-Field Liquidation Games with Market Drop-out

Abstract: We consider a novel class of portfolio liquidation games with market drop-out (“absorption”). More precisely, we consider mean-field and finite player liquidation games where a player drops out of the market once his position hits zero. In particular round-trips are not admissible. This can be viewed as a no statistical arbitrage condition on trading. In a model with only sellers we prove that the absorption condition is equivalent to a short selling constraint. We prove that equilibria (both in the mean-field and the finite player game) are given as solutions to integro differential equations with endogenous terminal conditions. We prove the existence of a unique solution to the differential equation from which we obtain the existence of an equilibrium in the MFG and the existence of a unique equilibrium in the $N$-player game. We establish the convergence of the equilbria in the finite player games to the obtained mean-field equilibrium and illustrate the impact of the drop-out constraint on equilibrium trading rates. 

The talk is based on joint work with Guanxing Fu and Paul Hager.

2/23/23

 

Sergey Nadtochiy (IIT)

Title. Consistency of MLE for partially observed diffusions, with application in market microstructure modeling.

Abstract. In this talk, I will present a tractable sufficient condition for the consistency of maximum likelihood estimators (MLEs) in partially observed diffusion models, stated explicitly via the stationary distribution of the fully observed system. This result is then applied to a model of market microstructure with latent (unobserved) price process, for which the estimation is performed using real market data for liquid NASDAQ stocks. In particular, we obtain an estimate of the price impact coefficient, as well as the micro-level volatility and the drift of the latent price process (the latter is responsible for the concavity of expected price impact of a large meta-order). Joint work with Y. Yin.

3/2/23

 

Melih Iseri (USC)

Title: Set Valued HJB Equations

Abstract:
In this talk, we introduce a notion of set valued PDEs. The set values have been introduced for many applications, such as time inconsistent stochastic optimization problems, multivariate dynamic risk measures, and nonzero sum games with multiple equilibria. One crucial property they enjoy is the dynamic programming principle (DPP). Together with the set valued Itô formula, which is a key component, DPP induces the PDE. In the context of multivariate optimization problems, we introduce the set valued Hamilton-Jacobi-Bellman equations and established its wellposedness. In the standard scalar case, our set valued PDE reduces back to the standard HJB equation. 

Our approach is intrinsically connected to the existing theory of surface evolution equations, where a well-known example is mean curvature flows. Roughly speaking, those equations can be viewed as first order set valued ODEs, and we extend them to second order PDEs. Another difference is that, due to different applications, those equations are forward in time (with initial conditions), while we consider backward equations (with terminal conditions).
 
The talk is based on a joint work with Prof. Jianfeng Zhang.

3/9/23

 

Per Mykland (U Chicago)

Title: Nonparametric Observed Standard Errors for High Frequency Data

Abstract: High frequency financial data has become an essential component of the digital world, giving rise to an increasing number of estimators. However, it is hard to reliably assess the uncertainty of such estimators. The Observed Asymptotic Variance (observed AVAR) is a non-parametric (squared) standard error for high- frequency-based estimators. We have earlier developed such an AVAR with time-discretization and two tuning parameters (per dimension). The current paper shows that these two parameters are confounded, and one can move to a single tuning parameter. This is shown by passing to continuous time (which is natural since observations are usually irregularly spaced). We show that the new time-continuous observed AVAR is a limit of the original observed AVAR. We also obtain a central limit theory for the new time-continuous observed AVAR, and the latter permits a sharper definition of our standard error. The device is related to observed information in likelihood theory, but in this case it is non-parametric and uses the high-frequency data structure. [With Lan Zhang, University of Illinois at Chicago.]

3/16/23

 

No Seminar – Spring Break

3/23/23

 

Dylan Possamai (ETH Zurich)

Title: Moral hazard for time-inconsistent agents, BSVIEs and stochastic targets

Abstract: We address the problem of Moral Hazard in continuous time between a Principal and an Agent that has time-inconsistent preferences. Building upon previous results on non-Markovian time-inconsistent control for sophisticated agents, we are able to reduce the problem of the principal to a novel class of control problems, whose structure is intimately linked to the representation of the problem of the Agent via a so-called extended Backward Stochastic Volterra Integral equation. We will present some results on the characterization of the solution to problem for different specifications of preferences for both the Principal and the Agent, and relate the general setting to control problems with Volterra stochastic target constraints.

3/30/23

Julien Guyon (Ecole des Ponts ParisTech)

Title: Volatility Is (Mostly) Path-Dependent

Abstract: We learn from data that volatility is mostly path-dependent: up to 90% of the variance of the implied volatility of equity indexes is explained endogenously by past index returns, and up to 65% for (noisy estimates of) future daily realized volatility. The path-dependency that we uncover is remarkably simple: a linear combination of a weighted sum of past daily returns and the square root of a weighted sum of past daily squared returns with different time-shifted power-law weights capturing both short and long memory. This simple model, which is homogeneous in volatility, is shown to consistently outperform existing models across equity indexes and train/test sets for both implied and realized volatility. It suggests a simple continuous-time path-dependent volatility (PDV) model that may be fed historical or risk-neutral parameters. The weights can be approximated by superpositions of exponential kernels to produce Markovian models. In particular, we propose a 4-factor Markovian PDV model which captures all the important stylized facts of volatility, produces very realistic price and (rough-like) volatility paths, and jointly fits SPX and VIX smiles remarkably well. We thus show that a continuous-time Markovian parametric stochastic volatility (actually, PDV) model can practically solve the joint SPX/VIX smile calibration problem. This is joint work with Jordan Lekeufack (UC Berkeley).

4/6/23

 

Lukas Wessels (Georgia Tech)

Title: Controlled SPDEs: Peng’s Maximum Principle and Numerical Methods

Abstract: In this talk, we consider a finite-horizon optimal control problem of stochastic reaction-diffusion equations. First we apply the spike variation method which relies on introducing the first and second order adjoint state. We give a novel characterization of the second order adjoint state as the solution to a backward SPDE. Using this representation, we prove the maximum principle for controlled SPDEs.
In the second part, we present a numerical algorithm that allows the efficient approximation of optimal controls in the case of stochastic reaction-diffusion equations with additive noise by first reducing the problem to controls of feedback form and then approximating the feedback function using finitely based approximations. Numerical experiments using artificial neural networks as well as radial basis function networks illustrate the performance of our algorithm.

This talk is based on joint work with Wilhelm Stannat and Alexander Vogler.

4/13/23

Andreas Sojmark (LSE)

Title: Convergence of Stochastic Integrals on Skorokhod Space under Skorokhod’s J1 and M1 topologies

Abstract: In this talk, I will introduce a new framework for weak convergence of stochastic integrals on Skorokhod space, building on the pioneering work of Jakubowski, Memin & Pages (PTRF ’89) and Kurtz & Protter (AOP ’91). This involves two main novelties. Firstly, we will see how one can rely on a simpler notion of good decompositions for the integrators instead of the rather intricate notions of predictable uniform tightness or uniformly controlled variations that have been the focus of all prior works on this theory and its applications. Secondly, I will show how to develop the theory in a way that incorporates both Skorokhod’s J1 and M1 topology, relying in part on some crucial ideas of Jakubowski (AOP ’96), and I will discuss the principal new subtleties that this entails. For the final part of the talk, I will present some applications concerned with limit theorems in insurance, finance, and econometrics for which we are able to provide novel results and clarify some apparent confusion in the literature. The talk is based on joint work with Fabrice Wunderlich (Oxford).

4/20/23

No seminar (Berkeley–Columbia Meeting)

4/27/23

 

Markus Pelger (Stanford)

Title: Deep Learning Statistical Arbitrage

Abstract: Statistical arbitrage exploits temporal price differences between similar assets. We develop a unifying conceptual framework for statistical arbitrage and a novel data driven solution. First, we construct arbitrage portfolios of similar assets as residual portfolios from conditional latent asset pricing factors. Second, we extract their time series signals with a powerful machine- learning time-series solution, a convolutional transformer. Lastly, we use these signals to form an optimal trading policy, that maximizes risk-adjusted returns under constraints. Our com- prehensive empirical study on daily US equities shows a high compensation for arbitrageurs to enforce the law of one price. Our arbitrage strategies obtain consistently high out-of-sample mean returns and Sharpe ratios, and substantially outperform all benchmark approaches.

Link:
https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3862004