Titles & Abstracts

Plenary Talks

Volatility estimation from a view point of entropy

Jiro Akahori
Ritsumeikan University
Plenary

We first revisit the volatility estimation approach proposed by N. Kunitomo and S. Sato, and second, we show that the volatility estimator proposed by P. Malliavin and M.E. Mancino can be understood in a unified way by the approach. The point of view connects the HFS to the framework of general statistical learning theory,

Dynamic Asset Pricing with α-MEU Model

Xuedong He
Chinese University of Hong Kong (CUHK)
Plenary

We study a dynamic asset pricing problem in which a representative agent is ambiguous about the aggregate endowment growth rate and trades a risky stock, human capital, and a risk-free asset to maximize her preference value of consumption represented by the α-maxmin expected utility model. This preference model is known to be dynamically inconsistent, so we consider intra-personal equilibrium strategies for the representative agent and define the market equilibrium as the one in which the strategy that clears the market is an intrapersonal equilibrium. We prove the existence and uniqueness of the market equilibrium and show that the asset prices in the equilibrium are the same as in the case when the agent does not perceive any ambiguity but believes in a particular probabilistic model of the endowment process. We show that with reasonable parameter values, the more ambiguity the agent perceives or the more ambiguity-averse she is, the lower the risk-free rate, the higher the stock price, the higher the stock risk premium, and the lower the stock volatility.

Can Dynamic Portfolio Theory Be Practically Implemented?

Hyeng Keun Koo
Ajou University
Plenary

The dynamic portfolio theory, despite its theoretical elegance and academic contributions, has had limited impact on real-world investment practice. To better understand this gap, I will review the historical development of the theory and critically evaluate Cochrane’s assessment. I will discuss the key advancements needed to make implementation feasible—including solutions to the general investment problem, empirical identification of intertemporal hedging components, and the design of sustainable payout policies. I will then briefly present my previous and ongoing work with coauthors that contributes to addressing these implementation challenges.

Invited Talks

A geometric model of synthetic filtrations

Takanori Adachi
Tokyo Metropolitan University (TMU)

The past is usually perceived as a single path leading to the present while the future is perceived as extending from the present. In this study, contrary to the above intuition, we take the position that the present is formed by the synthesis of multiple pasts. To construct such a model, we define the category $\Sigma$, which is an extension of the simplex category with context-dependent time, representing the time domain of generalized filtration. We then formalize the homology derived from this $\Sigma$-filtration as a tool for studying it. Finally, we propose the Dirichlet filtration as a more concrete example of $\Sigma$-filtration and examine how Bayesian statistical methods can be expressed.

Spike variational reinforcement learning of equilibrium mean-variance investment strategy

Mei Choi Chiu
Education University of Hong Kong (EdUHK)

Most reinforcement learning (RL) algorithms are based on the dynamic programming principle (DPP). Related algorithms, such as Q-learning, often involve examining the reward function during the training procedure. However, time inconsistency, prevalent in many finance problems, such as the mean-variance (MV) investment problem, violates the DPP and thus introduces challenges to continuous-time RL. To address time inconsistency, one approach uses the Nash subgame perfect equilibrium among the current and all future selves of an investor. Consequently, the task transforms from optimizing the reward function to finding the equilibrium. The spike variation technique is crucial in solving stochastic control problems in continuous time without assuming DPP. In this paper, we propose an open-loop spike variational RL for equilibrium MV investment subject to the Shannon entropy regularizer. We provide sufficient conditions to characterize the exploratory policy. Under these conditions, we introduce the concept of spike variational RL and show that the corresponding Gaussian exploratory policy is unique. Our results feature explicit solutions for exploratory time-consistent MV investment problems for both constant and state-dependent risk aversions. Numerical experiments indicate that our RL algorithm performs comparably to existing methods for the constant risk aversion case as well as the state-dependent risk aversion case.

Unified asymptotics for investment under illiquidity: transaction costs and search frictions

Jin Hyuk Choi
Ulsan National Institute of Science and Technology (UNIST)

We investigate the optimal investment problem in a market with two types of illiquidity: transaction costs and search frictions. We analyze a power-utility maximization problem where an investor encounters proportional transaction costs and trades only when a Poisson process triggers trading opportunities. We show that the optimal trading strategy is described by a no-trade region. We introduce a novel asymptotic framework applicable when both transaction costs and search frictions are small. Using this framework, we derive explicit asymptotics for the no-trade region and the value function along a specific parametric curve. This approach unifies existing asymptotic results for models dealing exclusively with either transaction costs or search frictions.

Data-Driven Merton’s Strategies via Policy Randomization

Yuchao Dong
Tongji University

We study Merton’s expected utility maximization problem in an incomplete market, characterized by a factor process in addition to the stock price process, where all the model primitives are unknown. The agent under consideration is a price taker who has access only to the stock and factor value processes and the instantaneous volatility.We propose an auxiliary problem in which the agent can invoke policy randomization according to a specific class of Gaussian distributions, and prove that the mean of its optimal Gaussian policy solves the original Merton problem. With randomized policies, we are in the realm of continuous-time reinforcement learning (RL) recently developed in Wang et al. (2020) and Jia and Zhou (2022a,b, 2023), enabling us to solve the auxiliary problem in a data-driven way without having to estimate the model primitives. Specifically, we establish a policy improvement theorem based on which we design both online and offline actor–critic RL algorithms for learning Merton’s strategies. A key insight from this study is that RL in general and policy randomization in particular are useful beyond the purpose for exploration – they can be employed as a technical tool to solve a problem that cannot be otherwise solved by mere deterministic policies. At last, we carry out both simulation and empirical studies in a stochastic volatility environment to demonstrate the decisive outperformance of the devised RL algorithms in comparison to the conventional model-based, plug-in method.

Optimal consumption and investment problem with delay under partial information

Hiroaki Hata
Hitotsubashi University

We consider an optimal consumption and investment problem with delay under a linear Gaussian stochastic factor model. A linear Gaussian stochastic factor model is a stochastic factor model in which the mean returns of risky assets depend linearly on underlying economic factors that are formulated as the solutions of linear stochastic differential equations. We consider the performance-related capital inflow/outflow, which implies that the wealth process is modeled by a stochastic differential delay equation. We also treat the partial information case where the investor cannot observe the factor process and can use only past information about risky assets. Under this setting, the investor tries to maximize the finite horizon discounted expected HARA utility of consumption, the terminal wealth, and the average wealth. A pair of forward-backward stochastic differential equations derived via the stochastic maximum principle have an explicit solution that can be obtained by solving a time-inhomogeneous Riccati differential equation. Thus, the optimal strategy and the optimal value can be obtained explicitly.

Monte Carlo Markov Processes

Chii-Ruey Hwang
Academia Sinica

When the underlying probability (density) is known up to a norming constant only, the usual direct (i.i.d.) sampling is not feasible. The dynamic Monte Carlo method (MCMC) is adopted instead. Mathematically it is to study asymptotic properties of Markov processes/chains with the same equilibrium under various criteria. One of the findings is ‘the non-reversibility performs better’.

Advancing Yield Curve Forecasting: Deep Learning Nelson-Siegel Models and Macroeconomic Insights

Bong-Gyu Jang
POSTECH

This study empirically examines whether incorporating macroeconomic information into a Deep Learning Nelson–Siegel (DLNS) framework improves yield‑curve forecasts, using monthly U.S. Treasury data. The benchmark DLNS model and its macro augmented variant, DLNS‑X, both outperform the traditional dynamic Nelson–Siegel (DNS) model across all forecast horizons. Integrating CPI inflation and M2 money supply into the DLNS‑X model further enhances medium‑term and long‑term predictive accuracy during periods of elevated short‑term interest rates, indicating that these variables convey incremental information beyond that contained in the yield curve. In contrast, the performance gap narrows in certain economic regimes, lending support to the macro‑spanning hypothesis that key macroeconomic news is already embedded in the curve. These findings confirm that deep‑learning term‑structure models can capture complex nonlinear dynamics while preserving economic interpretability, identify the conditions under which macro factors most improve forecast performance, and establish a foundation for the development of more refined forecasting methodologies.

Pontryagin-Guided Deep Learning for Large-Scale Constrained Dynamic Portfolio Choice

Jaegi Jeon
Chonnam National University

We present a Pontryagin-Guided Direct Policy Optimization (PG-DPO) method for constrained dynamic portfolio choice - incorporating consumption and multi-asset investment - that scales to thousands of risky assets. By combining neural-network controls with Pontryagin's Maximum Principle (PMP), it circumvents the curse of dimensionality that renders dynamic programming (DP) grids intractable beyond a handful of assets. Unlike value-based PDE or BSDE approaches, PG-DPO enforces PMP conditions at each gradient step, naturally accommodating no-short-selling or borrowing constraints and optional consumption bounds. A "one-shot" variant rapidly computes Pontryagin-optimal controls after a brief warm-up, leading to substantially higher accuracy than naive baselines. On modern GPUs, near-optimal solutions often emerge within just one or two minutes of training. Numerical experiments confirm that, for up to 1,000 assets, PG-DPO accurately recovers the known closed-form solution in the unconstrained case and remains tractable under constraints -- far exceeding the longstanding DP-based limit of around seven assets.

A local volatility correction to mean-reverting stochastic volatility model for pricing derivatives

Donghyun Kim
Gyeongsang National University

Generally, in the real market, empirical findings suggest that either local volatility (LV) or stochastic volatility (SV) models have a limit to capture the full dynamics and geometry of the implied volatilities of the given equity options. In this study, to overcome the disadvantage of such LV and SV models, we propose a special type of hybrid stochastic-local volatility (SLV∗) model in which the volatility is given by the squared logarithmic function of the underlying asset price added to a function of a fast mean-reverting process. By making use of asymptotic analysis and Mellin transform, we derive analytic pricing formulas for European derivatives with both smooth and non-smooth payoffs under the SLV∗ model. We run numerical experiments to verify the accuracy of the pricing formulas using a Monte-Carlo simulation method and to display that the proposed new model fits the geometry of the market implied volatility more closely than other models such as the Heston model, the stochastic elasticity of variance (SEV) model, the hybrid stochastic and CEV type local volatility (SVCEV) model and the multiscale stochastic volatility (MSV) model, especially for short time-to-maturity options.

Sharp estimates of transition densities of jump processes

Panki Kim
Seoul National University (SNU)

In this talk, we discuss transition densities of  symmetric pure jump Markov processes. Our focus is on the stability of estimates of transition densities when the jumping kernel exhibits mixed polynomial growth. Our results cover the processes on metric measure spaces under volume doubling conditions.

Reinforcement learning for jump-diffusions with financial applications

Lingfei Li
Chinese University of Hong Kong (CUHK)

We study continuous-time reinforcement learning (RL) for stochastic control of systems governed by jump-diffusion processes. We formulate an entropy- regularized exploratory control problem with stochastic policies to capture the exploration--exploitation balance essential for RL. Unlike the pure diffusion case initially studied by Wang et al. (2020), the exploratory dynamics under jump- diffusions calls for a careful formulation of the jump part. Through theoretical analysis, we find that one can simply use the same policy evaluation and q-learning algorithms in Jia and Zhou (2022, 2023), originally developed for controlled diffusions, without needing to check a priori whether the underlying data comes from a pure diffusion or a jump-diffusion. However, we show that the presence of jumps ought to affect parameterizations of actors and critics in general. We investigate as an application the mean--variance portfolio selection problem with stock price modelled as a jump-diffusion and show that both RL algorithms and parameterizations are invariant with respect to jumps. We also present a detailed study on applying the general theory to mean--variance hedging of options. In both applications, we show that RL can achieve better empirical results than the classical plug-in approach using model estimates.

Linear-Quadratic Stackelberg Games for Jump-Diffusion Systems

Jun Moon
Hanyang University

In this talk, we study linear-quadratic leader-follower Stackelberg differen- tial games for jump-diffusion stochastic differential equations. In Stackelberg games, the leader holds a dominating position; the leader chooses, and then, announces her/his optimal strategy by considering the rational behavior of the follower. The follower then chooses her/his optimal solution based on the choice of the leader. This describes a sequential and hierarchical optimization problem between the leader and the follower, which arises in various applications such as finance, engineering, and science. First, we study the case of complete information, where both the leader and the follower have access to the complete filtration F generated by the underlying stochastic processes. Then we study the case of asymmetric information, where the leader and the follower have access to partial filtrations G1 ⊂ F and G2 ⊂ F, respectively, with G2 ⊂ G1 reflecting the asymmetric information structure. In both problems, by the stochastic maximum principle as well as the four-step scheme, we obtain the explicit feedback-type Stackelberg equilibrium for the leader and the follower.

Backward Stochastic Volterra Integral Equations with General Nonlinearities

Chi Seng Pun
Nanyang Technological University (NTU Singapore)

We establish existence, uniqueness, and regularity results for multi-dimensional backward stochastic Volterra integral equations (BSVIEs) with generators that may be random and exhibit nonlinear dependence on the solution, its martingale integrand, and their diagonal processes. Our approach leverages Malliavin calculus, offering a novel framework to address the analytical challenges posed by diagonal terms, in contrast to traditional techniques. We further derive a probabilistic representation for classical solutions to corresponding semi-linear partial differential equations via adapted solutions of BSVIEs. As an application, we study the dynamically optimal mean-variance portfolios under stochastic market models, where the myopic and intertemporal hedging demands are characterized by the diagonal components of some BSVIE solutions.

Some studies of Merton’s Portfolio Optimization Problems

Shuenn-Jyi Sheu
National Central University (NCU)

I will give a brief review of my personal experience on the studies of continuous time Merton’s portfolio optimization problems. These include risk-sensitive portfolio optimization problems, downside risk probabilities minimization problems, and optimal consumption problems. For the study of the risk-sensitive portfolio optimization problems, we follow the ideas of Fleming (1995), in IMA Vol. He reformulates the risk-sensitive portfolio optimization problem as a risk-sensitive stochastic control problem. The dynamic programming principle is used to derive the HJB (Hamilton-Jacobi-Bellman) equation associated to the risk-sensitive stochastic control problem. This HJB equation is a nonlinear elliptic partial differential equation. An eigenvalue equation (a linear equation) is a particular case after a suitable transformation (the logarithmic transformation used by Fleming (1977)). Hence we may call this HJB equation a nonlinear eigenvalue problem. The structure of such nonlinear eigenvalue problems is given in Kaise-Sheu ( 2006). A class of downside risk probabilities optimization problem is studied in Hata-Nagai-Sheu (2010). This problem can be considered as a large time large deviation problems with control process. By a duality argument, these two problems are closely related to each other. Following this observation, the idea of changing probability measure used in large deviation theory can be applied to solve the problem in Hata-Nagai-Sheu (2010) for downside-risk probabilities minimization problem. How to choose the correct measure change is derived from the solution of the risk-sensitive portfolio optimization problem. In Hata-Sheu (2012), an optimal consumption problem in infinite time horizon is considered, and in Hata-Nagai-Sheu (2018) the optimal consumption problem in finite time is considered. For the infinite time problem, the HJB equation associated to the problem is different from the HJB equations from the one for risk-sensitive portfolio optimization problem. Therefore, a different analysis is developed to solve the problem. The studies are inspired by Fleming-Pang (2004). Another study follows from the observation that HJB equation can be rewritten as an inf-sup type Isaacs equation, suggested by the idea of market completion. This indicates a stochastic game interpretation of the portfolio optimization problem, and the saddle point searching for the solution. Eventually it leads to an iteration scheme that can find an approximation of the solution with exponential rate of convergence. This is shown in Fleming-Nagai-Sheu (2019) for risk seeking cases. However, it is still open for risk averse cases and it poses a question whether if the method can be applied to derive a similar result. Another approach by considering directly the HJB equation is used in Hata-Yasuda (2025) to derive another iteration scheme that can find an approximation of the solution with exponential rate of convergence. They prove the result only for the risk averse cases, but it is still open for risk seeking cases.

Partial Information in a Mean-Variance Portfolio Selection Game

Li-Hsien Sun
National Central University (NCU)

We consider finitely many investors who perform mean-variance portfolio selection under a relative performance criterion. That is, each investor is concerned about not only her terminal wealth, but how it compares to the average terminal wealth of all investors (i.e., the mean field). We derive such a Nash equilibrium explicitly in the idealized case of full information (i.e., the dynamics of the underlying stock is perfectly known), and semi-explicitly in the realistic case of partial information (i.e., the stock evolution is observed, but the expected return of the stock is not precisely known). The formula under partial information involves an additional state process that serves to filter the true state of the expected return. We comment on the effect of partial information through numerical analysis. Observe that partial information alone can reduce investors’ wealth significantly, thereby causing or aggravating systemic risk.

The Dynamics of Takeovers through Exchange Offers in the Presence of Competition

Teruyoshi Suzuki
Hokkaido University

This study examines the characteristics of takeovers in the presence of competition using a three-stage model. We first investigate the property of the equilibrium of mergers and takeovers in a frictionless market, and then analyze the effect that the existence of the competitors had on this process. We apply a general surplus function, rather than a specific one, eliminating the modifying effects of this factor from our study. Our model predicts that the existence of heavy competition in the takeover increases the number of unsuccessful or incomplete deals. Furthermore, we find that the shareholders of the target in a competitive market can choose the timing of accepting an offer, without the need to observe the surplus benefit of the deal. Our model shows that the presence of competition does not always provide the target shareholders with an advantage.

Design of Decentralized Exchanges under Concentrated Liquidity

Kohta Takehara
Tokyo Metropolitan University (TMU)

Decentralized exchanges (DEXs) play a pivotal role in the decentralized finance (DeFi) ecosystem. These platforms typically utilize automated market makers (AMMs), which are algorithms designed to pool liquidity and offer it to DEX users at automatically determined prices. In recent developments, the concept of concentrated liquidity has emerged as a groundbreaking innovation, enabling more efficient utilization of liquidity based on the views of liquidity providers regarding transactions within the DEX. This talk presents how to design DEXs under the concentrated liquidity, with a particular emphasis on fee structures and the allocation of fees among participants, liquidity providers and managers of the DEX. The insights provided can be helpful for the design of DEXs, especially under the upcoming upgrade to Uniswap.

Asset Prices and Sustainable Investing under Double Materiality

Tomonori Uchiyama
Tokyo Metropolitan University (TMU)

This study examines portfolio choice, expected returns, and environmental impact within a double materiality framework for sustainable investing. This framework accounts for both the impact on the environment (inside-out) and the impact of the environment (outside-in). The optimal portfolio comprises components that reflect investors’ environmental preferences and hedge environmental risks, alongside the conventional mean-variance component. In equilibrium, green assets (i.e., those generating relatively positive environmental externalities) and resilient assets (i.e., those less exposed to environmental risks) tend to exhibit lower expected returns. As investors place greater emphasis on environmental externalities, the cost of capital for green firms declines, thereby internalizing these externalities and encouraging sustainable practices by firms. However, due to the interaction between environmental externalities and environmental risks, this beneficial effect may be offset—partially or fully—for green firms that are vulnerable to environmental risks.

DeepMartingale: Duality of the optimal stopping problem with expressivity

Hoi Ying Wong
Chinese University of Hong Kong (CUHK)

Using a martingale representation, we introduce a novel deep-learning approach, which we call DeepMartingale, to study the duality of discrete-monitoring optimal stopping problems in continuous time. This approach provides a tight upper bound for the primal value function, even in high-dimensional settings. We prove that the upper bound derived from DeepMartingale converges under very mild assumptions. Even more importantly, we establish the expressivity of DeepMartingale: it approximates the true value function within any prescribed accuracy $\varepsilon$ under our architectural design of neural networks whose size is bounded by $\tilde{c}\,D^{\tilde{q}}\varepsilon^{-\tilde{r}}$, where the constants $\tilde{c}, \tilde{q}, \tilde{r}$ are independent of the dimension $D$ and the accuracy $\varepsilon$. This guarantees that DeepMartingale does not suffer from the curse of dimensionality. Numerical experiments demonstrate the practical effectiveness of DeepMartingale, confirming its convergence, expressivity, and stability.

Optimal Acquisition Contract Design under Asymmetric Information: The Role of Earnouts and Separating Menus

Kyoko Yagi
Tokyo Metropolitan University (TMU)

This paper analyzes the optimal design of acquisition contracts under asymmetric information and decomposes the resulting information rents. We consider three contractual forms: lump-sum acquisition, earnout contracts, and a separating menu offering both options. Under a lump-sum contract, low-type targets may mimic high-type ones to secure higher prices, generating mimicking rents, while high-type targets may reject the offer, leading to adverse selection. Earnout contracts can fully prevent adverse selection by ensuring high-type participation, eliminating mimicking rents but incurring higher contracting costs. A separating menu induces self-selection, mitigating adverse selection and extracting information rent from low-type targets. While more efficient than earnouts alone, separating menus may be less efficient than lump-sum contracts when earnout costs are sufficiently high, making it optimal in some cases to tolerate mimicking rents.

Risk-Sensitive Reinforcement Learning with Information Costs

Tingjin Yan
East China Normal University

We study a provably optimal exploration problem in risk-sensitive reinforcement learning (RL) with information acquisition costs. The agent can pay for signals revealing information about future states and act based on them in a finite-horizon Markov decision process (MDP). We investigate a general objective class named optimized certainty equivalence, which includes popular risk measures such as conditional value-at-risk, variance, and entropic risk. To handle both signal acquisition and action selection, we introduce a unified probabilistic decision policy and characterize the policy space under both constrained and unconstrained signal acquisition settings. We propose a novel bonus-driven value iteration algorithm for tabular MDPs and establish its regret bounds.

One-Step Survival Method for Barrier Option Price and Greeks Calculation under Heston Model

Kazuhiro Yasuda
Hosei University

In this presentation, we consider numerical computations of barrier option price and Greeks under the Heston model. We propose a one-step survival method for the Heston model, and through numerical experiments, demonstrate that this approach effectively reduces simulation variance.

An analytic pricing formula for timer options under constant elasticity of variance with stochastic volatility

Ji-Hun Yoon
Pusan National University (PNU)

Timer options, which were first introduced by Soci´et´ e G´en´ erale Corporate and Investment Banking in 2007, are financial securities whose payo s and exercise are determined by a random time associated with the accumulated realized variance of the underlying asset, unlike vanilla options exercised at the prescribed maturity date. In this paper, taking account of the correlation between the underlying asset price and volatility, we investigate the pricing of timer options under the constant elasticity of variance (CEV) model, proposed by Cox and Ross [10], taking advantage of the approach of asymptotic analysis. Additionally, we validate the pricing precision of the approximate formula for timer options using the Monte Carlo method. We conduct numerical experiments based on our corrected prices and analyze price sensitivities concerning various model parameters, with a focus on the value of elasticity.

Dynamic skew-t copula modelling for multivariate asset price movement

Toshinao Yoshiba
Tokyo Metropolitan University (TMU)

We summarize modelling and estimating two types of dynamic skew-t copulas. The first one is for DGHSktC (dynamic generalized hyperbolic skew-t copula) proposed by Ito and Nakamura (2019). The second one is for DACSktC (dynamic Azzalini-Capitanio skew-t copula) proposed by Ito and Yoshiba (2025). After summarizing the methods, we conduct empirical analysis using the fifteen-years weekly stock returns for two groups composed by three sectors from TOPIX 33 sectors. In results, we show the significance of the tail dependence and asymmetry. We also show that dynamic modelling for correlation matrix is more effective than static one in terms of information criteria.

Randomized multi-factor approximation for stochastic volterra equations with fractional kernel

Tomooki Yuasa
Tokyo Metropolitan University (TMU)

In this talk, we introduce a new simulation method for rough volatility models. In recent years, rough volatility models have attracted significant attention for modeling underlying asset prices. Unlike traditional volatility processes, which are modeled as solutions to stochastic differential equations (SDEs) driven by Brownian motion, the volatility processes in rough volatility models follow solutions to stochastic Volterra equations (SVEs) with the fractional kernel (i.e., driven by fractional Brownian motion). Since fractional Brownian motion lacks the Markov property, rough volatility models also exhibit non-Markovian behavior, which is one of the main challenges in this research area. Abi Jaber and El Euch (2019) proposed a method to approximate SVEs with systems of SDEs using the Laplace transform representation of the fractional kernel and numerical quadrature. This method, known as the Markovian approximation (or the multi-factor approximation), enables simulation techniques---previously difficult or inefficient to apply due to the lack of Markovianity at the SVE level---to be applied more effectively to the approximating Markovian SDE system. While all previous studies have used deterministic approaches for the integral approximation, the novelty of this work lies in incorporating a probabilistic approach (e.g., the Monte Carlo method). It has been shown that the probabilistic Markovian approximation developed in this study achieves higher accuracy than all existing deterministic Markovian approximations, except for the method of Bayer and Breneis (2023). Moreover, under certain conditions, our method has been observed to outperform even that of Bayer and Breneis (2023).

Characterization of the optimal solutions to the Rochet-Choné’s 2D model using duality

Kelvin Shuangjian Zhang
Fudan University

The principal-agent problem is one of the central problems in microeconomics with many applications. Existence, uniqueness, convexity/concavity, regularity, and characterization of the solutions have been widely studied after Mirrlees and Spence in the 1970s. For multidimensional spaces of agents and products, Rochet and Choné [1] reformulated this problem to a concave maximization over the set of convex functions, by assuming agent preferences combine bilinearity in the product and agent parameters with a quasilinear sensitivity to prices. We characterize solutions to this problem by identifying a dual minimization problem. This duality allows us to reduce the solution of the square example of Rochet-Choné to a novel free boundary problem, giving the first analytical description of an overlooked market segment, where the regularity built by Caffarelli-Lions plays a crucial role --- an extension of their regularity work to the quasilinear case is also recently studied. The results profoundly connect with the Optimal Transport theory, a powerful tool with potential applications in many areas.