Abstract:The rise of GenAI amplifies the economic phenomenon of positive spillovers. When creators contribute content that can be reused and adapted by Large Language Models (LLMs), each creator's effort can enhance the content quality of others by enabling easy imitation and recombination of existing content. On the one hand, such spillovers create value for the entire ecosystem; on the other hand, they risk undermining creators' incentives to invest genuine effort, as others may freely benefit from their contributions. To address this problem, we introduce the Content Creation with Spillovers (CCS) model. In our model, each creator chooses an effort level that, together with the efforts of others, determines her content quality. The platform aims to maximize the social welfare of consumers under stable behavior of the creators (pure Nash equilibrium), but can only observe the resulting qualities and not the underlying efforts. Interestingly, simple mechanisms like winner-takes-all and Tullock lead to the non-existence of equilibrium. In response, we propose the parametrized family of Provisional Allocation mechanisms, guaranteeing equilibrium existence and a unique Pareto-dominant equilibrium. While maximizing the social welfare under this family is NP-hard, we develop approximation algorithms that apply to a broad class of spillover structures and provide strong welfare guarantees. Specifically, in the worst-case analysis, we devise efficient algorithms for bounded spillovers and tree-structure spillovers. We also introduce Greedy Cost Selection, a linearithmic time algorithm that achieves approximately optimal results in the average case analysis. Together, our results provide game-theoretic foundations for sustaining human content creation in the era of GenAI.



Abstract:As GenAI platforms grow, their dependence on content from competing providers, combined with access to alternative data sources, creates new challenges for data-sharing decisions. In this paper, we provide a model of data sharing between a content creation firm and a GenAI platform that can also acquire content from third-party experts. The interaction is modeled as a Stackelberg game: the firm first decides how much of its proprietary dataset to share with GenAI, and GenAI subsequently determines how much additional data to acquire from external experts. Their utilities depend on user traffic, monetary transfers, and the cost of acquiring additional data from external experts. We characterize the unique subgame perfect equilibrium of the game and uncover a surprising phenomenon: The firm may be willing to pay GenAI to share the firm's own data, leading to a costly data-sharing equilibrium. We further characterize the set of Pareto improving data prices, and show that such improvements occur only when the firm pays to share data. Finally, we study how the price can be set to optimize different design objectives, such as promoting firm data sharing, expert data acquisition, or a balance of both. Our results shed light on the economic forces shaping data-sharing partnerships in the age of GenAI, and provide guidance for platforms, regulators and policymakers seeking to design effective data exchange mechanisms.




Abstract:The rise of Generative AI (GenAI) has significantly impacted human-based forums like Stack Overflow, which are essential for generating high-quality data. This creates a negative feedback loop, hindering the development of GenAI systems, which rely on such data to provide accurate responses. In this paper, we provide a possible remedy: A novel strategy we call selective response. Selective response implies that GenAI could strategically provide inaccurate (or conservative) responses to queries involving emerging topics and novel technologies, thereby driving users to use human-based forums like Stack Overflow. We show that selective response can potentially have a compounding effect on the data generation process, increasing both GenAI's revenue and user welfare in the long term. From an algorithmic perspective, we propose an approximately optimal approach to maximize GenAI's revenue under social welfare constraints. From a regulatory perspective, we derive sufficient and necessary conditions for selective response to improve welfare improvements.
Abstract:Principal-agent problems arise when one party acts on behalf of another, leading to conflicts of interest. The economic literature has extensively studied principal-agent problems, and recent work has extended this to more complex scenarios such as Markov Decision Processes (MDPs). In this paper, we further explore this line of research by investigating how reward shaping under budget constraints can improve the principal's utility. We study a two-player Stackelberg game where the principal and the agent have different reward functions, and the agent chooses an MDP policy for both players. The principal offers an additional reward to the agent, and the agent picks their policy selfishly to maximize their reward, which is the sum of the original and the offered reward. Our results establish the NP-hardness of the problem and offer polynomial approximation algorithms for two classes of instances: Stochastic trees and deterministic decision processes with a finite horizon.