Sarah Dean: Platforms as Dynamical Systems

This is a research digest. Ritam is thinking about reaching out to Sarah Dean — assistant professor in CS at Cornell, NSF CAREER awardee, AI2050 Fellow — and asked me to map out what she actually works on. So I went through her publications from 2023 to now and organized them by what she thinks about, not just what she publishes.

            The short version: Dean treats recommender systems and digital platforms the way a controls engineer treats a plant. Not as static optimization problems, but as dynamical systems with feedback loops, equilibria, and stability conditions. If you care about how algorithmic choices shape the behavior of the humans who use them — and how those humans reshape the algorithm in return — she is one of the clearest thinkers on this.


            Here is her work, grouped by theme.


            ## The main thing: recommender systems have dynamics


            This is Dean's signature contribution and the thread that ties most of her work together. The core argument is simple and, once you hear it, hard to un-hear: a recommendation algorithm doesn't just serve content. It changes what content gets made, which changes what data the algorithm trains on, which changes the recommendations, which changes the content again. It's a feedback loop. Treating it as a static prediction problem is like modeling a thermostat without modeling the room.


            The foundational paper here is ["Modeling Content Creator Incentives on Algorithm-Curated Platforms"](https://arxiv.org/abs/2206.13102) *(ICLR 2023, oral — top 5%)*. Dean and co-authors introduce an "exposure game" framework: creators compete for algorithmic exposure, and the algorithm's design determines the Nash equilibria of their behavior. The key finding is concrete and surprising. Whether the platform uses non-negative matrix factorization versus unconstrained factorization for its recommendations changes what creators produce at equilibrium. Not slightly — qualitatively. The algorithm's internal representation constrains the creative landscape.


            <div class="highlight">
                The design choice inside the algorithm becomes a policy choice over the ecosystem. This is the idea that runs through everything Dean does.


            </div>

            She's built out from there in several directions:


            ["Accounting for AI and Users Shaping One Another"](https://arxiv.org/abs/2404.12670) *(TMLR 2025)* is a position-paper-style argument for why the field needs formal interaction models. AI systems and users mutually shape each other, and recommender systems are the clearest example. She's not just observing this — she's arguing the ML community needs to build it into their problem formulations from the start.


            ["Policy Design for Two-sided Platforms with Participation Dynamics"](https://arxiv.org/abs/2502.07378) *(ICML 2025)* brings control theory and game theory together for two-sided platforms (think: creators and consumers). The main result: myopic-greedy approaches — optimize for today's metric, retrain, repeat — are provably suboptimal for long-term social welfare. She provides an algorithm that does better by reasoning about the dynamics explicitly.


            ["Harm Mitigation in Recommender Systems under User Preference Dynamics"](https://arxiv.org/abs/2406.09882) *(KDD 2024)* models how recommendations influence harmful content consumption over time. Users don't have fixed preferences — they drift toward what they're shown. The paper develops policies that balance click-through rate against harm reduction, taking this drift into account. This is the engagement-versus-harm tradeoff made formal.


            ["Ranking with Long-Term Constraints"](https://arxiv.org/abs/2406.02125) *(WSDM 2024)* uses control-based algorithms to handle long-term fairness and revenue objectives in ranking. Not just "be fair right now" but "maintain fairness as the system evolves."


            ## Multi-learner competition and market dynamics


            A related thread, but distinct enough to call out: what happens when multiple algorithms compete for users?


            ["Emergent Specialization from Participation Dynamics and Multi-Learner Retraining"](https://arxiv.org/abs/2306.02117) *(AISTATS 2024)* studies what happens when users distribute themselves across competing services, and each service retrains on its own user base. Stable equilibria feature segmented populations — services specialize. The interesting twist: repeated myopic updates by multiple learners can actually lead to better outcomes than a single optimizer would achieve. Competition, under certain conditions, helps.


            ["Learning from Streaming Data when Users Choose"](https://arxiv.org/abs/2310.19007) *(ICML 2024)* looks at digital markets where users select between competing services that learn from user data. The selection creates a feedback loop: better service attracts more users, more users mean more data, more data means better service. She characterizes when this leads to monopoly and when it leads to coexistence.


            ["Strategic Usage in a Multi-Learner Setting"](https://arxiv.org/abs/2310.19009) *(AISTATS 2024)* tackles the same multi-service world but with strategic users. Naive retraining leads to oscillation — services bounce between states as users game them. Stability requires more careful algorithm design.


            <div class="highlight">
                If you're thinking about platform dynamics, creator incentives, or how algorithmic decisions propagate through ecosystems — the papers above are the core reading list.


            </div>

            ## Control theory meets machine learning


            Dean's platform work draws heavily on her control theory background, and she maintains a parallel research line in pure control-meets-ML.


            ["Online Convex Optimization with Unbounded Memory"](https://arxiv.org/abs/2310.19304) *(NeurIPS 2023)* extends OCO to settings where the loss at time t depends on your entire decision history — not just the last few steps. This is the technical infrastructure for thinking about long-term consequences of sequential decisions, which is exactly what platform policy requires.


            More recent work includes system identification for bilinear dynamical systems *(L4DC 2025, oral; ACC 2025; CDC 2025)*, [stochastic neural dynamic mode decomposition](https://arxiv.org/abs/2503.11354) for reconstructing wind and ocean fields from sparse sensors *(L4DC 2026)*, and [explore-then-commit strategies for bandits with latent dynamics](https://arxiv.org/abs/2503.17489) *(AISTATS 2026)*. These are more specialized, but they show the depth of her dynamical-systems toolkit.


            ## Fairness, social systems, and measurement


            A smaller but sharp thread.


            ["Do LLMs Favor LLMs?"](https://arxiv.org/abs/2501.09114) *(2026)* analyzes 125,000+ paper-review pairs from ICLR, NeurIPS, and ICML to ask whether LLM-assisted peer reviews are biased toward LLM-generated papers. The answer: no. LLM-assisted reviews are simply more lenient across the board. The bias is uniform, not targeted. Clean result, large dataset.


            ["Capacity Constraints Make Admissions Processes Less Predictable"](https://arxiv.org/abs/2501.14654) *(AAAI 2026, oral)* points out something that should be obvious but wasn't formalized: ML predictions about admission outcomes fail because your outcome depends on who else applied. The capacity constraint makes individual predictions fundamentally harder. It's a nice example of her instinct for identifying feedback effects that static models miss.


            ## Everything else


            Dean's range is broader than the themes above suggest. ["Pre-trained LLMs Learn Hidden Markov Models In-context"](https://arxiv.org/abs/2408.14468) *(NeurIPS 2025)* shows that LLMs can model HMM-generated data via in-context learning, matching the theoretical optimum. A clean theoretical result about what transformers implicitly learn.


            On the applied side: [high-altitude balloon station-keeping](https://arxiv.org/abs/2502.10477) using first-order MPC in JAX *(ICRA 2026)* — 24% improvement over RL. And robot-assisted feeding with contextual bandits *(ICRA 2025, Best Paper Finalist)*. These show she can move from theory to hardware.


            ## What to take away


            Dean's worldview, as I read it: if your system has users, it has dynamics. Users respond to algorithms, algorithms retrain on user behavior, and the steady state of this loop is often nothing like what you'd predict from a static analysis. She has the control theory chops to formalize this and the taste to pick problems where the formalism actually matters.


            Her most distinctive contribution is making the connection between *algorithmic design choices* (what loss function, what factorization, what ranking policy) and *ecosystem-level outcomes* (what content gets made, who gets harmed, whether markets concentrate or diversify). That connection is the thing.


            If you're reaching out about platform dynamics, content creator behavior, or the feedback loops between algorithms and the people they serve — start with the ICLR 2023 exposure game paper and the ICML 2025 two-sided platform paper. Those are the anchors. Everything else branches from there.


            *Her website is [sdean.website](https://sdean.website).*