• Optimization for ML and AI Seminar: Transformers Learn Generalizable Chain-of-Thought Reasoning via Gradient Descent

    HDSI 123 and Virtual 3234 Matthews Ln, La Jolla, CA, United States

    Yuejie Chi, Yale Abstract: Transformers have demonstrated remarkable chain-of-thought reasoning capabilities, yet, the underlying mechanisms by which they acquire and extrapolate these capabilities remain limited. This talk presents a theoretical analysis of transformers trained via gradient descent for symbolic reasoning and state tracking tasks with increasing problem complexity. Our analysis reveals the coordination of multi-head […]

  • TILOS-SDSU Seminar: Autopilots Need Parachutes: Reliability Lessons from LLM-Automated Embedded AI Systems

    Lamden Hall 341 (SDSU) and Virtual San Diego, CA, United States

    Roberto Morabito, EURECOM Abstract: Embedded AI systems are becoming increasingly complex to develop and maintain, requiring specialized workflows that span data processing, model conversion, optimization, and deployment across heterogeneous hardware platforms. Recently, large language models have emerged as a promising tool to automate parts of this lifecycle. In this talk, I present recent work investigating […]

  • TILOS-Optimization for ML and AI Seminar: Implicit bias results for Muon, Adam, and Friends

    HDSI 123 and Virtual 3234 Matthews Ln, La Jolla, CA, United States

    Matus Telgarsky, New York University Abstract: This talk will give both an empirical overview and a few simple bonds controlling the optimization path, or implicit bias, of modern optimization methods such as Adam and Muon (and Friends). The talk will begin with empirical results demonstrating the implicit bias phenomenon with shallow networks and also transformers […]

  • Optimization for ML and AI Seminar: A survey of the mixing times of the Proximal Sampler algorithm

    HDSI 123 and Virtual 3234 Matthews Ln, La Jolla, CA, United States

    Andre Wibisono, Yale University Abstract: Sampling is a fundamental algorithmic task with many connections to optimization. In this talk, we survey a recent algorithm for sampling known as the Proximal Sampler, which can be seen as a proximal discretization of the continuous-time Langevin dynamics, and achieves the current state-of-the-art iteration complexity for sampling in discrete […]

  • ICLR 2026 Workshop: Principled Design for Trustworthy AI – Interpretability, Robustness, and Safety across Modalities

    ICLR 2026 Riocentro Convention and Event Center, Rio de Janiero, Brazil

    Modern AI systems, particularly large language models, vision-language models, and deep vision networks, are increasingly deployed in high-stakes settings such as healthcare, autonomous driving, and legal decisions. Yet, their lack of transparency, fragility to distributional shifts between train/test environments, and representation misalignment in emerging tasks and data/feature modalities raise serious concerns about their trustworthiness. This […]

  • TILOS-SDSU Seminar: A Modular AgenticAI Architecture for Commercially Scalable and Compliant Robotics

    TBA

    Sahil Rajesh Dhayalkar, Brain Corporation Abstract: Autonomous navigation in dynamic environments faces immense challenges. Traditional rigid, rules-based systems often fail due to a lack of semantic understanding needed to adapt to continuous environmental shifts. Conversely, emerging end-to-end Vision-Language-Action (VLA) models introduce a critical "black box" dilemma; they inherently lack the explicit application context, deterministic guardrails, […]

  • Optimization for ML and AI Seminar: Fantastic Pretraining Optimizers and Where to Find Them

    HDSI 123 and Virtual 3234 Matthews Ln, La Jolla, CA, United States

    Tengyu Ma, Stanford Abstract: AdamW has long been the dominant optimizer in language model pretraining, despite numerous claims that alternative optimizers offer 1.4 to 2x speedup. We posit that two methodological shortcomings have obscured fair comparisons and hindered practical adoption: (i) unequal hyperparameter tuning and (ii) limited or misleading evaluation setups. To address these two […]

  • Optimization for ML and AI Seminar with Nigel Goldenfeld (UC San Diego)

    HDSI 123 and Virtual 3234 Matthews Ln, La Jolla, CA, United States

    Nigel Goldenfeld, UC San Diego Abstract: TBA Nigel Goldenfeld holds the Chancellor's Distinguished Professorship in Physics at UC San Diego, which he joined in Fall 2021 after 36 years at the University of Illinois at Urbana-Champaign (UIUC). Nigel's research spans condensed matter theory, the theory of living systems, hydrodynamics and non-equilibrium statistical physics. Nigel received […]

  • CVPR 2026 Workshop: Trustworthy, Robust, Uncertainty-Aware, and Explainable Visual Intelligence and Beyond (TRUE-V)

    IEEE/CVF Conference on Computer Vision and Pattern Recognition Denver, CO, United States

    Contemporary vision models and vision–language models are increasingly deployed in high-stakes domains, yet remain opaque, fragile, and difficult to align across tasks and modalities. This workshop aim to foster dialogue on the urgent need for transparent, reliable, and safe computer vision systems, especially in critical domains such as healthcare, transportation, and legal decision making. It brings together […]