• TILOS Seminar: Amplifying human performance in combinatorial competitive programming

    Virtual

    Petar Veličković, Google DeepMind Abstract: Recent years have seen a significant surge in complex AI systems for competitive programming, capable of performing at admirable levels against human competitors. While steady progress has been made, the highest percentiles still remain out of reach for these methods on standard competition platforms such as Codeforces. In this talk, […]

  • TILOS Seminar: Optimal Quantization for LLMs and Matrix Multiplication

    HDSI 123 and Virtual 3234 Matthews Ln, La Jolla, CA, United States

    Yury Polyanskiy, MIT Abstract: The main building block of large language models is matrix multiplication, which is often bottlenecked by the speed of loading these matrices from memory. A number of recent quantization algorithms (SmoothQuant, GPTQ, QuIP, SpinQuant etc) address this issue by storing matrices in lower precision. We derive optimal asymptotic information-theoretic tradeoff between […]

  • TILOS-HDSI Seminar with Adam Klivans (UT Austin): A New Paradigm for Learning with Distribution Shift

    HDSI 123 and Virtual 3234 Matthews Ln, La Jolla, CA, United States

    Adam Klivans, The University of Texas at Austin Abstract: We revisit the fundamental problem of learning with distribution shift, where a learner is given labeled samples from training distribution D, unlabeled samples from test distribution D′ and is asked to output a classifier with low test error. The standard approach in this setting is to […]

  • Optimization for ML and AI Seminar with Courtney Paquette (McGill University): High-dimensional Optimization with Applications to Compute-Optimal Neural Scaling Laws

    CSE 1242 and Virtual 3235 Voigt Dr, La Jolla, CA, United States

    Courtney Paquette, McGill University Abstract: Given the massive scale of modern ML models, we now only get a single shot to train them effectively. This restricts our ability to test multiple architectures and hyper-parameter configurations. Instead, we need to understand how these models scale, allowing us to experiment with smaller problems and then apply those […]

  • TILOS-HDSI Seminar: AI safety theory: the missing middle ground

    HDSI 123 and Virtual 3234 Matthews Ln, La Jolla, CA, United States

    Adam Oberman, McGill University Abstract: Over the past few years, the capabilities of generative artificial intelligence (AI) systems have advanced rapidly. Along with the benefits of AI, there is also a risk of harm. In order to benefit from AI while mitigating the risks, we need a grounded theoretical framework. The current AI safety theory, […]

  • TILOS-SDSU Seminar: Certifiably Correct Machine Perception

    Lamden Hall 341 (SDSU) and Virtual San Diego, CA, United States

    David Rosen, Northeastern University Abstract: Many fundamental machine perception and state estimation tasks require the solution of a high-dimensional nonconvex estimation problem; this class includes (for example) the fundamental problems of simultaneous localization and mapping (in robotics), 3D reconstruction (in computer vision), and sensor network localization (in distributed sensing). Such problems are known to be […]

  • TILOS-SDSU Seminar: 95 Percent: Bridging the Gap Between Prototype and Product

    Lamden Hall 341 (SDSU) and Virtual San Diego, CA, United States

    Jeremy Schwartz, Zoox Abstract: When transitioning from the academic world to the professional world of engineering, one of the most common pitfalls is failing to understand the difference between a compelling prototype and a successful product. This talk will focus on that distinction. We will discuss the differences between them, and the work required to […]

  • Optimization for AI and ML Seminar: Training Neural Networks at Any Scale

    HDSI 123 and Virtual 3234 Matthews Ln, La Jolla, CA, United States

    Volkan Cevher, École Polytechnique Fédérale de Lausanne Abstract: At the heart of deep learning’s transformative impact lies the concept of scale--encompassing both data and computational resources, as well as their interaction with neural network architectures. Scale, however, presents critical challenges, such as increased instability during training and prohibitively expensive model-specific tuning. Given the substantial resources […]

  • Optimization for ML and AI Seminar: Stochastic-Gradient and Diagonal-Scaling Algorithms for Constrained Optimization and Learning

    HDSI 123 and Virtual 3234 Matthews Ln, La Jolla, CA, United States

    Frank E. Curtis, Lehigh University Abstract: I will motivate and provide an overview of recent efforts in my research group on the design and analysis of stochastic-gradient-based algorithms for solving constrained optimization problems. I will focus in particular on our motivation for informed supervised learning, where constraints in the training problem can be used to […]

  • TILOS-HDSI Seminar: Incentivizing Emergent Behaviors for LLMs via Reinforcement Learning

    Qualcomm Conference Center Room B (Jacobs Hall first floor) and Virtual 9736 Engineers Ln, La Jolla, CA, United States

    Yi Wu, Tsinghua University Abstract: Reinforcement Learning (RL) has become a powerful post-training method for eliciting advanced behaviors in large language models (LLMs). This talk presents recent results showing how RL can incentivize the emergence of LLM capabilities across three domains: (1) multi-player deduction game, Werewolf, where RL-trained LLM agents develop strategic behaviors and outperform […]