Upcoming Events

TBA Anish Agrawal, Traversal Entrepreneurship

7PM November 18, 2025 Davis Auditorium

AI Entrepreneurship: Obstacles, Opportunities, and Outcomes Anish Das Sarma, Reinforce Labs Entrepreneurship

7PM December 02, 2025 Davis Auditorium

TBA Ramin Hasani, Liquid AI Entrepreneurship

7PM December 16, 2025 451 Mudd

Past Events

Warehouse Adaptation of Robots and AI Systems Vivian Zhang, WarehouseRobot.ai Entrepreneurship

7PM November 04, 2025 Davis Auditorium

Autonomous warehouse systems represent one of the most advanced and large-scale examples of embodied multi-agent intelligence operating in the physical world. Thousands of mobile robots collaborate continuously to move, pick, and sort goods in dynamic, uncertain environments—achieving high efficiency, adaptability, and robustness in real time.

In this talk, I will discuss how warehouse robotics offers both a practical foundation and conceptual inspiration for the design of AI systems. I will introduce the market context and technical underpinnings of large-scale robotic automation, then explore how intelligence emerges through coordination, perception, and control across distributed agents. I will also discuss how large language models (LLMs) may extend this paradigm—enabling natural language tasking, high-level reasoning, and seamless human–AI collaboration in embodied systems.

Finally, I will reflect on how lessons from large-scale robotic coordination can inform the broader design of the warehouse automation ecosystems, highlighting new directions for research at the intersection of autonomy, learning, and collective intelligence.

Vivian Zhang is the Chief Executive Officer of WarehouseRobot.AI, where she leads research and development in intelligent robotics, large-scale multi-agent coordination, and autonomous warehouse systems. Her work spans planning under uncertainty, human–AI collaboration, and the integration of large language models into embodied intelligence. Vivian’s perspective bridges industrial-scale deployment and AI research, with a focus on how real-world robotics can inform the next generation of AI agents and collective intelligence systems. Her portfolio of AI startups and AI VC fund can be found at https://viviancompanies.com/ and at https://extelligenceinvest.com/.

Vivian was named one of the Top 100 Most Influential Chinese by Forbes and recognized among 50 top most well known Data Scientists by CBNData. Vivian’s always happy to hear from founders and fellows — feel free to reach out anytime!

How to Build Cursor for Law Thomas Bueler-Faudree, August Entrepreneurship

7PM October 21, 2025 Davis Auditorium

Modern document-heavy work in law, finance, and consulting is repetitive, bespoke, and review heavy. Code generation tools like Cursor have reshaped how developers work, but that shift has not happened in law. Legal tasks are hard to measure and even harder to “unit test,” which is why many AI tools miss the mark.

This talk shows how August is building agents that do better: the right indexing methods, visible state, disciplined tool use, and citations at every step to cleanly automate real professional workflows. We’ll walk through a multi-step research agent (e.g., over SEC filings) that plans the task, retrieves the right sources, routes to the relevant sections, and drafts grounded legal answers.

Attendees will leave with a lightweight design pattern for credence domains like law. We will also share our growth and go-to-market. We are productizing law-firm workflows and, in some cases, selling with the firm to their clients. This is a new model that is gaining traction across professional services.

Thomas Bueler-Faudree is the co-founder of August, where he leads engineering and product. He earned a bachelor’s degree in computer science and history from Columbia University. At Columbia he worked in the CRIS Lab and CML.

Disaggregated LLM Inference: Past, Present, and Future Junda Chen Seminar

12PM October 14, 2025 CSB 453

Large language model (LLM) serving faces a fundamental challenge: the prefill phase is compute-bound while the decode phase is memory-bound. Existing serving systems co-locate these two phases and batch their computation across users and requests, a strategy that not only introduces strong prefill–decode interference but also couples resource allocation and parallelism choices across both phases. We introduce DistServe, a low-latency serving engine that assigns prefill and decode computation to different GPUs, thereby eliminating interference between the two. DistServe co-optimizes resource allocation and parallelism strategies tailored for each phase, while placing them across the cluster to account for bandwidth constraints and minimize the communication overhead introduced by disaggregation. As a result, DistServe substantially improves serving performance, measured as the maximum request rate that can be sustained under strict time-to-first-token (TTFT) and time-per-output-token (TPOT) constraints. Since DistServe, disaggregation has moved into the spotlight of LLM inference, with many companies and open-source projects adopting or extending this paradigm in practice. We will also highlight related academic and open-source work that has emerged around disaggregated inference, and reflect on the open challenges faced today as well as the opportunities that lie ahead.

Junda Chen is a third-year Ph.D. student in Computer Science at the University of California, San Diego. His research focuses on large-scale LLM systems, with an emphasis on efficient inference and training of large language models. He is an author of DistServe, a disaggregated inference system adopted by major industry partners (including NVIDIA Dynamo, llm-d, etc.). His broader work spans GPU scheduling, distributed inference frameworks, and the design of next-generation serving systems for foundation models.

Agents at Scale: Lessons from LinkedIn's Agent Journey Xiaofeng Wang, LinkedIn Entrepreneurship

7PM October 07, 2025 Davis Auditorium

Join Xiaofeng Wang, Sr. Engineering Manager at LinkedIn’s Agents Platform, as he takes you behind the scenes of the platform powering all of LinkedIn’s member-facing agentic applications at scale. This talk will trace the GenAI product revolution, from simple prompt‑in‑string use cases to the latest LinkedIn Hiring Assistant. It will show how the Agent Platform evolved to enable these experiences. Xiaofeng will share the technical foundations, organizational challenges, hiring strategies, and hard‑won lessons that shaped LinkedIn’s journey, highlighting how the team bridged open‑source innovation with enterprise‑grade infrastructure, built seamless developer experiences, and enforced trust and responsible AI from day one. Whether you’re a researcher, engineer, product leader, or entrepreneur, this talk offers a front-row look at LinkedIn’s agentic journey, the latest trends in agent development, and what it takes to transform prototypes into impactful, production-grade agents at enterprise scale.

Xiaofeng Wang is a Senior Engineering Manager at LinkedIn, where he leads the Agents Platform team, building the foundational technologies that enable the entire engineering organization to deliver agentic experiences. His team is responsible for the agent runtime platform, agent framework, and Cognitive Memory Agent. He and his team have also been invited to speak at various leading industry conferences. In his previous role at LinkedIn, Xiaofeng led the first team dedicated to Generative AI Foundations, developing infrastructure, platforms, and tooling from the ground up to power all of LinkedIn’s member-facing GenAI products since 2022. Before that, he worked as an engineer and data scientist in the Data Science Productivity and Research team, building platforms to automate advanced data science analyses such as A/B testing, deep dives, and observational causal studies. Prior to LinkedIn, Xiaofeng worked at Bank of America Merrill Lynch, where he led the team building recommendation systems for the Merrill Edge’s personalized investing platform. He received his Ph.D. in Systems and Information Engineering from the University of Virginia.

Keep Computing Systems Safe under Disruptive Agents Tianyin Xu Seminar

12:30PM October 07, 2025 CSB 453

The brains of computing systems today have increasingly been realized by generative AI such as Large Language Models (LLMs) together with agentic technologies. While the AI approaches are arguably more creative and cheaper to develop, their safety becomes a fundamental challenge. Instead of using another LLM as a judge or waiting for a provably trustworthy model to exist, we may want to figure out how to build seatbelts and airbags for computing systems driven by LLMs/agents. In this talk, I will share our preliminary experience of developing a site reliability engineering (SRE) agent which attempts to autonomously mitigate production system failures. I would also discuss the potential of advanced system intelligence – AI that can understand “systems” beyond code and reason about their safety properties.

Tianyin Xu is an Associate Professor of Computer Science at the University of Illinois Urbana-Champaign (UIUC). His research focuses on building reliable computer systems that empower next-generation computing. He has been on the UIUC List of Teachers Ranked as Excellent for nine times. His work received two OSDI Jay Lepreau Best Paper Awards, two ASPLOS Best Paper Awards, two SIGSOFT Distinguished Paper Awards, a Gilles Muller Best Artifact Award, a CACM Research Highlight, and an ICML Spotlight. He is also a recipient of the C.W. Gear Outstanding Junior Faculty Award, a Dean’s Award for Excellence in Research, an Intel Rising Star Faculty Award, and a Facebook Distributed Systems Research Award.

Reality Checks Kyunghyun Cho Seminar

12PM September 30, 2025 CSB 453

Despite its amazing success, leaderboard chasing has become something researchers dread and mock. When implemented properly and executed faithfully, leaderboard chasing can lead to both faster and easily reproducible progress in science, as evident from the amazing progress we have seen with machine learning, or more broadly artificial intelligence, in recent decades. It does not however mean that it is easy to implement and execute leaderboard chasing properly. In this talk, I will go over four case studies demonstrating the issues that ultimately prevent leaderboard chasing from a valid scientific approach. The first case study is on the lack of proper hyperparameter tuning in continual learning, the second on the lack of consensus on evaluation metrics in machine unlearning, the third on the challenges of properly evaluating the evaluation metrics in free-form text generation, and the final one on wishful thinking. By going over these cases, I hope we can collectively acknowledge some of our own fallacies, think of underlying causes behind these fallacies and come up with better ways to approach artificial intelligence research.

Kyunghyun Cho is a professor of computer science and data science at New York University and an executive director of frontier research at the Prescient Design team within Genentech Research & Early Development (gRED). He became the Glen de Vries Professor of Health Statistics in 2025. He is also a CIFAR Fellow of Learning in Machines & Brains and an Associate Member of the National Academy of Engineering of Korea. He served as a (co-)Program Chair of ICLR 2020, NeurIPS 2022 and ICML 2022. He was one of the three founding Editors-in-Chief of the Transactions on Machine Learning Research (TMLR) until 2024. He was a research scientist at Facebook AI Research from June 2017 to May 2020 and a postdoctoral fellow at University of Montreal until Summer 2015 under the supervision of Prof. Yoshua Bengio, after receiving MSc and PhD degrees from Aalto University April 2011 and April 2014, respectively, under the supervision of Prof. Juha Karhunen, Dr. Tapani Raiko and Dr. Alexander Ilin. He received the Samsung Ho-Am Prize in Engineering in 2021. He tries his best to find a balance among machine learning, natural language processing, and life, but almost always fails to do so.

Lessons for Building Valuable Domain-Specific AI Products Tom Effland, Noetica Entrepreneurship

7PM September 23, 2025 Davis Auditorium

This talk shares lessons learned going from NLP researcher to building Noetica, a platform that analyzes populations of complex contracts to determine market standards for deal terms. Along the way I’ll discuss (1) how LLMs have enabled a new opportunity space for hybrid systems that combine extraction capabilities with formal analytical frameworks, (2) how building these systems still requires deep fusion of subject-matter and technical judgment to identify tractable decompositions of seemingly impossible problems, and (3) how this converges to a new version of the Technical Product Manager role. This role is highly suited to technical researchers willing to tackle sector-specific challenges.

Tom Effland is the technical founder and CTO of Noetica AI, a fast-growing VC-backed Series A startup. The company’s AI-powered knowledge platform helps many top law firms improve outcomes in valuable corporate debt, securities, and M&A transactions. Before founding Noetica, Tom earned his PhD in Computer Science from Columbia University. He was advised by Prof. Michael Collins and received support from a NSF Graduate Research Fellowship.

Simulation Environments for testing and training task-focused AI agents Mehdi Jamei Seminar

12PM September 16, 2025 CSB 453

AI agents hold great value for enterprises but are notoriously difficult to productionize in a robust and reliable way. Enterprise agents must internalize task context, constraints, and success metrics to be reliable, yet learning directly in production is risky and often infeasible. Veris is a simulation-first experiential learning platform for training and evaluating domain-specific AI agents through simulated experience. I will present how Veris constructs high-fidelity interactive environments that mirror target deployments, including tool APIs, data schemas, user behaviors, and organizational policies. Then I will discuss how Veris supports multiple post-training paths like RL with online rollouts and trajectory replay, and iterative prompt or policy optimization. Finally, I will outline current challenges and research opportunities.

Mehdi is the cofounder and CEO of Veris AI, an agentic infrastructure company based in NY, building simulation training grounds for enterprise AI agents. Veris is named on the Future50 list by the Generalist and is backed by Decibel Ventures and ACrew Capital.
Mehdi is an applied AI researcher, builder, and industry leader. He holds a PhD in Electrical Engineering and Computer Science and an M.S. in Physics, both from UC Berkeley. Most recently, he served as the Director of AI at System.com

Focus is Everything - How to Effectively Navigate the AI Landscape Aaron Vontell, Anthropic Entrepreneurship

7PM September 09, 2025 Davis Auditorium

In this talk, Aaron will walk through his journey building in the AI space, detailing his learnings from Instabase, his own AI startup Regression Games, and now his experience at Anthropic. This talk will focus on the importance of focusing on specific problems and navigating frequently-changing technological landscapes, and how to apply cutting-edge research in record time.

MassGen: Multi-Agent Scaling System Chi Wang Seminar

12PM September 09, 2025 CSB 453

Chi will present MassGen, a growing open-source multi-agent scaling system that tackles single-agent limitations through collaborative AI coordination. Built on AG2’s research foundation, the system implements novel coordination of frontier AI agents. Early benchmarking across reasoning and instruction-following tasks shows statistically significant improvements, with growing industrial and academic adoption demonstrating the research opportunity for scaling intelligence through collaborative emergence.
Chi Wang is founder of AutoGen (now AG2), the open-source AgentOS to support agentic AI, and its parent open-source project FLAML, a fast library for AutoML & tuning. He has received multiple awards such as best paper of ICLR’24 LLM Agents Workshop and SIGKDD Data Science/Data Mining PhD Dissertation Award. He has 15+ years of research experience in Computer Science from Google DeepMind, Microsoft Research, Meta, UIUC and Tsinghua.

From Serving LLMs to Serving Agents on the Cloud Xiaozhe Yao Seminar

12PM July 24, 2025

In this talk, I will discuss key challenges in building agentic AI systems in the cloud. I will highlight DeltaZip, our recent work on efficiently deploying multiple fine-tuned models – a step we believe is essential toward enabling future AI systems. The core insight behind DeltaZip is that fine-tuning often introduces small-magnitude changes to a pre-trained model. By co-designing the serving system and compression algorithm, DeltaZip achieves a 2x to 12x throughput improvement over state-of-the-art systems. In addition to this project, I will share some ongoing challenges we are tackling in this space. Xiaozhe Yao is a third-year doctoral student at Systems Group, Department of Computer Science, ETH Zurich advised by Prof. Dr. Ana Klimović. His research explores the complex and fundamental tensions between three pillars: from optimizing systems for efficient ML, to improving data quality and organization for ML, to developing frameworks that bridge the gap between algorithms and their practical deployment. Through this multi-faceted approach, his work aims to better understand and build AI systems.

The Road to High-Quality LLM Inference Services: System, Data, and Context Yizheng Jiao

12PM July 17, 2025

This talk is about sharing the experience of building enterprise LLM inference services including

  1. high-level principles of increasing performance and saving costs 2. a data selection algorithm to finetune LLM to increase the accuracy of domain-specific questions 3. a method to enhance users’ prompt with domain-specific knowledge bases. Yizheng Jiao graduated from UNC Chapel Hill with a doctoral degree in 2022. He joined Bytedance after graduation and am doing research on LLM systems. His goal is to build efficient and accurate LLM services with experience includes LLM inference systems, data selection for LLM finetuning, and prompt optimization.

Multi-Agent Systems in the Era of LLMs: Testbeds, Applications, and Beyond Yusen Zhang

12PM July 10, 2025

Autonomous agents powered by large language models (LLMs) are emerging as powerful tools for a wide range of tasks. However, a single agent often faces performance ceilings, especially when tackling complex workflows like running an AI company or AI4Research, and is inherently limited in scenarios that involve multiple instances, such as simulations, embodied agents, and digital twins. In this talk, I will present Multi-Agent Large Language Models (MA-LLMs), a promising paradigm designed to overcome the fundamental limitations of single-agent systems. I will begin by highlighting three threads of my previous work that lay the groundwork for MA-LLMs. Next, I’ll introduce our research on fairness summarization, which demonstrates challenges that a single agent struggles to handle well. Then, I will present how agents can collaborate in a chain-of-agent manner to solve difficult tasks, such as long-document summarization and multi-step reasoning. Finally, I will reflect on current limitations in MA-LLMs and outline my long-term vision of building Agent Societies: a human-centric society consisting of scalable, trustworthy, and collaborative intelligent agents and humans.
Yusen Zhang is a fourth-year CS PhD student at Penn State University, advised by Dr. Rui Zhang. He has done industry research internships at Amazon, Microsoft, and Google. He also worked closely with Dr. Dragomir Radev. He received his master’s degree from Emory University, advised by Dr. Jinho D. Choi.

Efficient Fine-Tuning and Compression of Large Language Models: Towards Low-bit and Ultra-Low Parameter Solutions Jiajun Zhou

12PM July 07, 2025

Efficient fine-tuning of Large Language Models (LLMs) is crucial due to their substantial memory and computational demands. This seminar discusses recent advancements in techniques aimed at significantly reducing these costs, enabling effective adaptation of large-scale models even on resource-constrained hardware. The talk will begin with an overview of current challenges and mainstream approaches to compressing and fine-tuning LLMs, highlighting trade-offs between model size, accuracy, and efficiency. Subsequently, the speaker will introduce novel approaches that enable fine-tuning at extremely low precision and ultra-low parameter regimes, significantly reducing memory requirements without compromising performance. Finally, the discussion will cover recent progress and future directions for achieving efficient deployment of LLMs in real-world applications.
Jiajun Zhou is currently a Ph.D. student in the Department of Electrical and Electronic Engineering at the University of Hong Kong (HKU), supervised by Prof. Ngai Wong, and a visiting scholar at the University of California, Santa Barbara (UCSB). He received his Master’s degree in IC Design Engineering from the Hong Kong University of Science and Technology (HKUST) in 2019. He previously worked as a Research Assistant at the Chinese University of Hong Kong (CUHK). His research primarily focuses on developing innovative frameworks for efficient training and inference of Large Language Models (LLMs), particularly through quantization, low-bit optimization, and tensor decomposition. He has published extensively in AI and hardware acceleration venues, including NAACL, IEEE FCCM, and IEEE TCAD.\n”

March 2025 Workshop: AI Agents for Work Workshop

12PM March 12, 2025

On March 12, 2025, DAPLab ran the first annual workshop at the Columbia Business School. The one-day workshop to brought together over 200 industry leaders, Columbia faculty and students, and technologists who are interested in the concept of AI agents. Speakers and panelists come from enterprises that are deploying agentic solutions, technologists and infrastructure leaders, and researchers at leading AI labs as well as Columbia. These include Jason Wei from OpenAI who led their chain-of-thought and agentic work, Danielle Perszyk from Amazon AGI, Jonathan Frankle from Databricks, Deepak Dastrala from Intellect, Cong Yu who leads AI at Celonis, and more.

Spring 2025 Agentic System Made Real Course Eugene Wu and Kostis Kaffes Course

10:10AM - 11:25AM TR January 21, 2025

LLMs have opened new possibilities of automated agents that plan and complete tasks on the user’s behalf. Such agents have the potential to usher in a new industrial revolution by automating organizational processes. This graduate-level course will cut across the technology stack to examine the research questions that need to be answered for agents to be possible in real tasks that matter. Each session will review 1-3 papers or systems, and discuss research opportunities that arise from the gap between existing research and enterprise requirements. Topics will span systems (data systems and ML systems), AI (LLMs, agent-based planning), HCI, and theory (reinforcement learning, markets).