Four-Week Machine Learning Symposium, Fall 2024

The CDS Machine Learning Symposium, developed by Assistant Professors Aldo Pacchiano and Xuezhou Zhang, and hosted by the Faculty of Computing and Data Sciences, brings together leading scholars in machine learning to delve into the cutting-edge developments and foundational technologies shaping the field of machine learning. By uniting experts from various technical disciplines, including algorithmic design, model architectures, and optimization techniques, the symposium aims to illuminate the latest advancements and challenges in core machine learning methodologies.


Kiante Brantley, Assistant Professor, Harvard UniversityPast Talks Fall 2024

Efficient Policy Optimization Techniques for LLMs

Kiante Brantley, Assistant Professor, Harvard University

Date: Friday, December 6, at 10:00 AM

Location: CDS 1646

Post-training is essential for enhancing large language model (LLM) capabilities and aligning them to human preferences. One of the most widely used post-training techniques is reinforcement learning from human feedback (RLHF). In this talk, I will first discuss the challenges of applying RL to LLM training. Next, I will introduce RL algorithms that tackle these challenges by utilizing key properties of the underlying problem. Additionally, I will present an approach that simplifies the RL policy optimization process for LLMs to relative reward regression. Finally, I will extend this idea to develop a policy optimization technique for multi-turn reinforcement learning from human feedback.

Kianté Brantley is an Assistant Professor in the Kempner Institute and School of Engineering and Applied Sciences (SEAS) at Harvard University. He completed his Ph.D. in Computer Science at the University of Maryland College Park, advised by Dr. Hal Daumé III. After graduating, he completed his postdoctoral studies at Cornell University, working with Thorsten Joachims. His research focuses on problems at the intersection of machine learning and interactive decision-making, with the goal of improving the decision-making capabilities of foundation models. He has received several awards with his colleagues, including spotlight talks at ICLR 2023 and ICLR 2019. He has also received multiple fellowships, including the NSF LSAMP BD Fellowship and the NSF CI Fellow Postdoctoral Fellowship. In his spare time, he enjoys playing sports; his favorite sport at the moment is powerlifting.


Synthetic Potential Outcomes and Causal Mixture Identifiability

Bijan Mazaheri, Postdoctoral Associate, Broad Institute of MIT and Harvard University; Incoming Assistnat Professor, DartmouthBijan Mazaheri, Postdoctoral Associate, Broad Institute of MIT and Harvard University; Incoming Assistant Professor, Dartmouth

Date: Friday, November 22, at 2:00 PM

Location: CDS 1646

Heterogeneous data from multiple populations, sub-groups, or sources can be represented as a "mixture model” with a single latent class influencing all of the observed covariates. Heterogeneity can be resolved at multiple levels by grouping populations according to different notions of similarity. This talk, presented by Dr. Bijan Mazaheri, proposes grouping with respect to the causal response of an intervention or perturbation on the system. Mazaheri will show that this definition is distinct from standard notions, such as similar covariate values (e.g. clustering) or similar correlations between covariates (e.g. Gaussian mixture models). To solve the problem, Mazaheri will describe “synthetically” sampling from a counterfactual distribution using higher-order multi-linear moments of the observable data. To understand how these "causal mixtures" fit in with more classical notions, a hierarchy of mixture identifiability will be developed. Reflecting on this hierarchy, Mazaheri will discuss the role of causal modeling as a guiding framework for data science.

Dr. Bijan Mazaheri is an Eric and Wendy Schmidt Postdoctoral Fellow at the Broad Institute of MIT and Harvard University. Bijan is broadly interested in the task of combining data and knowledge from multiple places, topics, and modalities. Before starting at the Broad, Bijan was an NSF Graduate Research Fellow and Amazon AI4Science Fellow at Caltech, supervised by Shuki Bruck and Leonard Schulman. Bijan also studied at the University of Cambridge under a Herschel Smith Fellowship and holds a BA from Williams College. Bijan is starting an assistant professorship at Dartmouth Engineering in January and is recruiting Ph.D. students.


Han Shao, Postdoctoral Associate, Harvard University; Incoming Assistant Professor, UMD

Learning from Strategic Data Sources

Han Shao, Postdoctoral Associate, Harvard University; Incoming Assistant Professor, UMD

Date: Friday, November 8, at 11:00 AM

Location: CDS 1101

Abstract: In contrast with standard classification tasks, strategic classification involves agents strategically modifying their features in an effort to receive favorable predictions. For instance, given a classifier determining loan approval based on credit scores, applicants may open or close their credit cards and bank accounts to fool the classifier. The learning goal is to find a classifier robust against strategic manipulations. Various settings, based on what and when information is known, have been explored in strategic classification. In this talk, Shao will focus on addressing a fundamental question: the learnability gaps between strategic classification and standard learning. This talk is based on joint work with Avrim Blum, Omar Montasser, Lee Cohen, Yishay Mansour, and Shay Moran (arxiv.org/abs/2305.16501 published at NeurIPS'23, arxiv.org/abs/2402.19303 published at COLT'24).

Bio: Han Shao is a CMSA postdoc at Harvard, hosted by Cynthia Dwork and Ariel Procaccia. She will join the Department of Computer Science at the University of Maryland in Fall 2025 as an Assistant Professor. She completed her PhD at TTIC, where she was advised by Avrim Blum. Her research focuses on the theoretical foundations of machine learning, particularly on fundamental questions arising from human social and adversarial behaviors in the learning process. She is interested in understanding how these behaviors affect machine learning systems and developing methods to enhance accuracy and robustness. Additionally, she is interested in gaining a theoretical understanding of empirical observations concerning adversarial robustness.


Learning in Strategic Environments: From Calibrated Agents to General Information Asymmetry

Chara Podimata, Assistant Professor, MIT

Date: Friday, November 15, at 12:00 PM

Location: CDS 1646

In this talk, Podimata will discuss learning in principal-agent games where there is information asymmetry between what the principal and the agent know about each other’s chosen actions. They will introduce a generalization of the standard Stackelberg Games (SGs) framework: Calibrated Stackelberg Games (CSGs). In CSGs, a principal repeatedly interacts with an agent who (contrary to standard SGs) does not have direct access to the principal’s action but instead best-responds to calibrated forecasts about it. Podimata will show that in CSGs, the principal can achieve utility that converges to the optimum Stackelberg value of the game (i.e., the value that they could achieve had the agent known the principal’s strategy all along) both in finite and continuous settings, and that no higher utility is achievable. Finally, they will discuss a meta-question: when learning in strategic environments, can agents overcome uncertainty about their preferences to achieve outcomes they could have achieved absent any uncertainty? And can they do this solely through interactions with each other?

Chara Podimata is a 1942 Career Development Professor of Operations Research and Statistics at MIT. Her research focuses on incentive-aware ML and more broadly on social computing both from a theoretical and a practical standpoint. Her research is supported by Amazon and the MacArthur foundation through an x-grant. She got her PhD from Harvard. In her free time, she runs and spends time with her dog, Terra.


Fall 2023/Past Speakers