Search this site
Embedded Files
(Ola)wale Salaudeen
  • Home
  • About
  • Publications
  • Talks and Presentations
  • CV
(Ola)wale Salaudeen
  • Home
  • About
  • Publications
  • Talks and Presentations
  • CV
  • More
    • Home
    • About
    • Publications
    • Talks and Presentations
    • CV

(Ola)wale Salaudeen

  • Visiting Scientist, Schmidt Sciences

  • Postdoctoral Associate, Massachusetts Institute of Technology

  • Postdoctoral Scholar, Eric and Wendy Schmidt Center at the Broad Institute of MIT and Harvard

  • Bio

Education

  • Ph.D. in Computer Science, 2024. University of Illinois at Urbana-Champaign

  • Visiting PhD Student in Computer Science, 2022-2024. Stanford University

  • B.S. with honors in Mechanical Engineering with minors in Mathematics and Computer Science, 2019. Texas A&M University

I am on 2025-26 academic job market and welcome inquiries about openings for tenure-track positions to begin Fall 2026!

LinkLinkedInLinkEmailGitHub

Please see CV for more info.

Research Interest

I am broadly interested in reliable and trustworthy AI. I primarily study questions related to the robustness of artificial intelligence (AI) in real-world decision-making. I work on developing methods that enable AI systems to generalize and adapt to new environments different from their training data (distribution shifts). I also work on developing the principles and practices of reliable AI evaluation. This includes studying the external validity of key benchmarks (ImageNet) in deep learning, the reliability of benchmarks for out-of-distribution generalization, and frameworks for valid evaluation of AI capabilities. Application areas of my work include biological imaging, algorithmic fairness, healthcare, and AI policy.

Research Themes

See Publications for more. * denotes equal contribution. α-β denotes alphabetical order.

AI Evaluation: AI evaluation is the rigorous science of scrutinizing the relationship between a measurement, like a benchmark score, and the real-world capabilities we claim it represents. This process aims to build a thorough understanding of a model’s strengths and failure points, ensuring that our claims about its capabilities and reliability are both meaningful and trustworthy.

  • Measurement to Meaning: A Validity-Centered Framework for AI Evaluation
    Olawale Salaudeen*, Anka Reuel*, Ahmed Ahmed, Suhana Bedi, Zachary Robertson, Sudharsan Sundar, Ben Domingue, Angelina Wang, Sanmi Koyejo
    Working Paper
    [arXiv] [webpage]

  • ImageNot: A contrast with ImageNet preserves model rankings
    Olawale Salaudeen, Moritz Hardt
    In Review
    [arXiv] [code] [webpage]

  • Toward an Evaluation Science for Generative AI Systems
    Laura Weidinger, Inioluwa Deborah Raji, Hanna Wallach, Margaret Mitchell, Angelina Wang, Olawale Salaudeen, Rishi Bommasani,  Deep Ganguli,  Sanmi Koyejo,  William Isaac
    In The Bridge 2025, National Academy of Engineering
    [arXiv]

Domain Generalization: Machine learning models often take shortcuts; consequently, they can fail in the real world when they associate a concept with irrelevant background details present in the training data. Domain generalization teaches a model to ignore this contextual noise and focus on the core, essential features, ensuring it can make reliable predictions in new and unfamiliar environments.

  • Are Domain Generalization Benchmarks with Accuracy on the Line Misspecified?
    A version (On Domain Generalization Datasets as Proxy Benchmarks for Causal Representation Learning) was presented as an Oral Presentation at the NeurIPS 2024 Causal Representation Learning Workshop
    Olawale Salaudeen, Nicole Chiou, Shiny Weng, Oluwasanmi Koyejo
    In TMLR 2025
    [arXiv] [code] [webpage] [news]

  • Causally Inspired Regularization Enables Domain General Representations
    Olawale Salaudeen, Oluwasanmi Koyejo
    In AISTATS 2024
    [arXiv] [code] [webpage]


Domain Adaptation: Often, we have some auxiliary information about the environment where the models will be deployed. Domain adaptation is the process of utilizing such information to adapt and maintain the model's effectiveness in the new context.

  • Proxy Methods for Domain Generalization
    Katherine Tsai, Stephen R. Pfohl, O. Salaudeen, Nicole Chiou, Matt J. Kusner, Alexander D’Amour, Sanmi Koyejo, Arthur Gretton.
    In AISTATS 2024
    [arXiv] [code]

  • Adapting to Latent Subgroup Shifts via Concepts and Proxies
    α–β. Ibrahim Alabdulmohsin*, Nicole Chiou*, Alexander D’Amour*, Arthur Gretton*, Sanmi Koyejo*, Matt J. Kusner*, Stephen R. Pfohl*, Olawale Salaudeen*, Jessica Schrouff*, Katherine Tsai*.
    In AISTATS 2023
    [arXiv] [code] [webpage]

I am very happy to discuss new research directions; please reach out if there is shared interest!

Selected Recent News

  • Summer 2025. Our [paper] on the limitations of domain generalization benchmarks and solutions – Are Domain Generalization Benchmarks with Accuracy on the Line Misspecified? – is accepted at TMLR!

  • Summer 2025. Our [preprint] on the limitations of evaluating AI systems with tests carefully designed for human populations – Stop Evaluating AI with Human Tests, Develop Principled, AI-specific Tests instead – is now available on arXiv!

  • Summer 2025. Our [preprint] on interpreting disaggregated evaluations of algorithm fairness – Understanding challenges to the interpretation of disaggregated evaluations of algorithmic fairness – is now available on arXiv!

  • Summer 2025. [service]. I am serving as a program chair for the Machine Learning for Health (ML4H) conference in San Diego, CA, in December. Please reach out if you are interested in sponsoring this great conference!

  • Summer 2025. [honors/appointment]. I will spend the next year at Schmidt Sciences in NYC as a Visiting Scientist (previously titled AI Institute Fellow) starting this summer! Please reach out if you are in NYC!

  • Spring 2025. Our [preprint] on AI evaluation and validity – Measurement to Meaning: A Validity-Centered Framework for AI Evaluation – is now available on arXiv!

  • Spring 2025. [honors/appointment]. I joined the Eric and Wendy Schmidt Center, led by Prof. Caroline Uhler at the Broad Institute of MIT and Harvard, as a postdoctoral scholar.

  • Spring 2025. Our [paper] Toward an Evaluation Science for Generative AI Systems appeared in the National Academy of Engineering's latest edition on "AI Promises & Risks."

  • Spring 2025. I gave a [talk] on addressing distribution shifts with varying levels of deployment distribution information at the MIT LIDS Postdoc NEXUS meeting!

  • Winter 2025. [service]. I am co-organizing the new AI for Society seminar at MIT.

  • Winter 2025. Our [paper] titled What’s in a Query: Examining Distribution-based Amortized Fair Ranking will appear at the International World Wide Web Conference (WWW), 2025.

  • Winter 2025. I was selected as an NYU Tandon Faculty First-Look Fellow; I look forward to visiting and giving a [honors/talk] on our work on distribution shifts at NYU in February; news!

  • Winter 2025. [service]. I am co-organizing the 30th Annual Sanjoy K. Mitter LIDS Student Conference at MIT.

  • Winter 2025. I was selected as a Georgia Tech FOCUS Fellow; I look forward to visiting and giving a [honors/talk] on our work on distribution shifts at Georgia Tech in January!

  • Fall 2024. Our [paper] titled On Domain Generalization Datasets as Proxy Benchmarks for Causal Representation Learning will appear at the Neurips 2024 workshop on causal representation learning as an Oral Presentation.

  • Fall 2024. [appointment]. I joined the Healthy ML Lab, led by Prof. Marzyeh Ghassemi, at MIT as a postdoctoral associate!

Older News

  • Spring 2025. Our [preprint] on domain generalization benchmarks – Are Domain Generalization Benchmarks with Accuracy on the Line Misspecified? – is now available on arXiv!

  • Summer 2024. I gave a talk on our work on distribution shift at Texas State's Computer Science seminar.

  • Summer 2024. I gave a [talk] on our work on distribution shift at UT Austin's Institute for Foundations of Machine Learning (IFML).

  • Summer 2024. I successfully defended my PhD dissertation titled “Towards External Valid Machine Learning: A Spurious Correlations Perspective”!

  • Spring 2024. I gave a [talk] on AI for critical systems at the MobiliT.AI forum (May 28-29)!

  • Spring 2024. I gave a [talk] at UIUC Machine Learning Seminar on our work on the external validity of ImageNet; artifacts here!

  • Spring 2024. Our [preprint] demonstrating the external validity of ImageNet model/architecture rankings – ImageNot: A contrast with ImageNet preserves model ranking – is now available on arXiv!

  • Winter 2024. Two [papers] on machine learning under distribution shift will appear at AISTATS 2024 (see Publications)!

  • Winter 2024. I have returned to Stanford from MPI!

  •  Fall 2023. I will join the Social Foundations of Computation department at the Max Planck Institute for Intelligent Systems in Tübingen, Germany this fall as a Research Intern working with Dr. Moritz Hardt! 

  • Spring 2023. I passed my PhD Preliminary Exam!

  • Spring 2023. I will join Cruise LLC's Autonomous Vehicles Behaviors team in San Francisco, CA this summer as a Machine Learning Intern! 

  • Fall 2022. I have moved to Stanford University as a "student of new faculty (SNF)" with Professor Sanmi Koyejo!

  • Summer 2022. I am honored to be selected as a top reviewer (10%) of ICML 2022!

  • Summer 2022. I will join Google Brain (now Google Deepmind) in Cambridge, MA this summer as a Research Intern!

  • Spring 2021. I gave a [talk] on leveraging causal discovery for fMRI denoising at the Beckman Institute Graduate Student Seminar!

  • Fall 2021. Our [paper] titled Exploiting Causal Chains for Domain Generalization was accepted at the 2021 NeurIPS Workshop on Distribution Shift!

  •  Fall 2021. I was selected as a Miniature Brain Machinery (MBM) NSF Research Trainee!

  • Summer 2021. I was selected to receive an Illinois GEM Associate Fellowship!

  • Spring 2021. I passed my Ph.D. qualifying exam!

  • Spring 2020. I was selected to receive a 2020 Beckman Institute Graduate Fellowship!

Mentorship

I am happy to mentor students with overlapping research interests. Particularly for undergrads at MIT, programs like UROP are a great mechanism for mentorship.

More generally, I am very happy and available to give advice and feedback on applying to and navigating both undergraduate and graduate programs in computer science and related disciplines – especially for those to whom this type of feedback and guidance would be otherwise unavailable.

About

Publications

Talks and Presentations

Curriculum Vitae

Office: Stata (32) D678

olawale [at] mit [dot] edu

LinkLinkedInLinkEmailGitHub
"The reasonable man adapts himself to the world: the unreasonable one persists in trying to adapt the world to himself. Therefore, all progress depends on the unreasonable man." - George Bernard Shaw, Man and Superman
Google Sites
Report abuse
Page details
Page updated
Google Sites
Report abuse