Carnegie Mellon University
AI Measurement Science & Engineering (AIMSEC)

CMU-NIST Cooperative Research Center

Media

Is It Time to Rethink LLM Pre-Training? [Aditi Raghunathan] - 747

CMU Experts: What If We Could Create a Trusted AI Ecosystem?

Publications


Explore publications from AIMSEC-affiliated faculty below.

How do I Acknowledge my work with the AIMSEC-NIST center?

2025

Table of AIMSEC faculty publications from 2025
Affiliated Faculty 

Title and Citation

Date

Hoda Heidari

What Can Generative AI Red-Teaming Learn from Cyber Red-Teaming?
Sinha, A., Lucassen, J., Grimes, K., Feffer, M., Soto, M., Heidari, H., & VanHoudnos, N. (2025, July 16). What Can Generative AI Red-Teaming Learn from Cyber Red-Teaming?. (Technical Report CMU/SEI-2025-TR-006). Retrieved July 22, 2025, from https://doi.org/10.1184/R1/29410136.

 July 2025

Kenneth Holstein

Measurement as Bricolage: Examining How Data Scientists Construct Target Variables for Predictive Modeling Tasks
Guerdan, L., Saxena, D., Chancellor, S., Wu, Z. S., & Holstein, K. (2025). Measurement as Bricolage: Examining How Data Scientists Construct Target Variables for Predictive Modeling Tasks. arXiv preprint arXiv:2507.02819.

3 July 2025

Nikhil George, Ramayya Krishnan, Rahul Telang

From Posting to Prediction: Building Validated Workforce Analytics
Charuvilparambil Titus, Nikhil George and Krishnan, Ramayya and Telang, Rahul, From Posting to Prediction: Building Validated Workforce Analytics
(November 05, 2024). Available at SSRN: https://ssrn.com/abstract=4906323

June 2025

Lujo Bauer

Perry: A High-level Framework for Accelerating Cyber Deception Experimentation
Singer, B., Saquib, Y., Bauer, L., & Sekar, V. (2025). Perry: A High-level Framework for Accelerating Cyber Deception Experimentation. arXiv preprint arXiv:2506.20770.

25 June 2025

Zico Kolter

OS-HARM: A Benchmark for Measuring Safety of Computer Use Agents
Kuntz, T., Duzan, A., Zhao, H., Croce, F., Kolter, Z., Flammarion, N., & Andriushchenko, M. (2025). OS-Harm: A Benchmark for Measuring Safety of Computer Use Agents. arXiv preprint arXiv:2506.14866.

17 June 2025

Hong Shen

From Margins to the Table: Charting the Potential for Public Participatory Governance of Algorithmic Decision Making
Eslami, M., Fox, S., Shen, H., Fan, B., Lin, Y. R., Farzan, R., & Schwanke, B. (2025, June). From Margins to the Table: Charting the Potential for Public Participatory Governance of Algorithmic Decision Making. In Proceedings of the 2025 ACM Conference on Fairness, Accountability, and Transparency (pp. 2657-2670).

 23 June 2025

Kenneth Holstein Making the Right Thing: Bridging HCI and Responsible AI in Early-Stage AI Concept Selection
Jung, J. Y., Saxena, D., Park, M., Kim, J., Forlizzi, J., Holstein, K., & Zimmerman, J. (2025, July). Making the Right Thing: Bridging HCI and Responsible AI in Early-Stage AI Concept Selection. In Proceedings of the 2025 ACM Designing Interactive Systems Conference (pp. 2992-3012).
20 Jun 2025
Ramayya Krishnan, Rema Padman Firm or Fickle? Evaluating Large Language Models Consistency in
Sequential Interactions
Li, Y., Miao, Y., Ding, X., Krishnan, R., & Padman, R. (2025). Firm or fickle? evaluating large language models consistency in sequential interactions. arXiv preprint arXiv:2503.22353.
5 Jun 2025
Ramayya Krishnan, Rema Padman Beyond Single-Turn: A Survey on Multi-Turn Interactions with Large Language Models
Li, Y., Shen, X., Yao, X., Ding, X., Miao, Y., Krishnan, R., & Padman, R. (2025). Beyond single-turn: A survey on multi-turn interactions with large language models. arXiv preprint arXiv:2504.04717.
14 May 2025
Aditi Raghunathan Memorization Sinks: Isolating Memorization during LLM Training
Ghosal, G. R., Maini, P., & Raghunathan, A. Memorization Sinks: Isolating Memorization during LLM Training. In Forty-second International Conference on Machine Learning.
01 May 2025

Past Webinars

Webinar recordings will be available here soon. Stay tuned for upcoming sessions and past event archives.

REGISTER FOR UPCOMING WEBINARS 

VIEW PAST EVENTS