Member Of Technical Staff Research Scientist at OpenAI
San Francisco Bay Area United States
Join Prog.AI to see contacts
Join Prog.AI to see contacts
Summary
👤
Senior
🎓
Top School
Micah Carroll is a research scientist and software engineer based in the San Francisco Bay Area, currently serving as a Member of Technical Staff on OpenAI's Safety Oversight team. He earned a PhD in Artificial Intelligence from UC Berkeley and a BA in Statistics, pairing rigorous academic training with hands-on experimentation and system-level AI safety work. His research spans multi-agent coordination, human-AI collaboration, and responsible AI deployment, including development of benchmarking environments like Overcooked-AI. Micah has collaborated with leaders in the field, including Anca Dragan and Rohin Shah, and has contributed across industry and academia through roles at Microsoft, Cambridge, and the Center for Human-Compatible AI. An active open-source contributor, he led QA and test automation for the Overcooked AI environment, implementing test suites for various agent types to study cooperative behavior. With a track record of translating research insights into practical, auditable tooling, he is focused on ensuring AI advances benefit society at scale.
10 years of coding experience
4 years of employment as a software developer
PhD Artificial Intelligence, PhD Artificial Intelligence at University of California, Berkeley
Graduated by completing the Esame Di Stato, Classical and Ancient Studies, 100 out of: 100, Graduated by completing the Esame Di Stato, Classical and Ancient Studies, 100 out of: 100 at Liceo Classico ISIS Niccolini Palli, Livorno, Italy
A benchmark environment for fully cooperative human-AI performance.
Role in this project:
QA Engineer / Test Automation Engineer
Contributions:68 reviews, 242 commits, 62 PRs in 3 years 7 months
Contributions summary:Micah implemented test files and test suites for the overcooked_ai environment, specifically focusing on agent behaviors and the OvercookedGridworld MDP. The commits demonstrate the creation of test cases for various agent types, including FixedPlanAgent, CoupledPlanningAgent, and GreedyHumanModel, and the testing of scenario-specific behaviors within the environment. The code changes include the addition of test fixtures, assertions, and the use of the unittest framework.
Codebase for "Targeted Manipulation and Deception Emerge in LLMs Trained on User Feedback". This repo implements a generative multi-turn RL environment with support for agent, user, user feedback, transition and veto models. It also implements KTO and expert iteration for training on user preferences.
Contributions:245 reviews, 107 PRs, 651 pushes in 3 months
Find and Hire Top DevelopersWe’ve analyzed the programming source code of over 60 million software developers on GitHub and scored them by 50,000 skills. Sign-up on Prog,AI to search for software developers.