Stanford University, 2018-2022
Dissertation: "Systematic Approaches to Machine Learning Security and Privacy Vulnerabilities"
Google DeepMind, 2022-Present
Focused on advanced research in AI safety, model robustness, and privacy preservation techniques
Lead collaborative research on AI security vulnerabilities, developing groundbreaking methodologies for identifying and mitigating potential risks in machine learning systems. Key achievements include:
Pioneering research on adversarial machine learning, developing cutting-edge methodologies for testing and improving the robustness of AI models across multiple domains. Focus areas include:
Conducting groundbreaking work on AI alignment and safety, developing proactive strategies to ensure responsible and ethical AI development. Key research domains:
Researching advanced techniques for privacy preservation and model interpretability, with significant contributions to understanding and mitigating risks in large language models. Primary focus areas:
A groundbreaking investigation into the fundamental privacy vulnerabilities of modern machine learning architectures, demonstrating unprecedented methods of extracting verbatim training data from neural network models. This research represents a critical milestone in understanding the potential privacy risks inherent in large-scale AI training processes.
This work fundamentally reshapes our understanding of AI privacy, compelling major tech companies to reconsider their model training and data protection strategies.
A comprehensive and meticulously designed framework for understanding, analyzing, and mitigating sophisticated adversarial attacks across diverse AI domains, including computer vision, natural language processing, and adaptive learning systems.
Established new paradigms for understanding AI model vulnerabilities, significantly influencing global AI security research and development strategies.