I lead the pretraining research team at Amazon AGI that supports Amazon Nova, focusing on principled scaling, model architecture, optimization, and novel pretraining objectives. My work centers on the co-evolution of algorithms and systems—the critical intersection that makes AI more capable, efficient, and accessible. My team also built the models behind Amazon Q, Titan, and the distributed training infrastructure powering Amazon Bedrock and SageMaker HyperPod.
I built this team from zero, starting in 2018 with a focus on distributed training and shared representations. What started as a small group of tech leaders and hackers grew into the engine behind foundation models serving millions through AWS.
Before that, I shaped the open-source AI ecosystem as VP and PMC Chair of Apache MXNet, where I co-authored the Gluon interface. I founded GluonNLP—the first toolkit to reproduce BERT with record-setting training speeds. I served on the ONNX Steering Committee and co-founded the Python Data API Standards Consortium. I believe accessible tools and open standards are essential for an AI future that benefits everyone.
Throughout this journey, I’ve maintained a core belief: AI should amplify human agency and ingenuity, not replace it. This principle guides my approach to both research and leadership. My leadership philosophy centers on coaching and enabling team members to grow into leaders themselves, creating a multiplier effect that has accelerated our innovation.
I hold an MS in Computer Science from the University of Maryland and a BS from Shanghai Jiao Tong University.