Senior Machine Learning Engineer @ Amazon AGI
Scaling multimodal foundation models β optimizing how they learn, generalize, and align through data and systems co-design.
I work at the intersection of machine learning and distributed systems,
designing large-scale learning pipelines and multimodal data systems that improve how foundation models learn from vast, diverse signals.
My focus areas:
- π§ Training dynamics & optimization β improving convergence, stability, and efficiency of large-scale multimodal models
- π§© Learning-centric systems β integrating data, architecture, and feedback to enhance representation learning and model alignment
- βοΈ Scalable orchestration β leveraging Ray, Spark, and Kubernetes to parallelize multimodal workloads across thousands of GPUs
- π Evaluation & feedback loops β automating model-driven data refinement and continual quality signals for alignment and adaptation
My work centers on how models learn, not just how theyβre trained.
1. Models and systems co-evolve.
The best architectures emerge when data, compute, and learning dynamics are designed together.
2. Scale reveals behavior.
Many learning problems only appear β and can only be solved β at massive scale.
3. Data is part of the model.
Every batch defines what the model becomes.
βAt scale, learning is a systems problem β and every system is a hypothesis about how intelligence forms.β
β Duo An



