+ years of professional software engineering experience, with a strong track record in AI/ML systems or large-scale infrastructure
Deep expertise in distributed systems architecture: microservices, event-driven architecture, data-intensive applications at massive scale Proven experience building and operating AI/ML platforms or infrastructure (model training, serving, feature engineering, or MLOps) Strong knowledge of deep learning frameworks (PyTorch, TensorFlow, JAX) and ML system internals Proficiency in 2+ programming languages (e.g., Python, Go, C++, Rust, Java, Scala) Extensive experience with cloud-native technologies at scale: Kubernetes, Docker, AWS/GCP/Azure, CI/CD pipelines Familiarity with GPU programming and high-performance computing concepts (CUDA, distributed training, model parallelism) Demonstrated ability to drive cross-team/cross-org technical alignment and influence without authority Excellent communication skills — ability to articulate complex technical concepts to diverse audiences Bachelor’s degree in Computer Science, Engineering, or equivalent practical experience