The Generative AI Innovation Center at AWS is seeking a Machine Learning Engineer to drive the development of custom Large Language Models (LLMs). This role sits at the intersection of cutting-edge AI technology and enterprise solutions, where you'll work on implementing distributed training pipelines and optimizing models for AWS's custom AI accelerators. You'll be responsible for fine-tuning state-of-the-art LLMs across various domains and modalities while ensuring high-performance deployment.
The position offers an exciting opportunity to work with AWS's top enterprise clients, helping them leverage generative AI for transformative business outcomes. You'll be part of a multidisciplinary team of strategists, scientists, engineers, and architects, collaborating to deliver customized AI solutions at scale. The role involves hands-on work with advanced technologies like Fully Sharded Data Parallel (FSDP) and DeepSpeed, as well as optimization for AWS Inferentia and Trainium hardware.
Key responsibilities include designing large-scale training pipelines, adapting LLMs through continued pre-training and fine-tuning, and implementing Reinforcement Learning with Human Feedback (RLHF). You'll also work closely with foundational model providers to optimize AI models for Amazon Silicon, enhancing both performance and efficiency.
The ideal candidate should have strong software development experience, deep understanding of machine learning methods, and hands-on experience with generative AI technology. This role offers competitive compensation ranging from $129,300 to $223,600 based on location, plus additional benefits including equity and comprehensive healthcare coverage. Join AWS to shape the future of AI technology and deliver innovative solutions that drive business transformation.