The Generative AI Innovation Center at AWS is at the forefront of AI innovation, empowering customers to leverage cutting-edge AI technologies for transformative business opportunities. As a Machine Learning Engineer, you'll join a multidisciplinary team of strategists, scientists, engineers, and architects working on state-of-the-art Large Language Models (LLMs).
Your role will involve developing custom LLMs across various languages, domains, and modalities, while optimizing them for deployment on AWS's custom AI accelerators. You'll be responsible for designing and implementing distributed training pipelines using advanced tools like Fully Sharded Data Parallel (FSDP) and DeepSpeed, ensuring scalability and efficiency at massive scale.
The position offers unique opportunities to work directly with enterprise customers and foundational model providers, understanding their business challenges and co-developing tailored generative AI solutions. You'll be involved in adapting LLMs through continued pre-training, fine-tuning, and Reinforcement Learning with Human Feedback (RLHF).
AWS values diverse experiences and maintains an inclusive team culture through employee-led affinity groups and ongoing learning experiences. The company offers strong career growth opportunities with mentorship programs and knowledge-sharing resources. Work-life harmony is prioritized, with flexibility as part of the working culture.
This role requires strong software development experience, deep understanding of machine learning methods, and hands-on experience with generative AI technology. You'll be working with AWS's custom silicon (Inferentia and Trainium) and the AWS Neuron SDK to optimize model performance, making this an ideal position for those passionate about pushing the boundaries of AI technology while delivering practical business solutions.