The Generative AI Innovation Center at AWS is at the forefront of empowering customers to leverage cutting-edge AI technologies for transformative business opportunities. As a Machine Learning Engineer, you'll join a multidisciplinary team of strategists, scientists, engineers, and architects who collaborate with customers across industries to fine-tune and deploy customized generative AI applications at scale.
Your role will focus on driving the development of custom Large Language Models (LLMs) across languages, domains, and modalities. You'll be responsible for designing and implementing distributed training pipelines using advanced tools like Fully Sharded Data Parallel (FSDP) and DeepSpeed, ensuring scalability and efficiency at massive scale.
Key responsibilities include adapting LLMs through continued pre-training, fine-tuning, and Reinforcement Learning with Human Feedback (RLHF). You'll work on optimizing AI models for deployment on AWS's custom AI accelerators, including Inferentia and Trainium, leveraging the AWS Neuron SDK and developing custom kernels for enhanced performance.
AWS values diverse experiences and work-life harmony. The company offers comprehensive benefits, mentorship opportunities, and a culture of continuous learning. You'll be part of AWS Global Services, working alongside technical experts from dozens of countries who help customers achieve more with the AWS cloud.
This role offers an exceptional opportunity to innovate at the cutting edge of AI technology, working with top AWS clients to deliver next-generation AI solutions. You'll be part of a team that values curiosity, inclusion, and continuous professional growth, with access to extensive knowledge-sharing resources and career advancement opportunities.