The Generative AI Innovation Center at AWS is at the forefront of empowering customers with cutting-edge AI technologies for transformative business opportunities. As a Machine Learning Engineer, you'll join a multidisciplinary team of strategists, scientists, engineers, and architects working on state-of-the-art Large Language Models (LLMs).
Your role will involve driving the development of custom LLMs across various languages, domains, and modalities. You'll be responsible for implementing distributed training pipelines using advanced tools like Fully Sharded Data Parallel (FSDP) and DeepSpeed, ensuring scalability and efficiency at massive scale. The position requires expertise in fine-tuning state-of-the-art LLMs and optimizing them for high-performance deployment on AWS's custom AI accelerators.
Working directly with enterprise customers and foundational model providers, you'll help understand and solve complex business and technical challenges through tailored generative AI solutions. The role offers unique opportunities to innovate with AWS's custom silicon (Inferentia and Trainium) and the AWS Neuron SDK.
AWS values diverse experiences and work-life harmony, offering comprehensive benefits including medical coverage, financial benefits, equity compensation, and sign-on payments. The company promotes an inclusive culture through employee-led affinity groups and ongoing learning experiences. Career growth is supported through knowledge-sharing, mentorship, and various professional development resources.
This position represents an opportunity to work at the cutting edge of AI technology while contributing to solutions that will shape the future of enterprise AI applications. The role combines technical expertise with customer interaction, offering a balance of hands-on development and strategic problem-solving in a rapidly evolving field.