AWS Neuron is seeking a Software Engineer to join their Machine Learning Applications team, focusing on the development of cloud-scale ML accelerators. This role combines software engineering with machine learning expertise, working on AWS Inferentia and Trainium platforms. The position involves developing and optimizing distributed training solutions for large language models like LLama3 and Mixtral, as well as other ML model families.
The role requires collaboration with chip architects and compiler engineers to build and enhance distributed training support in frameworks like PyTorch and JAX. You'll be responsible for performance tuning and maximizing efficiency of models running on AWS Trainium and Inferentia silicon. The team emphasizes work-life balance and offers significant opportunities for mentorship and career growth.
Amazon provides a comprehensive benefits package including medical coverage and financial benefits. The company strongly values diversity and inclusion, demonstrated through employee-led affinity groups and ongoing learning experiences. The team culture promotes knowledge sharing and professional development, with projects assigned to help team members grow their skills and tackle increasingly complex challenges.
Working at AWS Neuron means being at the forefront of machine learning infrastructure, developing solutions that power some of the most advanced AI models. The role offers competitive compensation based on geographic location and experience, along with equity and additional benefits as part of the total compensation package.