Apple's Video Computer Vision organization is seeking an Applied Research Engineer to join their team working on multimodal large language models (LLMs). This role sits at the intersection of cutting-edge AI research and practical applications, focusing on developing real-time on-device Computer Vision and Machine Perception technologies across Apple products. The ideal candidate will contribute to groundbreaking research in multimodal LLMs, integrating various data types including text, image, video, and audio.
The position offers an opportunity to work with state-of-the-art AI technologies while collaborating with cross-functional teams including hardware, software, and machine learning experts. You'll be part of a centralized applied research and engineering organization that balances research innovation with product development to deliver Apple-quality experiences.
The role combines theoretical research with practical implementation, requiring expertise in multimodal LLMs, computer vision, and machine learning. You'll have the chance to influence Apple's sensor and silicon roadmap while working on projects that directly impact product development. The position offers competitive compensation, including base salary, stock options, and comprehensive benefits.
This is an ideal opportunity for someone passionate about AI and machine learning who wants to work at the forefront of technology development. You'll be contributing to innovative solutions that will shape the future of Apple's products while working with some of the industry's best talents in AI and computer vision.