Taro Logo

Applied Research Engineer - Multimodal LLMs

Apple is a technology company known for innovative products and services, including iPhones, Macs, and AI research.
$143,100 - $264,200
Machine Learning
Senior Software Engineer
In-Person
5,000+ Employees
3+ years of experience
AI · Consumer
This job posting may no longer be active. You may be interested in these related jobs instead:

Description For Applied Research Engineer - Multimodal LLMs

We are seeking a highly motivated and skilled Applied Research Engineer to join our team in the Video Computer Vision org. This centralized applied research and engineering organization is responsible for developing real-time on-device Computer Vision and Machine Perception technologies across Apple products. The ideal candidate will have a strong background in developing and exploring multimodal large language models that integrate various types of data such as text, image, video, and audio.

Key responsibilities include: • Conducting research and development on multimodal large language models, focusing on exploring and utilizing diverse data modalities • Designing, implementing, and evaluating algorithms and models to enhance the performance and capabilities of our AI systems • Collaborating with cross-functional teams, including researchers, data scientists, and software engineers, to translate research into practical applications • Staying up-to-date with the latest advancements in AI, machine learning, and computer vision, and applying this knowledge to drive innovation within the company

We balance research and product to deliver Apple quality, state-of-the-art experiences, innovating through the full stack, and partnering with HW, SW, and ML teams to influence the sensor and silicon roadmap that brings our vision to life.

This role offers the opportunity to work on cutting-edge research projects to advance our AI and computer vision capabilities, contributing to both foundational research and practical applications. Join us in pushing the boundaries of what is possible with foundation models, LLMs, and multimodal LLMs!

Last updated 8 months ago

Responsibilities For Applied Research Engineer - Multimodal LLMs

  • Conduct research and development on multimodal large language models
  • Design, implement, and evaluate algorithms and models to enhance AI systems
  • Collaborate with cross-functional teams to translate research into practical applications
  • Stay up-to-date with the latest advancements in AI, machine learning, and computer vision

Requirements For Applied Research Engineer - Multimodal LLMs

Python
  • Experience in developing, training/tuning multimodal LLMs
  • Programming skills in Python and C++
  • Bachelor's Degree and a minimum of 3 years relevant industry experience
  • Expertise in one or more of: computer vision, NLP, multimodal fusion, Generative AI (preferred)
  • Experience with at least one deep learning framework such as JAX, PyTorch, or similar (preferred)
  • Publication record in relevant venues (preferred)
  • PhD in Computer Science, Electrical Engineering, or a related field with a focus on AI, machine learning, or computer vision (preferred)

Benefits For Applied Research Engineer - Multimodal LLMs

Medical Insurance
Dental Insurance
401k
Equity
Education Budget
  • Comprehensive medical and dental coverage
  • Retirement benefits
  • Discounted products and free services
  • Reimbursement for certain educational expenses
  • Opportunity to become an Apple shareholder through discretionary employee stock programs
  • Employee Stock Purchase Plan
  • Potential for discretionary bonuses or commission payments
  • Potential for relocation assistance

Interested in this job?