Multimodal Generative Modeling Research Engineer - SIML, ISE
Do you believe that generative models can fundamentally change how people interact with devices? Are you excited by the possibility of connecting people to unprecedented levels of creativity? Come join the team that turned cutting edge Generative AI research into compelling user experiences through Apple Intelligence. We are part of the Intelligent System Experiences team at Apple and work on the core technology behind Genmoji, Image Playground and Image Wand. From web-scale datasets to distributed training of foundation models, application specific adapters to on-device optimizations, you will have the unique opportunity to advance and ship state-of-the-art technology that impacts Apple users around the world!
We are looking for senior technical leaders experienced in architecting and deploying production scale multimodal ML systems. An ideal candidate has the ability to lead diverse cross functional efforts including prototyping, ML modeling, and integration into production workflows. Solid ML fundamentals, experience with applied research, and hands-on knowledge training and adapting image/video/audio/multimodal Generative AI models would be essential for this role!
SELECTED REFERENCES TO OUR TEAM’S WORK:
- https://machinelearning.apple.com/research/introducing-apple-foundation-models (https://machinelearning.apple.com/research/introducing-apple-foundation-models)(https://machinelearning.apple.com/research/introducing-apple-foundation-models)
- https://machinelearning.apple.com/research/stable-diffusion-coreml-apple-silicon (https://machinelearning.apple.com/research/stable-diffusion-coreml-apple-silicon)(https://machinelearning.apple.com/research/stable-diffusion-coreml-apple-silicon (https://machinelearning.apple.com/research/stable-diffusion-coreml-apple-silicon))
We are looking for a candidate with proven track record of leading applied ML research. Responsibilities in this role will include creating multimodal Generative AI models that can deliver high quality and enable new capabilities in support of production focussed user experiences. You will do this through distributed training of large scale models involving image/video/audio and research optimizations for deploying efficient models on device. Ensuring high quality in real world use, mitigating bias and preserving privacy are all core tenets. You will be interacting closely and cross-functionally with other ML researchers, software engineers, hardware & design teams.
- M.S. or PhD in Computer Science or a related field such as Electrical Engineering, Robotics, Statistics, Applied Mathematics, or equivalent experience.
- Hands on experience training or adapting image / video / audio generation models (eg: SDXL, Flux, etc) for downstream tasks
- Proficiency in ML frameworks e.g., PyTorch, Tensorflow
- Strong programming skills in Python / other high level languages
- Familiarity with distributed training
- Strong programming skills in C/C++/ObjC
Apple is an equal opportunity employer that is committed to inclusion and diversity. We take affirmative action to ensure equal opportunity for all applicants without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability, Veteran status, or other legally protected characteristics. Learn more about your EEO rights as an applicant.