Meta AR/VR Job | AI Research Scientist, Multimodal Machine Learning
Job(岗位): AI Research Scientist, Multimodal Machine Learning
Type(岗位类型): Artificial Intelligence
Citys(岗位城市): Redmond, WA
Date(发布日期): 2024-6-20
Summary(岗位介绍)
At Reality Labs Research (RL-R), our goal is to explore, innovate and design novel interfaces and hardware subsystems for the next generation of virtual, augmented, and mixed reality experiences. We are driving research towards a vision of an always-on augmented reality device that can enable high-quality contextually relevant interactions across a range of complex, dynamic, real-world tasks in natural environments; to achieve this goal, our team draws on and fundamentally advances methods and knowledge from artificial intelligence, multimodal machine learning, computer vision, and signal processing. We are looking for a skilled and motivated researcher with expertise in multimodal machine learning and large scale multimodal representation learning. More broadly, the chosen candidate will work with a diverse and highly interdisciplinary team of researchers and engineers and will have access to cutting edge technology, resources, and testing facilities.
Qualifications(岗位要求)
Bachelor's degree in Computer Science, Computer Engineering, relevant technical field, or equivalent practical experience.
PhD degree or equivalent experience in Computer Science, Mathematics, Engineering or a related field
Publication record in machine learning, AI, computer science, statistics, applied mathematics, data science, or related technical fields
3+ years of experience in one or more of the following areas: Deep Learning, NLP, Multimodal, Computer Vision, Speech
3+ years of experience developing end-to-end ML pipelines with a focus on dataset preprocessing, model development and evaluation, software integration, and real-time deployment
Experience writing software (Python and C/C++) and executing complex machine learning experiments
Research experience in one or more of these areas: NLP, computer vision, Multimodal, Speech
Description(岗位职责)
Define, collaborate, and execute on research that pushes forward the state of the art in multimodal context/content understanding.
Work collaboratively with other research scientists to develop novel solutions and models in service of contextualized AI for augmented reality.
Working on curation of large datasets across many modalities (e.g., vision, audio, eye gaze, IMU, sEMG).
Leading experimental design, implementation, evaluation, and reporting results to enable new capabilities for multimodal machine learning models.
Play a significant role in healthy cross-functional collaboration.
Additional Requirements(额外要求)
Publication record in AI conferences like NeurIPS, EMNLP, ACL, ICCV, CVPR, ICML, NAACL, ICLR, Interspeech, ICASSP.
Programming experience in Python and hands-on experience with frameworks such as PyTorch.
Experience with developing machine learning models at scale from inception to deployment.