Meta
Research Scientist – Acoustic and Multi-Modal Scene Understanding
Job Description
Research Scientist – Acoustic and Multi-Modal Scene Understanding Responsibilities:
- Design innovative solutions for challenging multi-modal egocentric recognition problems with resource constraints
- Communicate research results internally and externally in the form of technical reports and scientific publications
- Experience of consistently working under own initiative implementing state of the art models and techniques on Pytorch, Tensorflow or other platforms, seeking feedback and input where appropriate
- Identify, motivate, and execute on reasonable medium to large hypotheses (each with many tasks) for model improvements through data analysis, and domain knowledge, with capacity to communicate learnings effectively.
- Design, perform, and analyze online and offline experiments with specific and well thought-out hypotheses in mind.
- Generate reliable, correct training data with great attention to detail.
- Identify and debug common issues in training machine learning models such as overfitting/underfitting, leakage, offline/online inconsistency
- Aware of common systems considerations and modeling issues, and factor this into modeling choices.
- Design acoustic or audio-visual models which can have a small computational footprint on mobile devices and wearables such as smart glasses.
Minimum Qualifications:
- Currently has a PhD or a postdoctoral assignment in the field of deep learning, Machine Learning, Computer Vision, Computer Science, Computer Engineering or Statistics or a related field.
- 4+ years experience with development and implementation of signal processing and deep learning algorithms in the fields of acoustic and multi-modal detection, recognition and/or tracking problems.
- 4+ years experience with scientific programming languages such as Python, C++, or similar.
- 3+ years experience with research-oriented software engineering skills, including fluency with machine learning (e.g., PyTorch, TensorFlow, Scikit-learn, Pandas) and libraries for scientific computing (e.g. SciPy ecosystem).
- Must obtain work authorization in the country of employment at the time of hire, and maintain ongoing work authorization during employment.
- Demonstrated experience of implementing and evaluating end-to-end prototypical learning systems
- Ability to independently resolve most online and offline issues which affect the hypothesis testing
- Understand the model architecture used, and the consequences of this for different hypotheses tested. In general, you have a good understanding of computer vision from an applied perspective, even though you may not be up to date with the state of the art.
- Experience in communicating effectively with a broad range of stakeholders and collaborators at different levels
Preferred Qualifications:
- Experience with audio-visual learning, computer vision, source localization and tracking, audio and visual SLAM systems, egocentric multimodal learning, etc.
- Experience with building low-complexity models on acoustic and multi-modal problems aimed at low-power mobile devices and wearables
- Experience with integration of development models on real-time running mobile platforms with different levels of compute (on-sensor computation, system on chip, low power island, etc.)
- Experience with acoustic localization or visual multi-object tracking problems
- Proven track record of achieving significant results and innovation as demonstrated by first-authored publications and patents.
About Meta:
Meta builds technologies that help people connect, find communities, and grow businesses. When Facebook launched in 2004, it changed the way people connect. Apps like Messenger, Instagram and WhatsApp further empowered billions around the world. Now, Meta is moving beyond 2D screens toward immersive experiences like augmented and virtual reality to help build the next evolution in social technology. People who choose to build their careers by building with us at Meta help shape a future that will take us beyond what digital connection makes possible today—beyond the constraints of screens, the limits of distance, and even the rules of physics.
Individual compensation is determined by skills, qualifications, experience, and location. Compensation details listed in this posting reflect the base hourly rate, monthly rate, or annual salary only, and do not include bonus, equity or sales incentives, if applicable. In addition to base compensation, Meta offers benefits. Learn more about benefits at Meta.