What you can expect
As an Audio AI Engineer, you will research and develop algorithms for accent conversion, voice conversion, speech synthesis, and speech recognition on low-latency streaming architectures. Youโll prototype and refine end-to-end audio models that enhance intelligibility and naturalness while maintaining speaker identity. Working closely with product and platform teams, youโll help bring these models into real-time communication systems. You will also evaluate and optimize model performance across dimensions such as quality, latency, and scalability. Staying current with advances in speech processing, youโll contribute to innovation through patents and internal knowledge sharing.
โ
About the Team
Zoom's Audio team develops real-time audio features based on AI algorithms. Members of the team are spread worldwide, including the U.S., China and Singapore.
โ
โ
Responsibilities
- Researching, designing, and developing algorithms for accent conversion, voice conversion, speech synthesis, and automatic speech recognition, focusing on low-latency streaming architectures
- Prototyping end-to-end audio models that enhance intelligibility and naturalness while preserving speaker identity and expressiveness.
- Collaborating closely with product and platform teams to integrate models into real-time video and audio communication systems.
- Analyzing and optimizing model performance across speech quality, latency, robustness, and scalability dimensions.
- Staying current with the latest developments in speech processing research, and contribute to the community through patents, and internal knowledge sharing.
โ
โ
What weโre looking for
- Hold a PhD or equivalent experience in a relevant field in Streaming, Voice Conversion, TTS, or ASR.
- Show proficiency in deep learning frameworks like PyTorch or TensorFlow.
- Demonstrate effective programming skills in Python, C/C++, or similar languages.
- Have an understanding of sequence modeling architectures (Transformers, RNNs, diffusion models, or conformers).
- Demonstrate experience developing and deploying low-latency, real-time speech or audio models with streaming architectures and optimized pipelines.
- Show familiarity with model compression and acceleration techniques, including quantization, pruning, and distillation.
- Exhibit experience working with real-time audio systems in networked communication environments.
- Publish in top-tier conferences such as ICASSP, INTERSPEECH, NeurIPS, and ICLR.
- Must be fluent in Mandarin
โ