Rev is hiring a Machine Learning Runtime Engineer to translate research-oriented code into efficient, deployable software for AI/ML projects. You will work on core technologies like Automatic Speech Recognition, Speaker Diarization, and Large Language Models, leveraging deep learning and extensive audiovisual and textual data.
What You'll Do
- Work jointly with AI research and AI platform teams to productionize and optimize ML models.
- Implement fast and efficient production quality code.
- Improve and innovate on existing ASR, Diarization, LLM, and NLP infrastructure.
- Profile and optimize existing ASR, LLM, and NLP models.
- Collaborate with researchers to improve testing and deployment strategies.
- Automate and integrate workflows from diverse systems.
- Interact with other teams at Rev working towards shared goals.
What We're Looking For
- A university degree in Computer Science, Software Engineering, or a related field.
- Strong experience in C++.
- Literacy in Python and PyTorch to understand research code and algorithms.
- Experience debugging and identifying computational bottlenecks.
- Deep understanding of software testing, benchmarking, and continuous integration.
- Comfort with open-ended tasks, subjective evaluation, and piloting works for production trials.
- Experience working with AWS services.
- Comfort with hardware infrastructure (CPU/GPU, NVIDIA).
Nice to Have
- Knowledge of SIMD, matrix math, or OpenFst.
- Knowledge of C#.
- Knowledge of machine learning, speech or language processing, data science, or empirical evaluation.
- Knowledge of languages other than English or linguistics.
Technical Stack
- C++
- Python
- PyTorch
- AWS
- NVIDIA GPU
Team & Environment
You will work jointly with AI research and AI platform teams.
Work Mode
This role follows a hybrid work model and is based in Austin, TX.
Rev embraces bold thinking and diverse perspectives, giving teams the freedom and responsibility to innovate towards our mission of understanding the human voice.

