Why This Job is Featured on The SaaS Jobs
This role sits at the intersection of AI infrastructure and SaaS delivery, where model performance becomes a product capability rather than a research metric. Work focused on latency, throughput, and quality for real time and streaming audio inference reflects a common pressure in modern AI SaaS: serving demanding workloads reliably for developer and enterprise use cases.
From a SaaS career perspective, the position builds durable strengths in operating ML systems in production, especially where efficiency directly affects unit economics and customer experience. Collaboration across training and serving teams signals exposure to the full lifecycle from model development to deployment, a pattern that transfers well to other AI first SaaS companies building APIs, agents, search, or generation products.
The role is best suited to engineers who prefer deep systems work and measurable performance outcomes, and who enjoy tracing bottlenecks across software and hardware boundaries. It will appeal to professionals who want to stay close to product reality while working on technically rigorous problems in inference, streaming architectures, and GPU aware optimization.
The section above is editorial commentary from The SaaS Jobs, provided to help SaaS professionals understand the role in a broader industry context.
Job Description
Who are we?
Our mission is to scale intelligence to serve humanity. We’re training and deploying frontier models for developers and enterprises who are building AI systems to power magical experiences like content generation, semantic search, RAG, and agents. We believe that our work is instrumental to the widespread adoption of AI.
We obsess over what we build. Each one of us is responsible for contributing to increasing the capabilities of our models and the value they drive for our customers. We like to work hard and move fast to do what’s best for our customers.
Cohere is a team of researchers, engineers, designers, and more, who are passionate about their craft. Each person is one of the best in the world at what they do. We believe that a diverse range of perspectives is a requirement for building great products.
Join us on our mission and shape the future!
Why this role?
Our team is a fast-growing group of committed researchers and engineers. The mission of the team is to build reliable machine learning systems and optimize audio inference serving efficiency using innovative techniques. As an engineer on this team, you will work on advancing core audio model serving metrics, including latency, throughput, and quality by diving deep into our systems, identifying bottlenecks, and delivering creative solutions for audio processing and streaming workloads.
You’ll collaborate closely with both the training and serving infrastructure teams to ensure seamless integration between model development and deployment, with a special focus on real-time and streaming audio inference.
Please Note: We have offices in Toronto, Montreal, San Francisco, New York, Paris, Seoul and London. We embrace a remote-friendly environment, and as part of this approach, we strategically distribute teams based on interests, expertise, and time zones to promote collaboration and flexibility. You'll find the Model Efficiency team concentrated in the EST and PST time zones, these are our preferred locations.
You may be a good fit for the team if you have:
Significant experience developing high-performance audio or machine learning inference systems.
Proficiency with programming languages such as C++ and Python.
Hands-on experience with deep learning models for audio, speech, or language applications.
A bias for action and a strong results-oriented mindset.
It is a big plus if you also have considerable experience with:
GPU programming, low-level system optimization, model parallelization techniques over multiple GPUs
Have experience with duplex real-time streaming architectures.
Internals of machine learning frameworks for audio (such as PyTorch, TensorFlow, or specialized audio libraries).
Have experience with inference framework like vLLM, SGLang, Tensort-LLM, or custom distributed inference systems
Sequence modeling (e.g., transformers for audio/speech) and end-to-end audio pipeline optimization
If some of the above doesn’t line up perfectly with your experience, we still encourage you to apply!
We value and celebrate diversity and strive to create an inclusive work environment for all. We welcome applicants from all backgrounds and are committed to providing equal opportunities. Should you require any accommodations during the recruitment process, please submit an Accommodations Request Form, and we will work together to meet your needs.
Full-Time Employees at Cohere enjoy these Perks:
🤝 An open and inclusive culture and work environment
🧑💻 Work closely with a team on the cutting edge of AI research
🍽 Weekly lunch stipend, in-office lunches & snacks
🦷 Full health and dental benefits, including a separate budget to take care of your mental health
🐣 100% Parental Leave top-up for up to 6 months
🎨 Personal enrichment benefits towards arts and culture, fitness and well-being, quality time, and workspace improvement
🏙 Remote-flexible, offices in Toronto, New York, San Francisco, London and Paris, as well as a co-working stipend
✈️ 6 weeks of vacation (30 working days!)