Audio Inference Engineer, Model Efficiency
Cohere
About this role
The Audio Inference Engineer for Model Efficiency at Cohere is responsible for optimizing audio model serving efficiency by enhancing core metrics such as latency, throughput, and quality. This role involves identifying bottlenecks in existing systems and implementing innovative solutions for audio processing and streaming workloads. Collaboration with training and serving infrastructure teams is essential to ensure seamless integration of model development and real-time audio inference deployment. Candidates should have significant experience in high-performance audio or machine learning inference systems, with proficiency in C++ and Python, and a focus on achieving results.
Skills
About Cohere
cohere.comCohere is an AI company that builds large language models and enterprise AI platforms for businesses and developers.
Recent company news
Nvidia-Backed Cohere Forms AI Alliance With Telecom Firm BCE
15 hours ago
Enterprise AI startup Cohere tops revenue target as momentum builds to IPO: Investor memo
1 month ago
Cohere joins Aston Martin Aramco as Official Generative AI Partner to help accelerate AI innovation
2 weeks ago
The AI Model Race May Have Slowed Down for Cohere
Nov 17, 2025
Cohere Technologies drives ahead with innovation vision
1 week ago
About Cohere
Headquarters
San Francisco, CA
Company Size
201-500 employees
Founded
2018
Industry
Technology
Glassdoor Rating
4.2 / 5
Leadership Team
Sarah Johnson
Chief Executive Officer
Michael Chen
Chief Technology Officer
Emily Williams
VP of Engineering
David Rodriguez
VP of Product
Jessica Thompson
Chief Financial Officer
Andrew Park
VP of Sales
Unlock Company Insights
View leadership team, funding history,
and employee contacts for Cohere.
Salary
$179k – $239k
per year
More jobs at Cohere
Similar Jobs
Senior Software Engineer - Model Performance
Inference
Member of Technical Staff - ML Systems & Inference
Gimlet Labs
Principal Software Engineer - AI Inference
NVIDIA
Lead ML Inference Engineer, Advertising
Roku
Engineering Manager, Inference Routing and Performance
Anthropic
AI / ML Platform Engineer
Whatnot