About the Role
Volume AI is building a production-grade AI voice automation platform with a real-time audio pipeline (<800ms latency).
Our stack includes FastAPI, Node.js (Fastify/WebSockets), React, PostgreSQL, Redis, Celery, Docker, and integrations with multiple AI providers such as Groq, OpenRouter, Deepgram, ElevenLabs, Google Cloud, AWS, and Azure.
This role requires strong hands-on expertise, architectural thinking, and the ability to work across the stack in a fast-paced startup environment.
Responsibilities
Build & scale backend services using FastAPI, PostgreSQL, Redis, Celery, and Docker
Work on the real-time Node.js voice pipeline (Fastify/WebSockets, Twilio Media Streams)
Integrate & optimize LLMs, STT, and TTS providers (Groq, Deepgram, Whisper, ElevenLabs, etc.)
Improve & maintain the React dashboard
Work with AWS (EC2, RDS, S3, ElastiCache) and Google Cloud
Maintain and improve CI/CD workflows using GitHub Actions
Analyze and integrate open-source libraries
Implement monitoring, logging, and performance optimization
Reduce latency and improve the reliability of distributed systems
Take ownership of new technologies introduced as the product evolve
Required Skills
Strong Python expertise (FastAPI + async/await)
Strong ReactJS fundamentals
Advanced SQL & PostgreSQL schema design
Proficiency with Docker, GitHub Actions, and CI/CD pipelines
Experience with Node.js real-time systems (Fastify/WebSockets)
Redis, Celery, and distributed task queue experience
Experience with third-party APIs (Twilio, Google, AWS, Azure)
Understanding of LLMs, Whisper, TTS engines, and AI provider optimizations
Ability to understand & extend open-source codebases
High adaptability and ability to work independentl