AI/ML & Prompt Engineer LLM, RAG & Voice Agent at QuantumLoopAI

We are redirecting you to the source. If you are not redirected in 3 seconds, please click here.

AI/ML & Prompt Engineer LLM, RAG & Voice Agent at QuantumLoopAI. ABOUT QUANTUMLOOPAI. QuantumLoopAI is transforming how patients access healthcare. Our AI-powered reception system, EMMA, serves millions of patients, helping GP surgeries manage demand, reduce phone queues, and ensure patients get the right care at the right time. Backed by leading healthcare investors and working with stakeholders across England, we are building the future of primary care access. As we prepare for our Series A and expand our team, we are looking for exceptional people who want to make a genuine difference to healthcare.. ABOUT THE ROLE. As our AI/ML & Prompt Engineer, you will own the intelligence behind EMMA — the LLM and RAG systems that power our chatbot, voice agent and telephony services. You will design, build and optimise the AI models and prompt architectures that allow EMMA to understand patient needs, retrieve accurate clinical information, and respond naturally across text and voice channels. This is a hands-on, high-impact role where your work will directly shape patient outcomes across hundreds of GP surgeries.. WHAT YOU WILL DO. Architect and deploy LLM- and RAG-based systems that ingest data sources, generate accurate responses, and power both chatbot and voice agent interactions.. Design, iterate and optimise prompts and voice-agent dialogue flows to maximise response relevance, reduce latency and ensure clinical safety across text and voice channels.. Partner with frontend (Next.js, React) and backend (NestJS, Python) teams to integrate AI and voice agent components into our Azure-hosted microservices architecture.. Instrument AI and voice-agent pipelines, analyse logs and user feedback, troubleshoot edge cases, and implement continuous-learning improvements.. Stay abreast of the latest advances in LLMs, RAG, conversational AI frameworks, and (GDPR, DTAC) to inform our technical roadmap.. Define and execute rigorous test plans in collaboration with QA to validate model accuracy, voice-agent performance and compliance with healthcare standards.. Mentor junior engineers and advocate for best practices in MLOps, prompt engineering, voice-agent design and model governance.. ESSENTIAL. Demonstrable hands-on experience with prompt engineering, LLMs (e.g. GPT, LLaMA, Mistral), RAG frameworks, and voice-agent or dialogue-system design, with shipped projects or open-source contributions.. Advanced proficiency in Python for ML/AI development.. Experience building APIs in NestJS and familiarity with Next.js/React for frontend integration.. Strong understanding of Azure services (App Services, Functions, Cognitive Services), containerisation (Docker), and relational databases (MySQL).. Experience designing microservices, distributed architectures and RESTful or GraphQL APIs.. Excellent written and verbal communication skills, with the ability to articulate complex AI and voice-agent concepts to technical and non-technical stakeholders.. Comfortable working remotely and collaborating asynchronously with a UK-based team during UK business hours (9:00 AM to 6:00 PM GMT/BST).. DESIRABLE. Contributions to open-source LLM or RAG tooling.. Experience with MLOps pipelines, model monitoring and automated evaluation frameworks.. Background in speech-to-text, text-to-speech, or telephony integration.. LOGISTICS. Education: We value skills and experience over formal qualifications.. Working arrangements: Fully remote from India, aligned to UK business hours (9:00 AM to 6:00 PM GMT/BST).. Start date: Flexible.. Company Location: India.