Your search has found 2 jobs

Want to own the inference layer behind millions of real-world voice AI interactions every day?

You’ll join a profitable, founder-led enterprise conversational AI company powering billions of interactions annually across 30+ languages. Their systems sit behind major global brands and handle millions of customer conversations daily.

They’re now moving toward end-to-end multimodal and speech-to-speech architectures. You’ll own the inference stack powering both their multimodal speech-text LLM and their text reasoning LLM.

This goes well beyond tuning configs.

You will:

• Optimise production inference across A10, A100 and H100 GPUs
• Own scheduler design, KV cache allocation and batching logic
• Build serving systems tailored to multimodal audio-text workloads
• Support agentic, multi-step reasoning under real latency constraints
• Profile kernel-level bottlenecks and fix them properly

You’ve modified inference framework internals before, not just used them. You’re comfortable in Python and C++, and you’re happy diving into CUDA graphs, memory bandwidth limits or custom kernels when required.

This platform processes over 2 million interactions per day. Latency, throughput and cost are production realities, not lab metrics.

Package: €150,000 base + bonus + stock options
Location: Remote within Europe

If you want full ownership of inference performance at real enterprise scale, let’s talk.

All applicants will receive a response.

Location: Remote
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 03/03/2026
Job ID: 35239

How do you make a large language model genuinely human-centred, capable of reasoning, empathy, and nuance rather than just pattern-matching?

This team is built to answer that question. They’re a small, focused group of researchers and engineers working on the post-training challenges that matter most: RLHF, RLAIF, continual learning, multilingual behaviour, and evaluation frameworks designed for natural, reliable interaction.

You’ll work alongside a team from NVIDIA, Meta, Microsoft, Apple, and Stanford, in an environment that combines academic rigour with production-level delivery. Backed by over $400 million in funding, they have the freedom, compute, and scale to run experiments that push beyond the limits of standard alignment research.

This is a role where your work moves directly into deployed products. The team’s models are live, meaning every insight you develop, every method you refine, and every experiment you run has immediate, measurable impact on how large-scale conversational systems behave.

What you’ll work on

  • Developing post-training methods that improve alignment, reasoning, and reliability

  • Advancing instruction-tuning, RLHF/RLAIF, and preference-learning pipelines for deployed systems

  • Designing evaluation frameworks that measure human-centred behaviour, not just accuracy

  • Exploring continual learning and multilingual generalisation for long-lived models

  • Publishing and collaborating on research that informs real-world deployment

Who this role suits

  • Researchers or recent PhDs with experience in LLM post-training, alignment, or optimisation

  • A track record of rigorous work — published papers, open-source projects, or deployed research

  • Curiosity about how large models learn and behave over time, and how to steer that behaviour safely

  • Someone who values autonomy, clarity of purpose, and research that turns into impact

You’ll find a culture driven by technical depth rather than hype — where thoughtful research is backed by meaningful compute and where the best ideas scale fast.

Location: South Bay (on-site, collaborative setup)
Compensation: $200 000 – $250 000 base + equity + bonus

If you’re ready to work on post-training research that shapes how large language models behave, we’d love to hear from you.

All applicants will receive a response.

Location: Palo Alto
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 20/11/2025
Job ID: 33284