Your search has found 22 jobs

Training builds capability. Post-training decides what it becomes.

This team are rethinking how large multimodal models learn after pre-training — developing post-training and reinforcement learning methods that help models reason, plan, and interact in real time.

Founded by the researchers behind several of the most influential modern AI architectures, this lab are pushing alignment and learning efficiency beyond standard RLHF. They’re scaling preference-based training (RLHF, DPO, hybrid feedback loops) to new model types and creating systems that learn from interaction rather than static data.

You’ll work at the intersection of post-training, RL, and model architecture — designing reward models, scalable evaluation frameworks, and training strategies that make large-scale learning measurable and reliable. It’s applied research with direct impact, supported by serious compute and a tight researcher-to-GPU ratio.

You’ll bring experience in large-scale post-training or reinforcement learning (RLHF, DPO, or SFT pipelines), a solid grasp of LLM or multimodal training systems, and the curiosity to explore new optimisation and alignment methods. A publication record at top venues (NeurIPS, ICLR, ICML, CVPR, ACL) is a plus, but impact matters more than titles.

The team are based in San Francisco, working mostly in person. $1 million+ total compensation. Base salary circa $300K – $600K (negotiable) plus stock and bonus — exact package depends on experience.

If you want to work where post-training meets architecture — shaping how foundation models learn, reason, and adapt — this is that opportunity.

All applicants will receive a response.

Location: San Francisco
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 11/02/2026
Job ID: 34012

GPU Optimisation Engineer — Real-Time Inference

Want to push GPU performance to its limits — not in theory, but in production systems handling real-time speech and multimodal workloads?

This team is building low-latency AI systems where milliseconds actually matter. The target isn’t “faster than baseline.” It’s sub-50ms time-to-first-token at 100+ concurrent requests on a single H100 — while maintaining model quality.

They’re hiring a GPU Optimisation Engineer who understands GPUs at an architectural level. Someone who knows where performance is really lost: memory hierarchy, kernel launch overhead, occupancy limits, scheduling inefficiencies, KV cache behaviour, attention paths. The work sits close to the metal, inside inference execution — not general infra, not model research.

You’ll operate across the kernel and runtime layers, profiling large-scale speech and multimodal models end-to-end and removing bottlenecks wherever they appear.

What you’ll work on

  • Profiling GPU bottlenecks across memory bandwidth, kernel fusion, quantisation, and scheduling

  • Writing and tuning custom CUDA / Triton kernels for performance-critical paths

  • Improving attention, decoding, and KV cache efficiency in inference runtimes

  • Modifying and extending vLLM-style systems to better suit real-time workloads

  • Optimising models to fit GPU memory constraints without degrading output quality

  • Benchmarking across NVIDIA GPUs (with exposure to AMD and other accelerators over time)

  • Partnering directly with research to turn new model ideas into fast, production-ready inference

This is hands-on optimisation work across the stack. No layers of bureaucracy. No “platform ownership” theatre. Just deep performance engineering applied to models that are actively evolving.

What tends to work well

  • Strong experience with CUDA and/or Triton

  • Deep understanding of GPU execution (memory hierarchy, scheduling, occupancy, concurrency)

  • Experience optimising inference latency and throughput for large generative models

  • Familiarity with attention kernels, decoding paths, or LLM-style runtimes

  • Comfort profiling with low-level GPU tooling

The company is revenue-generating, its models are used by global enterprises, and the SF R&D team is expanding following a recent raise. This is growth hiring, not backfill.

Package & location

  • Base salary: up to ~$300,000 (negotiable based on depth)

  • Equity: Meaningful stock

  • Location: San Francisco preferred (relocation and visa sponsorship can be provided)

If you care about real-time constraints, GPU architecture, and squeezing every last millisecond out of large models, this is worth a conversation.

All applicants will receive a response.

Location: San Francisco, CA
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 11/02/2026
Job ID: 34843

Applied Scientist – Vision Language Models (Multimodal Reasoning)

Ready to build VLMs that go beyond captioning and simple grounding?

This role is centred on advancing vision-language models that power intelligent agents operating in complex, real-world environments. The focus is firmly on multimodal model design, training, and post-training, with a mix of computer vision.

As an Applied Scientist, you’ll work on large multimodal models that integrate visual inputs with language-based reasoning. You’ll explore how VLMs can move from recognition and description toward structured understanding, task execution, and agentic decision-making.

Your work will include designing model architectures, improving cross-modal alignment, and developing post-training strategies that strengthen reasoning, factual consistency, and controllability. You’ll contribute across the full lifecycle, from data curation and supervised fine-tuning through to preference optimisation and evaluation.

This is a research-heavy role with clear production impact. You’ll prototype new ideas, run rigorous experiments, and collaborate with engineering teams to deploy models into live agent workflows.

Your focus will include:

  • Training and fine-tuning large-scale vision-language models
  • Improving multimodal alignment between image and text representations
  • Applying post-training techniques such as SFT, RLHF, DPO, and reward modelling
  • Designing evaluation frameworks for reasoning quality, grounding accuracy, and robustness
  • Working with large multimodal datasets, including synthetic and proprietary data

Hands-on work with VLMs or multimodal foundation models is essential. Experience in post-training, alignment, or preference learning is highly valued.

A solid understanding of how to evaluate multimodal systems, including hallucination, grounding failures, and reasoning gaps, is important. You should be comfortable reading and implementing recent research, and designing experiments that move models forward in measurable ways.

You’ll have ownership over modelling decisions and the opportunity to influence how multimodal intelligence is shaped within a fast-growing AI team.

Compensation: $200,000 - $320,000 base (negotiable depending on level) + bonus + meaningful equity + benefits

Location: SF Bay Area (Hybrid). Remote flexibility in the short term.

If you’re motivated by pushing vision-language models toward deeper reasoning and real-world capability, we’d like to speak with you!

All applicants will receive a response.

Location: United States
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 11/02/2026
Job ID: 33847

Research Engineer – Computer Vision & Machine Learning

Want to build vision systems that let machines understand the physical world as naturally as we do?

This role sits within a highly technical team developing a new class of computing devices where perception, language, and interaction are tightly integrated. Vision is a core capability. Your work will directly influence how machines see, reason about space, and collaborate with humans in real-world environments.

You’ll join a specialist vision group working across 3D computer vision and machine learning. The problems sit at the boundary between learned models and physical reality, including gaze tracking, SLAM, multi-camera geometry, and systems that explicitly model optics, refraction, and light transport. The focus is on geometry-aware, physically grounded approaches rather than purely pixel-driven modelling.

This is a hands-on research engineering role. You’ll move between reading papers, building and training models, designing datasets, running controlled experiments, and deploying onto real hardware. You’ll work closely with firmware and hardware teams to ensure models operate reliably on-device.

Your work will include:

  • Developing ML models across 3D perception, tracking, and spatial understanding

  • Designing model architectures, training pipelines, evaluation frameworks, and inference systems

  • Working with large-scale, multi-camera and sensor-rich datasets

  • Translating state-of-the-art research into robust, production-ready systems

  • Creating new approaches when existing methods do not meet performance or physical constraints

You’ll have genuine technical ownership. The team values clear thinking, strong experimental discipline, and the ability to make informed bets on promising ideas.

You’ll likely bring end-to-end experience building computer vision and ML models, alongside strong familiarity with modern research in 3D or geometry-aware vision. Hands-on experience with PyTorch or JAX is expected, as is comfort working with complex datasets. The ability to operate independently in ambiguous environments is important, as is clear communication across research, hardware, and product teams.

A Bachelor’s degree or higher in computer science, machine learning, computer vision, applied mathematics, or a related field is required. A Master’s or PhD is a plus, particularly if you’ve worked on geometry-aware or physically informed modelling approaches. Experience deploying ML systems into real products or working in high-ownership startup environments would be valuable.

Compensation: $190,000 - $320,000 base (depending on experience) + equity
Benefits: 401(k) matching, 100% employer-paid health, vision, and dental insurance, unlimited PTO and sick time, medical FSA matching
Location: San Francisco, on-site collaboration required

If you’re motivated by building geometry-aware vision systems that connect AI to the physical world in meaningful ways, we’d like to hear from you!

All applicants will receive a response.

Location: San Francisco, CA
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 11/02/2026
Job ID: 34942

Lead Evaluation Engineer — Speech & Multimodal Models

How do you measure whether an AI voice truly sounds real — and prove it with data?

You’ll join an AI team developing large-scale speech and multimodal systems for real-time interaction — models that generate, clone, and understand voice with natural expression and precision.

This is a founding evaluation role, in a new dedicated Evals team defining how these models are measured, improved, and deployed safely at scale. You’ll design objective and subjective evaluation pipelines, run large-scale human studies, and build automated systems that turn perception into measurable signal.

Your work will span every stage of model development — from research to production — collaborating with speech, audio, and ML teams to close the loop between modelling, feedback, and user experience.

What you’ll do:
• Build and scale evaluation pipelines for TTS, voice conversion, and ASR systems
• Design human studies for subjective testing (e.g. MOS, ABX)
• Define and implement objective metrics (WER, intelligibility, naturalness, prosody)
• Automate evaluation dashboards and reporting systems
• Train auxiliary models to capture new evaluation dimensions
• Collaborate across data, model, and product teams to drive measurable improvement
• Establish and scale the evaluation function as the team grows

You’ll bring:
• Strong experience building or running eval systems for speech or multimodal models
• Familiarity with ASR, TTS, or voice cloning pipelines
• Experience designing user studies or subjective model evaluation
• Solid understanding of statistics and experimental design
• Proficiency in Python and ML frameworks (PyTorch, Hugging Face, etc.)
• Strong communication skills and cross-functional mindset

Why this role:
This is a rare chance to build the evaluation foundation for models already deployed globally — shaping how next-generation speech systems are measured and improved. You’ll have the autonomy to define standards, lead future hires, and see your work directly impact millions of real-world interactions.

Fully remote (EU timezones preferrred), global team. Competitive salary + meaningful stock options.

The company are well funded, with a 9 figure funding round and significant runway for meaningful growth, lots of compute and hiring! 

Apply today. Everyone will get a response.

Location: Remote
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 04/02/2026
Job ID: 34414

Teach AI how to reason — safely, transparently, and at scale.

How do we move beyond pattern-matching into true machine reasoning? This Applied Scientist role puts you at the centre of that challenge — developing models that can reason, explain their logic, and make verifiable decisions across complex, high-stakes industries.

You’ll join a well-funded startup building domain-specific reasoning systems and agentic AI for sectors like medtech, aerospace, advanced manufacturing  — where reliability and interpretability aren’t optional.

Your work will focus on post-training large multimodal models, applying the latest techniques in RLHF, DPO, and preference learning to make AI systems more consistent, factual, and aligned with human reasoning. You’ll design the frameworks that turn raw model potential into transparent, trustworthy intelligence.

You’ll develop and optimise post-training pipelines, implement reward modelling for reasoning depth and factual accuracy, and build evaluation frameworks for verifiable, human-aligned behaviour. Working with proprietary and synthetic datasets, you’ll run end-to-end experiments and deploy your methods directly into production.

You’ll bring a background in transformer-based model training (LLM, VLM, MLLM), post-training or alignment (RLHF, DPO, reward modelling), and strong practical skills in Python and PyTorch. Curiosity about reasoning agents, hybrid learning, and interpretability research will help you thrive here.

Bonus points for experience in multimodal reasoning, evaluation and verification, or prior research contributions in alignment or reasoning systems.

The company has raised $20M+ (Series A announcement imminent) and already partners with Fortune 100 and 500 customers. Founded by an entrepreneur with a prior billion-dollar exit, the AI team alone is scaling from 11 to 40+ this year.

Comp: $200K–$320K base (negotiable depending on experience) + bonus + stock + benefits
Location: SF Bay Area (remote for now; hybrid later in 2026)

If you’re excited about defining how AI systems reason, decide, and explain themselves — we’d love to hear from you.

All applicants receive a response.

Location: San Francisco Bay Area
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 26/01/2026
Job ID: 34909

Looking to push the boundaries of generative AI for real-time interaction?

You'll be joining a well-funded startup working on Multimodal AI where voice, vision, and language come together. 

They're building generative models for natural conversational experiences that need to perform in real-time.

Your mission

You'll be building and optimising diffusion or flow-matching models that power their speech and audio generation. 

This means developing production-ready architectures that can generate controllable, high-quality output at scale.

You'll own the full research-to-production pipeline - from architecture design and training through deployment and optimisation. 

Your work will directly impact how millions of AI characters sound and interact.

Your focus

  • Design and train large-scale diffusion or flow-matching models
  • Develop novel architectures and training techniques to improve controllability and quality
  • Build evaluation systems to measure generation quality and model behaviour
  • Work from low-level performance optimisations to high-level model design

What you'll bring

  • Proven track record building diffusion models or flow-matching systems
  • Experience training large models (3B+ parameters) with distributed systems

Nice to have

  • Experience with audio or speech generation
  • Publications or open-source contributions in diffusion models or generative AI

Remote in Europe with competitive comp + stock.

Location: Remote
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 26/01/2026
Job ID: 34280

Want to own complex 3D vision systems that operate outside controlled conditions?

This role is for someone who enjoys building perception systems that have to perform reliably on real data, under real constraints. 

You’ll join a small engineering group where senior ICs take responsibility for designing, evolving, and maintaining core 3D vision capabilities end-to-end.

There’s very little process overhead. Engineers are trusted to define solutions, make decisions, and move work forward quickly. The problems are open-ended, technically demanding, and require strong judgement as much as technical depth.

What you’ll do:

  • Own and evolve key parts of a 3D vision pipeline from early design through to deployment
  • Build and refine geometry-heavy vision systems, integrating learned models where appropriate
  • Prototype ideas quickly and evaluate them against realistic, imperfect data
  • Optimise models and pipelines for accuracy, performance, and robustness
  • Push successful approaches into production-quality implementations
  • Work closely with adjacent engineers while independently driving complex technical problems

What you’ll bring:

  • Experience building real 3D vision systems, not just experimenting with libraries
  • Strong Python and/or C++ skills, with care for performance and reliability
  • Comfort working across geometry, optimisation, and modern deep learning approaches
  • Experience tuning systems over time and making pragmatic technical trade-offs
  • Ability to operate effectively without detailed specifications or rigid roadmaps
  • A sharp, curious mindset and the ability to learn quickly in demanding environments

Why people join:

You’ll work alongside a small, elite group of engineers who take their craft seriously and value depth, speed, and ownership. 
The bar is high, the scope is meaningful, and good work is recognised quickly. If you enjoy being trusted with important systems and want to work with peers who are equally strong technically, this environment supports that.

Package

  • Compensation: $200,000 – $500,000 base (negotiable d.o.e) + equity
  • Benefits: Medical, dental, vision, 401(k)
  • Location: San Francisco (on-site)
  • Employment: Full-time

If this sounds like work that interests you, and you want to be working at the cutting edge of technology with an elite team, please apply now!

All applicants will receive a response.

Location: San Francisco
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 12/01/2026
Job ID: 34678

Build the 3D perception that gives AI agents real spatial intelligence.

How do AI systems truly see and reason about 3D geometry? This Applied Scientist role puts you at the centre of that challenge — developing models that bridge the physical world and intelligent reasoning systems.

You’ll join a well-funded startup building AI agents for advanced design and engineering workflows — across manufacturing, aerospace, and medtech. Your work will enable agents to understand CAD data, meshes, and point clouds deeply enough to plan, analyse, and make autonomous decisions.

This is a rare opportunity to establish the 3D foundation within the research team. You’ll define evaluation strategies, model objectives, and technical direction — building models that become the perception backbone for intelligent agents.

What you’ll do:
• Develop models that learn transferable 3D representations across CAD, mesh, and point cloud data
• Handle messy, lossy, real-world data — not just clean synthetic geometry
• Scale training across segmentation, classification, correspondence, and eventually generation
• Design robust evaluation pipelines for continuous performance monitoring
• Work toward a unified 3D foundation model supporting both discriminative and generative tasks

You’ll bring:
• Deep expertise in 3D computer vision (PhD or equivalent experience)
• Strong knowledge of modern 3D architectures (PointNet++, MeshCNN, Gaussian Splatting, Diffusion, VLMs)
• Proven ability training large-scale models in PyTorch
• Strong applied research instincts — turning papers into working systems
• Experience with multimodal or vision-language models

Bonus points:
• Background with CAD data or industrial design workflows
• Experience in robotics, autonomous driving, or AR/VR 3D perception
• Familiarity with SLAM, pose estimation, or differentiable rendering

You’ll join a small, research-driven team with full autonomy and major compute access — free to explore foundational methods while delivering practical impact.

Compensation & location:
• Base salary: $200K–$300K (negotiable by level)
• Up to 20% bonus + stock
• Full medical, dental, and vision coverage
• 401k (3% match) and 20+ vacation days

Based in the SF Bay Area (currently remote, moving hybrid soon).
Applicants must hold valid US work authorisation (US Citizen or Green Card).

If you’re excited about building the 3D understanding that will power the next generation of intelligent agents — we’d love to hear from you.
All applicants will receive a response.

Location: United States
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 07/01/2026
Job ID: 33515

Want to build the large-scale RL environments frontier labs use to train agents that can truly reason and act?

This team are creating complex reinforcement learning environments — simulations where advanced agents learn to plan, adapt, and solve multi-step problems that stretch beyond standard benchmarks. The focus isn’t on training the models themselves, but on building the worlds that make meaningful learning and evaluation possible — the foundation for more capable, aligned systems.

You’ll work end-to-end across environment design, reward dynamics, and scalable simulation — developing the feedback loops that define what “good” looks like for intelligent behaviour. It’s open-ended, research-driven work where the task definition, data, and reward structure are often the hardest and most important problems to solve.

You’ll collaborate closely with researchers tackling unsolved challenges in reinforcement learning and agent behaviour, shaping experiments, scaling infrastructure, and refining how agents learn in the loop.

It suits someone with strong ML and RL experience, deep intuition for agent dynamics, and the curiosity to explore problems that don’t come with clear instructions.

On-site in San Francisco. Compensation up to $300 K base (negotiable, depending on experience) plus equity.

If you want to help build the environments that teach the next generation of AI systems how to think, act, and adapt — we’d love to hear from you.

All applicants will receive a response.

Location: San Francisco, CA
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 06/01/2026
Job ID: 34645