Your search has found 27 jobs

Build the foundational models that will give AI agents true 3D understanding

Want to solve the fundamental challenge of how AI systems perceive and reason about 3D geometry?

This Lead Applied Scientist role puts you at the forefront of creating perception capabilities for the next generation of agentic AI systems. You'll lead a team building discriminative and generative models introduced into agentic workflows, solving complex challenges in agentic AI for industrial applications.

You'll be joining a well-funded startup developing AI agents for advanced design and manufacturing workflows. Your work will bridge the gap between the physical world and intelligent reasoning systems, creating models that understand CAD data, meshes, and point clouds at a level that enables autonomous decision-making.

This role offers the opportunity to hands on lead a team to build 3D computer vision capabilities from the ground up. You'll be establishing an entirely new domain within the research team, with significant autonomy to define evaluation strategies, model objectives, and technical direction. Your models will form the perception backbone that enables agents to truly understand and manipulate the 3D world.

Your technical challenges:

  • Build models that understand diverse 3D data types (CAD, mesh, point cloud) and learn transferable representations across formats
  • Handle messy, lossy, or incomplete real-world data - moving beyond clean synthetic geometry to tackle industrial reality
  • Scale training across multiple 3D tasks: segmentation, classification, correspondence, and eventually generation
  • Create evaluation pipelines that meaningfully assess model performance and enable continuous production monitoring
  • Work toward a foundational 3D model supporting both discriminative and generative tasks, integrated into broader agentic AI architecture

Your expertise should include:

  • Deep specialisation in 3D computer vision (ideally including a PhD in Computer Vision)
  • Strong knowledge of modern 3D architectures (PointNet++, MeshCNN, 3D Gaussian Splatting, diffusion models, VLMs)
  • Proven ability training large-scale deep learning models with PyTorch
  • Solid applied research skills - can implement novel architectures from papers and make them work in practice
  • Experience with multimodal or vision-language model development

Nice to have:

  • Background working with CAD data or industrial design workflows
  • Experience in complex topics such as robotics, autonomous driving, or AR/VR with 3D perception focus
  • Familiarity with SLAM, pose estimation, or differentiable rendering

You'll join a research team that values ownership and rapid iteration, with the resources to pursue ambitious technical goals. The company provides abundant compute resources and the freedom to explore foundational approaches whilst ensuring practical impact.

Package includes:

  • Base salary: $300,000
  • Performance bonus up to 20%
  • Medical, dental, and vision coverage
  • 401k with up to 3% company match
  • 20+ vacation days

You'll need to be based in SF Bay Area or Miami, with a collaborative team environment that encourages innovation and technical excellence.

You must have valid right to work in the US without sponsorship (US Citizenship or Green Card).

If you're excited about creating the 3D perception capabilities that will power the next generation of intelligent agents, we'd love to hear from you.

All applicants will receive a response.

Location: United States
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 15/09/2025
Job ID: 33847

Interested in building distributed training infrastructure that powers a frontier-scale superintelligence platform?

This role is about creating the systems that make breakthrough AI research possible. You’ll be working on large-scale training infrastructure for LLMs and multimodal models, building the backbone for models that generate new knowledge across multiple domains.

Your work will cover distributed training systems, performance optimisation, and scalable pipelines that enable complex experiments to run across thousands of GPUs. Instead of maintaining legacy stacks, you’ll be designing the infrastructure that pushes models further — and accelerates real-world progress.

You’ll work closely with researchers tackling problems at the cutting edge, ensuring the systems you build directly support new discoveries. This isn’t just another ML engineering role: it’s about creating the foundation for next-generation AI.

You should have:

  • Proven experience with distributed ML training frameworks.

  • Strong engineering background in Python and C++.

  • Understanding of large-scale model training techniques.

  • Experience in cloud or HPC environments.

Package: $250k–$350k base + equity, full benefits. Onsite in San Francisco, CA or Boston, MA.

If you want your engineering to enable genuine breakthroughs — not just optimise another product pipeline — this role is for you.
All applicants will receive a response.

Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 15/09/2025
Job ID: 33781

Want to build simulated RL environments that push frontier models to their limits?

This role is about advancing the science of post-training, reinforcement learning, and scalable evaluation. Instead of static benchmarks, you’ll create dynamic simulations that probe reasoning, planning, and long-horizon behaviour — work that defines how the next generation of AI will be trained and supervised.

You’ll design new post-training algorithms (RLHF, DPO, GRPO and beyond), develop reward models that move beyond exact-match signals, and publish your findings while seeing them deployed in production systems. The work spans both core research and practical implementation, giving you the chance to shape frameworks already being adopted by industry leaders.

We’re looking for:

  • Research experience in post-training or RL methods with LLMs.

  • Strong background in transformers and evaluation frameworks.

  • Publication record at top venues (NeurIPS, ICLR, ICML, ACL, EMNLP).

  • PhD in CS/ML/NLP/RL or equivalent research experience.

Package: Up to $300k base (DOE) + meaningful equity, with comprehensive benefits, 401k, unlimited PTO, relocation support and sponsorship available. Location is San Francisco preferred, with NYC also considered.

Ready to help define how AI learns and is evaluated in simulated environments?
All applicants will receive a response.

Location: San Francisco
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 22/08/2025
Job ID: 33119

Looking to solve complex ASR challenges at scale?

You'll be joining an established conversational AI company with proprietary in-house speech models that process billions of interactions annually, where your ASR expertise will directly impact real-world customer experiences.

You'll be tackling demanding ASR problems in production environments: streaming speech recognition in noisy conditions, robust accent handling and maintaining high performance at scale.

Working across key production environments, you'll enhance speech capabilities and bring new features to production that push the boundaries of what's possible in challenging acoustic environments.

 

Your focus

  • Maintain and iteratively improve existing ASR technology while introducing cutting-edge enhancements
  • Work end-to-end across speech processing components: speech enhancement, VAD, diarisation, and ASR (AM/LM modelling, ASR biasing)
  • Build streaming ASR systems optimised for challenging acoustic environments
  • Implement emotion detection and acoustic condition classification capabilities
  • Run extensive experiments to advance activity detection and speech processing performance

 

What you'll bring

  • Strong background in ASR model development and deployment
  • Hands-on experience with SOTA speech toolkits (Kaldi, K2, NVIDIA NeMo, Parakeet)
  • Proven streaming ASR experience in production environments

 

This is a fully remote role – must be close to EU timezone. 

Ready to make your mark on speech tech that millions rely on daily? Apply today.

Location: Remote
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 12/08/2025
Job ID: 33647

Want to work with AI agents used for good?

Join a well-funded AI company developing next-generation foundation models specifically for healthcare. With over $250 million backing from top-tier investors, they're creating agents, foundation language models, and real-time empathetic and nartural conversational AI that prioritise safety and medical accuracy.

You'll join a world-class research team comprised of ex-researchers from Microsoft, Meta, NVIDIA, Apple, Stanford, Johns Hopkins, and HuggingFace. This role offers the opportunity to work on AI technology that could transform healthcare outcomes for millions worldwide.

Your focus:

  • Developing and improving LLMs for healthcare applications, directly contributing to agent architecture and conversational AI systems
  • Implementing advanced post-training techniques including instruction tuning and RLHF
  • Building systems capable of real-time patient interactions with medical-grade accuracy
  • Advancing multilingual capabilities for global healthcare accessibility
  • Researching continual learning approaches that evolve through patient interactions
  • Collaborating with researchers from top universities and leading tech companies

Your expertise:

  • Hands-on experience training multi-billion parameter large language models
  • Strong background in NLP with deep understanding of transformer architectures
  • Experience as an Applied researcher with excellent Python and PyTorch coding skills
  • Expertise in reinforcement learning, RLHF, LLM post-training, and supervised fine-tuning
  • Experience with large-scale data processing and distributed computing

Nice to have:

  • Healthcare or life sciences background
  • Publications at top-tier AI conferences
  • Experience with multimodal AI systems

Your package:

  • Base salary: $200,000-$350,000 (based on experience)
  • Stock package
  • Comprehensive benefits including healthcare, dental, and vision
  • 401k with company contributions

Based near Palo Alto, CA with strong preference for on-site collaboration.

This isn't just another AI research role - it's the chance to build technology that could fundamentally improve healthcare delivery globally. Your work will contribute to creating AI systems that assist with patient care and improve healthcare accessibility.

Ready to help shape the future of healthcare through AI? Apply today. All applications will receive a response.

Location: Palo Alto
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 16/07/2025
Job ID: 33284

Ready to pioneer the speech intelligence behind the next generation of embodied AI?

Join a pioneering startup developing foundational technology for natural conversation in embodied agents. You'll advance the speech systems that power avatars with authentic behaviours, real-time expression, and conversational intelligence that handles interruptions and turn-taking just like humans.

This Lead Research Scientist role focuses on advancing real-time speech systems for interactive avatars. You'll develop full-duplex dialogue models and speech-to-speech architectures that enable natural conversational flow, interruption handling, and emotional expression.

Founded by ex-Googlers, they're building proprietary behaviour models that learn from two-way interactions, creating systems where speech timing, prosody, and contextual responses work in harmony with facial expressions and physical behaviours to drive authentic embodied intelligence.

Your focus:

  • Research & develop full-duplex speech systems with natural interruption handling
  • Develop expressive voice models with controllable prosody and timing
  • Build speech-to-speech architectures preserving identity and emotion
  • Create real-time audio generation systems for conversational avatars
  • Publish research while deploying systems in production
  • Collaborate across teams integrating speech with visual behaviour

Requirements:

  • PhD in Speech, Machine Learning, or related field
  • First-author publications at top conferences (Interspeech, ICASSP, NeurIPS, ICLR, etc)
  • Expertise in text-to-speech, speech-to-speech models, or voice cloning
  • Large-scale training experience
  • Experience in prosody modelling or real-time audio generation

Nice to have:

  • Experience with full-duplex speech research
  • Speech-visual alignment expertise (lip sync, expressions)
  • Real-time audio deployment optimisation

Package:

  • Competitive salary $200k- $300k base (based on experience)
  • Meaningful equity package
  • Comprehensive healthcare (90% covered)
  • Unlimited PTO
  • Fully remote work with regular team offsites
  • Life insurance and disability coverage

Location: Fully remote position, globally, with preference for Pacific Time alignment.

Ready to make AI conversations feel authentically human?

Contact Allys at Techire AI. All applicants will receive a response.

 
Location: Remote
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 02/07/2025
Job ID: 33482

Ready to architect the future of human-computer voice interaction?

Join a well established conversational AI company as they transition from traditional cascaded speech systems to cutting-edge E2E speech-to-speech technology. You'll lead this transformation, building multimodal systems that will redefine how millions interact with AI.

The opportunity

You'll be developing technology that directly impacts real users at massive scale. The company processes millions of daily interactions across major enterprise clients, meaning your research will shape real-world conversational experiences.

You'll spearhead the development of proprietary full-duplex speech systems, creating truly natural AI conversations that go far beyond current capabilities.

Your impact

  • Design and build next-generation speech language models from the ground up
  • Drive breakthroughs in speech-to-speech modeling and full-duplex conversation systems
  • Tackle turn-taking, interruption handling, and simultaneous speech processing
  • Bridge cutting-edge research with enterprise-grade production systems
  • Lead a growing team focused on speech-to-speech breakthroughs

What you'll bring

  • Deep understanding of SOTA speech models and neural audio processing
  • Experience building speech language models/multimodal systems
  • Strong background in speech AI research and modern speech architectures

With their established market position and proven track record, you'll have the resources and real-world testing ground to make transformative impact with your research.

The company has built everything in-house, giving you complete technical control and the freedom to explore any approach that delivers value.

Location

Remote (Must be close to EU timezone)

Location: Remote
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 28/06/2025
Job ID: 33350

Build production systems that bring 3D AI models to life in real-world applications

Ready to bridge cutting-edge 3D computer vision research with robust, scalable production systems? This ML Engineer role focuses on deploying 3D perception models into live agentic workflows where reliability and performance are paramount.

You'll be joining a well-funded startup developing AI agents for advanced design and manufacturing. Your role centres on creating the infrastructure that makes 3D understanding truly practical - from real-time inference pipelines to comprehensive monitoring systems that ensure geometry-aware agents perform reliably in production.

This position offers the opportunity to shape how 3D AI models integrate into agent decision-making pipelines. You'll work closely with applied scientists to productionise breakthrough research whilst building robust systems that handle the unique challenges of geometric data in mission-critical applications.

Your technical focus:

  • Architect inference pipelines for 3D vision models handling diverse data types (CAD, mesh, point cloud)
  • Build monitoring systems that meaningfully evaluate model performance on real-world, messy geometric data
  • Create robust deployment infrastructure scaling across multiple 3D tasks: segmentation, classification, correspondence, and generation
  • Implement model lifecycle management supporting both discriminative and generative 3D capabilities
  • Design observability frameworks enabling continuous production assessment of 3D model performance

Your background should include:

  • 3-10+ years industry experience as an ML Engineer / Computer Vision Engineer
  • Proven experience deploying models, especially vision or 3D models
  • Strong Python and PyTorch skills with engineering discipline around testing and performance profiling
  • Experience with observability tools and ML monitoring best practices
  • Deep understanding of challenges specific to deploying 3D models (geometric artifacts, mesh quality, robustness)

Valuable additional experience:

  • Working with CAD systems, robotics stacks, or AR/VR environments
  • Agent frameworks, planning pipelines, or LLM-integrated systems
  • 3D data evaluation methodologies and debugging tools
  • Any experience in 3D tools such as WebGL, Three.js, or Blender scripting for 3D visualisation would be useful but not essential.

You'll be establishing the infrastructure foundation for an entirely new capability domain, with high ownership and responsibility for defining production standards and deployment strategies.

Package includes:

  • Competitive salary: $180,000-$240,000 
  • Performance bonus up to 20%
  • Medical, dental, and vision coverage
  • 401k with up to 3% company match (after 3 months)
  • 20 vacation days, 10 sick days, and flexible working arrangements

Based in SF Bay Area or Miami, working alongside a research team that values practical impact and technical excellence.

You must have valid right to work in the US without sponsorship (US Citizenship or Green Card).

If building the systems that make breakthrough 3D AI research truly useful appeals to you, we'd love to discuss this opportunity. All applicants will receive a response.

Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 28/06/2025
Job ID: 33548

Build the foundational models that will give AI agents true 3D understanding

Want to solve the fundamental challenge of how AI systems perceive and reason about 3D geometry?

This Applied Scientist role puts you at the forefront of creating perception capabilities for the next generation of agentic AI systems. You'll see your models introduced into agentic workflows, solving complex challenges in agentic AI for industrial applications.

You'll be joining a well-funded startup developing AI agents for advanced design and manufacturing workflows. Your work will bridge the gap between the physical world and intelligent reasoning systems, creating models that understand CAD data, meshes, and point clouds at a level that enables autonomous decision-making.

This role offers the rare opportunity to build 3D computer vision capabilities from the ground up. You'll be establishing an entirely new domain within the research team, with significant autonomy to define evaluation strategies, model objectives, and technical direction. Your models will form the perception backbone that enables agents to truly understand and manipulate the 3D world.

Your technical challenges:

  • Build models that understand diverse 3D data types (CAD, mesh, point cloud) and learn transferable representations across formats
  • Handle messy, lossy, or incomplete real-world data - moving beyond clean synthetic geometry to tackle industrial reality
  • Scale training across multiple 3D tasks: segmentation, classification, correspondence, and eventually generation
  • Create evaluation pipelines that meaningfully assess model performance and enable continuous production monitoring
  • Work toward a foundational 3D model supporting both discriminative and generative tasks, integrated into broader agentic AI architecture

Your expertise should include:

  • Deep specialisation in 3D computer vision (ideally including a PhD in Computer Vision)
  • Strong knowledge of modern 3D architectures (PointNet++, MeshCNN, 3D Gaussian Splatting, diffusion models, VLMs)
  • Proven ability training large-scale deep learning models with PyTorch
  • Solid applied research skills - can implement novel architectures from papers and make them work in practice
  • Experience with multimodal or vision-language model development

Nice to have:

  • Background working with CAD data or industrial design workflows
  • Experience in complex topics such as robotics, autonomous driving, or AR/VR with 3D perception focus
  • Familiarity with SLAM, pose estimation, or differentiable rendering

You'll join a research team that values ownership and rapid iteration, with the resources to pursue ambitious technical goals. The company provides abundant compute resources and the freedom to explore foundational approaches whilst ensuring practical impact.

Package includes:

  • Base salary: $200,000 (mid/senior) - $250,000 (Staff - negotiable)
  • Performance bonus up to 20%
  • Medical, dental, and vision coverage
  • 401k with up to 3% company match
  • 20+ vacation days

You'll need to be based in SF Bay Area or Miami, with a collaborative team environment that encourages innovation and technical excellence.

You must have valid right to work in the US without sponsorship (US Citizenship or Green Card).

If you're excited about creating the 3D perception capabilities that will power the next generation of intelligent agents, we'd love to hear from you. All applicants will receive a response.

Location: United States
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 28/06/2025
Job ID: 33515

Ready to pioneer deep generative modeling for real-time video synthesis?

Join a pioneering startup developing the foundation layer for the next big AI unlock, video behaviour and naturalness of conversation in video generation. This in turn will change the game for embodied agents with natural behaviours, real-time expression, and conversational intelligence that goes far beyond current avatar technology.

This Research Scientist role focuses on advancing embodied AI through groundbreaking generative modelling research. While existing solutions rely on looped animations with basic lip-sync, this company is building behaviour driven models that drive authentic, real-time interactions capable of natural conversation flow, interruption handling, and emotional expression.

Founded 18 months ago by an exceptional team where 7 out of 12 members hold AI PhDs, they're solving fundamental challenges in visual generation for embodied intelligence. Their beta platform already demonstrates sophisticated real-time video generation systems with advanced generative models creating natural facial expressions and body movements.

The company is building foundational generative technology that creates dynamic visual content from multimodal inputs, developing systems that generate realistic human-like expressions and movements. Their research sits at the intersection of computer vision, deep generative modeling, and real-time video synthesis.

Your focus:

  • Conduct cutting-edge research in deep generative modeling for vision and video generation
  • Develop sophisticated generative models for facial expressions, body dynamics, and full avatar synthesis
  • Create novel architectures using diffusion models and flow matching for video generation
  • Build real-time generative pipelines for dynamic visual content creation
  • Advance state-of-the-art techniques in multimodal generative modeling
  • Collaborate with engineering to productionise generative models into real-time systems
  • Publish findings at top-tier conferences while deploying in real-world applications

Technical challenges: You'll work with cutting-edge techniques including diffusion models, flow matching, and advanced generative architectures for video synthesis. The focus is on creating high-quality, temporally consistent video generation that can power natural embodied agents, emphasising real-time performance and visual fidelity.

Requirements:

  • PhD in Computer Vision, Machine Learning, or related field
  • Strong publication record at top conferences (CVPR, NeurIPS, ICCV, ECCV, ICML, ICLR, SIGGRAPH)
  • Recent video generation or embodied agent/avatar research publications within the past 2 years (essential)
  • Expertise in flow matching and diffusion models
  • Experience with one or more: dyadic conversational avatars, behaviour modelling via LLMs, real-time multimodal generation
  • PyTorch proficiency and large-scale training experience

Nice to have:

  • Industry experience deploying generative models in real-time applications
  • Background in 3D generation, neural rendering, or Gaussian splatting
  • Experience with video generation frameworks and temporal consistency methods

Environment: You'll join a distributed team working primarily in Pacific Time zones, collaborating with specialists in generative modeling, computer vision, and video synthesis. The culture emphasises high ownership, velocity with purpose, and collaborative problem-solving in a fast-moving research environment.

Package:

  • Competitive salary $200k up to $300k base (flexible based on experience)
  • Meaningful equity package
  • Comprehensive healthcare (90% covered)
  • Unlimited PTO
  • Fully remote work with regular team offsites
  • Life and disability coverage

Location: Fully remote position with preference for Pacific Time alignment.

If you're excited about conducting pioneering research in deep generative modeling for vision while shaping the future of embodied agents, this offers an exceptional opportunity to work on genuinely transformative technology.

Ready to help create the next generation of visual AI?

Contact Marc Powell at Techire AI. All applicants will receive a response.

Location: Remote
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 18/06/2025
Job ID: 33449