Your search has found 25 jobs

Ready to tackle the number one problem facing widespread AI adoption?

Join a pioneering company developing the world's first Agent to supervise Agents, LLMs and AI systems. As we move toward an increasingly autonomous AI future, scalable oversight isn't just important—it's essential for safe AGI deployment.

You'll work on the world's first Agent which supervises Agents, LLMs and AI systems. Gone are the days of basic evaluations—this completely automates the evaluation process and automatically debugs AI agents, giving intelligent feedback on how to fix any failures, planning issues or systemic issues. While AI engineers currently spend hours combing through agent traces searching for planning mistakes and tool use errors, this technology makes the process instantaneous and comprehensive.

This role offers an opportunity to solve fundamental research problems at the intersection of LLM post-training, reinforcement learning, and agents. You'll work on post-training, creating realistic RL environments for agent evaluation, building the next generation of agentic evaluation systems and developing proprietary benchmarks for emerging AI capabilities.

Your research focus:

  • Core Research impacting their main Agent platform
  • Develop cutting-edge benchmarks that evaluate LLM and agent capabilities as they emerge
  • Create challenging RL environments that organically evaluate agents under realistic conditions
    Advance agentic evaluation methodologies beyond traditional LLM-as-a-Judge approaches
  • Build and improve evaluator models that provide systematic oversight of autonomous systems
  • Publish research findings in collaboration with top AI research labs

The role balances research with implementation, allowing you to see your theoretical contributions deployed in production systems used by industry leaders. You'll work directly with the founding team to shape the technical direction of AI supervision technology. They also co-publish papers with top research labs.

You should have:

  • Strong hands-on research experience in LLM post-training or RL, or agent research
  • Publication record at top conferences (NeurIPS, ICML, EMNLP, ACL, ICLR)
  • Experience with LLM fine-tuning and understanding of transformer architectures
  • Track record of following recent research and implementing state-of-the-art methods
  • PhD in Computer Science or related field (preferred but not required)
  • Ability to move quickly and thrive in a startup environment

You'll join a small but exceptionally high-calibre team working on problems that will define the future of responsible AI systems and AI safety. The company has proven market demand with existing enterprise customers (OpenAI, NVIDIA, Meta and more) and strong backing, allowing you to focus on breakthrough research with real-world impact.

Your package includes:

  • Salary up to $300,000 (negotiable based on experience)
  • Significant stock in a fast-growing company
  • Comprehensive benefits, 401k, and unlimited PTO
  • Relocation assistance available and sponsorship

You'll work from San Francisco, NYC also considered.

Ready to help ensure AI systems are safe and reliable as we approach AGI?

If you're excited about solving the most critical challenge in AI safety while publishing cutting-edge research, we'd love to hear from you. All applicants will receive a response.

Location: San Francisco Bay Area,
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 22/08/2025
Job ID: 33119

Looking to solve complex ASR challenges at scale?

You'll be joining an established conversational AI company with proprietary in-house speech models that process billions of interactions annually, where your ASR expertise will directly impact real-world customer experiences.

You'll be tackling demanding ASR problems in production environments: streaming speech recognition in noisy conditions, robust accent handling and maintaining high performance at scale.

Working across key production environments, you'll enhance speech capabilities and bring new features to production that push the boundaries of what's possible in challenging acoustic environments.

 

Your focus

  • Maintain and iteratively improve existing ASR technology while introducing cutting-edge enhancements
  • Work end-to-end across speech processing components: speech enhancement, VAD, diarisation, and ASR (AM/LM modelling, ASR biasing)
  • Build streaming ASR systems optimised for challenging acoustic environments
  • Implement emotion detection and acoustic condition classification capabilities
  • Run extensive experiments to advance activity detection and speech processing performance

 

What you'll bring

  • Strong background in ASR model development and deployment
  • Hands-on experience with SOTA speech toolkits (Kaldi, K2, NVIDIA NeMo, Parakeet)
  • Proven streaming ASR experience in production environments

 

This is a fully remote role – must be close to EU timezone. 

Ready to make your mark on speech tech that millions rely on daily? Apply today.

Location: Remote
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 12/08/2025
Job ID: 33647

Want to work with AI agents used for good?

Join a well-funded AI company developing next-generation foundation models specifically for healthcare. With over $250 million backing from top-tier investors, they're creating agents, foundation language models, and real-time empathetic and nartural conversational AI that prioritise safety and medical accuracy.

You'll join a world-class research team comprised of ex-researchers from Microsoft, Meta, NVIDIA, Apple, Stanford, Johns Hopkins, and HuggingFace. This role offers the opportunity to work on AI technology that could transform healthcare outcomes for millions worldwide.

Your focus:

  • Developing and improving LLMs for healthcare applications, directly contributing to agent architecture and conversational AI systems
  • Implementing advanced post-training techniques including instruction tuning and RLHF
  • Building systems capable of real-time patient interactions with medical-grade accuracy
  • Advancing multilingual capabilities for global healthcare accessibility
  • Researching continual learning approaches that evolve through patient interactions
  • Collaborating with researchers from top universities and leading tech companies

Your expertise:

  • Hands-on experience training multi-billion parameter large language models
  • Strong background in NLP with deep understanding of transformer architectures
  • Experience as an Applied researcher with excellent Python and PyTorch coding skills
  • Expertise in reinforcement learning, RLHF, LLM post-training, and supervised fine-tuning
  • Experience with large-scale data processing and distributed computing

Nice to have:

  • Healthcare or life sciences background
  • Publications at top-tier AI conferences
  • Experience with multimodal AI systems

Your package:

  • Base salary: $200,000-$350,000 (based on experience)
  • Stock package
  • Comprehensive benefits including healthcare, dental, and vision
  • 401k with company contributions

Based near Palo Alto, CA with strong preference for on-site collaboration.

This isn't just another AI research role - it's the chance to build technology that could fundamentally improve healthcare delivery globally. Your work will contribute to creating AI systems that assist with patient care and improve healthcare accessibility.

Ready to help shape the future of healthcare through AI? Apply today. All applications will receive a response.

Location: Palo Alto
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 16/07/2025
Job ID: 33284

Ready to pioneer the speech intelligence behind the next generation of embodied AI?

Join a pioneering startup developing foundational technology for natural conversation in embodied agents. You'll advance the speech systems that power avatars with authentic behaviours, real-time expression, and conversational intelligence that handles interruptions and turn-taking just like humans.

This Lead Research Scientist role focuses on advancing real-time speech systems for interactive avatars. You'll develop full-duplex dialogue models and speech-to-speech architectures that enable natural conversational flow, interruption handling, and emotional expression.

Founded by ex-Googlers, they're building proprietary behaviour models that learn from two-way interactions, creating systems where speech timing, prosody, and contextual responses work in harmony with facial expressions and physical behaviours to drive authentic embodied intelligence.

Your focus:

  • Research & develop full-duplex speech systems with natural interruption handling
  • Develop expressive voice models with controllable prosody and timing
  • Build speech-to-speech architectures preserving identity and emotion
  • Create real-time audio generation systems for conversational avatars
  • Publish research while deploying systems in production
  • Collaborate across teams integrating speech with visual behaviour

Requirements:

  • PhD in Speech, Machine Learning, or related field
  • First-author publications at top conferences (Interspeech, ICASSP, NeurIPS, ICLR, etc)
  • Expertise in text-to-speech, speech-to-speech models, or voice cloning
  • Large-scale training experience
  • Experience in prosody modelling or real-time audio generation

Nice to have:

  • Experience with full-duplex speech research
  • Speech-visual alignment expertise (lip sync, expressions)
  • Real-time audio deployment optimisation

Package:

  • Competitive salary $200k- $300k base (based on experience)
  • Meaningful equity package
  • Comprehensive healthcare (90% covered)
  • Unlimited PTO
  • Fully remote work with regular team offsites
  • Life insurance and disability coverage

Location: Fully remote position, globally, with preference for Pacific Time alignment.

Ready to make AI conversations feel authentically human?

Contact Allys at Techire AI. All applicants will receive a response.

 
Location: Remote
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 02/07/2025
Job ID: 33482

Ready to architect the future of human-computer voice interaction?

Join a well established conversational AI company as they transition from traditional cascaded speech systems to cutting-edge E2E speech-to-speech technology. You'll lead this transformation, building multimodal systems that will redefine how millions interact with AI.

The opportunity

You'll be developing technology that directly impacts real users at massive scale. The company processes millions of daily interactions across major enterprise clients, meaning your research will shape real-world conversational experiences.

You'll spearhead the development of proprietary full-duplex speech systems, creating truly natural AI conversations that go far beyond current capabilities.

Your impact

  • Design and build next-generation speech language models from the ground up
  • Drive breakthroughs in speech-to-speech modeling and full-duplex conversation systems
  • Tackle turn-taking, interruption handling, and simultaneous speech processing
  • Bridge cutting-edge research with enterprise-grade production systems
  • Lead a growing team focused on speech-to-speech breakthroughs

What you'll bring

  • Deep understanding of SOTA speech models and neural audio processing
  • Experience building speech language models/multimodal systems
  • Strong background in speech AI research and modern speech architectures

With their established market position and proven track record, you'll have the resources and real-world testing ground to make transformative impact with your research.

The company has built everything in-house, giving you complete technical control and the freedom to explore any approach that delivers value.

Location

Remote (Must be close to EU timezone)

Location: Remote
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 28/06/2025
Job ID: 33350

Build production systems that bring 3D AI models to life in real-world applications

Ready to bridge cutting-edge 3D computer vision research with robust, scalable production systems? This ML Engineer role focuses on deploying 3D perception models into live agentic workflows where reliability and performance are paramount.

You'll be joining a well-funded startup developing AI agents for advanced design and manufacturing. Your role centres on creating the infrastructure that makes 3D understanding truly practical - from real-time inference pipelines to comprehensive monitoring systems that ensure geometry-aware agents perform reliably in production.

This position offers the opportunity to shape how 3D AI models integrate into agent decision-making pipelines. You'll work closely with applied scientists to productionise breakthrough research whilst building robust systems that handle the unique challenges of geometric data in mission-critical applications.

Your technical focus:

  • Architect inference pipelines for 3D vision models handling diverse data types (CAD, mesh, point cloud)
  • Build monitoring systems that meaningfully evaluate model performance on real-world, messy geometric data
  • Create robust deployment infrastructure scaling across multiple 3D tasks: segmentation, classification, correspondence, and generation
  • Implement model lifecycle management supporting both discriminative and generative 3D capabilities
  • Design observability frameworks enabling continuous production assessment of 3D model performance

Your background should include:

  • 3-10+ years industry experience as an ML Engineer / Computer Vision Engineer
  • Proven experience deploying models, especially vision or 3D models
  • Strong Python and PyTorch skills with engineering discipline around testing and performance profiling
  • Experience with observability tools and ML monitoring best practices
  • Deep understanding of challenges specific to deploying 3D models (geometric artifacts, mesh quality, robustness)

Valuable additional experience:

  • Working with CAD systems, robotics stacks, or AR/VR environments
  • Agent frameworks, planning pipelines, or LLM-integrated systems
  • 3D data evaluation methodologies and debugging tools
  • Any experience in 3D tools such as WebGL, Three.js, or Blender scripting for 3D visualisation would be useful but not essential.

You'll be establishing the infrastructure foundation for an entirely new capability domain, with high ownership and responsibility for defining production standards and deployment strategies.

Package includes:

  • Competitive salary: $180,000-$240,000 
  • Performance bonus up to 20%
  • Medical, dental, and vision coverage
  • 401k with up to 3% company match (after 3 months)
  • 20 vacation days, 10 sick days, and flexible working arrangements

Based in SF Bay Area or Miami, working alongside a research team that values practical impact and technical excellence.

You must have valid right to work in the US without sponsorship (US Citizenship or Green Card).

If building the systems that make breakthrough 3D AI research truly useful appeals to you, we'd love to discuss this opportunity. All applicants will receive a response.

Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 28/06/2025
Job ID: 33548

Build the foundational models that will give AI agents true 3D understanding

Want to solve the fundamental challenge of how AI systems perceive and reason about 3D geometry?

This Applied Scientist role puts you at the forefront of creating perception capabilities for the next generation of agentic AI systems. You'll see your models introduced into agentic workflows, solving complex challenges in agentic AI for industrial applications.

You'll be joining a well-funded startup developing AI agents for advanced design and manufacturing workflows. Your work will bridge the gap between the physical world and intelligent reasoning systems, creating models that understand CAD data, meshes, and point clouds at a level that enables autonomous decision-making.

This role offers the rare opportunity to build 3D computer vision capabilities from the ground up. You'll be establishing an entirely new domain within the research team, with significant autonomy to define evaluation strategies, model objectives, and technical direction. Your models will form the perception backbone that enables agents to truly understand and manipulate the 3D world.

Your technical challenges:

  • Build models that understand diverse 3D data types (CAD, mesh, point cloud) and learn transferable representations across formats
  • Handle messy, lossy, or incomplete real-world data - moving beyond clean synthetic geometry to tackle industrial reality
  • Scale training across multiple 3D tasks: segmentation, classification, correspondence, and eventually generation
  • Create evaluation pipelines that meaningfully assess model performance and enable continuous production monitoring
  • Work toward a foundational 3D model supporting both discriminative and generative tasks, integrated into broader agentic AI architecture

Your expertise should include:

  • Deep specialisation in 3D computer vision (ideally including a PhD in Computer Vision)
  • Strong knowledge of modern 3D architectures (PointNet++, MeshCNN, 3D Gaussian Splatting, diffusion models, VLMs)
  • Proven ability training large-scale deep learning models with PyTorch
  • Solid applied research skills - can implement novel architectures from papers and make them work in practice
  • Experience with multimodal or vision-language model development

Nice to have:

  • Background working with CAD data or industrial design workflows
  • Experience in complex topics such as robotics, autonomous driving, or AR/VR with 3D perception focus
  • Familiarity with SLAM, pose estimation, or differentiable rendering

You'll join a research team that values ownership and rapid iteration, with the resources to pursue ambitious technical goals. The company provides abundant compute resources and the freedom to explore foundational approaches whilst ensuring practical impact.

Package includes:

  • Base salary: $200,000 (mid/senior) - $250,000 (Staff - negotiable)
  • Performance bonus up to 20%
  • Medical, dental, and vision coverage
  • 401k with up to 3% company match
  • 20+ vacation days

You'll need to be based in SF Bay Area or Miami, with a collaborative team environment that encourages innovation and technical excellence.

You must have valid right to work in the US without sponsorship (US Citizenship or Green Card).

If you're excited about creating the 3D perception capabilities that will power the next generation of intelligent agents, we'd love to hear from you. All applicants will receive a response.

Location: United States
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 28/06/2025
Job ID: 33515

Ready to pioneer deep generative modeling for real-time video synthesis?

Join a pioneering startup developing the foundation layer for the next big AI unlock, video behaviour and naturalness of conversation in video generation. This in turn will change the game for embodied agents with natural behaviours, real-time expression, and conversational intelligence that goes far beyond current avatar technology.

This Research Scientist role focuses on advancing embodied AI through groundbreaking generative modelling research. While existing solutions rely on looped animations with basic lip-sync, this company is building behaviour driven models that drive authentic, real-time interactions capable of natural conversation flow, interruption handling, and emotional expression.

Founded 18 months ago by an exceptional team where 7 out of 12 members hold AI PhDs, they're solving fundamental challenges in visual generation for embodied intelligence. Their beta platform already demonstrates sophisticated real-time video generation systems with advanced generative models creating natural facial expressions and body movements.

The company is building foundational generative technology that creates dynamic visual content from multimodal inputs, developing systems that generate realistic human-like expressions and movements. Their research sits at the intersection of computer vision, deep generative modeling, and real-time video synthesis.

Your focus:

  • Conduct cutting-edge research in deep generative modeling for vision and video generation
  • Develop sophisticated generative models for facial expressions, body dynamics, and full avatar synthesis
  • Create novel architectures using diffusion models and flow matching for video generation
  • Build real-time generative pipelines for dynamic visual content creation
  • Advance state-of-the-art techniques in multimodal generative modeling
  • Collaborate with engineering to productionise generative models into real-time systems
  • Publish findings at top-tier conferences while deploying in real-world applications

Technical challenges: You'll work with cutting-edge techniques including diffusion models, flow matching, and advanced generative architectures for video synthesis. The focus is on creating high-quality, temporally consistent video generation that can power natural embodied agents, emphasising real-time performance and visual fidelity.

Requirements:

  • PhD in Computer Vision, Machine Learning, or related field
  • Strong publication record at top conferences (CVPR, NeurIPS, ICCV, ECCV, ICML, ICLR, SIGGRAPH)
  • Recent video generation or embodied agent/avatar research publications within the past 2 years (essential)
  • Expertise in flow matching and diffusion models
  • Experience with one or more: dyadic conversational avatars, behaviour modelling via LLMs, real-time multimodal generation
  • PyTorch proficiency and large-scale training experience

Nice to have:

  • Industry experience deploying generative models in real-time applications
  • Background in 3D generation, neural rendering, or Gaussian splatting
  • Experience with video generation frameworks and temporal consistency methods

Environment: You'll join a distributed team working primarily in Pacific Time zones, collaborating with specialists in generative modeling, computer vision, and video synthesis. The culture emphasises high ownership, velocity with purpose, and collaborative problem-solving in a fast-moving research environment.

Package:

  • Competitive salary $200k up to $300k base (flexible based on experience)
  • Meaningful equity package
  • Comprehensive healthcare (90% covered)
  • Unlimited PTO
  • Fully remote work with regular team offsites
  • Life and disability coverage

Location: Fully remote position with preference for Pacific Time alignment.

If you're excited about conducting pioneering research in deep generative modeling for vision while shaping the future of embodied agents, this offers an exceptional opportunity to work on genuinely transformative technology.

Ready to help create the next generation of visual AI?

Contact Marc Powell at Techire AI. All applicants will receive a response.

Location: Remote
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 18/06/2025
Job ID: 33449

Ready to build speech AI that actually works in real-time?

A well-funded AI startup has developed new model architectures that make real-time conversational AI finally viable at scale. While most voice AI still suffers from delays and computational bottlenecks, they've solved the core efficiency problems that have held the field back. 

The role

As their Speech Research Scientist, you'll build the speech models that could define the next decade of voice interaction. You'll work on novel architectures that have immediate real-world impact for thousands of customers.

What you'll do

  • Design and implement SOTA speech synthesis models
  • Develop efficient algorithms for voice processing and audio understanding
  • Create scalable systems that handle massive audio workloads
  • Build comprehensive evaluation methods to validate model performance
  • Collaborate with engineering teams to transition research into production

What you'll bring

  • Deep expertise in modern speech technologies (Text-to-Speech, Speech LLMs, Voice Conversion/Cloning, Speech Synthesis, Speech Translation, Speech Restoration)
  • Strong background in generative modeling for audio and speech
  • Publications at leading conferences
  • Track record of implementing research ideas from concept to production
This role is based in the Bay Area.
 
If you're excited about building the foundational models that will power the Voice AI revolution, we'd love to hear from you.
 
Location: Bay Area
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 05/06/2025
Job ID: 33251

Build the ML infrastructure that powers cutting-edge AI across multiple domains

Ready to architect MLOps systems from the ground up for a fast-growing AI team? This greenfield opportunity offers complete autonomy to design and build training pipelines for LLMs, computer vision models, and other deep learning architectures that will power next-generation AI applications.

You'll join a well-funded startup ($20M+ raised, with a new round expected this year) developing production-grade AI solutions across regulated industries including healthcare, aerospace and manufacturing. Founded by a successful entrepreneur with a previous billion-dollar exit, they're already partnering with Fortune 100 and 500 clients where standard AI approaches fall short.

This role offers exceptional technical ownership - you'll build their ML infrastructure from current basic tooling to production-scale systems that support their rapidly expanding applied AI team. They have significant GPU resources with substantial budget growth expected. As the team scales to ~20 people within the year, there's high potential for you to lead future MLOps hires.

The challenge is substantial: creating infrastructure that supports training across multiple modalities - from LLMs to computer vision models. You'll work with large compute resources and have complete autonomy to select and implement the tooling that will define how the team operates for years to come. Your initial focus will be establishing robust training and evaluation pipelines, then scaling to enterprise-grade data workflows with versioning, monitoring, and automated deployment systems.

Your focus:

  • Build training and evaluation pipelines for LLMs, vision models, and other deep learning architectures
  • Design distributed training systems on multi-GPU clusters across model types
  • Create scalable data pipelines, versioning systems, and model checkpointing workflows
  • Implement model serving infrastructure with tools like vLLM, Triton, and TorchServe
  • Establish comprehensive monitoring, experiment tracking, and reproducibility systems
  • Support a rapidly growing applied AI team with robust CI/CD workflows for ML systems

You should have:

  • 3+ years building MLOps infrastructure or ML systems in production environments
  • Hands-on experience with training pipelines for deep learning models (LLMs, CNNs, transformers)
  • Strong expertise with AWS and Kubernetes (mandatory requirements)
  • Proficiency with Python, PyTorch/TensorFlow, and distributed training libraries
  • Experience with model tracking tools like Weights & Biases or MLflow
  • Understanding of modern ML architectures across multiple domains

Nice to have:

  • Experience with LLM inference tools (vLLM, SGLang, RayServer)
  • Ray experience for distributed computing
  • Knowledge of mixed-precision training, quantisation, and model optimisation
  • Computer vision workflow experience
  • Data versioning tools (DVC, LakeFS)
  • Early-stage startup experience

You'll receive:

  • Competitive base salary: circa $250K (based on experience)
  • Significant stock package in a fast-growing company
  • Access to substantial GPU budget with expected growth
  • Healthcare (medical, dental, vision) and 401k with matching
  • 20 vacation days plus flexible working arrangements

You must be based in SF Bay Area or Miami (relocation is provided to Florida only). At this time we can only consider US citizens or green card holders.nd.

Ready to build the infrastructure that powers the future of production AI? All applicants will receive a response.

Location: United States
Job type: Permanent
Emp type: Full-time
Salary type: Annual
Salary: negotiable
Job published: 05/06/2025
Job ID: 33218