Cognition Logo

Cognition

Research Engineer, Infrastructure

Posted 9 Days Ago
In-Office or Remote
Hiring Remotely in CA
Mid level
In-Office or Remote
Hiring Remotely in CA
Mid level
The Research Engineer, Infrastructure will build distributed training systems, optimize performance, manage data pipelines, and enhance research workflows, ensuring the infrastructure scales with AI advancements.
The summary above was generated by AI
Who We Are

We are an applied AI lab building end-to-end software agents. We're the team behind Devin, the first AI software engineer, and Windsurf, an AI-native IDE. These products represent our vision for AI that doesn't just assist engineers, but works alongside them as a genuine teammate.

Our team is small and talent-dense: world-class competitive programmers, former founders, and researchers from the frontier of AI, including Scale AI, Palantir, Cursor, Google DeepMind, and others.

Role Mission

Research moves at the speed of the infrastructure underneath it. Every training run, evaluation loop, and experimental iteration depends on systems that are fast, reliable, and built to scale. This role exists to make sure nothing in the stack becomes the bottleneck that slows down the frontier.

You will own the core systems that researchers depend on daily: distributed training infrastructure, experiment orchestration, data pipelines, and the tooling that turns raw compute into usable research velocity. This is not a support role. You will work directly alongside researchers, understand the science deeply enough to anticipate what they need next, and build systems that hold up under the pressure of training jobs running across thousands of GPUs. We don't distinguish between research and engineering; the best infrastructure engineers here are also the ones who understand why the research works.

What You'll Accomplish
  • Distributed Training Infrastructure: Build and own the systems that run large-scale training jobs reliably across GPU clusters. This includes job launchers, checkpointing and recovery, fault tolerance, and the monitoring that keeps researchers informed and unblocked.

  • Scaling Agent Rollouts: Own the infrastructure that runs hundreds of thousands of concurrent coding agent rollouts in VM sandboxes, from high-fidelity environment design to the distributed systems that hold up at our largest RL training scales.

  • Performance Optimization: Profile and improve training throughput end to end. Identify bottlenecks across data loading, communication overhead, memory utilization, and compute efficiency. Implement solutions that meaningfully improve step time and MFU at scale.

  • Experiment Orchestration and Tooling: Design and maintain the systems researchers use to launch, track, and analyze experiments. Reduce friction in the research loop so that more time is spent on ideas and less on waiting.

  • Data Pipeline Engineering: Build high-throughput, reliable data pipelines for training and evaluation. Ensure data quality, reproducibility, and efficiency at the scale our training runs demand.

  • Debugging and Reliability: Diagnose and resolve training failures across GPUs, networking, numerics, and data. Maintain detailed understanding of failure modes and build systems that fail gracefully and recover fast.

  • Parallelism and Systems Research: Implement and optimize parallelism strategies: data, tensor, pipeline, and sequence parallelism. Understand the tradeoffs deeply and apply them to get the most out of available hardware.

  • Scaling Infrastructure Ahead of Research: Anticipate what the research team will need next and build it before it becomes a constraint. The best infrastructure engineers here are proactive, not reactive.

Exceptional Candidates Have Demonstrated
  • Deep experience building and operating distributed training systems for large models; comfortable owning infrastructure end to end from the cluster level down to the training loop

  • Strong systems engineering fundamentals: distributed systems, networking, storage, and the ability to reason about performance across the full hardware-software stack

  • Proficiency in Python and C++; experience with PyTorch or equivalent deep learning frameworks at a systems level, not just API usage

  • Hands-on experience with GPU performance profiling, memory optimization, and compute efficiency; able to diagnose why a training run is underperforming and fix it

  • Experience implementing or optimizing parallelism strategies (data, tensor, pipeline, sequence) for large model training

  • Track record of building tooling and abstractions that meaningfully accelerate research workflows

  • Strong debugging instincts across complex, distributed systems where failures are non-deterministic and hard to reproduce

  • Enough ML knowledge to engage substantively with researchers: understand what they are training, why the architecture choices matter, and what the infrastructure needs to support

  • We care more about demonstrated capability than credentials. A PhD is one signal among many.

Resources & Environment
  • Small, highly selective team where research and product move together; prototypes reach real deployment quickly

  • You'll own and operate infrastructure running across thousands of GPUs; compute is not a constraint and neither is access to the systems you need to do the work well

  • The environment rewards speed, autonomy, and technical depth with minimal process overhead; this is one of the most competitive and fast-moving problems in AI

Equal Opportunity

Cognition is an equal opportunity employer. We do not discriminate on the basis of race, color, religion, sex, sexual orientation, gender identity, national origin, age, disability, veteran status, or any other protected characteristic under applicable law. We are committed to providing reasonable accommodations for candidates with disabilities throughout the hiring process - please let us know if you need any.

Top Skills

C++
Gpu
Python
PyTorch

Similar Jobs

21 Days Ago
Remote or Hybrid
CA
Mid level
Mid level
Artificial Intelligence • Information Technology • Software
The role involves designing scalable data pipelines for 3D, video, and sensor data, optimizing infrastructure, and productionizing ML models with researchers.
Top Skills: SparkAWSAzureDaskDvcFlyteGCPKubernetesMlflowPythonPyTorchRay
6 Days Ago
Remote or Hybrid
CA
Mid level
Mid level
Artificial Intelligence • Software
The role involves building and scaling data infrastructure for multimodal AI systems, developing high-throughput data pipelines, and collaborating with ML researchers and product teams to enhance data systems.
Top Skills: BeamPythonRaySpark
20 Days Ago
Remote or Hybrid
CA
Mid level
Mid level
Artificial Intelligence • Software
Seeking a Research Scientist/Engineer to design and optimize distributed training systems for large-scale multimodal AI models using thousands of GPUs.
Top Skills: Cloud InfrastructureContainerizationCudaDistributed SystemsLinuxMpiNcclOrchestrationPyTorch

What you need to know about the Ottawa Tech Scene

The capital city of Canada and the nation's fourth-largest urban area, Ottawa has proven a rapidly growing global tech hub. With over 1,800 tech companies, many of which are leaders in their sectors, the city's tech talent now makes up more than 13 percent of its total workforce. This growth is driven not only by the big players like UL Solutions and Dropbox, but also by a thriving startup ecosystem, as new businesses emerge to follow in the footsteps of those that came before them.

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account