← All Jobs
Posted Apr 21, 2026

Research Crawling Engineer

Apply Now ✨

Who We Are:

We build infrastructure that delivers massive amounts of web data to the companies training the world’s most powerful AI models.

We're the team that helps to power and support Grass, a bandwidth-sharing network that lets us operate a massive distributed crawler, giving us unique access to high-quality public web data at global scale. On top of that, we’ve built pipelines for ingesting, segmenting, and annotating billions of videos, transcripts, and audio files, powering dataset creation for frontier labs.

We’re lean, technical, and move fast. No red tape, no slow decision-making; just a team of builders pushing to expand what’s possible for open web data and AI.

Overview:
As a Research Crawling Engineer, you will design and operate large-scale web data acquisition systems for research and model development. You will work will span distributed systems, scraping infrastructure, and data pipelines.

Responsibilities:



Requirements:

Ability to debug unstable or adversarial environments

Preferred / Bonus:

What This Role Involves:

Evaluation Criteria:

Compensation:

Based on experience and demonstrated ability to operate at scale

Example Projects:

Why Work With Us: