Find your dream job faster with JobLogr
AI-powered job search, resume help, and more.
Try for Free
OpenRouter

OpenRouter

via Ashby

Apply Now
All our jobs are verified from trusted employers and sources. We connect to legitimate platforms only.

Senior Data Engineer

Anywhere
full-time
Posted 9/25/2025
Direct Apply
Key Skills:
Data Engineering
ETL
ELT
SQL
Python
Data Modeling
Performance Tuning
Data Warehousing
Data Pipelines
Data Infrastructure
Monitoring
Alerting
Customer-Facing Metrics
Privacy Standards
Event Streaming
CI/CD

Compensation

Salary Range

$Not specified

Responsibilities

Architect, build, and own the end-to-end data platform. Stand up data warehousing and pipelines, ensuring data is reliable and scalable for customers and internal use.

Requirements

4+ years of experience as a Data Engineer with expertise in SQL and Python. Experience in building end-to-end data infrastructure and a strong focus on customer needs is essential.

Full Description

OpenRouter operates the largest marketplace and routing layer for large-language models: 400+ models from 60+ providers are reachable through a single API, contract, and bill, eliminating infrastructure sprawl for developers and enterprises alike. We make AI “just work” so companies can focus on what they do best. We believe LLMs will be the largest software market ever, and today we’re the largest aggregator and discovery point. We serve more than a million developers and route over $100m in annualized inference spend, backed by a16z and Menlo Ventures. We are a fully-remote team of roughly 15 people. Our culture is engineering-first and velocity-obsessed; we prize clarity, autonomy, and rapid iteration. About the Role: We’re hiring our first Senior Data Engineer to architect, build, and own our end-to-end data platform. Our position in the AI marketplace gives us uniquely rich usage and performance data; your job is to turn that data into reliable and scalable datasets and products that deliver clear value to customers while informing our own product decisions. You’ll stand up our data warehousing and pipelines from scratch, connect product and platform data, and make high-leverage data accessible and trustworthy to our customers and across the company. Key Responsibilities: Data Platform & Warehousing: Stand up central analytics store. Define schemas, partitioning, retention, and performance strategies for scale. Establish data contracts and documentation to keep data consistent and discoverable. ETL/ELT Pipelines & Orchestration: Build, operate, and monitor robust ETL/ELT pipelines. Implement CDC and batch/stream ingestion patterns; ensure idempotency, safe backfills, and predictable reprocessing. Customer Data Products: Design and ship the data foundations for customer-facing metrics (latency, throughput, error rates, reliability/SLOs, cost), with strong handling of late/duplicate events. Build secure, multi-tenant datasets and APIs for customers; enforce isolation with row/column-level security, access controls, and privacy guardrails. End-to-End Ownership: Take data features from concept to production: design, implement, backfill/migrate, document, and support. About You: 4+ years as a Data Engineer (or similar), including owning production pipelines and a modern data warehouse (Clickhouse, Snowflake, Databricks, etc) Expert SQL and Python with deep experience with ETL/ELT design, data modeling, and performance tuning. Experience building end-to-end data infrastructure: storage, compute, networking, orchestration, CI/CD for data, monitoring/alerting, and cost management. Excellent communicator who can self-manage, set expectations, and partner across functions while working asynchronously. You’re customer-obsessed and product-minded, starting from the user problem and shipping pragmatically to deliver value fast. Biased to action, comfortable with ambiguity, and able to prioritize for impact. You default to simple, reliable solutions and iterate quickly. Bonus Points Experience as the first/early data hire building 0 to 1 Direct experience with Clickhouse, Postgres, GCP, or Terraform Experience standing up event streaming platforms (e.g. Kafka, Pub/Sub) Experience with database privacy/compliance standards (e.g., SOC 2, GDPR/CCPA) What We Offer: Cutting-Edge Work: Be at the forefront of AI and LLM infrastructure, building tools for developers and enterprises alike. Small, Talented Team: Work closely with a tight-knit group of top-notch engineers who value excellence and innovation. Growth Opportunities: As we scale, grow into leadership roles or specialize further in your area of interest. Competitive Compensation: Attractive salary, equity, and benefits that reflect the impact you’ll have on our success.

This job posting was last updated on 9/26/2025

Ready to have AI work for you in your job search?

Sign-up for free and start using JobLogr today!

Get Started »
JobLogr badgeTinyLaunch BadgeJobLogr - AI Job Search Tools to Land Your Next Job Faster than Ever | Product Hunt