Find your dream job faster with JobLogr
AI-powered job search, resume help, and more.
Try for Free
II

Innodata Inc

via Recruitee

All our jobs are verified from trusted employers and sources. We connect to legitimate platforms only.

Senior Data Engineer – Real-Time & Distributed Systems (GCP)

Anywhere
Full-time
Posted 2/19/2026
Direct Apply
Key Skills:
Python
Airflow
Apache Spark
Google Cloud Composer

Compensation

Salary Range

$90K - 140K a year

Responsibilities

Design and optimize scalable batch and real-time data pipelines and event-driven architectures for high-throughput systems.

Requirements

Advanced Python proficiency and experience with big data tools like Airflow, Apache Spark, and Pub/Sub for production pipeline management.

Full Description

Who we are: Innodata (NASDAQ: INOD) is a leading data engineering company. With more than 2,000 customers and operations in 13 cities around the world, we are the AI technology solutions provider-of-choice to 4 out of 5 of the world’s biggest technology companies, as well as leading companies across financial services, insurance, technology, law, and medicine. By combining advanced machine learning and artificial intelligence (ML/AI) technologies, a global workforce of subject matter experts, and a high-security infrastructure, we’re helping usher in the promise of clean and optimized digital data to all industries. Innodata offers a powerful combination of both digital data solutions and easy-to-use, high-quality platforms. Our global workforce includes over 3,000 employees in the United States, Canada, United Kingdom, the Philippines, India, Sri Lanka, Israel and Germany. We’re poised for a period of explosive growth over the next few years. Key Responsibilities Design, build, and optimize scalable data pipelines for batch and real-time processing Develop and maintain event-driven architectures for high-throughput systems Ensure data reliability, performance, and low-latency processing across distributed environments Collaborate with data scientists and application teams to enable analytics and AI use cases Implement best practices in performance tuning, monitoring, and cost optimization Advanced proficiency in Python for backend and large-scale data processing Strong experience building and managing big data pipelines in production environments Hands-on expertise with workflow orchestration tools such as Airflow or Google Cloud Composer Proven experience in batch and streaming data processing using: Apache Spark Apache Beam (Dataflow) Experience designing and operating event-driven systems using Pub/Sub Strong understanding of distributed systems architecture and scalability patterns Experience managing globally distributed, low-latency datasets Hands-on experience with NoSQL databases and/or Google Cloud Spanner Strong knowledge of system reliability, fault tolerance, and performance optimization Preferred Skills Proficiency in Go, Java, or Scala Experience with Kafka or Flume for streaming ingestion Deep familiarity with the Google Cloud Platform ecosystem Experience with production monitoring, logging, and observability frameworks Exposure to high-availability, multi-region deployments Please be aware of recruitment scams involving individuals or organizations falsely claiming to represent employers. Innodata will never ask for payment, banking details, or sensitive personal information during the application process. To learn more on how to recognize job scams, please visit the Federal Trade Commission’s guide at https://consumer.ftc.gov/articles/job-scams. If you believe you’ve been targeted by a recruitment scam, please report it to Innodata at verifyjoboffer@innodata.com and consider reporting it to the FTC at ReportFraud.ftc.gov. #LI-NS1

This job posting was last updated on 2/20/2026

Ready to have AI work for you in your job search?

Sign-up for free and start using JobLogr today!

Get Started »
JobLogr badgeTinyLaunch BadgeJobLogr - AI Job Search Tools to Land Your Next Job Faster than Ever | Product Hunt