Find your dream job faster with JobLogr
AI-powered job search, resume help, and more.
Try for Free
Qode

Qode

via Workable

All our jobs are verified from trusted employers and sources. We connect to legitimate platforms only.

Sr. Data Engineer

Anywhere
Full-time
Posted 3/11/2026
Direct Apply
Key Skills:
Python
PySpark
AWS Glue
AWS Lambda
Kafka
Data Quality Frameworks

Compensation

Salary Range

$70K - 120K a year

Responsibilities

Design and optimize scalable batch and streaming data pipelines using AWS and PySpark.

Requirements

Strong hands-on experience with Python, PySpark, AWS Glue, Lambda, and Kafka ecosystem with data quality framework expertise required.

Full Description

Job Title: Sr.Data Engineer (Mid–Senior Level) – AWS & StreamingExperience Level – 13-15+ YearsLocation: Fort Mill, SC (3 days hybrid)Role Summary: We are seeking a Mid–Senior Data Engineer with strong expertise in AWS-based data engineering, real-time streaming technologies, and enterprise-grade data quality frameworks. The ideal candidate will design, build, and optimize scalable batch and streaming data pipelines, implement robust data validation and monitoring processes, and support mission-critical analytics platforms. Key Responsibilities: Develop and maintain scalable ETL/ELT pipelines using AWS Glue, PySpark, and Python Build event-driven workflows using AWS Lambda Design and manage real-time streaming solutions using Kafka, KSQL, and Apache Flink Implement and enforce comprehensive data quality frameworks, including validation, profiling, monitoring, and reconciliation Optimize data processing performance, scalability, reliability, and cost in cloud environments Collaborate with cross-functional teams to deliver reliable, production-grade data platforms and ensure data integrity across the pipeline Must have Skills: Strong hands-on experience with Python and PySpark Proven expertise in AWS Glue, Lambda, and other cloud-native data services Solid experience with the Kafka ecosystem (topics, partitions, consumer groups, streaming patterns) Demonstrated experience building and supporting data quality frameworks (validation rules, reconciliation checks, profiling, anomaly detection) Strong understanding of distributed data processing and scalable architecture patterns Good-to-Have Skills: Experience with Apache Flink for real-time stream processing and stateful computations Knowledge of KSQL or other streaming SQL engines Exposure to CI/CD pipelines, IaC (Terraform/CloudFormation), and DevOps practices Familiarity with data lake/lakehouse architectures and table formats such as Iceberg, Delta, or Hudi Experience working in enterprise or financial data environments

This job posting was last updated on 3/11/2026

Ready to have AI work for you in your job search?

Sign-up for free and start using JobLogr today!

Get Started »
JobLogr badgeTinyLaunch BadgeJobLogr - AI Job Search Tools to Land Your Next Job Faster than Ever | Product Hunt