via Dice
$120K - 200K a year
Design, build, and maintain data pipelines, optimize workflows, and ensure data quality and security.
Proficiency in Python, SQL, data warehousing, and experience with big data tools and cloud platforms.
Key Responsibilities • Design, build, and maintain ETL/ELT data pipelines • Develop Python-based data processing applications • Work with structured and unstructured data at scale • Integrate data from multiple sources (APIs, databases, files, streams) • Optimize data workflows for performance and reliability • Ensure data quality, validation, and monitoring • Collaborate with data scientists, analysts, and backend teams • Manage and maintain data warehouses/lakes • Implement logging, error handling, and automation • Follow best practices for security and compliance Required Skills Programming • Strong Python (Pandas, NumPy, PySpark) • Writing clean, modular, and testable code Databases & Storage • SQL (PostgreSQL, MySQL, SQL Server) • NoSQL (MongoDB, Cassandra optional) • Data Warehouses (Snowflake, Redshift, BigQuery) Big Data & Processing • Apache Spark, Hadoop (preferred) • Batch and streaming data processing Cloud Platforms • AWS / Azure / Google Cloud Platform • S3, Lambda, Glue, Dataflow, BigQuery, etc. Data Engineering Tools • Airflow, Prefect, Luigi (orchestration) • Kafka / PubSub (streaming optional) • DBT (data transformation) DevOps & Other • Git, CI/CD • Docker, Kubernetes (nice to have) • Linux basics
This job posting was last updated on 1/8/2026