via Workday
$200K - 250K a year
Lead the development and migration of a unified enterprise data lake, ensuring data quality, governance, and stakeholder alignment.
Extensive experience in managing complex data platform initiatives in cloud environments, with strong technical skills in SQL, data orchestration, and modern data stacks.
Role Summary Ansira is consolidating and scaling a unified, enterprise Data Lake to integrate product, media, and business data, standardize reporting, and accelerate decision-making across the organization. We are seeking a Technical Program Manager (TPM) to lead this cross-functional program end-to-end — aligning product and engineering roadmaps, driving ingestion and migration from legacy systems, maturing data governance and quality, and ensuring business adoption of standardized, self-serve analytics. This leader will orchestrate work across Ansira’s Product solutions, Data Engineering, Data Science/BI, Media, and Client Partnership teams, with a clear mandate: deliver consistent, governed, and performant data to downstream products and reporting while deprecating redundant systems and minimizing operational cost. What You’ll Do Program Leadership & Delivery Own the multi-quarter program plan for the unified Data Lake: scope, roadmap, milestones, budgets (OPEX/CAPEX), risks, and dependencies. Stand up and run the operating model: weekly workstream standups, cross-functional syncs, monthly steering committee, and a transparent executive status rhythm. Build and maintain a single-source-of-truth for delivery: program charter, RACI, RAID log, decision log, intake/triage process, and dashboards for progress/risks. Drive the migration plan from legacy pipelines and tools (e.g., Alteryx, Insighter) to the target stack (e.g., Snowflake, Power BI embedded via platform connectors). Coordinate parallel workstreams (ingestion, modeling, governance, reporting cutover) to hit time-bound deliverables with predictable quality. Product Management & Roadmap Define and maintain the Data Lake program backlog, translating business use cases into technical epics, data contracts, and acceptance criteria. Partner with Product and Data Science teams to standardize media and product reporting packages and ensure they’re backed by governed, contract-driven data. Prioritize sources and domains for ingestion based on business value, client impact, and technical feasibility; establish clear go/no-go gates. Align with platform architecture to ensure scalable patterns for batch/stream ELT/CDC, cost control, observability, and reusability across domains. Data Governance, Quality, and Security Establish practical data contracts with upstream product and business owners; define schema, SLAs, lineage, and DQ checks at ingestion. Stand up governance ceremonies and roles (data owners, stewards) and implement data catalog/lineage practices to improve discoverability and trust. Define and monitor quality KPIs (completeness, timeliness, accuracy) and drive remediation plans with accountable teams. Ensure data privacy, compliance, and security best practices (e.g., PII handling, role-based access, data masking) across environments. Stakeholder Management & Change Adoption Serve as the connective tissue across Product, Engineering, Data Science, Media, Finance, and Client Partnership — communicating decisions, trade-offs, and timelines. Lead change management for reporting standardization (e.g., Media (AdTech/LBN)-based standard reports), business onboarding to the lake, and client-facing cutovers. Create enablement assets (runbooks, playbooks, onboarding guides) and training plans to accelerate adoption and reduce support burden. Technical Fluency Partner effectively with architects and data engineers on Snowflake/BigQuery/Databricks, Azure/AWS/GCP services, orchestration (ADF/Airflow), and transformation (dbt). Understand ELT/CDC patterns, API/file ingestion, schema design for analytics, and BI tooling (Power BI, Looker). Write and review basic SQL for validation. Apply FinOps and performance/cost optimization practices (storage tiers, compute sizing, job scheduling, caching strategies). Minimum Qualifications 8+ years in Program/Project/Product Management, with 5+ years leading complex data platform initiatives in a cloud environment. Proven delivery of cross-functional data programs involving multiple product lines and business stakeholders; strong executive communication. Hands-on experience with modern data stacks: one or more of Snowflake/BigQuery/Databricks; Azure Data Factory/Airflow; dbt; Kafka/Kinesis; Git/Terraform; REST/SFTP integrations. Strong grounding in data governance and quality practices, data contracts, catalog/lineage, and secure data access. Demonstrated expertise in Agile at scale (Scrum/Kanban), Jira/Confluence, dependency/risk management, and budget tracking (including CAPEX/OPEX). Competent SQL skills for validation/triage; fluency in reading pipeline/log artifacts and interpreting BI/semantic model requirements. Preferred Qualifications Background in marketing/media data and standardized performance reporting (e.g., Media (AdTech/LBN), campaign hierarchies, Power BI embedded). Prior experience migrating from legacy ETL/BI ecosystems (e.g., Alteryx/Insighter/Tableau) to a lakehouse with standardized semantic layers. Experience establishing data domains and productizing data (SLAs, contracts, versioning, lifecycle) to accelerate downstream analytics. Familiarity with privacy, security, and compliance standards (e.g., RBAC/ABAC, PII governance) and enterprise SSO/permissions models for embedded analytics. FinOps mindset: cost observability, unit economics, and right-sizing compute/storage. Success Metrics (What Great Looks Like) Sources Onboarded: number of prioritized sources/domains ingested to the lake with production-grade contracts and SLAs. Time-to-Data: cycle time from intake approval to governed data available for downstream consumption. Data Quality & Reliability: sustained improvement in DQ scorecards; incident MTTR reduction; SLA adherence for freshness and availability. Migration Progress: percent of targeted legacy pipelines and reports decommissioned; client reporting cutovers delivered on schedule. Adoption & Reuse: growth in standardized reporting/package usage; reduction in ad hoc one-off pipelines. Cost & Performance: measurable storage/compute cost per use case; query performance improvements aligned to agreed SLOs. 30/60/90-Day Plan 30 Days: Confirm program charter, governance model, and delivery rhythm. Baseline current-state architecture, sources, and reporting dependencies. Publish the first integrated roadmap/milestone plan and RAID. 60 Days: Land 1–2 ingestion patterns as reusable templates (contracts, lineage, observability). Execute the first cutover to standardized reporting for a targeted use case. Stand up DQ scorecards and weekly KPI review. 90 Days: Complete a tranche of source onboardings and at least one significant reporting migration. Retire targeted legacy jobs. Publish quarterly executive readout with outcomes, cost/performance improvements, and next-wave priorities. Tools You’ll Use Jira, Confluence, Power BI; Snowflake/BigQuery/Databricks; Azure/AWS/GCP services (e.g., ADF, IAM); Airflow; dbt; Git; catalog/lineage tools. Why This Role This is a high-visibility opportunity to unify Ansira’s data foundation and materially improve the speed, quality, and consistency of insights across products and services. You will own the operating model, drive cross-functional alignment, and deliver tangible business outcomes through a modern, governed data platform. Ansira is the leading marketing platform for brands that operate in distributed ecosystems. By orchestrating the entire brand-to-local value chain, Ansira synchronizes global companies’ internal and external teams across channels, markets and regions. Ansira leverages AI-powered technology, vertical expertise, local insights and strategic partnerships to optimize marketing performance and empower partner growth at scale.
This job posting was last updated on 1/9/2026