We use cookies. Find out more about it here. By continuing to browse this site you are agreeing to our use of cookies.
#alert
Back to search results

Senior Data Engineer

Intercontinental Exchange
United States, Georgia, Atlanta
5660 New Northside Drive Northwest (Show on map)
Sep 16, 2025
Overview

Job Purpose

We're seeking a talented Senior Data Engineer to join our Enterprise Architecture team in a cross-cutting role that will help define and implement our next-generation data platform. In this pivotal position, you'll lead the design and implementation of scalable, self-service data pipelines with a strong emphasis on data quality and governance. This is an opportunity to shape our data engineering practice from the ground up, working directly with key stakeholders to build mission-critical ML and AI data workflows.

About Our Technology Stack

You'll be working with a modern, on-premises data stack that includes:

  • Apache Airflow for workflow orchestration (self-hosted on Kubernetes)
  • dbt for data transformation and testing
  • Apache Flink for stream processing and real-time data workflows
  • Kubernetes for containerized deployment and scaling
  • Git-based version control and CI/CD for data pipelines
  • Oracle Exadata for data warehousing
  • Kafka for messaging and event streaming

We emphasize building systems that are maintainable, scalable, and focused on enabling self-service data access while maintaining high standards for data quality and governance.

Responsibilities

  • Design, build, and maintain our on-premises data orchestration platform using Apache Airflow, dbt, and Apache Flink
  • Create self-service capabilities that empower teams across the organization to build and deploy data pipelines without extensive engineering support
  • Implement robust data quality testing frameworks that ensure data integrity throughout the entire data lifecycle
  • Establish data engineering best practices, including version control, CI/CD for data pipelines, and automated testing
  • Collaborate with ML/AI teams to build scalable feature engineering pipelines that support both batch and real-time data processing
  • Develop reusable patterns for common data integration scenarios that can be leveraged across the organization
  • Work closely with infrastructure teams to optimize our Kubernetes-based data platform for performance and reliability
  • Mentor junior engineers and advocate for engineering excellence in data practices

Knowledge and Experience

  • 5+ years of professional experience in data engineering, with at least 2 years working on enterprise-scale data platforms
  • Deep expertise with Apache Airflow, including DAG design, performance optimization, and operational management
  • Strong understanding of dbt for data transformation, including experience with testing frameworks and deployment strategies
  • Experience with stream processing frameworks like Apache Flink or similar technologies
  • Proficiency with SQL and Python for data transformation and pipeline development
  • Familiarity with Kubernetes for containerized application deployment
  • Experience implementing data quality frameworks and automated testing for data pipelines
  • Knowledge of Git-based workflows and CI/CD pipelines for data applications
  • Ability to work cross-functionally with data scientists, ML engineers, and business stakeholders

Preferred Knowledge and Experience

  • Experience with self-hosted data orchestration platforms (rather than managed services)
  • Background in implementing data contracts or schema governance
  • Knowledge of ML/AI data pipeline requirements and feature engineering
  • Experience with real-time data processing and streaming architectures
  • Familiarity with data modeling and warehouse design principles
  • Prior experience in a technical leadership role

#LI-HR1 #LI-ONSITE

Applied = 0

(web-759df7d4f5-mz8pj)