| Overview 
 
 Job Purpose We're seeking a talented Senior Data Engineer to join ICE Data Services in a cross-cutting role that will help define and implement our next-generation data platform. In this pivotal position, you'll lead the design and implementation of scalable, self-service data pipelines with a strong emphasis on data quality and governance. This is an opportunity to shape our data engineering practice from the ground up, working directly with key stakeholders to build mission-critical ML and AI data workflows. You'll be working with a modern, on-premises data stack that includes: 
 Apache Airflow for workflow orchestration (self-hosted on Kubernetes)dbt for data transformation and testingApache Flink for stream processing and real-time data workflowsKubernetes for containerized deployment and scalingGit-based version control and CI/CD for data pipelinesOracle Exadata for data warehousingKafka for messaging and event streaming We emphasize building systems that are maintainable, scalable, and focused on enabling self-service data access while maintaining high standards for data quality and governance. The ideal candidate is a problem-solver who enjoys working on complex data systems and is passionate about data quality. You thrive in collaborative environments but can also work independently to deliver solutions. You're comfortable working directly with technical and non-technical stakeholders and can communicate complex technical concepts clearly. Most importantly, you're excited about creating systems that empower others to work with data efficiently and confidently. Responsibilities 
 Design, build, and maintain our on-premises data orchestration platform using Apache Airflow, dbt, and Apache FlinkCreate self-service capabilities that empower teams across the organization to build and deploy data pipelines without extensive engineering supportImplement robust data quality testing frameworks that ensure data integrity throughout the entire data lifecycleEstablish data engineering best practices, including version control, CI/CD for data pipelines, and automated testingCollaborate with ML/AI teams to build scalable feature engineering pipelines that support both batch and real-time data processingDevelop reusable patterns for common data integration scenarios that can be leveraged across the organizationWork closely with infrastructure teams to optimize our Kubernetes-based data platform for performance and reliabilityMentor junior engineers and advocate for engineering excellence in data practices Knowledge and Experience 
 5+ years of professional experience in data engineering, with at least 2 years working on enterprise-scale data platformsDeep expertise with Apache Airflow, including DAG design, performance optimization, and operational managementStrong understanding of dbt for data transformation, including experience with testing frameworks and deployment strategiesExperience with stream processing frameworks like Apache Flink or similar technologiesProficiency with SQL and Python for data transformation and pipeline developmentFamiliarity with Kubernetes for containerized application deploymentExperience implementing data quality frameworks and automated testing for data pipelinesKnowledge of Git-based workflows and CI/CD pipelines for data applicationsAbility to work cross-functionally with data scientists, ML engineers, and business stakeholders Preferred Knowledge and Experience  
 Experience with self-hosted data orchestration platforms (rather than managed services)Background in implementing data contracts or schema governanceKnowledge of ML/AI data pipeline requirements and feature engineeringExperience with real-time data processing and streaming architecturesFamiliarity with data modeling and warehouse design principlesPrior experience in a technical leadership role #LI-HR1 #LI-ONSITE |