Data Engineer - Senior
Cummins Turbo Technologies
Data Science
Columbus, IN, USA
Posted on Mar 14, 2026
We are looking for a talented Data Engineer- Senior to join our team specializing in Systems/Information Technology for our Corporate organization in Columbus, IN.
In this role, you will make an impact in the following ways:
- Streamlining Data Integration
You’ll design and automate scalable systems to ingest and transform data from diverse sources, ensuring seamless and efficient data flow across the organization. - Safeguarding Data Quality
By implementing robust monitoring frameworks, you’ll proactively detect and resolve data integrity issues, maintaining trust in analytics and reporting. - Establishing Data Governance
You’ll lead the development of governance processes to manage metadata, access, and retention, ensuring compliance and secure data usage for internal and external stakeholders. - Building Scalable Data Pipelines
You’ll architect reliable and high-performance ETL/ELT pipelines with built-in monitoring and alerts, enabling timely and accurate data delivery for business needs. - Optimizing Database Design and Performance
Through thoughtful physical data modeling and indexing strategies, you’ll enhance database efficiency and scalability for large-scale operations. - Modernizing Data Infrastructure
You’ll develop and operate advanced storage and processing solutions using distributed and cloud platforms, supporting big data initiatives and analytics. - Automating Data Workflows
By leveraging modern tools and techniques, you’ll reduce manual data preparation tasks, boosting productivity and minimizing errors. - Mentoring and Agile Collaboration
You’ll coach junior team members and contribute to agile practices like DevOps and Scrum, accelerating delivery of critical analytics projects and fostering team growth.
Cummins is an equal opportunity employer. Our policy is to provide equal employment opportunities to all qualified persons without regard to race, sex, color, disability, national origin, age, religion, union affiliation, sexual orientation, veteran status, citizenship, gender identity, or other status protected by law.
Leads projects for design, development and maintenance of a data and analytics platform. Effectively and efficiently process, store and make data available to analysts and other consumers. Works with key business stakeholders, IT experts and subject-matter experts to plan, design and deliver optimal analytics and data science solutions. Works on one or many product teams at a time.
Core Responsibilities / Activities:
- Design and implement scalable and efficient data pipelines using Apache Spark and Databricks on Azure.
- Lead the complex transformation and integration of unstructured data sources into structured Delta Lake formats, applying software engineering best practices to ensure reliability, modularity, and reusability.
- Troubleshoot and optimize Spark jobs for performance, reliability, and cost-efficiency in a production environment.
- Drive continuous improvement of data engineering solutions by leveraging AI/ML and LLM-based techniques to enhance observability, performance optimization, and long-term maintainability.
Skill, Education, or Experience Requirements:
- Minimum of 5 years of hands-on experience in data engineering with expertise in Azure Databricks and programming in Scala or Python.
- Proven experience in building and maintaining structured streaming pipelines using Spark.
- Strong knowledge of big data technologies, including Delta Lake, Apache Spark, Structured Streaming, and SQL.
- Experience with Git for version control and CI/CD pipeline management.
Nice to Have (Preferences):
- Data Engineering Certification (e.g., Databricks Certified Data Engineer, Apache Spark Professional Data Engineer, or equivalent).
- Exposure to real-time data ingestion frameworks and cloud-native data services (e.g., Azure Event Hub, Azure Data Lake, AWS SQS, etc).
- Familiarity with data governance, access control (e.g., Unity Catalog or Immuta), and performance monitoring tools in cloud environments.
To be successful in this role you will need the following:
- System Requirements Engineering - Uses appropriate methods and tools to translate stakeholder needs into verifiable requirements to which designs are developed; establishes acceptance criteria for the system of interest through analysis, allocation and negotiation; tracks the status of requirements throughout the system lifecycle; assesses the impact of changes to system requirements on project scope, schedule, and resources; creates and maintains information linkages to related artifacts.
- Collaborates - Building partnerships and working collaboratively with others to meet shared objectives.
- Communicates effectively - Developing and delivering multi-mode communications that convey a clear understanding of the unique needs of different audiences.
- Customer focus - Building strong customer relationships and delivering customer-centric solutions.
- Decision quality - Making good and timely decisions that keep the organization moving forward.
- Data Extraction - Performs data extract-transform-load (ETL) activities from variety of sources and transforms them for consumption by various downstream applications and users using appropriate tools and technologies.
- Programming - Creates, writes and tests computer code, test scripts, and build scripts using algorithmic analysis and design, industry standards and tools, version control, and build and test automation to meet business, technical, security, governance and compliance requirements.
- Quality Assurance Metrics - Applies the science of measurement to assess whether a solution meets its intended outcomes using the IT Operating Model (ITOM), including the SDLC standards, tools, metrics and key performance indicators, to deliver a quality product.
- Solution Documentation - Documents information and solution based on knowledge gained as part of product development activities; communicates to stakeholders with the goal of enabling improved productivity and effective knowledge transfer to others who were not originally part of the initial learning.
- Solution Validation Testing - Validates a configuration item change or solution using the Function's defined best practices, including the Systems Development Life Cycle (SDLC) standards, tools and metrics, to ensure that it works as designed and meets customer requirements.
- Data Quality - Identifies, understands and corrects flaws in data that supports effective information governance across operational business processes and decision making.
- Problem Solving - Solves problems and may mentor others on effective problem solving by using a systematic analysis process by leveraging industry standard methodologies to create problem traceability and protect the customer; determines the assignable cause; implements robust, data-based solutions; identifies the systemic root causes and ensures actions to prevent problem reoccurrence are implemented.
- Values differences - Recognizing the value that different perspectives and cultures bring to an organization.
Education, Licenses, Certifications:
- College, university, or equivalent degree in relevant technical discipline, or relevant equivalent experience required.
- This position may require licensing for compliance with export controls or sanctions regulations.
Experience:
- Intermediate experience in a relevant discipline area is required.
- Knowledge of the latest technologies and trends in data engineering are highly preferred and includes:
- Familiarity analyzing complex business systems, industry requirements, and/or data regulations
- Background in processing and managing large data sets
- Design and development for a Big Data platform using open source and third-party tools
- SPARK, Scala/Java, Map-Reduce, Hive, Hbase, and Kafka or equivalent college coursework
- SQL query language
- Clustered compute cloud-based implementation experience
- Experience developing applications requiring large file movement for a Cloud-based environment and other data extraction tools and methods from a variety of sources
- Experience in building analytical solutions
Intermediate experiences in the following are preferred:
- Experience with IoT technology
- Experience in Agile software development