DE Jobs

Search from over 2 Million Available Jobs, No Extra Steps, No Extra Forms, Just DirectEmployers

Job Information

Cummins Inc. Data Engineer - Senior in Pune, India

DESCRIPTION

GPP Database Link (https://cummins365.sharepoint.com/sites/CS38534/)

Job Summary:

Leads projects for the design, development, and maintenance of a data and analytics platform. Effectively and efficiently processes, stores, and makes data available to analysts and other consumers. Works with key business stakeholders, IT experts, and subject-matter experts to plan, design, and deliver optimal analytics and data science solutions. Works on one or many product teams at a time. Though the role category is generally listed as Remote, this specific position is designated as Hybrid.

Key Responsibilities:

  • Business Alignment & Collaboration – Partner with the Product Owner to align data solutions with strategic goals and business requirements.

  • Data Pipeline Development & Management – Design, develop, test, and deploy scalable data pipelines for efficient data transport into Cummins Digital Core (Azure DataLake, Snowflake) from various sources (ERP, CRM, relational, event-based, unstructured).

  • Architecture & Standardization – Ensure compliance with AAI Digital Core and AAI Solutions Architecture standards for data pipeline design and implementation.

  • Automation & Optimization – Design and automate distributed data ingestion and transformation systems, integrating ETL/ELT tools and scripting languages to ensure scalability, efficiency, and quality.

  • Data Quality & Governance – Implement data governance processes, including metadata management, access control, and retention policies, while continuously monitoring and troubleshooting data integrity issues.

  • Performance & Storage Optimization – Develop and implement physical data models, optimize database performance (indexing, table relationships), and operate large-scale distributed/cloud-based storage solutions (Data Lakes, Hadoop, HBase, Cassandra, MongoDB, Accumulo, DynamoDB).

  • Innovation & Tool Evaluation – Conduct proof-of-concept (POC) initiatives, evaluate new data tools, and provide recommendations for improvements in data management and integration.

  • Documentation & Best Practices – Maintain standard operating procedures (SOPs) and data engineering documentation to support consistency and efficiency.

  • Agile Development & Automation – Use Agile methodologies (DevOps, Scrum, Kanban) to drive automation in data integration, preparation, and infrastructure management, reducing manual effort and errors.

  • Coaching & Team Development – Provide guidance and mentorship to junior team members, fostering skill development and knowledge sharing.

RESPONSIBILITIES

Competencies:

  • System Requirements Engineering: Translates stakeholder needs into verifiable requirements, tracks status, and assesses impact changes.

  • Collaborates: Builds partnerships and works collaboratively with others to meet shared objectives.

  • Communicates Effectively: Delivers multi-mode communications tailored to different audiences.

  • Customer Focus: Builds strong customer relationships and provides customer-centric solutions.

  • Decision Quality: Makes good and timely decisions that drive the organization forward.

  • Data Extraction: Performs ETL activities from various sources using appropriate tools and technologies.

  • Programming: Develops, tests, and maintains code using industry standards, version control, and automation tools.

  • Quality Assurance Metrics: Measures and assesses solution effectiveness using IT Operating Model (ITOM) standards.

  • Solution Documentation: Documents knowledge gained and communicates solutions for improved productivity.

  • Solution Validation Testing: Validates configurations and solutions to meet customer requirements using SDLC best practices.

  • Data Quality: Identifies, corrects, and manages data flaws to support effective governance and decision-making.

  • Problem Solving: Uses systematic analysis to determine root causes and implement robust solutions.

  • Values Differences: Recognizes and leverages the value of diverse perspectives and cultures.

Education, Licenses, Certifications:

  • Bachelor's degree in a relevant technical discipline, or equivalent experience required.

  • This position may require licensing for compliance with export controls or sanctions regulations.

QUALIFICATIONS

Preferred Experience:

  • Technical Expertise – Intermediate experience in data engineering with hands-on knowledge of SPARK, Scala/Java, MapReduce, Hive, HBase, Kafka, and SQL.

  • Big Data & Cloud Solutions – Proven ability to design and develop Big Data platforms, manage large datasets, and implement clustered compute solutions in cloud environments.

  • Data Processing & Movement – Experience developing applications requiring large-scale file movement and utilizing various data extraction tools in cloud-based environments.

  • Business & Industry Knowledge – Familiarity with analyzing complex business systems, industry requirements, and data regulations to ensure compliance and efficiency.

  • Analytical & IoT Solutions – Experience building analytical solutions with exposure to IoT technology and its integration into data engineering processes.

  • Agile Development – Strong understanding of Agile methodologies, including Scrum and Kanban, for iterative development and deployment.

  • Technology Trends – Awareness of emerging technologies and trends in data engineering, with a proactive approach to innovation and continuous learning.

Technical Skills:

  • Programming Languages: Proficiency in Python, Java, and/or Scala.

  • Database Management: Expertise in SQL and NoSQL databases.

  • Big Data Technologies: Hands-on experience with Hadoop, Spark, Kafka, and similar frameworks.

  • Cloud Services: Experience with Azure, Databricks, and AWS platforms.

  • ETL Processes: Strong understanding of Extract, Transform, Load (ETL) processes.

  • Data Replication: Working knowledge of replication technologies like Qlik Replicate is a plus.

  • API Integration: Experience working with APIs to consume data from ERP and CRM systems.

Job Systems/Information Technology

Organization Cummins Inc.

Role Category Remote

Job Type Exempt - Experienced

ReqID 2410681

Relocation Package No

DirectEmployers