Short Description:
This position at Cummins is for a Data Engineer, focusing on supporting and maintaining their data and analytics platform. Key responsibilities include data processing, quality monitoring, data governance, and developing data pipelines using ETL/ELT tools. The role involves implementing data storage architectures, operating large-scale data solutions, and applying agile methodologies. Required skills include data extraction, solution documentation, quality assurance metrics, and proficiency in SQL. Preferred qualifications encompass exposure to big data technologies, cloud computing (Azure), and knowledge of IoT. The position is based in Pune, India, and offers flexibility in location, with a focus on remote work.
At our core, we believe in unlocking your potential. We offer global opportunities to advance your career, contribute to your community's betterment, and collaborate with today's most innovative minds to tackle the world's most pressing challenges. We prioritize flexibility, enabling you to explore your passions while making a meaningful impact within our inclusive workforce. This is the essence of #LifeAtCummins.
Position: Data Engineer
Description:
As a Data Engineer, you'll play a vital role in supporting, developing, and maintaining our data and analytics platform. Your responsibilities encompass efficiently processing, storing, and providing data access to analysts and other stakeholders. You'll collaborate closely with both Business and IT teams to understand requirements and leverage technology for agile data delivery at scale.
Key Responsibilities:
- Implement and automate the deployment of our distributed system, which ingests and transforms data from various sources (relational, event-based, unstructured).
- Continuously monitor and troubleshoot data quality and integrity issues.
- Establish data governance processes and metadata management for internal and external users.
- Develop robust, efficient, scalable data pipelines with monitoring and alerting mechanisms, incorporating diverse data sources using ETL/ELT tools or scripting languages.
- Create physical data models and implement data storage architectures according to design guidelines.
- Analyze complex data elements, systems, data flow, dependencies, and relationships to contribute to conceptual physical and logical data models.
- Participate in testing and troubleshooting of data pipelines.
- Develop and operate large-scale data storage and processing solutions using various distributed and cloud-based platforms (e.g., Data Lakes, Hadoop, Hbase, Cassandra, MongoDB, Accumulo, DynamoDB, etc.).
- Utilize agile development methodologies like DevOps, Scrum, Kanban, and continuous improvement cycles in data-driven applications.
Required Skills:
- Data Extraction: Perform ETL activities from various sources, transforming data for downstream applications and users.
- Solution Documentation: Document knowledge gained during product development for effective knowledge transfer.
- Quality Assurance Metrics: Apply measurement science to assess if solutions meet intended outcomes.
- Solution Validation Testing: Validate configuration changes to ensure they meet customer requirements.
- System Requirements Engineering: Translate stakeholder needs into verifiable requirements and establish acceptance criteria.
- Problem Solving: Employ systematic analysis and industry-standard methodologies to solve problems.
- Data Quality: Identify and correct data flaws to support information governance.
- Programming: Create, write, and test computer code, test scripts, and build scripts.
- Customer Focus: Build strong customer relationships and deliver customer-centric solutions.
- Decision Quality: Make timely decisions to propel the organization forward.
- Collaboration: Work collaboratively to meet shared objectives.
- Effective Communication: Deliver clear communications tailored to different audiences.
Education, Licenses, Certifications:
- Preferred: College, university, or equivalent degree in a relevant technical discipline.
- Licensing may be required for compliance with export controls or sanctions regulations.
Experience:
- Preferred: Relevant experience such as temporary student employment, internships, co-op programs, or extracurricular team activities.
Preferred Knowledge:
- Exposure to Big Data open-source technologies.
- Familiarity with SPARK, Scala/Java, Map-Reduce, Hive, Hbase, Kafka, or equivalent coursework.
- Proficiency in SQL query language.
- Experience with clustered compute cloud-based implementations.
- Knowledge of developing applications with large file movement in a Cloud-based environment.
- Exposure to Agile software development.
- Familiarity with building analytical solutions.
- Knowledge of IoT technology.
Key Skills:
- Cloud Computing (Azure): Azure Data Factory, Azure Databricks, Azure Synapse Analytics, Azure Data Lake Storage, Azure SQL Databases.
- Relational Databases: Oracle, SQL Server, PostgreSQL.
- Good to Have: Power BI, Snowflake.
Job Details:
- Job Type: Systems/Information Technology
- Location: India-Maharashtra-Pune-India, Pune, IOC Tower A
- Job Type: Experienced - Exempt / Office
- Recruitment Job Type: Exempt - Experienced
- Job Posting Date: Sep 14, 2023, 1:30:00 PM
- Unposting Date: Ongoing
- Organization: Corporate
- Role Category: Remote (Flexible Location)
- Relocation Package: Eligible
- Req ID: 230008UK
Please click here to apply.
