Design & Development
* Create and maintain ETL/ELT pipelines for large datasets in formats like CSV, JSON, Parquet, and Avro using Hadoop MapReduce and Spark frameworks. +2
Data Modeling
* Analyze existing schemas and recommend improvements to E-R and dimensional models. Develop logical and physical data models. +2
Performance Optimization
* Configure metadata, fine-tune pipeline performance, and optimize job execution. +10
Collaboration
* Work with cross-functional teams (DevOps, BI, stakeholders) to ensure smooth data workflows and scalable solutions. +10
Governance & Quality
* Apply best practices for data modeling, governance, and quality controls
Skills Required
Csv, Json, Devops, Bi, Etl