Job Title: Consultant | PySpark | Delhi | Engineering
Job Summary:
We are looking for a detail-oriented and technically proficient professional with strong expertise in Python and SQL as core competencies. The ideal candidate will also have working knowledge of Power BI or Tableau for data visualization, and exposure to PySpark is a plus. This role involves building robust data pipelines, performing advanced analytics, and delivering actionable insights through visual storytelling.
Key Responsibilities:
- Develop and maintain scalable data pipelines using Python and SQL.
- Perform data extraction, transformation, and loading (ETL) from various sources.
- Conduct exploratory data analysis and statistical modeling to support business decisions.
- Create dashboards and reports using Power BI or Tableau to visualize key metrics.
- Collaborate with cross-functional teams to understand data requirements and deliver solutions.
- Optimize query performance and ensure data integrity across systems.
- Contribute to big data processing tasks using PySpark (nice to have).
- Document processes, workflows, and technical specifications.
Required Skills:
Primary Skills:
- Python: Strong experience with data manipulation, scripting, and libraries such as pandas, numpy, and matplotlib.
- SQL: Advanced proficiency in writing and optimizing queries, joins, and stored procedures.
Secondary Skills:
- Power BI or Tableau: Ability to create interactive dashboards and visualizations.
- PySpark: Exposure to distributed data processing and Spark-based workflows (preferred but not mandatory).
Preferred Qualifications:
- Bachelor’s or Master’s degree in Computer Science, Data Science, or related field.
- Experience with cloud data platforms (Azure, AWS, GCP) is a plus.
- Familiarity with version control (Git) and Agile development practices.