DescriptionRole and responsibilities
• Strong technical, analytical, and problem-solving skills • Strong organizational skills, with the ability to work autonomously as well as in a team-based environment • Data pipeline template development • Data pipeline Monitoring development & support (operations)
Technical skills requirements
The candidate must demonstrate proficiency in,
• On-premise servers for data processing and extraction
• Ability to own and deliver on large, multi-faceted projects
• Fluency in complex SQL and experience with RDBMSs
• (Bigdata, Spark, PySpark, Scala, Python, NiFi, Hive, NoSql DBs)
• Experience designing and building big data pipelines
• Experience working on large scale, distributed systems
• Experience working on Azure Cloud services like (Azure Data Factory, Databricks, ADLS, Azure HD Insights, Azure Synapse, Azure Data warehouse)
• Strong hands-on experience of programming language like PySpark, Scala with Spark, Python.
• Exposure to various ETL and Business Intelligence tools
• Solid grounding in Agile methodologies
• Experience with git and other source control systems
• Strong communication and presentation skills
Nice-to-have skills
• Certification in Hadoop/Big Data – Hortonworks/Cloudera
• Azure Cloud certification
• Unix or Shell scripting
• Strong delivery background across the delivery of high-value, business-facing technical projects in major organizations
• Experience of managing client delivery teams, ideally coming from a Data Engineering / Data Science environment