Data warehousing , pyspark , Github, AWS data platform, Glue, EMR, RedShift, databricks,Data Marts. DBT/Glue/EMR or Matillion,Bigdata ecosystem (Hadoop), data architecture, data engineering, data modelling, data consumption
10-15 years of total experience and at least 3+ years of expertise in Cloud data warehouse technologies on AWS data platform covering - Glue, EMR, RedShift, databricks etc.
At least one End-to-end AWS data platform implementation is a must covering all aspects including architecture, design, data engineering, data visualization and data governance (specifically data quality and lineage).
Significant experience with data migrations and development of Operational Data Stores, Enterprise Data Warehouses, Data Lake and Data Marts.
Good hands-on knowledge on SQL and Data Warehousing life cycle is an absolute requirement.
Significant experience with data migrations and development, design, Operational Data Stores, Enterprise Data Warehouses and Data Marts.
Experience with cloud ETL and ELT in one of the tools like DBT/Glue/EMR or Matillion or any other ELT tool and exposure to Bigdata ecosystem (Hadoop).
Expertise with at least one of the Traditional data warehouses solutions on Oracle, Teradata, and Oracle Exadata.
Excellent communication skills to liaise with Business & IT stakeholders.
Expertise in planning execution of a project and efforts estimation.
Understanding of Data Vault, data mesh and data fabric architecture patterns.