5+ years of experience in designing and implementing large scale data processing/data storage/data distribution systems
Extensive experience working with large data sets with hands-on technology skills to design and build robust Big Data solutions using Spark framework, GCP Big data services and industry standard frameworks
Ability to work with multi-technology/cross-functional teams and key stakeholders to guide/manage a full life-cycle solution
Extensive experience in Relational and MPP database platforms like (GCP Bigquery/Hive/Cloud SQL etc)
Open source Hadoop stack/Spark framework
Strong understanding of Big Data Analytics platforms and ETL in the context of Big Data
Excellent problem solving, hands-on engineering skills and communication skills
Broad understanding and experience of real-time analytics
Participate in full Software Development Life Cycle (SDLC) of the Big Data Solution
Technical Skills Required
Any combination of below technical skills
Hadoop : HDFS, MapReduce, Hive, Airflow
DW : Bigquery, Hive
Languages : Python, PySpark, Shell Scripting, SQL Scripting