Location: El Segundo, CA (Onsite role from day 1, need local profiles)
Other Requirement
Job Description:
Has worked in the past as an Architect.
Professional experience in Data Engineering using Google Big Query (GBQ) and Google Cloud Platform (GCP) data sets and building data pipelines.
Hands on and deep experience working with Google Data Products (e.g. BigQuery, Dataflow, Dataproc, Dataprep, Cloud Composer, Airflow, DAG etc.).
Python Programming Expert PySpark, Pandas
Experience in Airflow (Create DAG, Configure the variables in Airflow, Scheduling)
Big Data technologies and solutions (Spark, Hadoop, Hive, MapReduce) and multiple scripting and languages (YAML, Python).
Experience in DBT to create the lineage in GCP. - Optional
Worked in Dev-Sec-Ops (CICD) environment.
Design and develop the ETL ELT framework using BigQuery, Expertise in Big Query concepts like Nested Queries, Clustering, Partitioning, etc.
Experience in Data Integration, Data Transformation, Data Quality and Data Lineage tools.
Should be able to automate the data load from Big Query using APIs or scripting language.
E2E Data Engineering and Lifecycle (including non-functional requirements and operations) management.
E2E Solution Design skills - Prototyping, Usability testing and data visualization literacy.
Experience with SQL and NoSQL modern data stores.
Job Description
(1.) To design and architect large-scale solutions, ensuring scalability, performance, and security.
(2.) To train and develop team so as to ensure that there is an adequate supply of trained manpower in the said technology and delivery risks are mitigated.
(3.) To continuously upskill with cutting-edge tech to deliver high-quality, future-proof solutions meeting client expectations and industry standards.
(4.) To leverage domainortech expertise to gather client needs, deliver solutions, and craft a technology strategy aligned with business goals.