Locations:. Hybrid - Onsite in McLean, VA Tuesday through Thursday.
Job Description:
Must Haves: Strong understand of SDLC, Python, R, proficient in SQL, Spark, SQL, Data processing frameworks- Apache Spark, Cucumber, Bitbucket and PySpark.
Notes:
This is a Data Engineering role mixed with Data Analyst (No Data Science)
CANDIDATES WITHOUT PYTHON EXPERIENCE WILL NOT BE CONSIDERED.
This team will be working with the business area and will help with business solutions .assist with gathering requirements, and work with engineering teams to help build the data solutions
Candidates will need to have Python, SQL, any Data Framework like Apache spark, DB management and Data Lakes experience
Prior fintech experience is preferred, especially in the process side (Like using Jenkins for deployments)
Be able to handle some ETL work
Role may require after hours prod support later on. For now, no afterhours work is needed
Qualifications:
• Bachelor's degree in computer science or related discipline; advanced Studies/ Degree preferred.
• 5-7 years of experience in data engineering with strong proficiency in Python programing language (Nice to have: Scala, Java).
• Strong proficiency in spark SQL, SQL, gremlin, GraphQL and database management (Snowflake cloud-base warehousing).
• Strong experience with data processing frameworks, DataFrames, Apache Spark, Graph DBs • Experience with cloud platform AWS (EMR, EKS, Lambda).
• Experience writing statistical and/or optimization programs to develop models and algorithms.
• Programming languages may include-but are not limited to-Python, R • Solid understanding of SDLC practices including development, testing, and release management.
• Experience with version control systems such as Git, BitBucket • Experience with RESTful API design and development.
• Familiarity with PySpark for large-scale data processing and analysis.