What are the responsibilities and job description for the Data Engineer position at ICONMA?
Job Description
Data Engineer:
Location: McLean, VA/ Hybrid
Duration: 9 months with possible extension/ possible contract to hire
Description:
Must haves: 6 years of hands-on Python development experience using PySpark, XML, Json. Database side they need to have experience with SQL using Hive. 6 years of experience with Hadoop is mandatory. Preference for Snowflake. Hands- on experience with Cloud technologies is preference and could result in contract extensions supporting future projects
They will be working with the development, helping the analyst, and testing.
As a client, you will do important work to build a better housing finance system and you’ll be part of a team helping to make homeownership and rental housing more accessible and affordable across the nation.
Your Impact:
You will be able to apply advanced Data Engineering and Machine learning skills to solve real world challenges in building applications that help the company build better models and do advanced reporting.
- Cleanse, manipulate and analyze large datasets (Structured and Unstructured data – XMLs, JSONs, PDFs) using Hadoop platform.
- Develop Python, PySpark, Spark scripts to filter/cleanse/map/aggregate data.
- Manage and implement data processes (Data Quality reports)
- Develop data profiling, deduping logic, matching logic for analysis
- Programming Languages experience in Python, PySpark and Spark for data ingestion.
- Programming experience in BigData platform using Hadoop platform.
- Present ideas and recommendations on Hadoop and other technologies best use to management.
Qualifications:
- Bachelor’s degree in Computer Science, Statistics, Data science or a related quantitative field.
- 5 years of experience in processing large volumes and variety of data (Structured and unstructured data, writing code for parallel processing, XMLS, JSONs, PDFs)
- 5 years of programming experience in Hadoop, Spark, Python for data processing and analysis.
- Strong SQL experience is a must
- 5 years of experience using Hadoop platform and performing analysis. Familiarity with Hadoop cluster environment and configurations for resource management for analysis work
- 2 Prior experience working in Cloud platforms AWS. Kubernetes experience is highly desirable
- Hands on Work experience using technologies for manipulating structured and unstructured big data. Big data technologies may include—but are not limited to—Hadoop, Hive, Spark, relational databases, and NoSQL .
- Desired prior experience with MPP Databases like SnowFlake.
- Detail oriented and superb communication and written skills
- Must be able to prioritize and meet deadlines
Keys to Success in this Role:
The candidate should be very analytical minded, have a good grasp of data architectures and keen in problem solving. We are looking for someone with good data Engineering skills along with good exploratory data analysis experience.
- Ability to think critically and logically.
- Solid communications skills – verbal and written.
- Detail oriented and excellent organization skills.
- Strong quantitative, data analytics, analytical, and problem-solving skills.