What are the responsibilities and job description for the GCP Data Engineer position at EPMA?
Our client is looking for a GCP Data engineer who has passion for data integration technologies and comes with data warehousing background. Someone who has experience in designing and coding ETL and one who wants to be part of a team that is actively delivering projects in Teradata, Big Data and working towards migrating to cloud technology.
Role: GCP Data Engineer
Duration: Long term Contract (12-18 months) With a high possibility of converting to full time
Location: Austin or Westlake, TX on-site in a hybrid model (3 days per week onsite)
W2 Employment only. No C2C/C2H
Job Description:
Job Description:
Data engineer who has passion for data integration technologies and comes with data warehousing background. Someone who has experience in designing and coding ETL and one who wants to be part of a team that is actively delivering projects in Teradata, Big Data and working towards migrating to cloud technology.
Must have Skills:
- 8 years of experience as a technical BA/Data Engineer
- Data warehousing/information management, preferably using Teradata database platform, Big Data, and Google Cloud Platform (GCP)
- Must have at least 1-year of recent experience using Google Cloud Big Query
- Large scale integration and cloud migration projects
- Expert-level SQL
- Experience in Data modeling
- Minimum of 7 years of experience in understanding of best practices for building and designing ETL code
- Strong SQL experience with the ability to develop, tune and debug complex SQL applications is required
- Knowledge in schema design, developing data models and proven ability to work with complex data is required
- Experience in ETL such as IICS/Informatica and Talend tools is required
- Strong Teradata coding and utilities experience is required
- Experience in working in large environments such as RDBMS, EDW, NoSQL, etc. is required
- Hands-on experience with Big Query, Hive, SPARK, and Kafka is a plus
- Understanding Hadoop file format and compressions is a plus
- Experience with scheduling tools (eg. Control M, ESP)
- Understanding of best practices for building Data Lake and analytical architecture on Hadoop is preferred
- Knowledge in real time data ingestion into Big Query is preferred
- Experience with Test Driven Code Development, SCM tools such as GIT, Jenkins is preferred
- Experience collaborating with business and technology partners and offshore development teams
- Good interpersonal, analytical, problem-solving, and organizational skills
- Excellent written/verbal communication skills
Responsibilities:
- Lead team in execution of development and delivery of the organization’s software products to QA, and ultimately to Production
- Manage day-to-day activities and influence team members to leverage, follow and support Agile software development practices
- Interface with product owner and key stakeholders in Project Management, Business, QA and Technology Operations to ensure high quality delivery of software projects on time and budget
- Work with Scrum Master and product owner to provide development sizing & cost analysis estimates
- Creating/updating ETL specifications and supporting documentation
- Developing ETL logic utilizing Informatica workflows, scripting and load utilities
- Building and maintaining code for big data ingestion using Talend, Sqoop, Hive etc
- Designing schemas, data models and data architecture for Teradata and Big Query environments
- Designing, building and support data processing pipelines to transform data in Big Query or Teradata platforms
- Developing and executing quality assurance and test scripts
- Work with business analysts to understand business requirements and use cases
- Problem solving and fixing technical issues
- What you have