What are the responsibilities and job description for the Sr. Data Lake Developer position at Expleo?
Overview
Trissential’s a trusted partner for end-to-end quality services and management consulting for digital transformation. As a part of our parent company Expleo, we’re a global organization partnering with major corporations and leading non-profits in over 30 countries. Guided by our mission and values, Trissential puts people at the heart of our organization.
Come join an experience. Add your talent to a team of forward-thinking game changers that make an impact by driving innovative solutions.
Trissential is currently seeking a Sr. Data Lake Developer to join our dynamic team in St Paul, MN (Remote).
Responsibilities
Job Overview:
- The client is seeking to build a data lake repository of project and portfolio management (PPM) data for the clients enterprise to enable the organization to make data-driven decisions for its IT projects.
- The database will be used as a single source of truth for both enterprise and the clients partner agency reporting needs, along with making data available for advanced analytics purposes.
- The work will be done in phases.
- The pilot phase will be focused on getting data from Sciforma and the statewide integrated financial tool (SWIFT) (project financials).
- The long term goal is to capture data from all PPM tools being used throughout clients partner agencies in the data lake.
- Additional phases may be added via amendment to the work order.
Responsibilities:
- Analyze and define data requirements
- Establish automated data extraction process for Sciforma, SWIFT and other PPM tools (i.e., API or other preferred methods)
- Develop data lake structure and populate with data
- Develop reports/queries for advanced data analytics
- Develop Power BI reports/data visualization using advanced queries
- Provide knowledge transfer
Qualifications
Required:
- Five years of experience in data lake development
- Two engagements with two different entities in a role involving data lake development
Preferred:
- Experience in data lake configuration, setup, and bringing in data from various data sources using ETL, API
- Experience in Azure Cloud services and solutions
- Experience working with enterprise data warehouse
- Experience as an ETL/ELT Developer using various ETL/ELT tools such as Azure Synapse Pipelines and Azure Data Factory, Apache Spark Pools using python scripts
- Experience in Azure DevOps Services using Azure Git Repos, Azure ¬data studio, Azure Analytics, data mapping, deployment artifacts and release packages for test & production environment
- Experience in building end-end scalable data solution, from sourcing raw data, transforming data to producing analytics reports
- Experience in Python (ETL and Data Visualization libraries)
- Experience in Azure SQL databases across SQL DB, Managed instance & Data warehouse
- Experience in Azure platform services such as blob storage, event hubs, monitoring services
- Experience in creating data structures optimized for storage and various query patterns, for example Parquet
- Experience in building secured Power BI reports, dashboards, paginated reports
- Experience in working in an Agile SDLC methodology
Skills
- Works independently or well within a team
- Wants to continuously grow knowledge base and skill set
- Collaborative, consultative mindset
- Works well in a fast paced environment
- Strong technical background
- Deep knowledge and curiosity about technology and systems
- Agile mindset