What are the responsibilities and job description for the Sr. Data Engineer position at Cars.com?
Data is the driver for our future at Cars. We’re searching for a collaborative, analytical, and innovative Sr Data Engineer to build scalable and highly performant platforms, systems and tools to enable innovations with data. If you are passionate about building large scale systems and data driven products, we want to hear from you.
Responsibilities Include:
- Build data pipelines and deriving insights out of the data using advanced analytic techniques, streaming and machine learning at scale
- Work within a dynamic, forward thinking team environment where you will design, develop, and maintain mission-critical, highly visible Big Data and Machine Learning applications
- Build, deploy and support data pipelines and ML models into production.
- Act as technology lead on projects to help drive best practice design decisions
- Work in close partnership with other Engineering teams, including Data Science, & cross-functional teams, such as Product Management & Product Design
Required Skills
- Ability to develop Spark jobs to cleanse/enrich/process large amounts of data.
- Experience with tuning Spark jobs for efficient performance including execution time of a job, execution memory, etc.
- Experience mentoring and leading developers and projects of medium to high complexity.
- Experience designing dimensional data models.
- Sound understanding of various file formats and compression techniques.
- Experience with source code management systems such as Github and developing CI/CD pipelines with tools such as Jenkins for data.
- Ability to understand deeply the entire architecture for a major part of the business and be able to articulate the scaling and reliability limits of that area; design, develop and debug at an enterprise level and design and estimate at a cross-project level.
- Excellent communication and collaboration skills.
Required Experience
- Software Engineering | 7 years of designing & developing complex, batch processes at enterprise scale; specifically utilizing Python and/or Scala.
- Big Data Ecosystem | 5 years of hands-on, professional experience with tools and platforms like PySpark, Airflow, and Redshift.
- AWS Cloud | 3 years of professional experience in developing Big Data applications in the cloud, specifically AWS.
Preferred:
- Experience working with Clickstream Data
- Experience working with digital marketing data
- Experience with developing REST APIs.
- Experience with Shell Scripting
- Experience in deploying ML models into production and integrating them into production applications for use.
- Experience with machine learning / deep learning using R, Python, Jupyter, Zeppelin, TensorFlow, etc.
#LI-JL #LI-REMOTE