What are the responsibilities and job description for the Data Engineer/Data Scientist position at MinIO?
MinIO’s first Data Engineer/Data Scientist will be a key member of the go to market team at MinIO doing innovative work that will fundamentally alter the trajectory of the organization. This unique position requires expertise in both data science and data engineering, combining analytical and programming skills to drive meaningful insights and build robust data infrastructure.
Responsibilities:
- Data Analysis and Modeling:
- Conduct advanced statistical analyses to extract actionable insights from large datasets.
- Develop and implement machine learning models to solve complex business problems.
- Collaborate with cross-functional teams to understand business requirements and translate them into analytical solutions.
- Database
- Ongoing SQL queries for ad hoc reporting
- Advanced SQL querying skills and optimization for fast queries (advanced joins and nested queries)
- Data Engineering:
- Design, develop, and maintain scalable data pipelines and ETL processes.
- Build and optimize data architectures, databases, and large-scale processing systems.
- Ensure data quality, integrity, and availability throughout the entire data lifecycle.
- Programming and Scripting:
- Proficient in programming languages such as Python and R for data manipulation and analysis.
- Develop and deploy code for data processing, feature engineering, and model deployment.
- Collaboration and Communication:
- Work closely with stakeholders to understand data needs and provide actionable insights.
- Communicate complex technical concepts to non-technical stakeholders in a clear and concise manner.
Qualifications:
- Bachelor's or Master's degree in Computer Science, Data Science, or a related field.
- Proven experience in both data science and data engineering roles.
- Strong proficiency in machine learning algorithms, statistical modeling, data visualization and data engineering
- Expertise in database design, data warehousing, and ETL processes.
- Advanced programming skills in languages such as Python and R
- Familiarity with big data technologies and cloud platforms (e.g., Hadoop, Spark, AWS, Azure).
- Experience with cloud infrastructure (Azure, Google Cloud Platform and AWS)
- Experience with datalakes
- Excellent communication and collaboration skills.
Health, Dental, Vision, 401K, Maternity/Paternity Paid Leave