Lead Cloud Data Engineer
Work From Home
Applications have closed
Protective
Protect your future with a life insurance policy or annuity from Protective. Discover options that give you and your family confidence for what's ahead.Protective is looking for a Lead Cloud Data Engineer to join the Emerging Technology team, which is creating a cloud native “Data Hub.” We are looking for someone that can help establish standards and procedures around architecture and Data Products, including code formatting, documentation, testing and automation of patterns established. A good understanding of cloud platforms and prior experience working with cloud-based ELT tools and PySpark frameworks is required. They must be comfortable working with a wide range of stakeholders and functional teams.
Primary Duties
- Technology lead who has experience and played a crucial role in successfully implementing enterprise solutions, in a cloud environment, using a vast range of data engineering and cloud technologies.
- Works on tasks independently and leads the work of junior engineers and data analysts across different business areas.
- Leads the design and build enterprise packages that help streamline development and running quality data pipelines.
- Leads and consults on the coordination of different in-progress data product development across the enterprise.
- Design documentation, test cases/tests ensuring reliability, quality and standardization (where needed)
- Leads conversations to evangelize the best practices for Data Engineer pipelines, DataOps and Testing.
- Shared ownership of teams’ agenda and goals while implementing required governance, monitors health.
- Responsible for the technical mentorship and guidance of junior team members and “super users” across different areas in the company.
Responsibilities
- Build out a Spark pipeline testing framework.
- ·Build and share best practices for Pipeline Development.
- Design and Implement Data Science Model Deployment.
- Maintain Protective data “packages” to utilize standard tooling and patterns.
- Build out internal monitoring for the DataHub.
- o What is in the Hub by data domain
- o How much is in the Hub by data domain
- o Who is using the Hub by data domain
- Build and Maintain data quality frameworks that leverage industry best practices.
Requirements
- Strong problem-solving skills with an emphasis solution design and development.
- 7-10 years of experience building data pipelines using ETL tools with at least 3 years of cloud data engineering background.
- 3-5 years of experience in using PySpark based solutions (Databricks etc.).
- 3+ years of CI/CD experience.
- 3+ years of experience on public cloud. (Azure experience a plus)
- 6+ years' experience with Python (Pandas, pytest) and SQL.
- Proficiency working with large data stores.
- Experience building cloud-based warehouses (Lakehouse, Snowflake, etc.).
- Excellent written and verbal communication skills for coordinating across teams.
- Good experience in end-to-end implementation of DW BI projects, especially in data warehouse and mart developments).
- Experience in preparing test scripts and test cases to validate data and maintaining data quality.
- Strong understanding and hands-on programming/scripting experience skills – UNIX shell, and Python Scripts.
- Good understanding of Data integration, Data Quality.
- Deep understanding of all aspects of database technologies - schema design, optimization, scalability, etc.
- Experience in SDLC, especially testing and deployment.
Employee Benefits: We aim to protect our employees' wellbeing through a broad benefits offering. For example, we protect physical wellbeing through health, dental and vision insurance. We protect mental wellbeing through mental health benefits and an employee assistance program. We protect time away from work with a variety of paid time away benefits (e.g., paid time off, paid parental leave, short-term disability, and a cultural observance day). We protect financial wellbeing through contributions to healthcare accounts, a pension plan, and a 401(k) plan with Company matching. All employees are encouraged to protect their overall wellbeing by engaging in ProHealth Rewards, Protective’s platform to improve wellbeing while earning cash rewards. Eligibility for certain benefits may vary by position in accordance with the terms of the Company’s benefit plans.
Diversity and Inclusion: At Protective, we are committed to providing an inclusive culture where all employees are able to fully contribute and thrive. Our goal is to grow and develop our people, attract diverse talent and support strong, diverse communities.
We support diversity, equity, and inclusion by working to develop a culture of inclusion and belonging led by leaders who develop potential and embrace unique skills and abilities. Our aim is to create an equitable and accountable environment for all leaders and employees that will drive performance and impact business strategy. In this way, we can increase overall diversity for leadership roles and pipelines of talent by maturing our hiring practices, robust development opportunities and focus on retention of key talent.
We are proud to be an equal opportunity employer committed to being inclusive and attracting, retaining, and growing the talents of a diverse and inclusive workforce.
Tags: Architecture Azure CI/CD Databricks DataOps Data pipelines Data quality Data warehouse ELT Engineering ETL Model deployment Pandas Pipelines PySpark Python SDLC Snowflake Spark SQL Testing
Perks/benefits: 401(k) matching Career development Equity Health care Insurance Parental leave
More jobs like this
Explore more AI, ML, Data Science career opportunities
Find even more open roles in Artificial Intelligence (AI), Machine Learning (ML), Natural Language Processing (NLP), Computer Vision (CV), Data Engineering, Data Analytics, Big Data, and Data Science in general - ordered by popularity of job title or skills, toolset and products used - below.
- Open Lead Data Analyst jobs
- Open MLOps Engineer jobs
- Open Data Science Manager jobs
- Open Senior Business Intelligence Analyst jobs
- Open Data Engineer II jobs
- Open Data Manager jobs
- Open Principal Data Engineer jobs
- Open Sr Data Engineer jobs
- Open Power BI Developer jobs
- Open Business Intelligence Developer jobs
- Open Data Analytics Engineer jobs
- Open Junior Data Scientist jobs
- Open Data Scientist II jobs
- Open Product Data Analyst jobs
- Open Senior Data Architect jobs
- Open Sr. Data Scientist jobs
- Open Business Data Analyst jobs
- Open Big Data Engineer jobs
- Open Data Analyst Intern jobs
- Open Manager, Data Engineering jobs
- Open Azure Data Engineer jobs
- Open Data Product Manager jobs
- Open Data Quality Analyst jobs
- Open Junior Data Engineer jobs
- Open Principal Data Scientist jobs
- Open GCP-related jobs
- Open Data quality-related jobs
- Open Business Intelligence-related jobs
- Open ML models-related jobs
- Open Java-related jobs
- Open Data management-related jobs
- Open Privacy-related jobs
- Open Data visualization-related jobs
- Open Finance-related jobs
- Open Deep Learning-related jobs
- Open APIs-related jobs
- Open PhD-related jobs
- Open TensorFlow-related jobs
- Open PyTorch-related jobs
- Open Consulting-related jobs
- Open NLP-related jobs
- Open Snowflake-related jobs
- Open CI/CD-related jobs
- Open Kubernetes-related jobs
- Open LLMs-related jobs
- Open Generative AI-related jobs
- Open Data governance-related jobs
- Open Hadoop-related jobs
- Open Airflow-related jobs
- Open Docker-related jobs