DataOps Engineer
Poland - Warsaw
Applications have closed
Veeva Systems
Veeva Systems Inc. is a leader in cloud-based software for the global life sciences industry. Committed to innovation, product excellence, and customer success, Veeva has more than 1,100 customers, ranging from the world's largest...The Role
Veeva OpenData supports the industry by providing real-time reference data across the complete healthcare ecosystem to support commercial sales execution, compliance, and business analytics. We drive value to our customers through constant innovation, using cloud-based solutions and state-of-the-art technologies to deliver product excellence and customer success. The OpenData Global Data Tools team delivers the tools and data processing pipelines to build the global data core for life sciences in 100+ countries. As a DataOps engineer of the Global Data Tools team, you will design the data assembly line that allows deriving insights from data faster and with fewer errors. You will be responsible for creating the tools and the processes used to store, manage and process all compiled data to build the OpenData Reference.
What You'll Do
- Build DataOps tools for data workflows automation and streamline data processing (e.g., reusable software libraries, tools to orchestrate data processing tasks and its dependencies and components to enable CI/CD integrations)
- Adopt solutions and tools that adhere to the DataOps best practices
- Continually strive to reduce wasted effort, identify gaps and correct them, and improve data development and deployment processes
- Develop the ingestion pipelines for raw data
- Put in place the building blocks to deliver the data core for life sciences
Requirements
- Proficient in Python programming language and PySpark
- 3+ years of experience working with Apache Spark
- Previous experience building tools and libraries to automate and streamline data processing workflows
- Experience running data workflows through DevOps pipelines
- Experience orchestrating data workflows using state-of-the-art tools (e.g., Airflow, AWS Steps, or similar from other cloud vendors), spawning jobs in a Spark cloud-managed cluster (e.g., EMR, Databricks)
- Experience with the Delta Lake Architecture and the delta format
Nice to Have
- Hands-on experience using DevOps tools to deploy and administer clusters in a managed Apache Spark platform in the cloud (e.g., Databricks, AWS EMR)
- Valuable previous experience with Scala or Kotlin programming languages
- Experience with Amazon Redshift
- Previous experience in the Life Sciences sector
Perks & Benefits
- Benefits package including Restricted Stock Units (RSUs), family health insurance, and contributions to private pension plans
- Annual allocations for continuous learning, development & charitable contributions
- Fitness reimbursement
- Work from anywhere
Veeva’s headquarters is located in the San Francisco Bay Area with offices in more than 15 countries around the world.
Veeva is committed to fostering a culture of inclusion and growing a diverse workforce. Diversity makes us stronger. It comes in many forms. Gender, race, ethnicity, religion, politics, sexual orientation, age, disability and life experience shape us all into unique individuals. We value people for the individuals they are and the contributions they can bring to our teams.
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Airflow Architecture AWS Business Analytics CI/CD Consulting Databricks DataOps DevOps Pipelines PySpark Python Redshift Scala Spark
Perks/benefits: Career development Fitness / gym Health care Insurance
More jobs like this
Explore more AI, ML, Data Science career opportunities
Find even more open roles in Artificial Intelligence (AI), Machine Learning (ML), Natural Language Processing (NLP), Computer Vision (CV), Data Engineering, Data Analytics, Big Data, and Data Science in general - ordered by popularity of job title or skills, toolset and products used - below.
- Open Lead Data Analyst jobs
- Open MLOps Engineer jobs
- Open Data Science Manager jobs
- Open Senior Business Intelligence Analyst jobs
- Open Data Manager jobs
- Open Data Engineer II jobs
- Open Power BI Developer jobs
- Open Principal Data Engineer jobs
- Open Sr Data Engineer jobs
- Open Data Analytics Engineer jobs
- Open Business Intelligence Developer jobs
- Open Junior Data Scientist jobs
- Open Data Scientist II jobs
- Open Product Data Analyst jobs
- Open Senior Data Architect jobs
- Open Sr. Data Scientist jobs
- Open Business Data Analyst jobs
- Open Big Data Engineer jobs
- Open Data Analyst Intern jobs
- Open Azure Data Engineer jobs
- Open Manager, Data Engineering jobs
- Open Data Product Manager jobs
- Open Data Quality Analyst jobs
- Open Junior Data Engineer jobs
- Open Principal Data Scientist jobs
- Open Data quality-related jobs
- Open Business Intelligence-related jobs
- Open GCP-related jobs
- Open ML models-related jobs
- Open Data management-related jobs
- Open Java-related jobs
- Open Privacy-related jobs
- Open Finance-related jobs
- Open Data visualization-related jobs
- Open APIs-related jobs
- Open Deep Learning-related jobs
- Open PyTorch-related jobs
- Open Snowflake-related jobs
- Open TensorFlow-related jobs
- Open Consulting-related jobs
- Open PhD-related jobs
- Open CI/CD-related jobs
- Open NLP-related jobs
- Open Kubernetes-related jobs
- Open Data governance-related jobs
- Open Airflow-related jobs
- Open Hadoop-related jobs
- Open LLMs-related jobs
- Open Databricks-related jobs
- Open Data warehouse-related jobs