Lead Data Engineer (3847)

Bogotá, Colombia

Applications have closed

Company Description

At Publicis Groupe, we are looking for a Data Engineer fluent in English to join Publicis Global Delivery, the outstanding platform that we created to become a global interconnected network and provide offshore & nearshore solutions for our partners - sister companies´ business worldwide.

We are a never sleeps machine of creation that continuously grows and mutates to become a more efficient and collaborative system. A cross-media transformation agent, based in Argentina, Colombia, Costa Rica and Mexico, that provides centralized expertise of all Publicis Global Services' capabilities to enable consistent and standardized delivery across Media, Production, Commerce, Content, Data & Technology.

Job Description

What you’ll do:

We are looking for a savvy Data Engineer to join our growing team. The hire will be responsible for expanding and optimizing our data and data pipeline architecture, as well as optimizing data flow and collection for cross functional teams. The ideal candidate is an experienced data pipeline builder and data wrangler who enjoys optimizing data systems and building them from the ground up. The Data Engineer will support data analysts and data scientists on data initiatives and will ensure optimal data delivery architecture is consistent throughout ongoing projects. They must be self-directed and comfortable supporting the data needs of multiple teams, systems, and products. The right candidate will be excited by the prospect of optimizing or even re-designing client data architecture to support next generation of products and data initiatives.

Responsibilities

  • Create and maintain optimal data pipeline architecture, including table design.
  • Assemble large, complex data sets that meet functional / non-functional business requirements.
  • Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc.
  • Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL and (GCP, AWS, Azure, Snowflake)  ‘big data’ technologies.
  • Build analytics tools that utilize the data pipeline to provide actionable insights into customer acquisition, operational efficiency, and other key business performance metrics.
  • Work with stakeholders including the Executive, Product, Data and Design teams to assist with data-related technical issues and support their data infrastructure needs.
  • Create data tools for analytics and data scientist team members that assist them in building and optimizing advanced analytics and data models.
  • Work with data and analytics experts to strive for greater functionality in our data systems.

Qualifications

We’re looking for strong, impactful work experience, which typically includes:

  • Advanced working SQL knowledge and experience working with relational databases, query authoring (SQL) as well as working familiarity with a variety of databases.
  • Experience building and optimizing ‘big data’ data pipelines, architectures, and data sets.
  • Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement.
  • Build processes supporting data transformation, data structures, metadata, dependency, and workload management.
  • A successful history of manipulating, processing and extracting value from large disconnected datasets.
  • Strong project management and organizational skills.
  • Experience supporting and working with cross-functional teams in a dynamic environment.
  • We are looking for a candidate with 3-5 years of experience in a Data Engineer role, who has attained an undergraduate degree in Computer Science, Statistics, Informatics, Information Systems or another quantitative field. They should also have experience using the following software/tools:
    • Experience with relational SQL and NoSQL databases, including Postgres and Cassandra.
    • Experience with data pipeline and workflow management tools: Azkaban, Luigi, Airflow, etc.
    • Experience with big data tools: Hadoop, Spark, Kafka, etc.
    • Experience with AWS cloud services: EC2, EMR, RDS, Redshift, Athena
    • Experience with GCP cloud services: BigQuery, GCS, Cloud SQL, ML, Pub/Sub, DataProc
    • Experience with Azure cloud services: DataBricks, 
    • Experience with stream-processing systems: Storm, Spark-Streaming, etc.
    • Experience with object-oriented/object function scripting languages: Python, Java, C++, Scala, etc.
    • Experience with data architecture creating staging and final table design.
    • Experience creating ERD / Data Dictionary documentation.

Additional Information

Benefits

  • Access to Prepaid Medical Plan
  • Flexible schedule
  • 100% Remote work
  • 14 business days of vacation
  • English lessons
  • Discounts on courses, trainings and universities
  • Access to E-Learning platforms
  • Technical trainings & soft skills development
  • Certification programs
  • Level up program
  • Engagement activities and events
  • A mentor who´ll do a coaching process with you to develop your professional career!

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Tags: Airflow Architecture Athena AWS Azkaban Azure Big Data BigQuery Cassandra Computer Science Databricks Data pipelines Dataproc EC2 GCP Hadoop Kafka Machine Learning NoSQL Pipelines PostgreSQL Python RDBMS Redshift Scala Snowflake Spark SQL Statistics Streaming

Perks/benefits: Career development Flex hours Flex vacation Team events

Region: South America
Country: Colombia
Job stats:  1  0  0

More jobs like this

Explore more AI, ML, Data Science career opportunities

Find even more open roles in Artificial Intelligence (AI), Machine Learning (ML), Natural Language Processing (NLP), Computer Vision (CV), Data Engineering, Data Analytics, Big Data, and Data Science in general - ordered by popularity of job title or skills, toolset and products used - below.