Data Engineer (m/f/d)

Almaty-office

Applications have closed
Company Description

AUTODOC is a technology company with a leading e-commerce platform for vehicle parts and accessories in Europe. Founded in 2008 in Berlin, we are now 5,000 people from 68 nationalities, collaborating to make mobility easy and sustainable. By placing technology at the core of everything we do, we manage to serve customers in 27 European countries in 23 languages. In 2022, AUTODOC achieved sales of over 1,1 billion euros; stocks 5.2 million products, processes around 50,000 orders per day with over 6 million active customers.


Curious minds, adventurous experts and tech-savvy professionals - one team, one billion euros revenue. Catch the ride!

Job Description

Data Engineer plays a critical role in data processing, responsible for designing and developing DBMS to transform and connect various data sources into analytical datasets. He is responsible for improving the efficiency and reliability of the ETL process infrastructure, including fault tolerance. He is also responsible for data quality assurance and project management.



Responsibilities:

  • Development and optimization of architecture of database processing systems
  • Development and support of infrastructure in the Google Cloud Platform environment
  • Improving the quality and reliability of data that drives end-to-end cleansing processes
  • Create efficient CI/CD and other automated solutions
  • Maintenance and enhancement of existing infrastructure
  • Infrastructure security and monitoring


Qualifications

  • In-depth knowledge of Data Engineering methodologies
  • Ability to work with large datasets, ETL tools and databases
  • Experience with cloud tools for data warehousing and processing such as AWS, GCP or Azure
  • Ability to design, build and maintain robust and scalable data pipelines
  • Interacting with other departments and participating in strategic decision making at the company level in the field of data engineering
  • Ability to find innovative ways to solve problems
  • Developing solutions for system monitoring

Experience:

  • Experience in developing fault tolerance mechanisms - approaches to clustering, replication, scaling, etc.
  • Strong understanding and use of programming languages such as Python, SQL, and professional knowledge of Big Data technologies such as Hadoop, Spark.
  • Experience with GCP cloud infrastructure
  • Experience in Data Science for 3 years or more.
  • Understanding of services: GCP, JupyterHub, AirFlow, ClickHouse, Spark, MLFlow
  • Finding your way out of the Vim text editor

What do we offer?
  • Competitive salaries based on your professional experience
  • Fast growing international company with stable employment
  • Annual vacation of 24 days and 1 additional day off on your birthday
  • Monthly Allowance for cover the costs of medical insurance expenses
  • Mental Wellbeing Program – the opportunity for free psychological counseling for you and your family members 24/7 hotline and online sessions
  • Opportunities for advancement, further trainings (over 650 courses on soft and hard skills on our e-learning platform) and coaching
  • Free English and German language classes
  • Flexible working hours and hybrid work

Join us today and let’s create a success story together!

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Tags: Airflow Architecture AWS Azure Big Data CI/CD Clustering Data pipelines Data quality Data Warehousing E-commerce Engineering ETL GCP Google Cloud Hadoop MLFlow Pipelines Python Security Spark SQL

Perks/benefits: Career development Flex hours Flex vacation Wellness

Region: Asia/Pacific
Country: Kazakhstan
Job stats:  6  2  0
Category: Engineering Jobs

More jobs like this

Explore more AI, ML, Data Science career opportunities

Find even more open roles in Artificial Intelligence (AI), Machine Learning (ML), Natural Language Processing (NLP), Computer Vision (CV), Data Engineering, Data Analytics, Big Data, and Data Science in general - ordered by popularity of job title or skills, toolset and products used - below.