Senior Data Engineer
Bengaluru, Karnataka, India - Remote
Applications have closed
FairMoney
Digital banking and Instant Loans in Nigeria providing collateral-free personal loans, a bank account with free bank transfers, and zero convenience fee on b...About FairMoney
FairMoney is a credit-led mobile bank for emerging markets. The company was launched in 2017, operates in Nigeria & India, and raised close to €50m from global investors like Tiger Global, DST & Flourish Ventures. The company has offices in France, Nigeria, and India.
Role and responsibilities
At FairMoney, we are making a lot of data driven decisions in real time: risk scoring, fraud detection as examples.
Our data is mainly produced by our backend services, and is being used by data science team, BI team, and management team. We are building more and more real time data driven decision making processes, as well as a self serve data analytics layer.
As a senior data engineer at FairMoney, you will help building our Data Platform:
- Ensure data quality and availability for all data consumers, mainly data science and BI teams.
- Ingest raw data into our DataWarehouse (BigQuery / Snowflake)
- Make sure data is processed and stored efficiently:
- Work with backend teams to offload data from backend storage
- Work with data scientists to build a machine learning feature store
- Spread best practices in terms of data architecture across all tech teams
- Effectively form relationships with the business in order to help with the adoption of data-driven decision-making.
You will be part of the Datatech team, sitting right between data producers and data consumers. You will help building the central nervous system of our real time data processing layer by building an ecosystem around data contracts between producers and consumers.
Our current stack is made of
- Batch processing jobs (Apache Spark in Python or Scala)
- Streaming jobs (Apache Flink deployed on Kinesis Data Analytics - Apache Beam deployed on Google Dataflow)
- REST apis (Python FastApi)
Our tool stack
- Programming language: Python, SQL
- Streaming Applications: Flink, Kafka
- Databases: MySQL, DynamoDB
- DWH: BigQuery, Snowflake
- BI: Tableau, Metabase, dbt
- ETL: Hevo, Airflow
- Production Environment: Python API deployed on Amazon EKS (Docker, Kubernetes, Flask)
- ML: Scikit-Learn, LightGBM, XGBoost, shap
- Cloud: AWS, GCP
Requirements
You will work on a daily basis with the below tools, so you need working experience on
- Languages: Python and Scala.
- Big data processing frameworks: all or one of Apache Spark (batch/streaming) - Apache Flink (streaming) - Apache Beam.
- Streaming services: Apache Kafka / AWS Kinesis.
- Managed cloud services: one of AWS EMR / AWS Kinesis Data Analytics / Google Dataflow.
- Docker.
- Building REST APIs.
Ideally, you have experience with:
- deployment/management of stateful streaming jobs.
- the Kafka ecosystem: Kafka connects mainly.
- infrastructure as code frameworks (Terraform).
- architecture around data contracts: Avro Schemas management, schema registries (Confluent Kafka / AWS Glue).
- Kubernetes.
Overall experience required for this role: 6+ Years.
Benefits
- Training & Development
- Family Leave (Maternity, Paternity)
- Paid Time Off (Vacation, Sick & Public Holidays)
- Remote Work
Recruitment Process
- A screening interview with one of the members of the Talent Acquisition team for 30 minutes.
- Takeaway assignment to be done at home.
- Technical design interview for 60-90 minutes.
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Airflow APIs Architecture Avro AWS AWS Glue Big Data BigQuery Data Analytics Dataflow Data quality Docker DynamoDB ETL FastAPI Flask Flink GCP Kafka Kinesis Kubernetes LightGBM Machine Learning Metabase MySQL Python Scala Scikit-learn Snowflake Spark SQL Streaming Tableau Terraform XGBoost
Perks/benefits: Career development Parental leave
More jobs like this
Explore more AI, ML, Data Science career opportunities
Find even more open roles in Artificial Intelligence (AI), Machine Learning (ML), Natural Language Processing (NLP), Computer Vision (CV), Data Engineering, Data Analytics, Big Data, and Data Science in general - ordered by popularity of job title or skills, toolset and products used - below.
- Open MLOps Engineer jobs
- Open Data Science Manager jobs
- Open Lead Data Analyst jobs
- Open Data Manager jobs
- Open Senior Business Intelligence Analyst jobs
- Open Data Engineer II jobs
- Open Power BI Developer jobs
- Open Sr Data Engineer jobs
- Open Principal Data Engineer jobs
- Open Business Intelligence Developer jobs
- Open Data Analytics Engineer jobs
- Open Junior Data Scientist jobs
- Open Product Data Analyst jobs
- Open Data Scientist II jobs
- Open Sr. Data Scientist jobs
- Open Senior Data Architect jobs
- Open Business Data Analyst jobs
- Open Data Analyst Intern jobs
- Open Big Data Engineer jobs
- Open Manager, Data Engineering jobs
- Open Data Product Manager jobs
- Open Junior Data Engineer jobs
- Open Data Quality Analyst jobs
- Open Azure Data Engineer jobs
- Open Principal Data Scientist jobs
- Open GCP-related jobs
- Open Data quality-related jobs
- Open Business Intelligence-related jobs
- Open Java-related jobs
- Open ML models-related jobs
- Open Data management-related jobs
- Open Privacy-related jobs
- Open Data visualization-related jobs
- Open Finance-related jobs
- Open Deep Learning-related jobs
- Open PhD-related jobs
- Open APIs-related jobs
- Open TensorFlow-related jobs
- Open PyTorch-related jobs
- Open NLP-related jobs
- Open Consulting-related jobs
- Open Snowflake-related jobs
- Open CI/CD-related jobs
- Open LLMs-related jobs
- Open Kubernetes-related jobs
- Open Generative AI-related jobs
- Open Data governance-related jobs
- Open Hadoop-related jobs
- Open Airflow-related jobs
- Open Docker-related jobs