Staff, Data Engineer (Coupang Play)
Singapore, Singapore
Coupang
Join us to innovate. Rocket your career. Collaborate with teams across the globe. Find your role and learn more about our culture.Staff, Data Engineer (Coupang Play)
About Coupang Play
Launched in December 2020, Coupang Play is Coupang's OTT (over-the-top) service. Coupang Play aims to become the definitive destination for digital content. Backed by our strong Product and Engineering teams in Seoul, Singapore, and Mountain View — Coupang Play has been creating and optimizing the viewing experience for our customers.
Coupang Play is available on mobile devices, tablet PCs, smart TVs, and your preferred browsers. We are building a competitive content library that includes TV for children and all ages, movies, live sports, educational content, and more. Coupang Play also creates original content, catered specifically to our customers.
If you like to work backwards and solve problems for our customers, join us. We are committed to redefining how our customers enjoy digital entertainment as one of the fastest-growing OTT services in Korea.
Role Overview
This is a critical role to Coupang Play's Data Platform org. You will build a world-class data platform to power mission critical features such as search, recommendation, and personalization. The team is responsible for developing the platform and the data pipelines for generating all the source data needed to power Coupang Play search, including both the ground truth data and ranking signals/ML features. You will identify and solve hard problems to improve the system scalability, enhance the reliability, as well as optimize the resource efficiency.
What You Will Do
- Design and build Coupang Play's next-generation data platform.
- Design and build self-service data ingestion frameworks for enterprise-wide use.
- Design and build reliable data pipelines with the ability to scale to very large data using modern distributed processing technologies such as Spark and Hive.
- Develop, mentor and implement industry best practices in data architecture design, design patterns, and coding standards.
- Design and build source and sync connectors for data ingestion.
Basic Qualifications
- Bachelor's degree or higher in computer science or related technology.
- 8+ years of software development experience with deep understanding of algorithms, data structures and design patterns
- Experience with one or more programming languages such as Java, Scala, Python, Go, or Kotlin.
- Experience in using distributed systems to deal with big data platform (Hadoop, Spark, Spark Streaming, HBase, Cassandra, etc) to write ETL pipelines.
- Solid knowledge of Database System, Data Warehousing concepts and languages (SQL, NoSQL)
- Ability to proactively understand and evaluate your business to drive development
- Ability to proactively diagnose problems and solve problems innovatively
- Software engineering fundamentals: version control systems (i.e Git, Github) and workflows, and the ability to write production-ready code
- Knowledge and ability to train and mentor fellow engineers
- Strong communication skills, written and verbal; ability to present the work and gather feedbacks from internal developers
Preferred Qualifications
- Experience with cloud computing platforms and understanding of scalability/reliability issues.
- Experience in maintaining and monitoring ETL pipelines
- Experience in analyzing large scale dataset and making data driven decisions
- Experience with stream processing systems (Kafka, Spark-Streaming or equivalent) preferred.
- Experience deploying highly scalable software supporting millions or more users.
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Architecture Big Data Cassandra Computer Science Data pipelines Data Warehousing Distributed Systems Engineering ETL Git GitHub Hadoop HBase Kafka Machine Learning NoSQL Pipelines Python Scala Spark SQL Streaming
More jobs like this
Explore more AI, ML, Data Science career opportunities
Find even more open roles in Artificial Intelligence (AI), Machine Learning (ML), Natural Language Processing (NLP), Computer Vision (CV), Data Engineering, Data Analytics, Big Data, and Data Science in general - ordered by popularity of job title or skills, toolset and products used - below.
- Open MLOps Engineer jobs
- Open Lead Data Analyst jobs
- Open Data Science Manager jobs
- Open Senior Business Intelligence Analyst jobs
- Open Data Engineer II jobs
- Open Data Manager jobs
- Open Sr Data Engineer jobs
- Open Principal Data Engineer jobs
- Open Data Analytics Engineer jobs
- Open Power BI Developer jobs
- Open Business Intelligence Developer jobs
- Open Junior Data Scientist jobs
- Open Data Scientist II jobs
- Open Senior Data Architect jobs
- Open Product Data Analyst jobs
- Open Sr. Data Scientist jobs
- Open Business Data Analyst jobs
- Open Manager, Data Engineering jobs
- Open Big Data Engineer jobs
- Open Data Analyst Intern jobs
- Open Data Quality Analyst jobs
- Open Data Product Manager jobs
- Open Azure Data Engineer jobs
- Open Junior Data Engineer jobs
- Open ETL Developer jobs
- Open Data quality-related jobs
- Open Business Intelligence-related jobs
- Open ML models-related jobs
- Open GCP-related jobs
- Open Data management-related jobs
- Open Privacy-related jobs
- Open Java-related jobs
- Open Finance-related jobs
- Open Data visualization-related jobs
- Open APIs-related jobs
- Open Deep Learning-related jobs
- Open PyTorch-related jobs
- Open Consulting-related jobs
- Open Snowflake-related jobs
- Open TensorFlow-related jobs
- Open PhD-related jobs
- Open CI/CD-related jobs
- Open NLP-related jobs
- Open Kubernetes-related jobs
- Open Data governance-related jobs
- Open Airflow-related jobs
- Open Hadoop-related jobs
- Open Databricks-related jobs
- Open LLMs-related jobs
- Open DevOps-related jobs