Data Engineer

Remote - Chennai, Tamil Nadu, India

Applications have closed

Packt

Packt is the online library and learning platform for professional developers. Learn Python, JavaScript, Angular and more with eBooks, videos and courses

View company page

Are you an aspiring data enthusiast who is looking to make it big in the field of data science? Data engineering is an excellent starting point to not only build and improve your data skills but also work on cloud technology to make crucial architectural decisions that are the building block for todays technology.

We are helping our partners Aureus Tech Systems to hire for their Data Engineers. They are on a mission to make our generation and beyond data literate by consistently making big leaps in developing data solutions that make some of the most complicated problems seem simple. As they grow their portfolio of products there is need for individuals with unique capabilities to help them build great products.

Why is this role great for you?

  • Join Aureus Tech on their mission to positively exploit abundant data opportunities by building unique yet robust solutions that you feel great about!
  • Work with a highly skilled and agile team that are constantly evolving in terms capability.
  • Expert data veterans who are always ready to guide you and nurture your skills.
  • Abundant opportunities to learn, apply and upskill yourself.

What will you be doing?

  • Assisting in quality control of quantitative and qualitative research projects.
  • Designing, building and support data processing pipelines to transform data using Hadoop technologies.
  • Working with business analysts to understand business requirements and use cases.
  • Developing and executing quality assurance and test scripts.
  • Building survey and voter file data pipelines.
  • Designing, building data assets in HIVE.
  • Designing and developing tools for internal team use to simplify complex task flows.
  • Designing schemas, data models and data architecture for Hadoop and HBase environments.
  • Implementing data flow scripts using Unix / Hive QL / Oozie scripting.
  • Creating scripts and workflows to automate repeated data processing tasks.
  • Working with data team to identify opportunities for task automation and helpful tools.

Requirements

You will be a great match if you...

  • Understand some of the best practices for building Data Lake and analytical architecture on Hadoop.
  • Familiarity with MapR distribution of Hadoop.
  • Scripting / programming with UNIX, Java, Python, Scala etc.
  • Experience with Graph database.
  • Experience in working in large environments such as RDBMS, EDW, NoSQL, etc.
  • Knowledge in schema design, developing data models and proven ability to work with complex data.
  • Understanding Hadoop file format and compressions.
  • Firsthand experience with Hadoop, MapReduce, Hive, Oozie, Airflow, ElasticSearch.
  • Knowledge in real time data ingestion into Hadoop.
  • Experience with Test Driven Code Development, SCM tools such as Git, Jenkins
  • Have familiar with cloud architecture and any of the cloud platforms (AWS, Azure and GCP)

Benefits

You will also have access to a range of flexible benefits such as:

  • Performance Bonus
  • Mediclaim
  • Opportunity to gain experience and develop your skills

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Tags: Agile Airflow AWS Azure Data pipelines Elasticsearch Engineering GCP Git Hadoop HBase NoSQL Oozie Pipelines Python RDBMS Research Scala

Perks/benefits: Career development Flex hours

Regions: Remote/Anywhere Asia/Pacific
Country: India
Job stats:  20  2  0
Category: Engineering Jobs

More jobs like this

Explore more AI, ML, Data Science career opportunities

Find even more open roles in Artificial Intelligence (AI), Machine Learning (ML), Natural Language Processing (NLP), Computer Vision (CV), Data Engineering, Data Analytics, Big Data, and Data Science in general - ordered by popularity of job title or skills, toolset and products used - below.