Data Engineer
Remote - Chennai, Tamil Nadu, India
Packt
Packt is the online library and learning platform for professional developers. Learn Python, JavaScript, Angular and more with eBooks, videos and coursesAre you an aspiring data enthusiast who is looking to make it big in the field of data science? Data engineering is an excellent starting point to not only build and improve your data skills but also work on cloud technology to make crucial architectural decisions that are the building block for todays technology.
We are helping our partners Aureus Tech Systems to hire for their Data Engineers. They are on a mission to make our generation and beyond data literate by consistently making big leaps in developing data solutions that make some of the most complicated problems seem simple. As they grow their portfolio of products there is need for individuals with unique capabilities to help them build great products.
Why is this role great for you?
- Join Aureus Tech on their mission to positively exploit abundant data opportunities by building unique yet robust solutions that you feel great about!
- Work with a highly skilled and agile team that are constantly evolving in terms capability.
- Expert data veterans who are always ready to guide you and nurture your skills.
- Abundant opportunities to learn, apply and upskill yourself.
What will you be doing?
- Assisting in quality control of quantitative and qualitative research projects.
- Designing, building and support data processing pipelines to transform data using Hadoop technologies.
- Working with business analysts to understand business requirements and use cases.
- Developing and executing quality assurance and test scripts.
- Building survey and voter file data pipelines.
- Designing, building data assets in HIVE.
- Designing and developing tools for internal team use to simplify complex task flows.
- Designing schemas, data models and data architecture for Hadoop and HBase environments.
- Implementing data flow scripts using Unix / Hive QL / Oozie scripting.
- Creating scripts and workflows to automate repeated data processing tasks.
- Working with data team to identify opportunities for task automation and helpful tools.
Requirements
You will be a great match if you...
- Understand some of the best practices for building Data Lake and analytical architecture on Hadoop.
- Familiarity with MapR distribution of Hadoop.
- Scripting / programming with UNIX, Java, Python, Scala etc.
- Experience with Graph database.
- Experience in working in large environments such as RDBMS, EDW, NoSQL, etc.
- Knowledge in schema design, developing data models and proven ability to work with complex data.
- Understanding Hadoop file format and compressions.
- Firsthand experience with Hadoop, MapReduce, Hive, Oozie, Airflow, ElasticSearch.
- Knowledge in real time data ingestion into Hadoop.
- Experience with Test Driven Code Development, SCM tools such as Git, Jenkins
- Have familiar with cloud architecture and any of the cloud platforms (AWS, Azure and GCP)
Benefits
You will also have access to a range of flexible benefits such as:
- Performance Bonus
- Mediclaim
- Opportunity to gain experience and develop your skills
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Agile Airflow AWS Azure Data pipelines Elasticsearch Engineering GCP Git Hadoop HBase NoSQL Oozie Pipelines Python RDBMS Research Scala
Perks/benefits: Career development Flex hours
More jobs like this
Explore more AI, ML, Data Science career opportunities
Find even more open roles in Artificial Intelligence (AI), Machine Learning (ML), Natural Language Processing (NLP), Computer Vision (CV), Data Engineering, Data Analytics, Big Data, and Data Science in general - ordered by popularity of job title or skills, toolset and products used - below.
- Open AI Engineer jobs
- Open MLOps Engineer jobs
- Open Data Science Manager jobs
- Open Data Manager jobs
- Open Senior Business Intelligence Analyst jobs
- Open Data Engineer II jobs
- Open Power BI Developer jobs
- Open Sr Data Engineer jobs
- Open Principal Data Engineer jobs
- Open Data Analytics Engineer jobs
- Open Junior Data Scientist jobs
- Open Business Intelligence Developer jobs
- Open Data Scientist II jobs
- Open Senior Data Architect jobs
- Open Product Data Analyst jobs
- Open Business Data Analyst jobs
- Open Sr. Data Scientist jobs
- Open Big Data Engineer jobs
- Open Data Analyst Intern jobs
- Open Manager, Data Engineering jobs
- Open Data Quality Analyst jobs
- Open Junior Data Engineer jobs
- Open Azure Data Engineer jobs
- Open Data Product Manager jobs
- Open ETL Developer jobs
- Open Data quality-related jobs
- Open Business Intelligence-related jobs
- Open ML models-related jobs
- Open Data management-related jobs
- Open GCP-related jobs
- Open Java-related jobs
- Open Privacy-related jobs
- Open Finance-related jobs
- Open Data visualization-related jobs
- Open APIs-related jobs
- Open Deep Learning-related jobs
- Open PyTorch-related jobs
- Open Snowflake-related jobs
- Open Consulting-related jobs
- Open TensorFlow-related jobs
- Open PhD-related jobs
- Open CI/CD-related jobs
- Open NLP-related jobs
- Open Data governance-related jobs
- Open Kubernetes-related jobs
- Open Airflow-related jobs
- Open Data warehouse-related jobs
- Open LLMs-related jobs
- Open Databricks-related jobs
- Open Hadoop-related jobs