Databricks (Remote)
Bengaluru, KA, India
PradeepIT Consulting Services Pvt Ltd
PradeepIT, supported by Asia's largest tech professional network, revolutionizing global talent acquisition. Discover the potential of hiring top Asian tech talents at ten times the speed, starting today!Roles & responsibilities
- Developing Modern Data Warehouse solutions using Databricks and AWS/ Azure Stack
- Ability to provide solutions that are forward-thinking in data engineering and analytics space
- Collaborating with DW/BI leads to understanding new ETL pipeline development requirements.
- Triage issues to find gaps in existing pipelines and fix the issues
- Work with businesses to understand the need in the reporting layer and develop a data model to fulfill
- reporting needs
- Help joiner team members to resolve issues and technical challenges.
- Drive technical discussion with client architects and team members
- Orchestrate the data pipelines in the scheduler via Airflow
- Qualification & experience
- Bachelor's and/or master's degree in computer science or equivalent experience.
- Must have a total of 6+ yrs. of IT experience and 3+ years' experience in Data warehouse/ETL projects.
- Deep understanding of Star and Snowflake dimensional modeling.
- Strong knowledge of Data Management principles
- Good understanding of Databricks Data & AI platform and Databricks Delta Lake Architecture
- Should have hands-on experience in SQL, Python, and Spark (PySpark)
- Candidate must have experience in AWS/ Azure stack
- Desirable to have ETL with batch and streaming (Kinesis).
- Experience in building ETL / data warehouse transformation processes
- Experience with Apache Kafka for use with streaming data / event-based data
- Experience with other Open-Source big data products Hadoop (incl. Hive, Pig, Impala)
- Experience with Open Source non-relational / NoSQL data repositories (incl. MongoDB,
- Cassandra, Neo4J)
- Experience working with structured and unstructured data including imaging & geospatial data.
- Experience working in a Dev/Ops environment with tools such as Terraform, CircleCI, and GIT.
- Proficiency in RDBMS, complex SQL, PL/SQL, Unix Shell Scripting, performance tuning, an troubleshooting.
- Databricks Certified Data Engineer Associate/Professional Certification (Desirable).
- Comfortable working in a dynamic, fast-paced, innovative environment with several ongoing
- concurrent projects
- Should have experience working in Agile methodology.
- Strong verbal and written communication skills.
- Strong analytical and problem-solving skills with a high attention to detail. Mandatory Skills:
- Python/ PySpark / Spark with Azure/ AWS Databricks
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Agile Airflow Architecture AWS Azure Big Data Cassandra Computer Science Databricks Data management Data pipelines Data warehouse Engineering ETL Git Hadoop Kafka Kinesis MongoDB Neo4j NoSQL Open Source Pipelines PySpark Python RDBMS Shell scripting Snowflake Spark SQL Streaming Terraform Unstructured data
More jobs like this
Explore more AI, ML, Data Science career opportunities
Find even more open roles in Artificial Intelligence (AI), Machine Learning (ML), Natural Language Processing (NLP), Computer Vision (CV), Data Engineering, Data Analytics, Big Data, and Data Science in general - ordered by popularity of job title or skills, toolset and products used - below.
- Open Data Engineer II jobs
- Open Business Intelligence Developer jobs
- Open Data Science Manager jobs
- Open BI Analyst jobs
- Open Data Scientist II jobs
- Open Principal Data Scientist jobs
- Open Business Data Analyst jobs
- Open Business Intelligence Engineer jobs
- Open Sr Data Engineer jobs
- Open Senior Business Intelligence Analyst jobs
- Open Data Science Intern jobs
- Open Lead Data Analyst jobs
- Open Sr. Data Scientist jobs
- Open MLOps Engineer jobs
- Open Junior Data Scientist jobs
- Open Software Engineer, Machine Learning jobs
- Open Azure Data Engineer jobs
- Open Marketing Data Analyst jobs
- Open Data Analytics Engineer jobs
- Open Manager, Data Engineering jobs
- Open Data Analyst II jobs
- Open Junior Data Engineer jobs
- Open Data Engineer III jobs
- Open Product Data Analyst jobs
- Open Data Quality Analyst jobs
- Open Data quality-related jobs
- Open Data management-related jobs
- Open GCP-related jobs
- Open Privacy-related jobs
- Open Data pipelines-related jobs
- Open ML models-related jobs
- Open PhD-related jobs
- Open APIs-related jobs
- Open Business Intelligence-related jobs
- Open Data visualization-related jobs
- Open PyTorch-related jobs
- Open LLMs-related jobs
- Open Finance-related jobs
- Open TensorFlow-related jobs
- Open Deep Learning-related jobs
- Open Consulting-related jobs
- Open Generative AI-related jobs
- Open NLP-related jobs
- Open Data governance-related jobs
- Open CI/CD-related jobs
- Open DevOps-related jobs
- Open Kubernetes-related jobs
- Open Snowflake-related jobs
- Open Hadoop-related jobs
- Open Git-related jobs