Data Engineer
London, United Kingdom
Applications have closed
Informa Group Plc.
Company Description
Citeline, is part of the Norstella group of Pharma information solutions , is one of the world's leading providers of data and intelligence on clinical trials, drug treatments, medical devices and what's new in the regulatory and commercial landscape. Relying on us to deliver vital advantage when making critical R&D and commercial decisions, our customers come from over 3000 of the world’s leading pharmaceutical, contract research organizations (CROs), medical technology, biotechnology and healthcare service providers, including the top 10 global pharma and CROs.
From drug and device discovery and development to regulatory approval, and from product launch to lifecycle management, we provide the intelligence and insight to help our customers seize opportunities, mitigate risk and make business-critical decisions, faster. As the pharma and healthcare sector faces unparalleled upheaval, customers rely on our independent advice, enabling them to cut through the clutter and make sense of changing drug development, regulatory and competitive landscapes.
Job Description
Summary
An opportunity has arisen to join Citeline as a Data Engineer. The Data Engineer will be responsible for designing, implementing, deploying, and supporting various data management technologies and architectures.
You will be a critical team member and developing the means to collect and ingest data, developing data models and data architectures, creating automated data pipelines, and taking the lead in making these Production-ready. You will assist with integrating with existing applications and will have the opportunity to accelerate the delivery of and improve the quality of the Enterprise Insight. We are interested in hearing from individuals with a background in data - both SQL and non-SQL, with very strong Python scripting background.
Responsibilities
- Analyze data sources, and acquire data
- Create data pipelines, and integrate to final data destinations
- Create appropriate data models, architectures and pipelines
- Move the models and pipelines into Production
- Perform data loads and optimize data for extraction and reporting use
- Design and implement ETL jobs and transformations to populate a data lake and data warehouse
- Maintain complex databases by performing appropriate database management functions (e.g., maintain space availability, rebuilds indexes, file cleanup, runs utilities to check database integrity) to ensure optimum capacity and application performance
- Monitor, report, and analyze usage trends and statistical output in order to maintain quality control and high performance of the data retrieval from a database or other data storage
- Developing templates and accelerators, across a variety of libraries and platforms.
- Participating in data workshops as necessary
- Collaborating with business and technology partners to grow and develop the data engineering practice.
Qualifications
- Experience working in technology especially AWS stack
- Apache airflow experience is a must, AWS Data stack is a must
- Excellent data-related development skills, preferably in mainstream versions of SQL and NoSQL
- Experience in data warehouse concepts like fact tables, lookup tables, both designing and creating database objects like tables, views, stored procedures, functions, row level security
- Experience working with Data Lakehouse especially Snowflake
- ETL tools specifically Apache Airflow or AWS Glue
- Hands-on Python development skills for building and redefining data pipelines
- Coding in C#, T-SQL to support existing data components
- Experience with databases, modelling, data flows
- Exposure to the full Software Development Lifecycle, and experience of working in a modern development tram
- Good analytics skills
- Verbal and written communication skills. The successful candidate will be expected to communicate effectively with both business and technical teams when troubleshooting issues
- Experience to support other engineers, infrastructure, and configuration
- Excellent debugging, problem solving and testing skills
- Experience using reporting tools such as Power BI and Qlik
- Good interpersonal/management skills and capable of working individually and as part of a team
- Awareness of Apache Hadoop, HDFS, Hive
- BS degree in Computer Science, Information Systems or equivalent experience.
Advantage But not Essential
- Real-time data processing with Apache Kafka
- Experience in working with / supplying data to visualization tools such as Qlik, Tableau, PowerBI or similar
- Good understanding of data integration patterns.
- Experience with / exposure to software development for analytic applications
- Experience in projects involving cross-functional teams.
Additional Information
Why work for Citeline?
- 25 days annual leave plus bank holidays, 4 days for volunteering and a day off for your Birthday!
- Hybrid working opportunity balanced between home and central London office
- Work with a high quality of specialist products and service
- Bright and friendly staff who are all “expert’s experts” and additional training and development for helping you achieve your career aspiration
- Life assurance
- Pension match and a range of other entirely flexible benefits to suit your needs
We know that sometimes the 'perfect candidate' doesn't exist, and that people can be put off applying for a job if they don't fit all the requirements. If you're excited about working for us and have most of the skills or experience we're looking for, please go ahead and apply. You could be just what we need! We believe strongly in the value of diversity and creating supportive, inclusive environments where our colleagues can succeed. We do not discriminate on the basis of race, colour, ancestry, national origin, religion, or religious creed, mental or physical disability, medical condition, genetic information, sex (including pregnancy, childbirth, and related medical conditions), sexual orientation, gender identity, gender expression, age, marital status, citizenship, or other protected characteristics under local law.
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Airflow Architecture AWS Computer Science Data management Data pipelines Data warehouse Engineering ETL Hadoop HDFS Kafka NoSQL Pipelines Power BI Python Qlik R R&D Research Security Snowflake SQL Statistics Tableau Testing T-SQL
Perks/benefits: Career development Flex hours Flex vacation Medical leave
More jobs like this
Explore more AI, ML, Data Science career opportunities
Find even more open roles in Artificial Intelligence (AI), Machine Learning (ML), Natural Language Processing (NLP), Computer Vision (CV), Data Engineering, Data Analytics, Big Data, and Data Science in general - ordered by popularity of job title or skills, toolset and products used - below.
- Open Lead Data Analyst jobs
- Open MLOps Engineer jobs
- Open Data Science Manager jobs
- Open Senior Business Intelligence Analyst jobs
- Open Data Manager jobs
- Open Data Engineer II jobs
- Open Power BI Developer jobs
- Open Principal Data Engineer jobs
- Open Sr Data Engineer jobs
- Open Business Intelligence Developer jobs
- Open Data Analytics Engineer jobs
- Open Junior Data Scientist jobs
- Open Data Scientist II jobs
- Open Product Data Analyst jobs
- Open Senior Data Architect jobs
- Open Sr. Data Scientist jobs
- Open Business Data Analyst jobs
- Open Big Data Engineer jobs
- Open Data Analyst Intern jobs
- Open Manager, Data Engineering jobs
- Open Azure Data Engineer jobs
- Open Data Product Manager jobs
- Open Data Quality Analyst jobs
- Open Junior Data Engineer jobs
- Open Principal Data Scientist jobs
- Open GCP-related jobs
- Open Data quality-related jobs
- Open Business Intelligence-related jobs
- Open ML models-related jobs
- Open Java-related jobs
- Open Data management-related jobs
- Open Privacy-related jobs
- Open Data visualization-related jobs
- Open Finance-related jobs
- Open Deep Learning-related jobs
- Open APIs-related jobs
- Open PhD-related jobs
- Open TensorFlow-related jobs
- Open PyTorch-related jobs
- Open Consulting-related jobs
- Open NLP-related jobs
- Open Snowflake-related jobs
- Open CI/CD-related jobs
- Open Kubernetes-related jobs
- Open LLMs-related jobs
- Open Generative AI-related jobs
- Open Data governance-related jobs
- Open Hadoop-related jobs
- Open Airflow-related jobs
- Open Docker-related jobs