Data Engineer

London, United Kingdom

Applications have closed

Company Description

Citeline, is part of the Norstella group of Pharma information solutions , is one of the world's leading providers of data and intelligence on clinical trials, drug treatments, medical devices and what's new in the regulatory and commercial landscape. Relying on us to deliver vital advantage when making critical R&D and commercial decisions, our customers come from over 3000 of the world’s leading pharmaceutical, contract research organizations (CROs), medical technology, biotechnology and healthcare service providers, including the top 10 global pharma and CROs. 

From drug and device discovery and development to regulatory approval, and from product launch to lifecycle management, we provide the intelligence and insight to help our customers seize opportunities, mitigate risk and make business-critical decisions, faster. As the pharma and healthcare sector faces unparalleled upheaval, customers rely on our independent advice, enabling them to cut through the clutter and make sense of changing drug development, regulatory and competitive landscapes. 

Job Description

Summary

An opportunity has arisen to join Citeline as a Data Engineer. The Data Engineer will be responsible for designing, implementing, deploying, and supporting various data management technologies and architectures.

You will be a critical team member and developing the means to collect and ingest data, developing data models and data architectures, creating automated data pipelines, and taking the lead in making these Production-ready. You will assist with integrating with existing applications and will have the opportunity to accelerate the delivery of and improve the quality of the Enterprise Insight.  We are interested in hearing from individuals with a background in data - both SQL and non-SQL, with very strong Python scripting background.

Responsibilities

  • Analyze data sources, and acquire data
  • Create data pipelines, and integrate to final data destinations
  • Create appropriate data models, architectures and pipelines
  • Move the models and pipelines into Production
  • Perform data loads and optimize data for extraction and reporting use
  • Design and implement ETL jobs and transformations to populate a data lake and data warehouse
  • Maintain complex databases by performing appropriate database management functions (e.g., maintain space availability, rebuilds indexes, file cleanup, runs utilities to check database integrity) to ensure optimum capacity and application performance
  • Monitor, report, and analyze usage trends and statistical output in order to maintain quality control and high performance of the data retrieval from a database or other data storage
  • Developing templates and accelerators, across a variety of libraries and platforms.
  • Participating in data workshops as necessary
  • Collaborating with business and technology partners to grow and develop the data engineering practice.

 

Qualifications

  • Experience working in technology especially AWS stack
  • Apache airflow experience is a must, AWS Data stack is a must
  • Excellent data-related development skills, preferably in mainstream versions of SQL and NoSQL
  • Experience in data warehouse concepts like fact tables, lookup tables, both designing and creating database objects like tables, views, stored procedures, functions, row level security
  • Experience working with Data Lakehouse especially Snowflake
  • ETL tools specifically Apache Airflow or AWS Glue
  • Hands-on Python development skills for building and redefining data pipelines
  • Coding in C#, T-SQL to support existing data components
  • Experience with databases, modelling, data flows
  • Exposure to the full Software Development Lifecycle, and experience of working in a modern development tram
  • Good analytics skills
  • Verbal and written communication skills. The successful candidate will be expected to communicate effectively with both business and technical teams when troubleshooting issues
  • Experience to support other engineers, infrastructure, and configuration
  • Excellent debugging, problem solving and testing skills
  • Experience using reporting tools such as Power BI and Qlik
  • Good interpersonal/management skills and capable of working individually and as part of a team
  • Awareness of Apache Hadoop, HDFS, Hive
  • BS degree in Computer Science, Information Systems or equivalent experience.

Advantage But not Essential

  • Real-time data processing with Apache Kafka
  • Experience in working with / supplying data to visualization tools such as Qlik, Tableau, PowerBI or similar
  • Good understanding of data integration patterns.
  • Experience with / exposure to software development for analytic applications
  • Experience in projects involving cross-functional teams.

Additional Information

Why work for Citeline?

  • 25 days annual leave plus bank holidays, 4 days for volunteering and a day off for your Birthday!
  • Hybrid working opportunity balanced between home and central London office
  • Work with a high quality of specialist products and service
  • Bright and friendly staff who are all “expert’s experts” and additional training and development for helping you achieve your career aspiration
  • Life assurance
  • Pension match and a range of other entirely flexible benefits to suit your needs

We know that sometimes the 'perfect candidate' doesn't exist, and that people can be put off applying for a job if they don't fit all the requirements. If you're excited about working for us and have most of the skills or experience we're looking for, please go ahead and apply. You could be just what we need! We believe strongly in the value of diversity and creating supportive, inclusive environments where our colleagues can succeed. We do not discriminate on the basis of race, colour, ancestry, national origin, religion, or religious creed, mental or physical disability, medical condition, genetic information, sex (including pregnancy, childbirth, and related medical conditions), sexual orientation, gender identity, gender expression, age, marital status, citizenship, or other protected characteristics under local law.

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Tags: Airflow Architecture AWS Computer Science Data management Data pipelines Data warehouse Engineering ETL Hadoop HDFS Kafka NoSQL Pipelines Power BI Python Qlik R R&D Research Security Snowflake SQL Statistics Tableau Testing T-SQL

Perks/benefits: Career development Flex hours Flex vacation Medical leave

Region: Europe
Country: United Kingdom
Job stats:  9  1  0
Category: Engineering Jobs

More jobs like this

Explore more AI, ML, Data Science career opportunities

Find even more open roles in Artificial Intelligence (AI), Machine Learning (ML), Natural Language Processing (NLP), Computer Vision (CV), Data Engineering, Data Analytics, Big Data, and Data Science in general - ordered by popularity of job title or skills, toolset and products used - below.