Senior Data Engineer, Life Sciences

Remote or Boston, MA or Pittsburgh, PA

Arcadia

Top providers and payers use Arcadia's healthcare data platform to unify data sources, leverage analytics, gain insight, and drive action across the care continuum.

View company page

Based on the evolving situation with Covid–19 we are mostly remote, this includes the hiring process. We are in the process of moving to a hybrid of onsite and remote with some positions remaining 100% remote.
Why this role is important to ArcadiaReporting to the SVP, Life Sciences, in this position you will work within the burgeoning Life Sciences organization to design, build, install and test data processes which ingest, transform, and analyze data at petabyte scale.  Leveraging technologies like Apache Spark, AWS Glue, and Snowflake, you will build the toolset to enable Arcadia’s data aggregation projects on cloud platforms such as AWS. The ideal candidate would be passionate about using technologies to solve complex problems in a distributed, non-homogeneous environment.The ideal candidate will be motivated by bleeding edge technologies and methodologies with an exciting curiosity to learn and apply them to the healthcare space.  They bring a strong track record in developing and delivering software through all phases of the SDLC, including cross-functional collaboration to achieve business outcomes.  We are seeking self-motivated individuals with the drive to be entrepreneurial who are passionate about data and problem solving and excited about the potential of that data to make meaningful change in healthcare. 
What Success Looks LikeIn 3 months- Understand the high-level approaches and architecture employed by Arcadia to integrate and aggregate data within the Life Sciences product line- Contribute to product development on high priority features across multiple software components managed by the team- Complete onboarding requirements- Complete LS team specific trainingIn 6 months- Achieve a deep understanding of the approaches and architecture employed by Arcadia to integrate and aggregate data by contributing to all components managed by the team- Contribute to the design and build of more complex and impactful features- Actively contribute to discussions around feature/system design and groomingIn 12 months- Continued excellence in building and deploying product features and technical improvements- Identify opportunities to eliminate technical debt and improve development and delivery practices and efficiencies

What You’ll Be Doing

  • Designing and building software components that underpin the large-scale data ingestion and transformation pipelines at Arcadia
  • Recognizing common patterns in data integration and analysis and building the tools to apply those patterns at scale
  • Collaborating with and supporting data engineers who are building and supporting data pipelines to improve the developer experience and increase productivity
  • Performing code reviews
  • Developing unit and integration tests
  • Managing code repositories
  • Utilizing CI/CD pipelines to manage software delivery and deployment
  • Providing, analyzing, and responding to software development metrics such as Feature Lifecycle and Burn Down
  • Providing feedback and recommendations to improve software development processes
  • Interfacing with Product to groom features (epic definition, story estimates, task breakdown)

What You'll Bring

  • BS degree in Computer Science or equivalent hands-on experience
  • Experience in at least one of Scala, Java, and/or Python
  • Strong object oriented and/or functional programming fundamentals
  • Experience with distributed systems and technologies (Spark, Hadoop, Kafka, etc.)
  • Familiarity with the challenges of working with large-scale, complex data sets and data pipelines
  • Familiarity with relational and/or NoSQL databases
  • Proven ability to deliver production software, including the use of Git/GitHub and CI tools
  • Commitment to learning new skills and improving upon existing ones

Would Love For You To Have

  • Healthcare data experience, such as EHR or Claims data
  • Expertise in Python
  • DevOps experience, such as automating infrastructure and services in the AWS cloud

What You Get

  • Be a part of a mission driven company that is transforming the healthcare industry by changing the way patients receive care
  • Chance to be surrounded by a team of extremely talented and dedicated individuals driven to succeed
  • Competitive compensation and amazing benefits including unlimited FTO
  • A flexible, remote friendly company with personality and heart
  • Employee driven programs and initiatives for personal and professional development
  • Be a member of the Arcadian and Barkadian Community
About ArcadiaArcadia.io helps innovative healthcare systems and health plans around the country transform healthcare to reduce cost while improving patient health.   We do this by aggregating massive amounts of clinical and claims data, applying algorithms to identify opportunities to provide better patient care, and making those opportunities actionable by physicians at the point of care in near-real time.  We are passionate about helping our customers drive meaningful outcomes. We are growing fast and have emerged as the market leader in the highly competitive population health management software and value-based care services markets, and we have been recognized by industry analysts KLAS, IDC, Forrester and Chilmark for our leadership. For a better sense of our brand and products, please explore our website, our online resources, and our interactive Data Gallery.
This position is responsible for following all Security policies and procedures in order to protect all PHI under Arcadia's custodianship as well as Arcadia Intellectual Properties.  For any security-specific roles, the responsibilities would be further defined by the hiring manager.

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Tags: Architecture AWS CI/CD Computer Science Data pipelines DevOps Distributed Systems Git GitHub Hadoop Kafka NoSQL Pipelines Python Scala SDLC Security Snowflake Spark

Perks/benefits: Career development Competitive pay Flex hours Health care Unlimited paid time off

Regions: Remote/Anywhere North America
Country: United States
Job stats:  14  1  0
Category: Engineering Jobs

More jobs like this

Explore more AI, ML, Data Science career opportunities

Find even more open roles in Artificial Intelligence (AI), Machine Learning (ML), Natural Language Processing (NLP), Computer Vision (CV), Data Engineering, Data Analytics, Big Data, and Data Science in general - ordered by popularity of job title or skills, toolset and products used - below.