Data Engineer
Crane, Indiana
Applications have closed
Barbaricum
All-inclusive government contracting firm that develops innovative strategiesBarbaricum is a rapidly growing government contractor providing leading-edge support to federal customers, with a particular focus on Defense and National Security mission sets. We leverage more than 14 years of support to stakeholders across the federal government, with established and growing capabilities across Intelligence, Analytics, Engineering, Mission Support, and Communications disciplines. Founded in 2008, our mission is to transform the way our customers approach constantly changing and complex problem sets by bringing to bear the latest in technology and the highest caliber of talent. Headquartered in Washington, DC's historic Dupont Circle neighborhood, Barbaricum also has a corporate presence in Tampa, FL and Dayton, OH, with team members across the United States and around the world. As a leader in our space, we partner with firms in the private sector, academic institutions, and industry associations with a goal of continually building our expertise and capabilities for the benefit of our employees and the customers we support. Through all of this, we have built a vibrant corporate culture diverse in expertise and perspectives with a focus on collaboration and innovation. Our teams are at the frontier of the Nation's most complex and rewarding challenges. Join us.
Barbaricum is seeking a Data Engineer to support a Naval Surface Warfare Center (NSWC) Crane Division initiative to stand up a hybrid cloud-based System Integration Lab (SIL). This SIL will aggregate, store, and further process ingested vehicle data to increase maintenance availability. This data will then be transmitted to external analytics organizations for predictive analysis. This initiative will support the United States Marine Corps (USMC) Condition Based Maintenance (CBM+) capability, which is a key component of USMC’s strategy to reduce maintenance workloads, costs, and downtime, improve supply chain visibility and management and overall military readiness, and sustain lethality. In order to realize the CBM+ capability, data will be collected from USMC ground vehicles transmitted through electronic means so it can be aggregated, stored, and further processed and analyzed using diagnostic, reliability, predictive trend, prognostic, and condition monitoring analysis. This analysis will be leveraged to reevaluate maintenance strategies, capture demand, conductive preemptive maintenance, and create decision support tools.
This position may be performed in a hybrid capacity, with routine in-person requirements dictated by project/customer needs. As such, candidates must reside within a reasonable commutable distance to NSWC-Crane, to include candidates in the Indianapolis or Louisville metro areas.
Responsibilities
- Responsible for establishment and operation of a data pipeline to support transfer of USMC ground vehicle CAN bus data from USMC vehicle fleet to System Integration Lab (SIL) hybrid cloud environment hosted at NSWC Crane.
- Utilize data engineering tools and generate scripts to enable transformation of raw CAN bus data into processed data ready for ingestion into SIL cloud environment.
- Develop automation processes for ingestion of processed data into SIL cloud environment and export of data to external predictive analytics organizations.
- Identify methods to improve data collection, data integrity, data security, analysis, validation, scalability, and reporting.
- Design and implement data solutions using industry best practices.
- Maintain and monitor data pipelines to ensure high service availability.
- Work with Data Scientists to understand analysis needs and mathematical models and optimize data solutions accordingly.
Qualifications
- Active DoD Secret clearance required
- Bachelor’s degree in Computer or Electrical Engineering, Computer Science, Data Science or Information Systems
- 5+ years of demonstrated experience working in a data engineering / data scientist role
- Demonstrated experience with data workflow/orchestration
- Demonstrated experience with data access/analysis languages/tools (e.g., Python, SQL, MongoDB Aggregation Framework)
- Demonstrated experience with relational databases (e.g., Oracle, SQL Server, PostgreSQL)
- Demonstrated experience with NoSQL databases (e.g., MongoDB, Redis)
- Demonstrated knowledge of extract, transform, and load (ETL) best practices in an enterprise environment
- Demonstrated experience with Linux and container-based deployment environments
- Demonstrated knowledge of software engineering best practices (Agile principles, code standards, code reviews, source code management, build processes, testing, and operations)
- Demonstrated ability to communicate effectively with individuals in both technical and leadership roles
Preferred Qualifications
- Demonstrated experience building out data pipelines by leveraging Amazon Web Services (AWS) or Microsoft Azure cloud computing services (preferably AWS)
- Demonstrated experience with distributed systems utilizing tools such as Kafka, Apache Hadoop, Spark
- Demonstrated experience in shell scripting using Linux, CLI commands
- Demonstrated experience with Big Data file formats such as HDF5, Parquet, Avro
- Demonstrated experience working with data logging devices to capture CAN bus data from heavy-duty vehicles
- Demonstrated experience implementing Continuous Integration / Continuous Deployment (CI/CD) pipeline using DevOps tools to automate software builds, software quality, cybersecurity scanning, testing, user access, and deployment
- Familiar with ASAM MDF (Measurement Data Format), NASA Common Data format (CDF), and Army Bulk CBM+ Data (ABCD) format for logging of CAN Bus data
- Familiar with Controller Area Network (CAN bus) standards including: ISO 11898 / SAE J1939 / SAE 1587 / SAE J1708
- Current or previous work experience supporting USMC or other Navy customers
For more information about Barbaricum, please visit our website at www.barbaricum.com. We will contact candidates directly to schedule interviews. No phone calls please.
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Agile Avro AWS Azure Big Data CI/CD Computer Science Data pipelines DevOps Distributed Systems Engineering ETL Hadoop Kafka Linux MongoDB NoSQL Oracle Parquet Pipelines PostgreSQL Python RDBMS Security Shell scripting Spark SQL Testing
More jobs like this
Explore more AI, ML, Data Science career opportunities
Find even more open roles in Artificial Intelligence (AI), Machine Learning (ML), Natural Language Processing (NLP), Computer Vision (CV), Data Engineering, Data Analytics, Big Data, and Data Science in general - ordered by popularity of job title or skills, toolset and products used - below.
- Open MLOps Engineer jobs
- Open Data Science Manager jobs
- Open Lead Data Analyst jobs
- Open Data Manager jobs
- Open Senior Business Intelligence Analyst jobs
- Open Data Engineer II jobs
- Open Power BI Developer jobs
- Open Sr Data Engineer jobs
- Open Principal Data Engineer jobs
- Open Business Intelligence Developer jobs
- Open Data Analytics Engineer jobs
- Open Junior Data Scientist jobs
- Open Product Data Analyst jobs
- Open Data Scientist II jobs
- Open Sr. Data Scientist jobs
- Open Senior Data Architect jobs
- Open Business Data Analyst jobs
- Open Data Analyst Intern jobs
- Open Big Data Engineer jobs
- Open Manager, Data Engineering jobs
- Open Data Product Manager jobs
- Open Junior Data Engineer jobs
- Open Data Quality Analyst jobs
- Open Azure Data Engineer jobs
- Open Principal Data Scientist jobs
- Open GCP-related jobs
- Open Data quality-related jobs
- Open Business Intelligence-related jobs
- Open Java-related jobs
- Open ML models-related jobs
- Open Data management-related jobs
- Open Privacy-related jobs
- Open Data visualization-related jobs
- Open Finance-related jobs
- Open Deep Learning-related jobs
- Open PhD-related jobs
- Open APIs-related jobs
- Open TensorFlow-related jobs
- Open PyTorch-related jobs
- Open NLP-related jobs
- Open Consulting-related jobs
- Open Snowflake-related jobs
- Open CI/CD-related jobs
- Open LLMs-related jobs
- Open Kubernetes-related jobs
- Open Generative AI-related jobs
- Open Data governance-related jobs
- Open Hadoop-related jobs
- Open Airflow-related jobs
- Open Docker-related jobs