Data Engineer
Remote - San Francisco, California, United States
Applications have closed
WorkStep
WorkStep is the leading employee engagement platform for the frontline workforce, helping enterprise organizations increase operational effectiveness and productivity through holistic employee listening.The Company:
WorkStep is the leading software provider of workforce retention and hiring solutions for the supply chain industry. We’re a Series B startup (backed by leading investors) who is disrupting the industry and changing the way companies have traditionally hired and retained their frontline supply chain workers.
Our mission is simple: to make the supply chain a better place to work. How? By helping companies within e-commerce, manufacturing, retail, transportation, and logistics make better-fit hires and improve their frontline workforce satisfaction and retention.
The Role:
WorkStep is a data-first company and putting data to work is an aspect of everything we do. Our growing Data team is a specialized, cross-functional group that collaborates with a variety of departments to help drive decisions and provide direction for the company. As an early member and the first data engineer of the team, you will be instrumental in continuing to build out our existing analytics and data pipeline capabilities in the short term and will have a tremendous impact as we lay the foundational data practices that will enable the business to benefit from data science/analytics at scale.
You will work closely with our Director of Data Engineering and Chief Technology Officer to design and build scalable and resilient end to end data pipelines that ingest, transform and deliver data using a combination of industry leading open-source and vendor-managed solutions. You'll serve as the primary owner for building and evolving our core datasets that serve as unique sources of truth for product and business functions. This role will collaborate with product managers, product engineers, data scientists, and bizops to accommodate stakeholder needs by building reliable and scalable data pipelines.
Responsibilities:
o Engineer efficient, adaptable, and scalable data pipelines to process structured and unstructured data.
o Develop ETL workflows for data profiling, matching, cleansing, integration, and management
o Own and grow our data lake/warehouse and data governance strategy
o Design and maintain tools and systems to empower our data subject matter experts (analysts and scientists) to explore and derive insights from the data
o Serve as an internal expert for data sources, governance, and mathematical/analytical rigor
o Develop and own the software and infrastructure for APIs and other interfaces for internal and external stakeholders to access our data products
Requirements
o Bachelor's degree in a quantitative field (computer science, applied math, economics, statistics, engineering) or equivalent industry experience
o Strong command of SQL, Python, object-oriented programming paradigms, and common analytical libraries
o Experience building modern and scalable data pipelines and ETL processes
o Experience with data streaming/workflow/orchestration technologies, like Spark, Kafka, Airflow
o Experience with version control (git) and CI/CD tooling/processes for data pipelines and infrastructure configurations.
o Experience with agile software development process
o Familiarity with dbt, Hadoop/EMR, PySpark
o Experience with containerization tools and workflows/execution (i.e. Docker, Kubernetes)
o Comfortable in a smaller rapidly evolving early stage start-up environment while working remotely.
o Desire and comfort in experimenting and deriving POCs to iteratively converge on the right tool for the task.
o Passion for the WorkStep mission and leveraging data to bring insights various business operations/processes
Preferred experience:
Experience working with (or in) our primary industries: HR, warehousing, trucking, manufacturing and production
- Experience with deploying/leveraging production-grade MLOps and ML Pipelines
- Strong command of advanced SQL in a modern data warehouse environment (BigQuery, Redshift, Snowflake)
Benefits
WorkStep is a fully remote company, meaning our team can work from where it suits them—whether that's East Coast or West Coast, in the mountains, or at the beach. We're a collaborative bunch who are focused on helping our customers succeed and deliver results, FAST. But we also know how to have fun and enjoy each other's company. Our benefits include:
- Remote working environment
- Flexible PTO
- Top-notch technology
- Annual team building on-sites (when safe to resume)
- Workspace, wellness, and professional development stipends
- Internet and phone reimbursement
- Competitive company-sponsored health, vision, and dental benefits package
- Opportunity to join a passionate, motivated, and fun team at an early stage to help shape and execute on our mission
If you’re a collaborator who likes a challenge, who doesn’t mind rolling up their sleeves, and wants to join a fast growing company at an early stage, we want to hear from you!
WorkStep is an EEO employer. We do not discriminate against any applicant for employment, or any employee because of race, color, religion, national origin, age, sex, sexual orientation, gender identity, gender, disability, age, or protected veteran status. We are committed to building a safe, inclusive environment for people of all backgrounds.
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Agile Airflow APIs BigQuery CI/CD Computer Science Data pipelines Docker E-commerce Economics Engineering ETL Git Hadoop Kafka Kubernetes Machine Learning MLOps OOP Pipelines PySpark Python Redshift Snowflake Spark SQL Statistics Streaming Unstructured data
Perks/benefits: Career development Flex hours Flex vacation Health care Startup environment Wellness
More jobs like this
Explore more AI, ML, Data Science career opportunities
Find even more open roles in Artificial Intelligence (AI), Machine Learning (ML), Natural Language Processing (NLP), Computer Vision (CV), Data Engineering, Data Analytics, Big Data, and Data Science in general - ordered by popularity of job title or skills, toolset and products used - below.
- Open MLOps Engineer jobs
- Open Lead Data Analyst jobs
- Open Data Science Manager jobs
- Open Senior Business Intelligence Analyst jobs
- Open Data Manager jobs
- Open Data Engineer II jobs
- Open Power BI Developer jobs
- Open Principal Data Engineer jobs
- Open Sr Data Engineer jobs
- Open Business Intelligence Developer jobs
- Open Junior Data Scientist jobs
- Open Data Analytics Engineer jobs
- Open Data Scientist II jobs
- Open Product Data Analyst jobs
- Open Sr. Data Scientist jobs
- Open Senior Data Architect jobs
- Open Business Data Analyst jobs
- Open Data Analyst Intern jobs
- Open Big Data Engineer jobs
- Open Manager, Data Engineering jobs
- Open Azure Data Engineer jobs
- Open Data Quality Analyst jobs
- Open Data Product Manager jobs
- Open Junior Data Engineer jobs
- Open Principal Data Scientist jobs
- Open GCP-related jobs
- Open Data quality-related jobs
- Open Business Intelligence-related jobs
- Open ML models-related jobs
- Open Java-related jobs
- Open Data management-related jobs
- Open Privacy-related jobs
- Open Finance-related jobs
- Open Data visualization-related jobs
- Open Deep Learning-related jobs
- Open PhD-related jobs
- Open APIs-related jobs
- Open TensorFlow-related jobs
- Open PyTorch-related jobs
- Open NLP-related jobs
- Open Consulting-related jobs
- Open Snowflake-related jobs
- Open CI/CD-related jobs
- Open Kubernetes-related jobs
- Open LLMs-related jobs
- Open Generative AI-related jobs
- Open Data governance-related jobs
- Open Hadoop-related jobs
- Open Airflow-related jobs
- Open Docker-related jobs