We’re looking for a Data Engineer I to join our healthcare analytics platform team. In this role, you’ll help build and support scalable data pipelines that power data-driven healthcare decisions for millions of members. You’ll work closely with experienced engineers in a collaborative, learning-oriented environment where curiosity and growth are encouraged.
Primary Responsibilities:
Build & Maintain Data Pipelines
- Develop and maintain data pipelines using Spark SQL within a modern lakehouse architecture
- Support ETL/ELT processes for healthcare claims data (medical, pharmacy, eligibility)
- Write clear, well-documented code and participate in peer code reviews
Collaborate & Learn
- Partner with senior engineers to learn data architecture and design patterns
- Work with analysts and stakeholders to translate data needs into technical solutions
- Contribute to technical documentation and team knowledge sharing
- Participate in agile practices such as sprint planning, standups, and retrospectives
Support Quality & Reliability
- Monitor pipeline performance and help troubleshoot data issues
- Assist with production support alongside senior team members
- Learn and apply quality, risk, and data validation standards
Required Qualifications
Technical Foundation
- Bachelor’s degree in Computer Science, Engineering, Information Systems, or a related field — or equivalent practical experience
- 0–2 years of relevant experience in data engineering, software engineering, or a related technical role
- Working knowledge of SQL and relational databases
- Experience with Python or similar programming languages
- Familiarity with version control tools such as Git
Ways of Working
- Strong problem-solving and analytical skills
- Care and attention to detail, especially with data accuracy
- Clear communication skills and a collaborative mindset
- Motivation to learn, grow, and build technical expertise
Nice to Have (Not Required)
- Exposure to Apache Spark or Spark SQL
- Experience with cloud platforms (Azure, AWS, or GCP)
- Familiarity with data warehousing or lakehouse concepts
- Experience working with healthcare data or other regulated industries
- Exposure to CI/CD pipelines, automated testing, or Databricks
- Interest in AI-assisted development tools and modern engineering practices
Technical Environment
- Languages: Python, SQL, Spark SQL
- Data Platforms: Azure Databricks, Delta Lake
- Cloud & DevOps: Azure, Azure DevOps, Git, CI/CD pipelines
- Data Domains: Healthcare claims, risk adjustment, population health