Back to Jobs
Zapcom

Data Engineer at Zapcom

Zapcom Boston, MA

Job Description

 Design and deploy scalable ETL/ELT pipelines to ingest, transform,and load clinical data from diverse sources (EMRs,  labs, IoT devices, datalakes, FHIR/HL7 APIs), including Azure and Snowflake.Architect and optimize Microsoft Azure and Snowflakeenvironments for clinical data storage, extraction, transformation, and loading, machinelearning operations (MLOps) performance tuning, cost management, and securedata sharing. Ensure compliance with healthcare regulations (HIPAA, GDPR) byimplementing data anonymization, encryption, and audit trails.Collaborate with clinical stakeholders to translate businessrequirements into technical solutions for analytics and reporting.Develop and maintain data governance frameworks, including metadatamanagement, data lineage, and quality checks (e.g., validation of lab results,patient demographics). Automate data pipelines using orchestration tools (e.g., ApacheAirflow, Prefect) and integrate real -time streaming solutions (e.g., Kafka)where applicable. Build and maintain documentation for data models,pipelines, and processes to ensure reproducibility and transparency.Advancedproficiency in Snowflake (Snowpipe, Time Travel, Zero -Copy Cloning) andSQL for complex transformations.Hands -onexperience with ETL/ELT tools (Apache Spark, AWS Glue, Azure DataFactory) and cloud platforms (AWS, Azure, GCP).Strongprogramming skills in Python/Scala (Pandas, PySpark) for data scriptingand automation. Familiaritywith healthcare data formats (OMOP, FHIR, HL7, DICOM) and clinical workflows.Expertise infederated learning and running large jobs on high -performance computing serversis a plus.Data Governance: Ability to implementdata quality frameworks (e.g., Great Expectations) and metadata managementtools.Regulatory Compliance: Proven experiencesecuring PHI/PII data and adhering to HIPAA/GDPR requirements.Problem -Solving: Ability totroubleshoot pipeline failures, optimize query performance, and resolve datadiscrepancies.Collaboration: Strong communicationskills to work with cross -functional teams (clinicians, analysts, IT).RequirementsSnowflakeSnowPro Core Certification (or higher).AWS/Azure/GCPData Engineering Certification (e.g., AWS Certified Data Analytics, Azure DataEngineer Associate).Runningjobs on high -performance computing serversHealthcare -specificcertifications (e.g., HL7 FHIR Certification, Certified Health Data Analyst(CHDA)).Securitycertifications (e.g., CISSP, CIPP) for handling sensitive clinical data.3+ years of experience in data engineering, with 2+ years focusedon healthcare/clinical data (e.g., hospitals, EMR systems, clinicaltrials).2+ years of hands -on experience with Snowflake inproduction environments.Proven track record of building ETL pipelines for large -scale clinicaldatasets.Experience with OMOP CDM, Epic/Cerner EHR systems, or clinicaldata lakes.Exposure to DevOps practices (CI/CD, Terraform) and Agilemethodologies.Some front -end development is preferred.

Resume Suggestions

Highlight relevant experience and skills that match the job requirements to demonstrate your qualifications.

Quantify your achievements with specific metrics and results whenever possible to show impact.

Emphasize your proficiency in relevant technologies and tools mentioned in the job description.

Showcase your communication and collaboration skills through examples of successful projects and teamwork.

Explore More Opportunities