Back to Jobs
Codvo.ai

(US) Fullstack Data Engineer - Databricks at Codvo.ai

Codvo.ai Dallas, TX

Job Description

Job Description:Data Engineer – Databricks Role OverviewWe are looking for a highly skilled Full Stack Data Engineer with expertise in Databricks to design, develop, and optimize end-to-end data pipelines, data platforms, and analytics solutions. This role combines strong data engineering, cloud platform expertise, and software engineering skills to deliver scalable, production-grade solutions.Key ResponsibilitiesDesign and develop ETL/ELT pipelines on Databricks (PySpark, Delta Lake, SQL).Architect data models (batch and streaming) for analytics, ML, and reporting.Optimize performance of large-scale distributed data processing jobs.Implement CI/CD pipelines for Databricks workflows using GitHub Actions, Azure DevOps, or similar.Build and maintain APIs, dashboards, or applications that consume processed data (full-stack aspect).Collaborate with data scientists, analysts, and business stakeholders to deliver solutions.Ensure data quality, lineage, governance, and security compliance. Required Skills & QualificationsCore Databricks Skills:Strong in PySpark, Delta Lake, Databricks SQL.Experience with Databricks Workflows, Unity Catalog, and Delta Live Tables.Programming & Full Stack:Python (mandatory), SQL (expert).Exposure to Java/Scala (for Spark jobs).Knowledge of APIs, microservices (FastAPI/Flask), or basic front-end (React/Angular) is a plus.Cloud Platforms:Proficiency with at least one: Azure Databricks, AWS Databricks, or GCP Databricks.Knowledge of cloud storage (ADLS, S3, GCS), IAM, networking.DevOps & CI/CD:Git, CI/CD tools (GitHub Actions, Azure DevOps, Jenkins).Containerization (Docker, Kubernetes is a plus).Data Engineering Foundations:Data modeling (OLTP/OLAP).Batch & streaming data processing (Kafka, Event Hub, Kinesis).Data governance & compliance (Unity Catalog, Lakehouse security). Nice-to-HaveExperience with machine learning pipelines (MLflow, Feature Store).Knowledge of data visualization tools (Power BI, Tableau, Looker).Exposure to Graph databases (Neo4j) or RAG/LLM pipelines.QualificationsBachelor’s or Master’s in Computer Science, Data Engineering, or related field.4–7 years of experience in data engineering, with at least 2 years on Databricks.Soft SkillsStrong problem-solving and analytical skills.Ability to work in fusion teams (business + engineering + AI/ML).Clear communication and documentation abilities. About UsAt Codvo, we are committed to building scalable, future-ready data platforms that power business impact. We believe in a culture of innovation, collaboration, and growth, where engineers can experiment, learn, and thrive. Join us to be part of a team that solves complex data challenges with creativity and cutting-edge technology.

Resume Suggestions

Highlight relevant experience and skills that match the job requirements to demonstrate your qualifications.

Quantify your achievements with specific metrics and results whenever possible to show impact.

Emphasize your proficiency in relevant technologies and tools mentioned in the job description.

Showcase your communication and collaboration skills through examples of successful projects and teamwork.

Explore More Opportunities