Back to Jobs
Mercor

Remote Senior Data Engineer / Analytics Engineer (India-Based) - AI Trainer ($35-$70 per hour) at Mercor

Mercor Remote - Coconut Creek, FL

Job Description

Mercor is partnering with a cutting-edge AI research lab to hire a **Senior Data/Analytics Engineer** with expertise across **DBT** and **Snowflake’s Cortex CLI**. In this role, you will build and scale Snowflake-native data and ML pipelines, leveraging Cortex’s emerging AI/ML capabilities while maintaining production-grade DBT transformations. You will work closely with data engineering, analytics, and ML teams to prototype, operationalise, and optimise AI-driven workflows—defining best practices for Snowflake-native feature engineering and model lifecycle management.This is a high-impact role within a modern, fully cloud-native data stack. * * * ## **Responsibilities** - Design, build, and maintain **DBT models, macros, and tests** following modular data modeling and semantic best practices. - Integrate **DBT workflows with Snowflake Cortex CLI**, enabling: - Feature engineering pipelines - Model training & inference tasks - Automated pipeline orchestration - Monitoring and evaluation of Cortex-driven ML models - Establish best practices for **DBT–Cortex architecture and usage patterns**. - Collaborate with data scientists and ML engineers to **produce Cortex workloads** in Snowflake. - Build and optimise **CI/CD pipelines** for dbt (GitHub Actions, GitLab, Azure DevOps). - Tune Snowflake compute and queries for **performance and cost efficiency**. - Troubleshoot issues across DBT arti-facts, Snowflake objects, lineage, and data quality. - Provide guidance on **DBT project governance, structure, documentation, and testing frameworks**. * * * ## **Required Qualifications** - **3+ years** experience with **DBT Core or DBT Cloud**, including macros, packages, testing, and deployments. - Strong expertise with **Snowflake** (warehouses, tasks, streams, materialised views, performance tuning). - Hands-on experience with **Snowflake Cortex CLI**, or strong ability to learn it quickly. - Strong SQL skills; working familiarity with **Python** for scripting and DBT automation. - Experience integrating DBT with orchestration tools (Airflow, Dagster, Prefect, etc.). - Solid understanding of **modern data engineering, ELT patterns, and version-controlled analytics development**. * * * ## **Nice-to-Have Skills** - Prior experience operationalising **ML workflows inside Snowflake**. - Familiarity with **Snow-park**, Python UDFs/UDTFs. - Experience building **semantic layers** using DBT metrics. - Knowledge of **MLOps / DataOps** best practices. - Exposure to **LLM workflows, vector search, and unstructured data pipelines**. * * * ## **Why Join** - You will be an **hourly contractor through Mercor**, working **20–40 hours per week** with flexibility. - Direct opportunity to build **next-generation Snowflake AI/ML systems** with Cortex. - High-impact ownership of DBT and Snowflake architecture across production pipelines. - Work alongside top-tier ML engineers, data scientists, and research teams. - Fully remote, high-autonomy environment focused on innovation, velocity, and engineering excellence.

Resume Suggestions

Highlight relevant experience and skills that match the job requirements to demonstrate your qualifications.

Quantify your achievements with specific metrics and results whenever possible to show impact.

Emphasize your proficiency in relevant technologies and tools mentioned in the job description.

Showcase your communication and collaboration skills through examples of successful projects and teamwork.

Explore More Opportunities