Apptad Inc.

Senior Analytics Engineer (DBT & Databricks)

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior Analytics Engineer (DBT & Databricks) for a 6-month remote contract (US & Canada). Requires 5+ years in data engineering, expertise in SQL, DBT, Databricks on AWS, and proficiency in Python and orchestration tools.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
April 25, 2026
🕒 - Duration
More than 6 months
-
🏝️ - Location
Remote
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
United States
-
🧠 - Skills detailed
#Scala #Data Quality #Airflow #GIT #Data Engineering #Data Mart #Python #AWS (Amazon Web Services) #Documentation #Monitoring #Version Control #SQL (Structured Query Language) #Data Governance #dbt (data build tool) #Data Modeling #Data Pipeline #Databricks #"ETL (Extract #Transform #Load)" #Datasets
Role description
Job Title: Senior Analytics Engineer (DBT & Databricks) Location: Remote (US & Canada) Duration: 6 Months About the Role We are seeking a Senior Analytics Engineer to design and build scalable data transformation solutions using DBT and Databricks. This role focuses on modernizing data pipelines, improving data models, and enabling high-quality, self-service analytics. Key Responsibilities • Migrate legacy data pipelines into a modern Databricks + DBT (Lakehouse) architecture • Analyze existing workflows and redesign them using best-in-class data modeling practices • Develop and maintain scalable DBT models and transformation pipelines • Design dimensional and relational data models (fact/dimension, star schema) • Own transformation layers (Silver/Gold) and ensure data readiness for analytics • Implement data quality checks, monitoring, and SLA-driven pipelines • Collaborate with analytics, product, and data engineering teams • Support data governance, documentation, and lineage • Enable self-service reporting through well-structured data marts Required Skills & Experience • 5+ years in data/analytics engineering or data modeling • Strong hands-on expertise in SQL (advanced level) • Proven experience with DBT (models, tests, incremental loads, optimization) • Experience with Databricks on AWS (must-have) • Strong understanding of data warehousing & dimensional modeling • Experience handling large-scale datasets and performance tuning • Hands-on with Python and orchestration tools (Airflow or similar) • Familiarity with CI/CD, Git, and version control practices