Vaniam Group

Independent Contractor, Data Engineer, Business Operations

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is an Independent Contractor, Data Engineer for Business Operations, lasting "contract length". Pay rate is "pay rate". Requires 2+ years of data engineering experience, proficiency in Databricks, SQL, and cloud platforms (AWS/Azure). Familiarity with data governance and healthcare compliance preferred.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
Unknown
-
πŸ—“οΈ - Date
February 6, 2026
πŸ•’ - Duration
Unknown
-
🏝️ - Location
Unknown
-
πŸ“„ - Contract
1099 Contractor
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
United States
-
🧠 - Skills detailed
#Tableau #Automation #Cloud #GraphQL #Microsoft Power BI #BI (Business Intelligence) #Data Modeling #Databases #Looker #AI (Artificial Intelligence) #SaaS (Software as a Service) #Databricks #PySpark #Spark (Apache Spark) #dbt (data build tool) #Scala #GitHub #Airflow #"ETL (Extract #Transform #Load)" #GIT #Data Governance #Metadata #Datasets #CRM (Customer Relationship Management) #Data Catalog #Azure #Data Pipeline #REST (Representational State Transfer) #Observability #Compliance #AWS (Amazon Web Services) #Spark SQL #ML (Machine Learning) #Delta Lake #Forecasting #Data Engineering #SQL (Structured Query Language)
Role description
Independent Contractor, Data Engineer, Business Operations What You’ll Do We’re seeking a hands-on Data Engineer to partner with Business Operations in building a reliable, scalable data foundation. You’ll centralize operational data across core systems, develop automated pipelines and models that power self-serve analytics, and replace manual reporting with scheduled, high-trust data products. You’ll also lay the groundwork for advanced use cases such as forecasting, decision support, and natural-language access to trusted dataβ€”accelerating better, faster decisions across the organization. A Day in the Life Deliverables β€’ Stand up a scalable Databricks lakehouse to ingest, model, and serve business operations data (finance, resourcing, project delivery, CRM, marketing, and time tracking). β€’ Design and maintain automated ELT/ETL pipelines that move data from SaaS tools, databases, and files into bronze/silver/gold layers. β€’ Build the core semantic layer (cleaned, conformed, documented tables) that powers self-serve BI and executive dashboards. β€’ Replace legacy/manual engagement and utilization reports with scheduled, monitored jobs and SLAs. β€’ Partner with Business Operations, Finance, and People Operations leaders to define source-of-truth metrics (e.g., revenue, margin, utilization, velocity, pipeline, engagement health). β€’ Lay groundwork for AI use cases (RAG over operational data, agentic processes, querying company data) by implementing robust lineage, metadata, and access controls. β€’ Architecture & Modeling: Design lakehouse architecture, dimensional/medallion models, and data contracts across systems. β€’ Pipeline Automation: Implement CI/CD for data (branching, PRs, jobs, environments), with observability and reproducibility. β€’ Data Governance: Enforce PII/PHI handling, role-based access, auditability, and retention aligned to healthcare-adjacent standards. β€’ Enablement: Document datasets, publish a data catalog, and enable self-serve usage via BI and SQL. β€’ Reporting Modernization: Decommission manual spreadsheets and one-off extracts; consolidate to certified, scheduled outputs. β€’ AI Readiness: Capture lineage/metadata and vector-friendly document stores to support future ML and RAG initiatives. What You Must Have Education and Experience β€’ 2+ years in data engineering or analytics engineering, including building production data pipelines at scale. β€’ Expert with Databricks (Delta Lake, SQL, PySpark) and cloud data platforms (AWS or Azure). Skills and Competencies β€’ Proficient with dbt and/or Delta Live Tables; strong SQL and data modeling fundamentals. β€’ Experience orchestrating jobs (Airflow, Databricks Workflows, or equivalent) β€’ Comfortable with PowerBI and semantic modeling for self-serve analytics. β€’ Familiarity with data governance (RBAC/ABAC, secrets management, token-based auth) and healthcare-adjacent compliance (e.g., HIPAA concepts) is a plus. β€’ Strong stakeholder skills. Can translate business needs into reliable data products and clear SLAs. β€’ Databricks, Delta Lake, PySpark, SQL, dbt, REST/GraphQL APIs, Git/GitHub, Power BI/Tableau/Looker.