DHI Group, Inc.

Data Engineer - Databricks

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer specializing in Azure Databricks for a 6-month contract, offering hourly compensation. Remote or hybrid work is available. Key skills include Databricks, Azure Cloud, Python, and experience with financial systems or agricultural finance.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
February 6, 2026
🕒 - Duration
More than 6 months
-
🏝️ - Location
Hybrid
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
United States
-
🧠 - Skills detailed
#API (Application Programming Interface) #Cloud #Documentation #GraphQL #BI (Business Intelligence) #Data Lake #DevOps #Databricks #PySpark #Spark (Apache Spark) #Data Warehouse #Data Integration #Scala #Data Accuracy #GitHub #"ETL (Extract #Transform #Load)" #Batch #GIT #Storage #Data Governance #Vault #Azure #Azure DevOps #Data Quality #Programming #CLI (Command-Line Interface) #TypeScript #REST (Representational State Transfer) #Data Processing #Azure Databricks #Docker #React #Compliance #Azure cloud #Spark SQL #Azure CLI (Azure Command Line Interface) #PostgreSQL #Delta Lake #Security #Classification #Redis #Data Engineering #Python #SQL (Structured Query Language)
Role description
Azure Databricks Data Engineer Contract Duration: 6 months Location: Remote or Hybrid (Kansas City area) Compensation: Hourly (Contract Position) Work Authorization: Must be legally authorized to work in the U.S. without sponsorship About the Project Join CFA’s initiative to extend our industry-renowned lending platform through Azure Databricks-powered data integration. You will architect and build a centralized data lake that unifies data across systems including Salesforce, loan servicing platforms, document management systems (e.g., DocuSign, Conga), and more — delivering a single source of truth for agricultural finance operations. Core Responsibilities • Design and implement Azure Databricks-based data lake architecture integrating multiple enterprise data sources • Build real-time streaming and batch pipelines for Salesforce, loan servicing, and document management systems • Develop robust data quality, validation, and cleansing processes • Create analytics-ready data structures optimized for BI and operational reporting • Implement data governance and security aligned with SOC 2 Type II compliance • Collaborate with API/service teams to expose unified data via REST and GraphQL endpoints Required Technical Skills • Databricks Expertise: Delta Lake, Spark SQL, PySpark • Azure Cloud Stack: Data Factory, Event Hubs, Blob Storage, Key Vault, Azure AD • ETL/Data Engineering: Pipeline design, dimensional/star schema modeling, data warehouse patterns • Programming: Python (primary), SQL; Scala is a plus • Integration Experience: Salesforce APIs, REST/SOAP, document system connectors • Data Processing: Real-time ingestion & scheduled sync strategies • Security: AES-256 encryption, TLS 1.3, RBAC, data classification Preferred Qualifications • Experience with loan servicing or financial systems • Familiarity with DocuSign, Conga, and email archival systems • GraphQL experience for data exposure • Azure or Databricks certifications (e.g., Data Engineer Associate) • Background in agricultural finance or fintech • Experience with SOC 2 or similar compliance frameworks Technical Environment • Cloud: Azure (Databricks, Data Factory, Blob Storage, Key Vault) • Tools: Git/GitHub, Azure DevOps, GitHub Actions, Docker, Azure CLI • Data Sources: Salesforce, NLS, DocuSign, Conga, email archives • Backend: Node.js, Python APIs, PostgreSQL, Redis • Frontend: React 18+, TypeScript (basic understanding helpful) Success Criteria • Fully integrated, synchronized data lake across all source systems • 95% data accuracy and completeness • <5-minute latency on real-time data streams • Thorough documentation to ensure maintainability • SOC 2 Type II-compliant security implementation