

American Unit, Inc
Big Data Developer 3
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Big Data Developer 3 on an 8-month remote contract, paying "rate". Requires 8 years of experience in ETL/ELT workflows, Azure, Python, R, and Spark. Candidates may need to relocate to Texas for occasional meetings.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
October 14, 2025
🕒 - Duration
More than 6 months
-
🏝️ - Location
Remote
-
📄 - Contract
Fixed Term
-
🔒 - Security
Unknown
-
📍 - Location detailed
Texas, United States
-
🧠 - Skills detailed
#Spark SQL #MLflow #Libraries #Delta Lake #SQL (Structured Query Language) #R #Debugging #Storage #Data Pipeline #Data Governance #Scala #Data Quality #Data Science #ML (Machine Learning) #Security #Azure #Cloud #Deployment #Azure Data Factory #Compliance #Data Catalog #Big Data #Data Storage #Apache Spark #"ETL (Extract #Transform #Load)" #Data Lineage #ADLS (Azure Data Lake Storage) #Data Engineering #Data Security #Python #Version Control #Data Warehouse #Azure cloud #TensorFlow #Data Lake #Data Processing #Spark (Apache Spark) #DevOps #ADF (Azure Data Factory) #Metadata #Databricks #Data Management #Programming #Agile #Data Manipulation #Azure ADLS (Azure Data Lake Storage)
Role description
Title/Role: Data Scientist (Big Data Engineer) 3
Worksite Address: Remote.
Duration : 8 Months Contract
The Worker is responsible for developing, maintaining, and optimizing big data solutions using the Databricks Unified Analytics Platform. This role supports data engineering, machine learning, and analytics initiatives within this organization that relies on large-scale data processing.
Two of the three required professional references must include managers or project managers from prior engagements.
Candidates selected for this opportunity may be required to physically attend meetings in Austin, TX, with minimal notice.
Candidates currently residing outside of the State of Texas may be required to relocate to Texas, if selected for this position and the offer is accepted.
Duties include:
• Designing and developing scalable data pipelines
• Implementing ETL/ELT workflows
• Optimizing Spark jobs
• Integrating with Azure Data Factory
• Automating deployments
• Collaborating with cross-functional teams
• Ensuring data quality, governance, and security.
II. CANDIDATE SKILLS AND QUALIFICATIONS
Minimum Requirements:
Candidates that do not meet or exceed the minimum stated requirements (skills/experience) will be displayed to customers but may not be chosen for this opportunity.
Years
Required/Preferred
Experience
8
Required
Implement ETL/ELT workflows for both structured and unstructured data
8
Required
Automate deployments using CI/CD tools
8
Required
Collaborate with cross-functional teams including data scientists, analysts, and stakeholders
8
Required
Design and maintain data models, schemas, and database structures to support analytical and operational use cases
8
Required
Evaluate and implement appropriate data storage solutions, including data lakes (Azure Data Lake Storage) and data warehouses
8
Required
Implement data validation and quality checks to ensure accuracy and consistency
8
Required
Contribute to data governance initiatives, including metadata management, data lineage, and data cataloging
8
Required
Implement data security measures, including encryption, access controls, and auditing; ensure compliance with regulations and best practices
8
Required
Proficiency in Python and R programming languages
8
Required
Strong SQL querying and data manipulation skills
8
Required
Experience with Azure cloud platform
8
Required
Experience with DevOps, CI/CD pipelines, and version control systems
8
Required
Working in agile, multicultural environments
8
Required
Strong troubleshooting and debugging capabilities
5
Required
Design and develop scalable data pipelines using Apache Spark on Databricks
5
Required
Optimize Spark jobs for performance and cost-efficiency
5
Required
Integrate Databricks solutions with cloud services (Azure Data Factory)
5
Required
Ensure data quality, governance, and security using Unity Catalog or Delta Lake
5
Required
Deep understanding of Apache Spark architecture, RDDs, DataFrames, and Spark SQL
5
Required
Hands-on experience with Databricks notebooks, clusters, jobs, and Delta Lake
1
Preferred
Knowledge of ML libraries (MLflow, Scikit-learn, TensorFlow)
1
Preferred
Databricks Certified Associate Developer for Apache Spark
1
Preferred
Azure Data Engineer Associate
III. TERMS OF SERVICE
Services are expected to start 12/03/2025 and are expected to complete by 08/31/2026. Total estimated hours per Candidate shall not exceed 1496 hours. This service may be amended, renewed, and/or extended providing both parties agree to do so in writing.
Title/Role: Data Scientist (Big Data Engineer) 3
Worksite Address: Remote.
Duration : 8 Months Contract
The Worker is responsible for developing, maintaining, and optimizing big data solutions using the Databricks Unified Analytics Platform. This role supports data engineering, machine learning, and analytics initiatives within this organization that relies on large-scale data processing.
Two of the three required professional references must include managers or project managers from prior engagements.
Candidates selected for this opportunity may be required to physically attend meetings in Austin, TX, with minimal notice.
Candidates currently residing outside of the State of Texas may be required to relocate to Texas, if selected for this position and the offer is accepted.
Duties include:
• Designing and developing scalable data pipelines
• Implementing ETL/ELT workflows
• Optimizing Spark jobs
• Integrating with Azure Data Factory
• Automating deployments
• Collaborating with cross-functional teams
• Ensuring data quality, governance, and security.
II. CANDIDATE SKILLS AND QUALIFICATIONS
Minimum Requirements:
Candidates that do not meet or exceed the minimum stated requirements (skills/experience) will be displayed to customers but may not be chosen for this opportunity.
Years
Required/Preferred
Experience
8
Required
Implement ETL/ELT workflows for both structured and unstructured data
8
Required
Automate deployments using CI/CD tools
8
Required
Collaborate with cross-functional teams including data scientists, analysts, and stakeholders
8
Required
Design and maintain data models, schemas, and database structures to support analytical and operational use cases
8
Required
Evaluate and implement appropriate data storage solutions, including data lakes (Azure Data Lake Storage) and data warehouses
8
Required
Implement data validation and quality checks to ensure accuracy and consistency
8
Required
Contribute to data governance initiatives, including metadata management, data lineage, and data cataloging
8
Required
Implement data security measures, including encryption, access controls, and auditing; ensure compliance with regulations and best practices
8
Required
Proficiency in Python and R programming languages
8
Required
Strong SQL querying and data manipulation skills
8
Required
Experience with Azure cloud platform
8
Required
Experience with DevOps, CI/CD pipelines, and version control systems
8
Required
Working in agile, multicultural environments
8
Required
Strong troubleshooting and debugging capabilities
5
Required
Design and develop scalable data pipelines using Apache Spark on Databricks
5
Required
Optimize Spark jobs for performance and cost-efficiency
5
Required
Integrate Databricks solutions with cloud services (Azure Data Factory)
5
Required
Ensure data quality, governance, and security using Unity Catalog or Delta Lake
5
Required
Deep understanding of Apache Spark architecture, RDDs, DataFrames, and Spark SQL
5
Required
Hands-on experience with Databricks notebooks, clusters, jobs, and Delta Lake
1
Preferred
Knowledge of ML libraries (MLflow, Scikit-learn, TensorFlow)
1
Preferred
Databricks Certified Associate Developer for Apache Spark
1
Preferred
Azure Data Engineer Associate
III. TERMS OF SERVICE
Services are expected to start 12/03/2025 and are expected to complete by 08/31/2026. Total estimated hours per Candidate shall not exceed 1496 hours. This service may be amended, renewed, and/or extended providing both parties agree to do so in writing.