American Unit, Inc

Big Data Developer 3

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Big Data Developer 3 on an 8-month remote contract, paying "rate". Requires 8 years of experience in ETL/ELT workflows, Azure, Python, R, and Spark. Candidates may need to relocate to Texas for occasional meetings.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
October 14, 2025
🕒 - Duration
More than 6 months
-
🏝️ - Location
Remote
-
📄 - Contract
Fixed Term
-
🔒 - Security
Unknown
-
📍 - Location detailed
Texas, United States
-
🧠 - Skills detailed
#Spark SQL #MLflow #Libraries #Delta Lake #SQL (Structured Query Language) #R #Debugging #Storage #Data Pipeline #Data Governance #Scala #Data Quality #Data Science #ML (Machine Learning) #Security #Azure #Cloud #Deployment #Azure Data Factory #Compliance #Data Catalog #Big Data #Data Storage #Apache Spark #"ETL (Extract #Transform #Load)" #Data Lineage #ADLS (Azure Data Lake Storage) #Data Engineering #Data Security #Python #Version Control #Data Warehouse #Azure cloud #TensorFlow #Data Lake #Data Processing #Spark (Apache Spark) #DevOps #ADF (Azure Data Factory) #Metadata #Databricks #Data Management #Programming #Agile #Data Manipulation #Azure ADLS (Azure Data Lake Storage)
Role description
Title/Role: Data Scientist (Big Data Engineer) 3 Worksite Address: Remote. Duration : 8 Months Contract The Worker is responsible for developing, maintaining, and optimizing big data solutions using the Databricks Unified Analytics Platform. This role supports data engineering, machine learning, and analytics initiatives within this organization that relies on large-scale data processing. Two of the three required professional references must include managers or project managers from prior engagements. Candidates selected for this opportunity may be required to physically attend meetings in Austin, TX, with minimal notice. Candidates currently residing outside of the State of Texas may be required to relocate to Texas, if selected for this position and the offer is accepted. Duties include: • Designing and developing scalable data pipelines • Implementing ETL/ELT workflows • Optimizing Spark jobs • Integrating with Azure Data Factory • Automating deployments • Collaborating with cross-functional teams • Ensuring data quality, governance, and security. II. CANDIDATE SKILLS AND QUALIFICATIONS Minimum Requirements: Candidates that do not meet or exceed the minimum stated requirements (skills/experience) will be displayed to customers but may not be chosen for this opportunity. Years Required/Preferred Experience 8 Required Implement ETL/ELT workflows for both structured and unstructured data 8 Required Automate deployments using CI/CD tools 8 Required Collaborate with cross-functional teams including data scientists, analysts, and stakeholders 8 Required Design and maintain data models, schemas, and database structures to support analytical and operational use cases 8 Required Evaluate and implement appropriate data storage solutions, including data lakes (Azure Data Lake Storage) and data warehouses 8 Required Implement data validation and quality checks to ensure accuracy and consistency 8 Required Contribute to data governance initiatives, including metadata management, data lineage, and data cataloging 8 Required Implement data security measures, including encryption, access controls, and auditing; ensure compliance with regulations and best practices 8 Required Proficiency in Python and R programming languages 8 Required Strong SQL querying and data manipulation skills 8 Required Experience with Azure cloud platform 8 Required Experience with DevOps, CI/CD pipelines, and version control systems 8 Required Working in agile, multicultural environments 8 Required Strong troubleshooting and debugging capabilities 5 Required Design and develop scalable data pipelines using Apache Spark on Databricks 5 Required Optimize Spark jobs for performance and cost-efficiency 5 Required Integrate Databricks solutions with cloud services (Azure Data Factory) 5 Required Ensure data quality, governance, and security using Unity Catalog or Delta Lake 5 Required Deep understanding of Apache Spark architecture, RDDs, DataFrames, and Spark SQL 5 Required Hands-on experience with Databricks notebooks, clusters, jobs, and Delta Lake 1 Preferred Knowledge of ML libraries (MLflow, Scikit-learn, TensorFlow) 1 Preferred Databricks Certified Associate Developer for Apache Spark 1 Preferred Azure Data Engineer Associate III. TERMS OF SERVICE Services are expected to start 12/03/2025 and are expected to complete by 08/31/2026. Total estimated hours per Candidate shall not exceed 1496 hours. This service may be amended, renewed, and/or extended providing both parties agree to do so in writing.