Data Lake Developer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Lead Data Lake Developer with a contract length of "unknown," offering a pay rate of "$XX/hour." Required skills include Data Lake, Snowflake, SQL, AWS, and DataStage, along with experience in ETL processes and data governance.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
640
-
πŸ—“οΈ - Date discovered
August 28, 2025
πŸ•’ - Project duration
Unknown
-
🏝️ - Location type
Unknown
-
πŸ“„ - Contract type
Unknown
-
πŸ”’ - Security clearance
Unknown
-
πŸ“ - Location detailed
Columbus, OH
-
🧠 - Skills detailed
#Scala #Apache Spark #Compliance #DataStage #Data Privacy #GDPR (General Data Protection Regulation) #Data Quality #Azure #Data Lake #SQL (Structured Query Language) #AWS S3 (Amazon Simple Storage Service) #Data Catalog #Data Management #"ETL (Extract #Transform #Load)" #Spark (Apache Spark) #Storage #RDBMS (Relational Database Management System) #S3 (Amazon Simple Storage Service) #Databricks #Snowflake #Cloud #GCP (Google Cloud Platform) #Data Science #AWS (Amazon Web Services) #Data Security #Metadata #BI (Business Intelligence) #Data Processing #AWS Glue #IoT (Internet of Things) #Security
Role description
We are seeking a highly skilled Lead Data Lake Developer to design, develop, and manage our enterprise data lake infrastructure. In this role, you will be responsible for leading the architecture, development, and optimization of scalable, high-performance data lake solutions that support analytics, data science, and business intelligence needs across the organization. Must Have: β€’ Must have Lead experience β€’ Data Lake β€’ Snowflake or equivalent tool β€’ SQL β€’ AWS β€’ DataStage Key Responsibilities: β€’ Lead the design and implementation of enterprise-scale Data Lake solutions using cloud-native technologies (e.g., AWS S3, Azure Data Lake, GCP Storage). β€’ Ingest, process, and manage large volumes of structured and unstructured data from diverse sources (e.g., RDBMS, APIs, IoT, logs). β€’ Collaborate with data scientists, analysts, and business stakeholders to understand data needs and translate them into scalable data lake architectures. β€’ Develop and manage ETL/ELT pipelines using tools like Apache Spark, AWS Glue, Databricks, or similar. β€’ Ensure data quality, metadata management, lineage tracking, and governance using tools like AWS Lake Formation, Apache Atlas, or Data Catalogs. β€’ Optimize performance of data processing jobs and storage layout (e.g., partitioning, file formats like Parquet/ORC). β€’ Enforce data security, access controls, and compliance with data privacy regulations (GDPR, HIPAA, etc.). β€’ Mentor junior engineers and enforce engineering best practices for code quality, testing, and CI/CD in data workflows.