Advanced Software Talent

Senior Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior Data Engineer, lasting 6 to 12 months, with a pay rate of "X" at a remote location. Requires 7+ years of experience, strong SQL and Python skills, and expertise in cloud-based data platforms, preferably AWS.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
560
-
πŸ—“οΈ - Date
May 16, 2026
πŸ•’ - Duration
More than 6 months
-
🏝️ - Location
Unknown
-
πŸ“„ - Contract
W2 Contractor
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
South San Francisco, CA
-
🧠 - Skills detailed
#Data Strategy #Compliance #Lambda (AWS Lambda) #Datasets #Leadership #Cloud #Agile #Redshift #Data Quality #SQL (Structured Query Language) #Deployment #Scala #Data Modeling #Data Pipeline #Code Reviews #Data Governance #API (Application Programming Interface) #Observability #ML (Machine Learning) #Data Engineering #Data Science #Computer Science #Data Security #Docker #Strategy #Athena #Monitoring #Metadata #AWS (Amazon Web Services) #Data Profiling #Data Architecture #S3 (Amazon Simple Storage Service) #Python #Security #Data Management #"ETL (Extract #Transform #Load)" #AI (Artificial Intelligence)
Role description
Direct W2 contractors only! We are not able to sponsor any kind of Visa! No 3rd party agencies! Duration: 6 to 12 months As a Lead Data Engineer, you will act as a hands-on technical leader- designing and building data solutions while guiding engineering best practices across the team. This is a player-coach role, where you will actively contribute to development while mentoring others and driving high-quality delivery. You will be a pivotal member of our team, responsible for: Key Responsibilities: β€’ Hands-on Data Engineering & Delivery: β€’ Design, build, and maintain scalable data pipelines to ingest, transform, and curate structured and unstructured data. β€’ Write production-quality code in SQL and Python and actively contribute to day-to-day development. β€’ Troubleshoot, optimize, and improve performance of data workflows and systems. β€’ Technical Leadership & Data Architecture Ownership: β€’ Lead by example through hands-on contributions to critical projects. β€’ Provide technical guidance, code reviews, and mentorship to other data engineers. β€’ Help drive implementation of best practices in data engineering, testing, and deployment. β€’ Design and Build Scalable Data Pipelines: β€’ Architect, develop and oversee development of pipelines to ingest, transform, and curate structured and unstructured data from internal and external sources. β€’ Ensure high performance, scalability, and reliability of data systems β€’ Data Profiling, Mapping & Standardization: β€’ Profile data, identify quality issues, and align disparate datasets. Define data models and standardization frameworks to support scalable, reusable, and AI/ML-ready data products. β€’ Data Product Engineering & API Development: β€’ Build and maintain reusable data products and APIs to support analytics and AI use cases. β€’ Ensure solutions are well-documented, secure, and scalable. β€’ AI/ML Enablement: β€’ Work closely with data scientists to prepare and deliver high-quality datasets for ML and AI use cases. β€’ Support data pipelines for LLM and AI-driven workflows. β€’ Metadata Management & Data Governance: β€’ Champion data governance, lineage, and metadata management practices β€’ Ensure compliance with enterprise data security and privacy standards. β€’ Monitoring and Event Frameworks: β€’ Implement monitoring, alerting, and event-driven frameworks for data pipelines. β€’ Ensure robustness, observability, and reliability of data systems. β€’ Container and Workflow Orchestration: β€’ Lead adoption of containerized and orchestrated data workloads (e.g., Docker, Amazon EKS) β€’ Guide orchestration of complex AI/data workflows using modern tooling. β€’ Cross-functional Collaboration & Influence β€’ Partner with business, product, and external stakeholders to align data strategy with organizational goals. β€’ Translate business needs into scalable technical solutions. β€’ Continuous Improvement: β€’ Identify opportunities to improve tooling, processes, and performance. β€’ Stay hands-on with new technologies and bring practical improvements to the team. Qualifications Basic Qualifications: β€’ Bachelor's or Master s degree in Computer Science, Engineering, Data Science, or a related technical field. β€’ 7+ years of experience in data engineering or similar roles, including experience leading projects or initiatives. β€’ Proven track record designing and scaling cloud-based data platforms (preferably AWS). β€’ Strong proficiency in SQL, Python, and advanced data modeling techniques. β€’ Experience leading architecture decisions and implementing best practices. β€’ Strong understanding of data quality, integration, transformation, and governance. β€’ Excellent communication skills with the ability to influence technical and non-technical stakeholders Preferred Qualifications: β€’ Experience acting as a technical lead or senior individual contributor on data engineering projects. β€’ Hands-on experience with AWS data services (Glue, Redshift, S3, Lambda, Athena, etc.). β€’ Experience supporting AI/ML data pipelines and workflows. β€’ Familiarity with metadata management and data governance frameworks. β€’ Experience in healthcare/life sciences or partnering domains. β€’ Experience working in Agile environments.