Trilyon, Inc.

Data Pipeline Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Pipeline Engineer with a 6-month contract, remote work (PST hours), offering W2 pay. Key skills include 4+ years in data engineering, proficiency in Python, ETL/ELT experience, and knowledge of data privacy techniques.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
200
-
🗓️ - Date
April 29, 2026
🕒 - Duration
More than 6 months
-
🏝️ - Location
Remote
-
📄 - Contract
W2 Contractor
-
🔒 - Security
Unknown
-
📍 - Location detailed
California, United States
-
🧠 - Skills detailed
#IAM (Identity and Access Management) #S3 (Amazon Simple Storage Service) #GCP (Google Cloud Platform) #"ETL (Extract #Transform #Load)" #HTML (Hypertext Markup Language) #Data Pipeline #Datasets #Scala #Data Engineering #JSON (JavaScript Object Notation) #Data Storage #Data Processing #Storage #XML (eXtensible Markup Language) #Python #Monitoring #Data Quality #AWS (Amazon Web Services) #Cloud #Data Privacy
Role description
Job Title: Data Pipeline Engineer (Python, Privacy & Data Processing) Location: Remote (PST Hours) Duration: 6 Months Contract (Possible Extension) Looking for a W2 candidate. About the Role We are seeking a detail-oriented Data Pipeline Engineer to design, build, and maintain scalable data processing pipelines. In this role, you will transform raw, multi-format datasets (text, code, documents, structured exports) into clean, structured, and privacy-compliant data for downstream research and analysis. Key Responsibilities • Build and maintain data pipelines for ingestion, processing, and transformation • Normalise and structure data from diverse formats (JSON, NDJSON, HTML/XML, archives) • Ensure high data quality through validation, QA checks, and continuous improvements • Implement privacy safeguards, including PII detection and anonymisation • Collaborate with internal teams to identify issues, iterate, and improve pipeline performance • Develop tools for data monitoring, auditing, and exploration • Manage data storage, access controls, and pipeline reliability Required Skills & Experience • 4+ years of experience in software engineering or data engineering • Strong proficiency in Python • Hands-on experience building and scaling data pipelines (ETL/ELT) • Experience working with semi-structured or unstructured data • Understanding of data privacy, PII handling, and anonymisation techniques • Familiarity with cloud platforms (AWS/GCP) and storage systems (S3, IAM) • Strong attention to detail and problem-solving skills