

Sr. Data Engineer - W2
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Sr. Data Engineer in Dallas, TX, or NYC, on a contract basis for over 6 months, offering competitive pay. Requires 10+ years of experience in data engineering, expertise in time series data, and proficiency in Python.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
-
ποΈ - Date discovered
May 30, 2025
π - Project duration
More than 6 months
-
ποΈ - Location type
On-site
-
π - Contract type
W2 Contractor
-
π - Security clearance
Unknown
-
π - Location detailed
Dallas, TX
-
π§ - Skills detailed
#NumPy #AWS (Amazon Web Services) #Libraries #Spark (Apache Spark) #Data Pipeline #ML (Machine Learning) #Storage #Cloud #Scala #Data Engineering #Kafka (Apache Kafka) #Pandas #Data Ingestion #Python #GCP (Google Cloud Platform) #Data Architecture #Monitoring #Batch #Datasets #Data Quality #PyTorch #Data Governance #Data Design #Azure #"ETL (Extract #Transform #Load)" #Data Science #Time Series #Consul #Observability #Consulting #Indexing #Databases
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
Trident Consulting is seeking a " Sr. Data Engineerβ for one of our clients in Dallas TX
Role: Sr. Data Engineer
Location: Dallas TX / NYC NY βOnsite
Type: Contract
Experience level: 10 Years
β’
β’
β’
β’ The candidate should be local to TX or NY
β’
β’
β’
β’
β’ Description:
As a Senior Data Engineer, youβll architect and implement large-scale time series data pipelines that support high-throughput ingestion, real-time querying, and seamless integration with Python-based machine learning workflows. Youβll work closely with engineering, analytics, and data science teams to ensure data systems are reliable, high-performance, and optimized for large volumes and low-latency workloads.
Your work will enable model training, evaluation, and inference on dynamic, continuously evolving datasets that drive real-time insight and innovation.
What Youβll Do
As a Senior Data Engineer, you will:
β’ Design, build, and optimize high-performance data pipelines for large-scale time series data
β’ Implement scalable data infrastructure using tools such as KDB+, TimeSet (Googleβs large time series database), or Kronos
β’ Develop efficient data ingestion and transformation workflows that handle real-time and historical time series data
β’ Connect time series data systems with Python-based model pipelines to support machine learning training and inference
β’ Collaborate with data scientists and ML engineers to ensure data availability, quality, and accessibility for experimentation and production
β’ Design data models and schemas optimized for time series use cases, including downsampling, aggregation, and indexing strategies
β’ Ensure system reliability, scalability, and performance through monitoring, testing, and tuning
β’ Establish data governance, lineage, and observability best practices in large-scale environments
β’ Mentor junior engineers on large-scale data design, distributed processing, and real-time system architecture
β’ Partner with product, engineering, and infrastructure teams to align data systems with business goals
Requirements
Youβll bring:
β’ 8+ years of experience in data engineering, with a focus on large-scale and high-throughput systems
β’ Deep experience working with time series data and purpose-built storage systems (e.g., KDB+, TimeSet, Kronos)
β’ Strong experience building streaming and batch data pipelines using tools like Kafka, Flink, or Spark
β’ Proficiency in Python and integrating data pipelines with machine learning workflows and libraries (e.g., pandas, NumPy, scikit-learn, PyTorch)
β’ Experience designing efficient, scalable data models and partitioning strategies for time series data
β’ Knowledge of distributed systems, columnar databases, and parallel processing
β’ Familiarity with cloud-native data architectures (AWS, GCP, or Azure) and containerized data infrastructure
β’ Strong understanding of data quality, lineage, monitoring, and observability tools
β’ Excellent communication skills and a proactive, consultative mindset in client-facing environments
β’ Bonus: Experience with multiple time series systems (e.g., KDB+ and Kronos) or contributing to open-source data infrastructure projects
About Trident: Trident Consulting is an award-winning IT/engineering staffing company founded in 2005 and headquartered in San Ramon, CA. We specialize in placing high-quality vetted technology and engineering professionals in contract and full-time roles. Trident's commitment is to deliver the best and brightest individuals in the industry for our clients' toughest requirements.
Some of our recent awards include: 2022, 2021, 2020 Inc. 5000 fastest-growing private companies in America 2022, 2021 SF Business Times 100 fastest-growing private companies in Bay Area