

Senior DataOps Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior DataOps Engineer in Dearborn, MI, with a contract length of unspecified duration and a competitive pay rate. Key skills include GCP, DataOps principles, Python, and data governance. A Bachelor's or Master's degree in a related field is required.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
-
ποΈ - Date discovered
July 30, 2025
π - Project duration
Unknown
-
ποΈ - Location type
On-site
-
π - Contract type
Unknown
-
π - Security clearance
Unknown
-
π - Location detailed
Dearborn, MI
-
π§ - Skills detailed
#Dataflow #Batch #Cloud #Security #Terraform #Data Encryption #Consulting #REST (Representational State Transfer) #Schema Design #DataOps #Observability #Data Governance #"ETL (Extract #Transform #Load)" #Documentation #Automation #Version Control #BigQuery #GCP (Google Cloud Platform) #Data Engineering #Computer Science #Data Modeling #Data Quality #Network Security #Python #Scala #Programming #Data Architecture #Data Pipeline #Automated Testing #GIT #Monitoring #VPC (Virtual Private Cloud) #Infrastructure as Code (IaC) #Compliance #Logging
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
Details:
Job Description
Stefanini Group is hiring!
Stefanini is looking for Senior DataOps Engineer, Dearborn, MI (Onsite)
For quick apply, please reach out Pawan Rawat Singh at 248-213-3605/pawansingh.rawat@stefanini.com
We are seeking a highly skilled and experienced Senior DataOps Engineer to join our EPEO DataOps team. This role will be pivotal in designing, building, and maintaining robust, scalable, and secure telemetry data pipelines on Google Cloud Platform (GCP). The ideal candidate will have a strong background in DataOps principles, deep expertise in GCP data services, and a solid understanding of IT operations, especially within the security and network domains. You will enable real-time visibility and actionable insights for our security and network operations centers, contributing directly to our operational excellence and threat detection capabilities.
Responsibilities
β’ Lead the design, development, and implementation of high-performance, fault-tolerant telemetry data pipelines for ingesting, processing, and transforming large volumes of IT operational data (logs, metrics, traces) from diverse sources, with a focus on security and network telemetry.
β’ Architect and manage data solutions using a comprehensive suite of GCP services, ensuring optimal performance, cost-efficiency, and scalability. This includes leveraging services like Cloud Pub/Sub for messaging, Dataflow for real-time and batch processing, BigQuery for analytics, Cloud Logging for log management, and Cloud Monitoring for observability.
β’ Drive the adoption and implementation of DataOps best practices, including automation, CI/CD for data pipelines, version control (e.g., Git), automated testing, data quality checks, and robust monitoring and alerting.
β’ Develop specialized pipelines for critical security and network data sources such as VPC Flow Logs, firewall logs, intrusion detection system (IDS) logs, endpoint detection and response (EDR) data, and Security Information and Event Management (SIEM) data (e.g., Google Security Operations / Chronicle).
β’ Implement and enforce data governance, compliance, and security measures, including data encryption (at rest and in transit), access controls (RBAC), data masking, and audit logging to protect sensitive operational data.
β’ Continuously monitor, optimize, and troubleshoot data pipelines for performance, reliability, and cost-effectiveness, identifying and resolving bottlenecks.
Job Requirements
Details:
Experience Required
β’ Code Assessment, GCP, Data Architecture, Endpoint Security, Google Cloud Platform, Data Governance, Cloud Infrastructure, Extract Transform Load (Etl), Big Query, Network Security, Python
β’ Proven experience as a DataOps Engineer, Data Engineer, or similar role, with a strong focus on operationalizing data pipelines.
β’ Expertise in designing, building, and optimizing large-scale data pipelines for both batch and real-time processing.
β’ Strong understanding of DataOps principles, including CI/CD, automation, data quality, data governance, and monitoring.
β’ Proficiency in programming languages commonly used in data engineering, such as Python.
β’ Experience with Infrastructure as Code (IaC) tools (e.g., Terraform) for managing cloud resources.
β’ Solid understanding of data modeling, schema design, and data warehousing concepts (e.g., star schema).
Experience Preferred
β’ 8+ years of experience in data engineering, with at least 4 years in a Senior or Lead role focused on DataOps or cloud-native data platforms.
β’ Collaborate closely with IT operations, security analysts, network engineers, and other data stakeholders to understand data requirements and deliver solutions that meet business needs. Mentor junior engineers and contribute to the team's technical growth.
β’ Create and maintain comprehensive documentation for data pipelines, data models, and operational procedures
Education
β’ Bachelor's or Master's degree in Computer Science, Data Engineering, Information Technology, or a related quantitative field.
β’ Listed salary ranges may vary based on experience, qualifications, and local market. Also, some positions may include bonuses or other incentives
β’
β’
β’ Stefanini takes pride in hiring top talent and developing relationships with our future employees. Our talent acquisition teams will never make an offer of employment without having a phone conversation with you. Those face-to-face conversations will involve a description of the job for which you have applied. We also speak with you about the process, including interviews and job offers.
About Stefanini Group
The Stefanini Group is a global provider of offshore, onshore and near shore outsourcing, IT digital consulting, systems integration, application, and strategic staffing services to Fortune 1000 enterprises around the world. Our presence is in countries like the Americas, Europe, Africa, and Asia, and more than four hundred clients across a broad spectrum of markets, including financial services, manufacturing, telecommunications, chemical services, technology, public sector, and utilities. Stefanini is a CMM level 5, IT consulting company with a global presence. We are a CMM Level 5 company.