

Satguru Technosoft Consulting Group Inc. (STCG)
AWS Data Operations - DataOps/DataLake
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Operations Engineer on a long-term contract in SFO, CA. Key skills include cloud infrastructure management, data lifecycle management, and Pharma/Biotech data workflows. Requires 4+ years of experience, strong Unix/Linux skills, and proficiency in scripting languages.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
Unknown
-
ποΈ - Date
February 4, 2026
π - Duration
Unknown
-
ποΈ - Location
On-site
-
π - Contract
Unknown
-
π - Security
Unknown
-
π - Location detailed
San Francisco, CA
-
π§ - Skills detailed
#Linux #Compliance #Python #Data Quality #AWS (Amazon Web Services) #Migration #CLI (Command-Line Interface) #Bash #Data Management #Data Engineering #Unix #Data Lifecycle #DevOps #SQL (Structured Query Language) #Data Governance #DataOps #API (Application Programming Interface) #Deployment #S3 (Amazon Simple Storage Service) #Business Analysis #Security #Automation #Jira #Scala #Project Management #Scripting #Cloud #Storage #Data Science
Role description
Please go through the JD and share your resume.
Job Title: Data Operations Engineer
Duration : Contract long term
Location: SFO, CA (Onsite)
Need Cube Formation
About the Job
As a Data Operations Engineer, you will manage and optimize cloud-based data collections, ensuring the availability, performance, and security of data services and applications. You will focus on the entire data lifecycle. The ideal candidate will possess a deep understanding of cloud platforms, data management, and operational best practices.
Your problem-solving mindset will guide you in collaborating with domain experts, software engineers, and data engineers to create integrated solutions that enable scientific research and discovery while supporting data operations activities with increased data value and support reuse according to FAIR principles.
Key Responsibilities
Cloud Infrastructure Management
β’ Support the design, deployment, and management of scalable, reliable, and secure cloud infrastructure
β’ Monitor and report on cloud-based systems, including instances and storage services
β’ Optimize resource utilization and cost management
Data Lifecycle Management
β’ Oversee the end-to-end lifecycle of data from ingestion, processing, and storage to archiving and purging
β’ Co-develop and implement data governance practices to ensure data quality, consistency, and accessibility throughout its lifecycle
β’ Develop and maintain policies and procedures for data retention, archival, and disposal in compliance with regulatory requirements
Security and Compliance
β’ Implement and manage security policies and practices in accordance with industry standards
β’ Ensure compliance with data protection regulations and company policies
Performance Optimization
β’ Monitor system performance and implement tuning recommendations
β’ Perform root cause analysis and provide solutions for performance issues
β’ Perform data analytics to forecast data growth and anticipate future storage capacity requirements
β’ Continuously seek opportunities for process improvement and automation
Collaboration and Support
β’ Work closely with cross-functional teams including data scientists, software engineers, and business analysts
β’ Provide technical support and troubleshooting for data-related issues
β’ Collaborate with data engineers to build new capabilities and integrations aimed at increasing the value and reusability of scientific data
β’ Document processes, best practices, and operational guidelines
Required Qualifications
Successful candidates will meet many of the following requirements:
β’ Bachelorβs degree or equivalent practical experience
β’ 4+ years of experience supporting IT storage platforms and/or data systems
β’ Competent understanding of Pharma/Biotech scientific data workflows, needs, and requirements
β’ Competent understanding of Unix, NTFS, and SMB/Samba permission schemas and semantics
β’ Moderate expertise with scale-out network attached storage (NAS) solutions such as Dell EMC Isilon or NetApp
β’ Experience with data discovery tools such as Starfish, InsightIQ, S3 Storage Lens, or home-brewed solutions
β’ Experience with object storage platforms and protocols such as Amazon S3
β’ Strong technical Unix/Linux and project management background
β’ Strong scripting and dynamic language expertise: Bash, Python, SQL (preferred)
β’ Proficiency in software development with a strong focus on API integration and management
β’ Proficient in the use of Atlassian Jira and Confluence tools
β’ Proficient with Google Workspace tools (Docs, Sheets, Slides)
β’ Strong analytical and problem-solving skills
β’ Excellent communication and collaboration abilities
β’ High attention to detail and a proactive approach
Nice to Have / Plus Skills
Previous experience with any of the following is a plus:
β’ Isilon OneFS CLI and/or UI
β’ Parallel filesystem platforms such as IBM GPFS / Spectrum Scale
β’ Petabyte-scale data movement and migration
β’ Familiarity with CI/CD pipelines and DevOps practices related to API deployment and management
β’ AWS Cloud Solutions Architect Associate or equivalent AWS certifications
Thanks & Regards
Wilson Kadam
STCG Inc
Email: wilson@stcgusa.com
Direct Number :Β 609-250-9877
Please go through the JD and share your resume.
Job Title: Data Operations Engineer
Duration : Contract long term
Location: SFO, CA (Onsite)
Need Cube Formation
About the Job
As a Data Operations Engineer, you will manage and optimize cloud-based data collections, ensuring the availability, performance, and security of data services and applications. You will focus on the entire data lifecycle. The ideal candidate will possess a deep understanding of cloud platforms, data management, and operational best practices.
Your problem-solving mindset will guide you in collaborating with domain experts, software engineers, and data engineers to create integrated solutions that enable scientific research and discovery while supporting data operations activities with increased data value and support reuse according to FAIR principles.
Key Responsibilities
Cloud Infrastructure Management
β’ Support the design, deployment, and management of scalable, reliable, and secure cloud infrastructure
β’ Monitor and report on cloud-based systems, including instances and storage services
β’ Optimize resource utilization and cost management
Data Lifecycle Management
β’ Oversee the end-to-end lifecycle of data from ingestion, processing, and storage to archiving and purging
β’ Co-develop and implement data governance practices to ensure data quality, consistency, and accessibility throughout its lifecycle
β’ Develop and maintain policies and procedures for data retention, archival, and disposal in compliance with regulatory requirements
Security and Compliance
β’ Implement and manage security policies and practices in accordance with industry standards
β’ Ensure compliance with data protection regulations and company policies
Performance Optimization
β’ Monitor system performance and implement tuning recommendations
β’ Perform root cause analysis and provide solutions for performance issues
β’ Perform data analytics to forecast data growth and anticipate future storage capacity requirements
β’ Continuously seek opportunities for process improvement and automation
Collaboration and Support
β’ Work closely with cross-functional teams including data scientists, software engineers, and business analysts
β’ Provide technical support and troubleshooting for data-related issues
β’ Collaborate with data engineers to build new capabilities and integrations aimed at increasing the value and reusability of scientific data
β’ Document processes, best practices, and operational guidelines
Required Qualifications
Successful candidates will meet many of the following requirements:
β’ Bachelorβs degree or equivalent practical experience
β’ 4+ years of experience supporting IT storage platforms and/or data systems
β’ Competent understanding of Pharma/Biotech scientific data workflows, needs, and requirements
β’ Competent understanding of Unix, NTFS, and SMB/Samba permission schemas and semantics
β’ Moderate expertise with scale-out network attached storage (NAS) solutions such as Dell EMC Isilon or NetApp
β’ Experience with data discovery tools such as Starfish, InsightIQ, S3 Storage Lens, or home-brewed solutions
β’ Experience with object storage platforms and protocols such as Amazon S3
β’ Strong technical Unix/Linux and project management background
β’ Strong scripting and dynamic language expertise: Bash, Python, SQL (preferred)
β’ Proficiency in software development with a strong focus on API integration and management
β’ Proficient in the use of Atlassian Jira and Confluence tools
β’ Proficient with Google Workspace tools (Docs, Sheets, Slides)
β’ Strong analytical and problem-solving skills
β’ Excellent communication and collaboration abilities
β’ High attention to detail and a proactive approach
Nice to Have / Plus Skills
Previous experience with any of the following is a plus:
β’ Isilon OneFS CLI and/or UI
β’ Parallel filesystem platforms such as IBM GPFS / Spectrum Scale
β’ Petabyte-scale data movement and migration
β’ Familiarity with CI/CD pipelines and DevOps practices related to API deployment and management
β’ AWS Cloud Solutions Architect Associate or equivalent AWS certifications
Thanks & Regards
Wilson Kadam
STCG Inc
Email: wilson@stcgusa.com
Direct Number :Β 609-250-9877






