

SGS Consulting
Data Engineer
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer on a contract basis, focusing on a data platform migration from Salesforce. Requires 3+ years of experience, proficiency in SQL, ETL pipelines, and real-time data systems. Location: "Remote". Pay rate: "$/hour".
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
680
-
🗓️ - Date
May 15, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
Unknown
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
United States
-
🧠 - Skills detailed
#Monitoring #Presto #Databricks #SaaS (Software as a Service) #SQL (Structured Query Language) #Data Migration #Snowflake #Data Engineering #MySQL #CRM (Customer Relationship Management) #GraphQL #Schema Design #GitHub #Data Quality #PHP #BigQuery #Trino #"ETL (Extract #Transform #Load)" #Spark (Apache Spark) #Kafka (Apache Kafka) #Data Pipeline #Data Management #Data Warehouse #Computer Science #Batch #AI (Artificial Intelligence) #Data Architecture #Migration #Data Documentation #Data Ingestion #Documentation
Role description
Job Description
• The Data Engineer will support a strategic data platform migration initiative, building the data pipelines and warehouse infrastructure required to transition business-critical operations from a third-party SaaS CRM (Salesforce) onto an internal entity-modeled data platform.
• The engineer will design and maintain bidirectional ETL flows between Salesforce, distributed data warehouses, and operational entity stores, ensuring data consistency, freshness, and auditability throughout the migration.
• Work spans pipeline development, schema design, data quality engineering, and orchestration of scheduled batch and streaming jobs.
Job Responsibilities
• Design and build bidirectional data pipelines between Salesforce, distributed data, and internal operational entity-modeled data stores.
• Build streaming data pipelines using distributed event-streaming systems (Kafka or equivalent pub/sub log architectures) for real-time data propagation.
• Design and maintain dimensional and entity schemas in the internal data platform, including access permissions, mutation rules.
• Build data validation, reconciliation, and quality monitoring frameworks to ensure parity between legacy and replacement systems during transition.
• Develop and maintain data documentation including schema definitions, transformation logic, source-to-target mappings, and data dictionaries.
• Use AI-assisted development tools (e.g., Claude Code, Cursor, Copilot) as a core part of the development workflow to accelerate pipeline development, SQL authoring, and schema migrations.
• Partner with software engineers, Salesforce administrators, and business stakeholders to define data contracts, SLAs, and migration cutover criteria.
• Investigate and resolve data quality incidents, pipeline failures, and reconciliation discrepancies.
Skills
• Experience with Hack/PHP coding and GraphQL.
• Strong proficiency in SQL.
• Comfort with ORM frameworks.
• Hands-on experience building and operating large-scale ETL/ELT pipelines in distributed data warehouse environments (Hive, Spark, Presto/Trino, BigQuery, Snowflake, or Databricks).
• Familiarity with distributed event-streaming systems (Kafka, Pulsar, Kinesis, or equivalent) for real-time data ingestion.
• Hands-on experience using AI coding assistants (Claude Code, Cursor, GitHub Copilot, etc.) as part of a daily development workflow.
• Strong understanding of data quality engineering -- testing, validation, monitoring, and reconciliation patterns.
• Experience with data migrations between heterogeneous systems is highly desirable.
• Verbal and written communication skills, problem solving skills, customer service and interpersonal skills.
• Strong ability to work independently and manage one's time.
• Strong ability to troubleshoot data issues and make system changes as needed to resolve them.
Education/Experience
• Bachelor's degree in computer science, data engineering, information systems, or relevant field required.
• 3+ years of professional data engineering experience preferred.
Typical Day-to-Day in the role
• Perform data migration of all tables (approximately 80 objects of data) from Salesforce to the internal data platform.
• Clean up and restructure data to improve organization and usability.
• Define and implement data architecture decisions for how data should be structured in the new system.
• Prepare and transform data to be AI-ready for downstream tooling and applications.
• Build out and maintain the data management layer using internal tools including Scribe (Client's internal Kafka equivalent).
Must-Have Skills
• Entity/schema modeling and GraphQL.
• ORM-framework & MySQL.
• Real time data propagation (Kafka, Pulsar, Kinesis, or equivalent).
Nice-to-have Skills
• Salesforce Knowledge.
• AI development workflows.
• Events and Subscription handling (Iris).
Job Description
• The Data Engineer will support a strategic data platform migration initiative, building the data pipelines and warehouse infrastructure required to transition business-critical operations from a third-party SaaS CRM (Salesforce) onto an internal entity-modeled data platform.
• The engineer will design and maintain bidirectional ETL flows between Salesforce, distributed data warehouses, and operational entity stores, ensuring data consistency, freshness, and auditability throughout the migration.
• Work spans pipeline development, schema design, data quality engineering, and orchestration of scheduled batch and streaming jobs.
Job Responsibilities
• Design and build bidirectional data pipelines between Salesforce, distributed data, and internal operational entity-modeled data stores.
• Build streaming data pipelines using distributed event-streaming systems (Kafka or equivalent pub/sub log architectures) for real-time data propagation.
• Design and maintain dimensional and entity schemas in the internal data platform, including access permissions, mutation rules.
• Build data validation, reconciliation, and quality monitoring frameworks to ensure parity between legacy and replacement systems during transition.
• Develop and maintain data documentation including schema definitions, transformation logic, source-to-target mappings, and data dictionaries.
• Use AI-assisted development tools (e.g., Claude Code, Cursor, Copilot) as a core part of the development workflow to accelerate pipeline development, SQL authoring, and schema migrations.
• Partner with software engineers, Salesforce administrators, and business stakeholders to define data contracts, SLAs, and migration cutover criteria.
• Investigate and resolve data quality incidents, pipeline failures, and reconciliation discrepancies.
Skills
• Experience with Hack/PHP coding and GraphQL.
• Strong proficiency in SQL.
• Comfort with ORM frameworks.
• Hands-on experience building and operating large-scale ETL/ELT pipelines in distributed data warehouse environments (Hive, Spark, Presto/Trino, BigQuery, Snowflake, or Databricks).
• Familiarity with distributed event-streaming systems (Kafka, Pulsar, Kinesis, or equivalent) for real-time data ingestion.
• Hands-on experience using AI coding assistants (Claude Code, Cursor, GitHub Copilot, etc.) as part of a daily development workflow.
• Strong understanding of data quality engineering -- testing, validation, monitoring, and reconciliation patterns.
• Experience with data migrations between heterogeneous systems is highly desirable.
• Verbal and written communication skills, problem solving skills, customer service and interpersonal skills.
• Strong ability to work independently and manage one's time.
• Strong ability to troubleshoot data issues and make system changes as needed to resolve them.
Education/Experience
• Bachelor's degree in computer science, data engineering, information systems, or relevant field required.
• 3+ years of professional data engineering experience preferred.
Typical Day-to-Day in the role
• Perform data migration of all tables (approximately 80 objects of data) from Salesforce to the internal data platform.
• Clean up and restructure data to improve organization and usability.
• Define and implement data architecture decisions for how data should be structured in the new system.
• Prepare and transform data to be AI-ready for downstream tooling and applications.
• Build out and maintain the data management layer using internal tools including Scribe (Client's internal Kafka equivalent).
Must-Have Skills
• Entity/schema modeling and GraphQL.
• ORM-framework & MySQL.
• Real time data propagation (Kafka, Pulsar, Kinesis, or equivalent).
Nice-to-have Skills
• Salesforce Knowledge.
• AI development workflows.
• Events and Subscription handling (Iris).






