

Infojini Inc
Data Engineer - III
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer - III in New York, NY, with a contract length of "unknown" and a pay rate of "unknown." Key skills include Databricks ETL, AWS services, SQL, Python, and cloud infrastructure. Computer Science degree required.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
November 5, 2025
🕒 - Duration
Unknown
-
🏝️ - Location
On-site
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
New York, NY
-
🧠 - Skills detailed
#Data Lake #SaaS (Software as a Service) #Data Integration #AWS Databases #Data Ingestion #Automation #Databricks #Scala #SQL (Structured Query Language) #GitLab #Programming #Agile #Data Architecture #DynamoDB #"ETL (Extract #Transform #Load)" #PySpark #Python #Storage #Cloud #Data Quality #Terraform #Apache Spark #Code Reviews #Spark (Apache Spark) #Debugging #Data Pipeline #EC2 #Security #Deployment #Data Analysis #S3 (Amazon Simple Storage Service) #Monitoring #SQL Queries #AWS (Amazon Web Services) #Docker #RDS (Amazon Relational Database Service) #Trino #Computer Science #Data Engineering #Kubernetes #Databases #Data Processing
Role description
Role- Data Engineer - III
Location-New York, NY, 10038
Qualifications
What we are looking for:
• Hands-on experience in building ETL using Databricks SaaS infrastructure.
• Experience in developing data pipeline solutions to ingest and exploit new and existing data sources.
• Expertise in leveraging SQL, programming language like Python and ETL tools like Databricks
• Perform code reviews to ensure requirements, optimal execution patterns and adherence to established standards. Computer Science or Equivalent
• Expertise in AWS Compute (EC2, EMR), AWS Storage (S3, EBS), AWS Databases (RDS, DynamoDB), AWS Data Integration (Glue).
• Advanced understanding of Container Orchestration services including Docker and Kubernetes, and a variety of AWS tools and services.
• Good understanding of AWS Identify and Access management, AWS Networking and AWS Monitoring tools.
• Proficiency in CI/CD and deployment automation using GITLAB pipeline.
• Proficiency in Cloud infrastructure provisioning tools e.g., Terraform.
• Proficiency in one or more programming languages e.g., Python, Scala.
• Experience in Starburst, Trino and building SQL queries in federated architecture.
• Good knowledge of Lake house architecture.
• Design, develop, and optimize scalable ETL/ELT pipelines using Databricks and Apache Spark (PySpark and Scala).
• Build data ingestion workflows from various sources (structured, semi-structured, and unstructured).
• Develop reusable components and frameworks for efficient data processing.
• Implement best practices for data quality, validation, and governance.
• Collaborate with data architects, analysts, and business stakeholders to understand data requirements.
• Tune Spark jobs for performance and scalability in a cloud-based environment.
• Maintain robust data lake or Lakehouse architecture.
• Ensure high availability, security, and integrity of data pipelines and platforms.
• Support troubleshooting, debugging, and performance optimization in production workloads.
Responsibilities
Your role as a Senior Data Engineer
• Work on migrating applications from an on-premises location to the cloud service providers.
• Develop products and services on the latest technologies through contributions in development, enhancements, testing and implementation.
• Develop, modify, extend code for building cloud infrastructure, and automate using CI/CD pipeline.
• Partners with business and peers in the pursuit of solutions that achieve business goals through an agile software development methodology.
• Perform problem analysis, data analysis, reporting, and communication.
• Work with peers across the system to define and implement best practices and standards.
• Assess applications and help determine the appropriate application infrastructure patterns.
• Use the best practices and knowledge of internal or external drivers to improve products or services.
Role- Data Engineer - III
Location-New York, NY, 10038
Qualifications
What we are looking for:
• Hands-on experience in building ETL using Databricks SaaS infrastructure.
• Experience in developing data pipeline solutions to ingest and exploit new and existing data sources.
• Expertise in leveraging SQL, programming language like Python and ETL tools like Databricks
• Perform code reviews to ensure requirements, optimal execution patterns and adherence to established standards. Computer Science or Equivalent
• Expertise in AWS Compute (EC2, EMR), AWS Storage (S3, EBS), AWS Databases (RDS, DynamoDB), AWS Data Integration (Glue).
• Advanced understanding of Container Orchestration services including Docker and Kubernetes, and a variety of AWS tools and services.
• Good understanding of AWS Identify and Access management, AWS Networking and AWS Monitoring tools.
• Proficiency in CI/CD and deployment automation using GITLAB pipeline.
• Proficiency in Cloud infrastructure provisioning tools e.g., Terraform.
• Proficiency in one or more programming languages e.g., Python, Scala.
• Experience in Starburst, Trino and building SQL queries in federated architecture.
• Good knowledge of Lake house architecture.
• Design, develop, and optimize scalable ETL/ELT pipelines using Databricks and Apache Spark (PySpark and Scala).
• Build data ingestion workflows from various sources (structured, semi-structured, and unstructured).
• Develop reusable components and frameworks for efficient data processing.
• Implement best practices for data quality, validation, and governance.
• Collaborate with data architects, analysts, and business stakeholders to understand data requirements.
• Tune Spark jobs for performance and scalability in a cloud-based environment.
• Maintain robust data lake or Lakehouse architecture.
• Ensure high availability, security, and integrity of data pipelines and platforms.
• Support troubleshooting, debugging, and performance optimization in production workloads.
Responsibilities
Your role as a Senior Data Engineer
• Work on migrating applications from an on-premises location to the cloud service providers.
• Develop products and services on the latest technologies through contributions in development, enhancements, testing and implementation.
• Develop, modify, extend code for building cloud infrastructure, and automate using CI/CD pipeline.
• Partners with business and peers in the pursuit of solutions that achieve business goals through an agile software development methodology.
• Perform problem analysis, data analysis, reporting, and communication.
• Work with peers across the system to define and implement best practices and standards.
• Assess applications and help determine the appropriate application infrastructure patterns.
• Use the best practices and knowledge of internal or external drivers to improve products or services.






