Tential Solutions

Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer with a contract length of "unknown," offering a pay rate of "unknown," and is fully remote. Key skills include Python, Java, SQL, AWS, Spark, and Databricks, with preferred experience in banking or financial services.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
Unknown
-
πŸ—“οΈ - Date
December 11, 2025
πŸ•’ - Duration
Unknown
-
🏝️ - Location
Remote
-
πŸ“„ - Contract
Unknown
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
District of Columbia, United States
-
🧠 - Skills detailed
#Cloud #Spark SQL #GIT #SQL (Structured Query Language) #Data Science #PySpark #Python #IAM (Identity and Access Management) #Version Control #Datasets #Code Reviews #Documentation #Complex Queries #Lambda (AWS Lambda) #Databricks #Data Analysis #Delta Lake #Data Processing #Data Engineering #Batch #Scala #Consulting #AWS (Amazon Web Services) #Kafka (Apache Kafka) #Spark (Apache Spark) #S3 (Amazon Simple Storage Service) #Data Modeling #Java #Data Pipeline #"ETL (Extract #Transform #Load)" #Airflow #BI (Business Intelligence)
Role description
Overview We’re partnering with a Big 4 consulting firm to add a Data Engineer to their team supporting a major banking and credit organization. This role focuses on building and optimizing scalable, cloud-based data pipelines using Python, Java, SQL, AWS, Spark, Databricks, and EMR. You’ll work across consulting and client teams to deliver reliable data solutions that power analytics, risk, and credit decisioning use cases. This position is fully remote. Responsibilities β€’ Design, build, and maintain scalable data pipelines and ETL/ELT processes using Python, Java, and SQL. β€’ Develop and optimize distributed data processing workloads using Spark (batch and/or streaming) on AWS. β€’ Build and manage data workflows on AWS, leveraging services such as EMR, S3, Lambda, Glue, and related components as appropriate. β€’ Use Databricks to develop, schedule, and monitor notebooks, jobs, and workflows supporting analytics and data products. β€’ Implement data models and structures that support banking/credit analytics, reporting, and downstream applications (e.g., risk, fraud, portfolio, customer insights). β€’ Monitor, troubleshoot, and tune pipeline performance, reliability, and cost in a production cloud environment. β€’ Collaborate with consultants, client stakeholders, data analysts, and data scientists to understand requirements and translate them into technical solutions. β€’ Apply best practices for code quality, testing, version control, and CI/CD within the data environment. β€’ Contribute to documentation, standards, and reusable components to improve consistency and speed across the data engineering team. Required Qualifications β€’ Strong hands-on experience with Python and Java for data engineering, ETL/ELT, or backend data services. β€’ Advanced SQL skills, including complex queries, performance tuning, and working with large, relational datasets. β€’ Production experience on AWS, ideally with services such as EMR, S3, Lambda, Glue, IAM, and CloudWatch. β€’ Practical experience building and optimizing Spark jobs (PySpark, Spark SQL, or Scala). β€’ Hands-on experience with Databricks (notebooks, clusters, jobs, and/or Delta Lake). β€’ Proven experience building and supporting reliable, performant data pipelines in a modern cloud environment. β€’ Solid understanding of data warehousing concepts, data modeling, and best practices for structured and semi-structured data. β€’ Experience working in collaborative engineering environments (Git, code reviews, branching strategies). β€’ Strong communication skills and comfort working in a consulting/client-facing environment. Preferred Qualifications (Nice To Have) β€’ Experience in banking, credit, financial services, or highly regulated environments. β€’ Background with streaming data (e.g., Spark Streaming, Kafka, Kinesis) and real-time or near–real-time data processing. β€’ Familiarity with orchestration tools (e.g., Airflow, Databricks jobs scheduler, Step Functions). β€’ Experience supporting analytics, BI, or data science teams (e.g., building curated datasets, feature stores, or semantic layers). #Remote