

Data Engineer - Databricks (Mid Level) - US Citizens/US Green Card Holders with 3 Years
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer - Databricks (Mid Level) with a contract length of over 6 months, offering a pay rate of $90,000 - $130,000 per year. Candidates must have 3+ years of experience, Databricks certification, and expertise in data migration, ETL processes, and AWS.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
590.9090909091
-
ποΈ - Date discovered
August 11, 2025
π - Project duration
More than 6 months
-
ποΈ - Location type
Remote
-
π - Contract type
Unknown
-
π - Security clearance
Yes
-
π - Location detailed
Remote
-
π§ - Skills detailed
#GIT #Data Reconciliation #PySpark #SQL (Structured Query Language) #Distributed Computing #Data Processing #RDS (Amazon Relational Database Service) #Compliance #Database Systems #"ETL (Extract #Transform #Load)" #Informatica PowerCenter #Data Migration #Computer Science #Programming #Spark (Apache Spark) #Azure ADLS (Azure Data Lake Storage) #Data Modeling #Databricks #Datasets #Scala #Data Storage #Informatica #Data Science #DynamoDB #Data Conversion #AWS (Amazon Web Services) #AWS S3 (Amazon Simple Storage Service) #Security #S3 (Amazon Simple Storage Service) #Data Ingestion #Version Control #Data Integration #ML (Machine Learning) #Azure #Storage #Oracle #Delta Lake #Documentation #Shell Scripting #Migration #ADLS (Azure Data Lake Storage) #Big Data #Python #Data Manipulation #Greenplum #Scripting #Apache Spark #Data Engineering #Kafka (Apache Kafka) #Data Lake #Cloud #DevOps
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
---- Project requirements mandate role open only for US Citizens/US Green Card Holders with a minimum of 3 years on Green Card . IRS MBI Clearance a plus/ Active Secret or Top Secret a Plus. All candidates will have to go through Clearance process before being able to start on the project.--(No exceptions to this requirement)
Job Description
Infobahn Solutions is hiring Databricks Data Engineering professionals in the Washington DC Metro Area for a US Government Federal Project with the Department of Treasury .
The Data Engineers will be part of a Data Migration & Conversion Team on a large DataLake being implemented on AWS Gov Cloud .
Data will be migrated from on premise Main Frame /Legacy database systems using Informatica PowerCenter to the AWS Landing Zone on S3.
Further conversion will be done using Databricks (PySpark) in AWS.
The Data Engineer should have prior Data Migration experience and understand all the intricacies required of developing data integration routines for moving data from multiple source systems to a new target system with a different data model.
The Data Engineer should have experience in converting Oracle PL/SQL and/or Greenplum code to Databricks.
Must have experience - Experience with Data Migrations and Conversion using Databricks .
Experience of using Databricks on AWS and managing a Databricks production system is critical and a must have for the project.
What youβll be doing:
Databricks Environment Setup: Configure and maintain Databricks clusters, ensuring optimal performance and scalability for big data processing and analytics.
ETL (Extract, Transform, Load): Design and implement ETL processes using Databricks notebooks or jobs to process and transform raw data into a usable format for analysis.
Data Lake Integration: Work with data lakes and data storage systems to efficiently manage and access large datasets within the Databricks environment.
Data Processing and Analysis: Develop and optimize Spark jobs for data processing, analysis, and machine learning tasks using Databricks notebooks.
Collaboration: Collaborate with data scientists, data engineers, and other stakeholders to understand business requirements and implement solutions.
Performance Tuning: Identify and address performance bottlenecks in Databricks jobs and clusters to optimize data processing speed and resource utilization.
Security and Compliance: Implement and enforce security measures to protect sensitive data within the Databricks environment, ensuring compliance with relevant regulations.
Documentation: Maintain documentation for Databricks workflows, configurations, and best practices to facilitate knowledge sharing and team collaboration.
Skills:
Apache Spark: Strong expertise in Apache Spark, which is the underlying distributed computing engine in Databricks.
Databricks Platform: In-depth knowledge of the Databricks platform, including its features, architecture, and administration.
Programming Languages: Proficiency in languages such as Python or Scala for developing Spark applications within Databricks.
SQL: Strong SQL skills for data manipulation, querying, and analysis within Databricks notebooks.
ETL Tools: Experience with ETL tools and frameworks for efficient data processing and transformation.
Data Lake and Storage: Familiarity with data lakes and storage systems, such as Delta Lake, AWS S3, or Azure Data Lake Storage.
Collaboration and Communication: Effective communication and collaboration skills to work with cross-functional teams and stakeholders.
Problem Solving: Strong problem-solving skills to troubleshoot issues and optimize Databricks workflows.
Version Control: Experience with version control systems (e.g., Git) for managing and tracking changes to Databricks notebooks and code.
Role Requirements:
Bachelor/Masterβs degree in computer science, Engineering, or related field
7-8 plus years of development experience on ETL tools (4+ years of Databricks is a must have)
5+ years of experience as a Databricks Engineer or similar role.
Strong expertise in Apache Spark and hands-on experience with Databricks.
More than 7 years of experience performing data reconciliation, data validation, ETL testing, deploying ETL packages and automating ETL jobs, developing reconciliation reports.
Working knowledge of message-oriented middleware/streaming data technologies such as Kafka, Confluent
Proficiency in programming languages such as Python or Scala for developing Spark applications.
Solid understanding of ETL processes and data modeling concepts.
Experience with data lakes and storage systems, such as Delta Lake, AWS S3, or Azure Data Lake Storage.
Strong SQL skills for data manipulation and analysis.
Good experience in shell scripting, AutoSys
Strong Data Modeling Skills
Strong analytical skills applied to business software solutions maintenance and/or development
Must be able to work with a team to write code, review code, and work on system operations.
Past project experience with Data Conversion and Data Migration
Communicate analysis, results and ideas to key decision makers including business and technical stakeholders.
Experience in developing and deploying data ingestion, processing, and distribution systems with AWS technologies
Experience with using AWS datastores, including RDS Postgres, S3, or DynamoDB
Dev-ops experience using GIT, developing, deploying code to production
Proficient in using AWS Cloud Services for Data Engineering tasks
Proficient in programming in Python/shell or other scripting languages for the purpose of data movement
Eligible for a US Government issued IRS MBI (candidates with active IRS MBIs will be preferred)
Preferred Qualifications
Cloud Data Migration and Conversion projects
Experience on AWS
Databricks industry certifications
Job Types: Full-time, Contract
Pay: $90,000.00 - $130,000.00 per year
Benefits:
Dental insurance
Flexible schedule
Health insurance
Life insurance
Paid time off
Vision insurance
Education:
Bachelor's (Preferred)
License/Certification:
Databricks Certified Data Engineer Professional (Required)
Security clearance:
Secret (Preferred)
Work Location: Remote