

Infobahn Solutions Inc.
Data Engineer - Databricks (Mid Level) - US Citizens Only
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer - Databricks (Mid Level) with a contract length of over 6 months, offering a pay rate of $90,000 - $130,000 per year. Remote work is available. Key skills include Databricks, Apache Spark, and ETL processes. A Bachelor's degree and Databricks certification are required, along with 7+ years of relevant experience.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
590
-
ποΈ - Date
January 7, 2026
π - Duration
More than 6 months
-
ποΈ - Location
Remote
-
π - Contract
Unknown
-
π - Security
Yes
-
π - Location detailed
Remote
-
π§ - Skills detailed
#DynamoDB #Cloud #ADLS (Azure Data Lake Storage) #Computer Science #Compliance #Data Science #Apache Spark #Database Systems #PySpark #Data Conversion #Greenplum #Data Lake #Python #Scripting #Oracle #Distributed Computing #S3 (Amazon Simple Storage Service) #Data Manipulation #Programming #Documentation #SQL (Structured Query Language) #Storage #Version Control #Data Processing #Data Migration #AWS (Amazon Web Services) #Data Storage #Informatica PowerCenter #"ETL (Extract #Transform #Load)" #Data Integration #Scala #ML (Machine Learning) #Spark (Apache Spark) #Data Modeling #Databricks #Shell Scripting #GIT #DevOps #RDS (Amazon Relational Database Service) #Data Ingestion #Informatica #AWS S3 (Amazon Simple Storage Service) #Security #Data Engineering #Azure #Azure ADLS (Azure Data Lake Storage) #Migration #Datasets #Delta Lake #Big Data #Data Reconciliation #Kafka (Apache Kafka)
Role description
---- Project requirements mandate role open only for US Citizens. IRS MBI Clearance a plus/ Active Secret or Top Secret a Plus. All candidates will have to go through Clearance process before being able to start on the project.--(No exceptions to this requirement)
Job Description
Infobahn Solutions is hiring Databricks Data Engineering professionals in the Washington DC Metro Area for a US Government Federal Project with the Department of Treasury .
The Data Engineers will be part of a Data Migration & Conversion Team on a large DataLake being implemented on AWS Gov Cloud .
Data will be migrated from on premise Main Frame /Legacy database systems using Informatica PowerCenter to the AWS Landing Zone on S3.
Further conversion will be done using Databricks (PySpark) in AWS.
The Data Engineer should have prior Data Migration experience and understand all the intricacies required of developing data integration routines for moving data from multiple source systems to a new target system with a different data model.
The Data Engineer should have experience in converting Oracle PL/SQL and/or Greenplum code to Databricks.
Must have experience - Experience with Data Migrations and Conversion using Databricks .
Experience of using Databricks on AWS and managing a Databricks production system is critical and a must have for the project.
What youβll be doing:
Databricks Environment Setup: Configure and maintain Databricks clusters, ensuring optimal performance and scalability for big data processing and analytics.
ETL (Extract, Transform, Load): Design and implement ETL processes using Databricks notebooks or jobs to process and transform raw data into a usable format for analysis.
Data Lake Integration: Work with data lakes and data storage systems to efficiently manage and access large datasets within the Databricks environment.
Data Processing and Analysis: Develop and optimize Spark jobs for data processing, analysis, and machine learning tasks using Databricks notebooks.
Collaboration: Collaborate with data scientists, data engineers, and other stakeholders to understand business requirements and implement solutions.
Performance Tuning: Identify and address performance bottlenecks in Databricks jobs and clusters to optimize data processing speed and resource utilization.
Security and Compliance: Implement and enforce security measures to protect sensitive data within the Databricks environment, ensuring compliance with relevant regulations.
Documentation: Maintain documentation for Databricks workflows, configurations, and best practices to facilitate knowledge sharing and team collaboration.
Skills:
Apache Spark: Strong expertise in Apache Spark, which is the underlying distributed computing engine in Databricks.
Databricks Platform: In-depth knowledge of the Databricks platform, including its features, architecture, and administration.
Programming Languages: Proficiency in languages such as Python or Scala for developing Spark applications within Databricks.
SQL: Strong SQL skills for data manipulation, querying, and analysis within Databricks notebooks.
ETL Tools: Experience with ETL tools and frameworks for efficient data processing and transformation.
Data Lake and Storage: Familiarity with data lakes and storage systems, such as Delta Lake, AWS S3, or Azure Data Lake Storage.
Collaboration and Communication: Effective communication and collaboration skills to work with cross-functional teams and stakeholders.
Problem Solving: Strong problem-solving skills to troubleshoot issues and optimize Databricks workflows.
Version Control: Experience with version control systems (e.g., Git) for managing and tracking changes to Databricks notebooks and code.
Role Requirements:
Bachelor/Masterβs degree in computer science, Engineering, or related field
7-8 plus years of development experience on ETL tools (4+ years of Databricks is a must have)
5+ years of experience as a Databricks Engineer or similar role.
Strong expertise in Apache Spark and hands-on experience with Databricks.
More than 7 years of experience performing data reconciliation, data validation, ETL testing, deploying ETL packages and automating ETL jobs, developing reconciliation reports.
Working knowledge of message-oriented middleware/streaming data technologies such as Kafka, Confluent
Proficiency in programming languages such as Python or Scala for developing Spark applications.
Solid understanding of ETL processes and data modeling concepts.
Experience with data lakes and storage systems, such as Delta Lake, AWS S3, or Azure Data Lake Storage.
Strong SQL skills for data manipulation and analysis.
Good experience in shell scripting, AutoSys
Strong Data Modeling Skills
Strong analytical skills applied to business software solutions maintenance and/or development
Must be able to work with a team to write code, review code, and work on system operations.
Past project experience with Data Conversion and Data Migration
Communicate analysis, results and ideas to key decision makers including business and technical stakeholders.
Experience in developing and deploying data ingestion, processing, and distribution systems with AWS technologies
Experience with using AWS datastores, including RDS Postgres, S3, or DynamoDB
Dev-ops experience using GIT, developing, deploying code to production
Proficient in using AWS Cloud Services for Data Engineering tasks
Proficient in programming in Python/shell or other scripting languages for the purpose of data movement
Eligible for a US Government issued IRS MBI (candidates with active IRS MBIs will be preferred)
Databricks industry certifications - Associate / Professional Level
Preferred Qualifications
Cloud Data Migration and Conversion projects
Experience on AWS
Job Types: Full-time, Contract
Pay: $90,000.00 - $130,000.00 per year
Benefits:
Dental insurance
Flexible schedule
Health insurance
Life insurance
Paid time off
Vision insurance
Education:
Bachelor's (Preferred)
License/Certification:
Databricks Certified Data Engineer Professional (Required)
Security clearance:
Secret (Preferred)
Work Location: Remote
---- Project requirements mandate role open only for US Citizens. IRS MBI Clearance a plus/ Active Secret or Top Secret a Plus. All candidates will have to go through Clearance process before being able to start on the project.--(No exceptions to this requirement)
Job Description
Infobahn Solutions is hiring Databricks Data Engineering professionals in the Washington DC Metro Area for a US Government Federal Project with the Department of Treasury .
The Data Engineers will be part of a Data Migration & Conversion Team on a large DataLake being implemented on AWS Gov Cloud .
Data will be migrated from on premise Main Frame /Legacy database systems using Informatica PowerCenter to the AWS Landing Zone on S3.
Further conversion will be done using Databricks (PySpark) in AWS.
The Data Engineer should have prior Data Migration experience and understand all the intricacies required of developing data integration routines for moving data from multiple source systems to a new target system with a different data model.
The Data Engineer should have experience in converting Oracle PL/SQL and/or Greenplum code to Databricks.
Must have experience - Experience with Data Migrations and Conversion using Databricks .
Experience of using Databricks on AWS and managing a Databricks production system is critical and a must have for the project.
What youβll be doing:
Databricks Environment Setup: Configure and maintain Databricks clusters, ensuring optimal performance and scalability for big data processing and analytics.
ETL (Extract, Transform, Load): Design and implement ETL processes using Databricks notebooks or jobs to process and transform raw data into a usable format for analysis.
Data Lake Integration: Work with data lakes and data storage systems to efficiently manage and access large datasets within the Databricks environment.
Data Processing and Analysis: Develop and optimize Spark jobs for data processing, analysis, and machine learning tasks using Databricks notebooks.
Collaboration: Collaborate with data scientists, data engineers, and other stakeholders to understand business requirements and implement solutions.
Performance Tuning: Identify and address performance bottlenecks in Databricks jobs and clusters to optimize data processing speed and resource utilization.
Security and Compliance: Implement and enforce security measures to protect sensitive data within the Databricks environment, ensuring compliance with relevant regulations.
Documentation: Maintain documentation for Databricks workflows, configurations, and best practices to facilitate knowledge sharing and team collaboration.
Skills:
Apache Spark: Strong expertise in Apache Spark, which is the underlying distributed computing engine in Databricks.
Databricks Platform: In-depth knowledge of the Databricks platform, including its features, architecture, and administration.
Programming Languages: Proficiency in languages such as Python or Scala for developing Spark applications within Databricks.
SQL: Strong SQL skills for data manipulation, querying, and analysis within Databricks notebooks.
ETL Tools: Experience with ETL tools and frameworks for efficient data processing and transformation.
Data Lake and Storage: Familiarity with data lakes and storage systems, such as Delta Lake, AWS S3, or Azure Data Lake Storage.
Collaboration and Communication: Effective communication and collaboration skills to work with cross-functional teams and stakeholders.
Problem Solving: Strong problem-solving skills to troubleshoot issues and optimize Databricks workflows.
Version Control: Experience with version control systems (e.g., Git) for managing and tracking changes to Databricks notebooks and code.
Role Requirements:
Bachelor/Masterβs degree in computer science, Engineering, or related field
7-8 plus years of development experience on ETL tools (4+ years of Databricks is a must have)
5+ years of experience as a Databricks Engineer or similar role.
Strong expertise in Apache Spark and hands-on experience with Databricks.
More than 7 years of experience performing data reconciliation, data validation, ETL testing, deploying ETL packages and automating ETL jobs, developing reconciliation reports.
Working knowledge of message-oriented middleware/streaming data technologies such as Kafka, Confluent
Proficiency in programming languages such as Python or Scala for developing Spark applications.
Solid understanding of ETL processes and data modeling concepts.
Experience with data lakes and storage systems, such as Delta Lake, AWS S3, or Azure Data Lake Storage.
Strong SQL skills for data manipulation and analysis.
Good experience in shell scripting, AutoSys
Strong Data Modeling Skills
Strong analytical skills applied to business software solutions maintenance and/or development
Must be able to work with a team to write code, review code, and work on system operations.
Past project experience with Data Conversion and Data Migration
Communicate analysis, results and ideas to key decision makers including business and technical stakeholders.
Experience in developing and deploying data ingestion, processing, and distribution systems with AWS technologies
Experience with using AWS datastores, including RDS Postgres, S3, or DynamoDB
Dev-ops experience using GIT, developing, deploying code to production
Proficient in using AWS Cloud Services for Data Engineering tasks
Proficient in programming in Python/shell or other scripting languages for the purpose of data movement
Eligible for a US Government issued IRS MBI (candidates with active IRS MBIs will be preferred)
Databricks industry certifications - Associate / Professional Level
Preferred Qualifications
Cloud Data Migration and Conversion projects
Experience on AWS
Job Types: Full-time, Contract
Pay: $90,000.00 - $130,000.00 per year
Benefits:
Dental insurance
Flexible schedule
Health insurance
Life insurance
Paid time off
Vision insurance
Education:
Bachelor's (Preferred)
License/Certification:
Databricks Certified Data Engineer Professional (Required)
Security clearance:
Secret (Preferred)
Work Location: Remote






