

Lead Pyspark/Iceberg Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Lead Pyspark/Iceberg Engineer in the Banking/Financial industry, based in Charlotte, NC. The contract lasts 24+ months with a competitive pay rate. Key skills include 5+ years in SQL and 3+ years in PySpark, Iceberg, and Airflow.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
680
-
ποΈ - Date discovered
August 8, 2025
π - Project duration
More than 6 months
-
ποΈ - Location type
Hybrid
-
π - Contract type
Unknown
-
π - Security clearance
Unknown
-
π - Location detailed
Charlotte, NC
-
π§ - Skills detailed
#Web Services #Java #"ETL (Extract #Transform #Load)" #GIT #Scala #Grafana #Data Modeling #Kafka (Apache Kafka) #SQL (Structured Query Language) #Python #Ab Initio #Batch #Airflow #Compliance #Data Pipeline #Kubernetes #Agile #S3 (Amazon Simple Storage Service) #PySpark #Prometheus #Visual Studio #Docker #Data Engineering #Scrum #Deployment #Documentation #Migration #Monitoring #Data Ingestion #Spark (Apache Spark) #Jira #GitHub #Kanban #Data Governance
Role description
Please find details for this position below:
Client: Banking/Financial Industry
Title: Lead Pyspark/Iceberg Engineer / Lead Data Engineer - Pyspark/Iceberg
Location: Charlotte, NC β Hybrid Roles (In Person Interview)
Duration: 24+ Month (s) Extend or Convert based on performances
Job Descriptions:
β’ We are seeking a highly skilled and adaptable Lead Software Engineer to join our Counterparty Credit Risk organization.
β’ This role supports: Data Services, with a focus on modernizing legacy systems, managing high-volume data pipelines, and contributing to full-stack application development.
β’ You will a team member on business-as-usual (BAU) processes, while also contributing to the development of a new data platform over the next two years. The ideal candidate is a strong communicator, a proactive problem-solver, and comfortable working in a Kanban Agile environment.
Key Responsibilities:
β’ Lead Agile Development: Guide and support multiple Agile teams focused on data Extract , Ingestion, and transformation.
β’ Modernize Legacy Systems: Migrate data pipelines from Ab Initio and Filesystem to modern technologies such as PySpark, S3, Airflow, Parquet, and Iceberg.
β’ Full-Stack Engineering: Design and develop scalable backend services using PySpark and Python.
β’ Data Platform Enablement: Support ingestion of 300+ data feeds into the platform to ensure timely nightly batch processing.
β’ Cross-Functional Collaboration: Partner with business stakeholders and product owners to understand requirements and deliver effective solutions.
β’ Agile Execution: Working with both Kanban and scrum teams and should be familiar with both and check-ins and managing tasks via Jira.
β’ Platform Transition Support: Contribute to the migration from legacy systems to a new data platform over the next two years.
β’ BAU and Strategic Support: Balance business-as-usual responsibilities while contributing to long-term platform modernization.
β’ Documentation and Data Modeling: Maintain clear technical documentation and demonstrate a strong understanding of columnar data structures.
β’ Experience on the different file format systems (Parquet, ORC, AVRO).
β’ Experience on the code containerization deployments using Docker / Kubernetes.
β’ Java background would be a plus.
β’ Good Knowledge on large scale ETL Based frameworks
β’ Experience on ETL tool (AbInitio).
Top Skills:
β’ 5 years of experience as an engineer
β’ 5 years of SQL engineering
β’ 2-3 years Pyspark
β’ 2-3 years Iceberg
β’ 2-3 years Parquet
β’ S3
β’ Airflow
β’ Application is legacy Abinitio, transitioning to Python and Pyspark. Also replacing Autosys with Airflow for job scheduling. Trying to upscale to growing demand for data with the asset cap lifting. This person will be working on the new project after first learning the existing platform. On a team with 7-8 onshore engineers and 6-7 offshore.
β’ Will do heads down engineering but needs to be able to interact with the existing team and coordinate with them on needs when implementing the new system. Will by doing daily scrum calls, reviewing work orders in order.
β’ Agile is not necessary but good to know.
β’ Banking background is good but not a must.
β’ Certifications are a plus.
Required Skills & Experience
Top Technical Skills
β’ 3+ years of experience with PySpark, S3, Iceberg, Git, Python, Airflow, and Parquet
β’ 5+ years of experience with SQL
β’ Experience with Agile methodologies and tools like Jira
β’ Familiarity with Kafka
β’ Experience with GitHub Copilot, Web Services, Visual studio, IntelliJ, and Gradle
β’ Experience in monitoring tools like Grafana or Prometheus
Preferred Qualifications
β’ Proven experience leading Agile teams and mentoring junior developers
β’ Strong communication skills and the ability to collaborate with business stakeholders
β’ Comfortable working in both Scrum and Kanban model with frequent scrum check-ins
β’ Ability to identify blockers and proactively seek help when needed
β’ Experience working in a regulated environment with a focus on compliance and data governance.
β’ 2+ years of working with Ab Initio graphs and plans
Team Structure & Projects
β’ You will be part of a team that handles over 300+ data feeds, ensuring timely ingestion for nightly batch processing.
β’ Role will focus on Data Services, modernizing data ingestion pipelines.
EEO:
Mindlance is an Equal Opportunity Employer and does not discriminate in employment on the basis of β Minority/Gender/Disability/Religion/LGBTQI/Age/Veterans.
Please find details for this position below:
Client: Banking/Financial Industry
Title: Lead Pyspark/Iceberg Engineer / Lead Data Engineer - Pyspark/Iceberg
Location: Charlotte, NC β Hybrid Roles (In Person Interview)
Duration: 24+ Month (s) Extend or Convert based on performances
Job Descriptions:
β’ We are seeking a highly skilled and adaptable Lead Software Engineer to join our Counterparty Credit Risk organization.
β’ This role supports: Data Services, with a focus on modernizing legacy systems, managing high-volume data pipelines, and contributing to full-stack application development.
β’ You will a team member on business-as-usual (BAU) processes, while also contributing to the development of a new data platform over the next two years. The ideal candidate is a strong communicator, a proactive problem-solver, and comfortable working in a Kanban Agile environment.
Key Responsibilities:
β’ Lead Agile Development: Guide and support multiple Agile teams focused on data Extract , Ingestion, and transformation.
β’ Modernize Legacy Systems: Migrate data pipelines from Ab Initio and Filesystem to modern technologies such as PySpark, S3, Airflow, Parquet, and Iceberg.
β’ Full-Stack Engineering: Design and develop scalable backend services using PySpark and Python.
β’ Data Platform Enablement: Support ingestion of 300+ data feeds into the platform to ensure timely nightly batch processing.
β’ Cross-Functional Collaboration: Partner with business stakeholders and product owners to understand requirements and deliver effective solutions.
β’ Agile Execution: Working with both Kanban and scrum teams and should be familiar with both and check-ins and managing tasks via Jira.
β’ Platform Transition Support: Contribute to the migration from legacy systems to a new data platform over the next two years.
β’ BAU and Strategic Support: Balance business-as-usual responsibilities while contributing to long-term platform modernization.
β’ Documentation and Data Modeling: Maintain clear technical documentation and demonstrate a strong understanding of columnar data structures.
β’ Experience on the different file format systems (Parquet, ORC, AVRO).
β’ Experience on the code containerization deployments using Docker / Kubernetes.
β’ Java background would be a plus.
β’ Good Knowledge on large scale ETL Based frameworks
β’ Experience on ETL tool (AbInitio).
Top Skills:
β’ 5 years of experience as an engineer
β’ 5 years of SQL engineering
β’ 2-3 years Pyspark
β’ 2-3 years Iceberg
β’ 2-3 years Parquet
β’ S3
β’ Airflow
β’ Application is legacy Abinitio, transitioning to Python and Pyspark. Also replacing Autosys with Airflow for job scheduling. Trying to upscale to growing demand for data with the asset cap lifting. This person will be working on the new project after first learning the existing platform. On a team with 7-8 onshore engineers and 6-7 offshore.
β’ Will do heads down engineering but needs to be able to interact with the existing team and coordinate with them on needs when implementing the new system. Will by doing daily scrum calls, reviewing work orders in order.
β’ Agile is not necessary but good to know.
β’ Banking background is good but not a must.
β’ Certifications are a plus.
Required Skills & Experience
Top Technical Skills
β’ 3+ years of experience with PySpark, S3, Iceberg, Git, Python, Airflow, and Parquet
β’ 5+ years of experience with SQL
β’ Experience with Agile methodologies and tools like Jira
β’ Familiarity with Kafka
β’ Experience with GitHub Copilot, Web Services, Visual studio, IntelliJ, and Gradle
β’ Experience in monitoring tools like Grafana or Prometheus
Preferred Qualifications
β’ Proven experience leading Agile teams and mentoring junior developers
β’ Strong communication skills and the ability to collaborate with business stakeholders
β’ Comfortable working in both Scrum and Kanban model with frequent scrum check-ins
β’ Ability to identify blockers and proactively seek help when needed
β’ Experience working in a regulated environment with a focus on compliance and data governance.
β’ 2+ years of working with Ab Initio graphs and plans
Team Structure & Projects
β’ You will be part of a team that handles over 300+ data feeds, ensuring timely ingestion for nightly batch processing.
β’ Role will focus on Data Services, modernizing data ingestion pipelines.
EEO:
Mindlance is an Equal Opportunity Employer and does not discriminate in employment on the basis of β Minority/Gender/Disability/Religion/LGBTQI/Age/Veterans.