

Lead Data Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Lead Data Engineer with a contract length of "unknown," offering a pay rate of "$/hour." Required skills include SQL, Python, Spark, Databricks, and cloud migration experience (AWS/Azure preferred). Must have 7+ years of relevant experience.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
-
ποΈ - Date discovered
June 10, 2025
π - Project duration
Unknown
-
ποΈ - Location type
Unknown
-
π - Contract type
W2 Contractor
-
π - Security clearance
Unknown
-
π - Location detailed
Columbus, Ohio Metropolitan Area
-
π§ - Skills detailed
#SQL (Structured Query Language) #PySpark #Spark (Apache Spark) #Data Migration #dbt (data build tool) #Data Cleansing #Data Architecture #AWS (Amazon Web Services) #Leadership #Python #Databricks #Terraform #Cloud #Migration #Database Design #DevOps #Azure #Data Quality #Scripting #"ETL (Extract #Transform #Load)" #Data Modeling #JSON (JavaScript Object Notation) #Data Engineering
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
β’
β’
β’ Candidate Must Hold Current Authorization To Work W2, Without The Need Of Future Sponsorship, Cannot Work C2C
β’
β’
β’ Agility Partners is seeking a qualified Data Engineer. This is an exciting opportunity to utilize your skills in Data modeling and advanced SQL techniques. This will give an engineer the opportunity to use their skills in cloud migration methodologies and processes.
Key Responsibilities:
β’ SQL, Python, Spark, Databricks,DBT
β’ Working with varied data file formats (Avro, json, csv) using PySpark for ingesting and transformation
β’ DevOps process and Terraform scripting
β’ Leadership skills: must be able to put together and deliver presentation.
β’ Acting as the lead on a data engineering team
Qualifications:
β’ Experience in Data modeling and advanced SQL techniques
β’ Experience working on cloud migration methodologies (AWS and Azure Preferred)
β’ Expert in SQL, Python, Spark, Databricks
β’ Experience working with varied data file formats (Avro, json, csv) using PySpark for ingesting and transformation
β’ Experience with DevOps process and understanding of Terraform scripting
β’ Understanding the benefits of data warehousing, data architecture, data quality processes, data warehousing design and implementation, table structure, fact and dimension tables, logical and physical database design
β’ Experience designing and implementing ingestion processes for unstructured and structured data sets
β’ Experience designing and developing data cleansing routines utilizing standard data operations
β’ Minimum of 7 years of hands-on experience working on design, configuration, implementation, and data migration for medium to large sized enterprise data platforms.
Reasons to Love It
β’ Work within a collaborative team environment where ideas and creativity are welcomed! Family and Work Life balance are important to this organization and valued for the employees.
β’ Working for an organization that focuses on company culture, inclusion and diversity
β’ 50% medical coverage for you and your entire family, short/long term disability and life insurance options
β’ 401(k)
β’ Life Insurance
β’ Disability coverage