

Galent
Data Engineer with AI
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer with AI, offering a contract length of "unknown" and a pay rate of "unknown." Key skills include Python, Apache Spark, AWS services, SQL, and experience with large-scale data systems.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
February 27, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
Unknown
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
Columbus, OH
-
🧠 - Skills detailed
#Databases #S3 (Amazon Simple Storage Service) #Python #AWS (Amazon Web Services) #"ETL (Extract #Transform #Load)" #Airflow #AI (Artificial Intelligence) #Apache Spark #Lambda (AWS Lambda) #Version Control #Kubernetes #Data Modeling #PySpark #Data Engineering #Data Processing #SQL (Structured Query Language) #Docker #GIT #Delta Lake #Spark (Apache Spark) #Redshift #Kafka (Apache Kafka) #Athena
Role description
Required Skills & Qualifications:
• Strong proficiency in Python for data processing and pipeline development
• Hands-on experience with Apache Spark (PySpark preferred)
• Solid experience with AWS services such as S3, Glue, EMR, Redshift, Athena, Lambda
• Experience with SQL and relational/non-relational databases
• Knowledge of data modeling, data warehousing concepts, and ETL frameworks
• Experience working with large-scale, distributed data systems
• Familiarity with CI/CD pipelines and version control tools (Git)
• Strong problem-solving and communication skills
Preferred / Nice to Have:
• Experience with Airflow or other workflow orchestration tools
• Knowledge of Kafka, Kinesis, or streaming data platforms
• Experience with Docker/Kubernetes
• Exposure to Delta Lake, Iceberg, or HuD.
Required Skills & Qualifications:
• Strong proficiency in Python for data processing and pipeline development
• Hands-on experience with Apache Spark (PySpark preferred)
• Solid experience with AWS services such as S3, Glue, EMR, Redshift, Athena, Lambda
• Experience with SQL and relational/non-relational databases
• Knowledge of data modeling, data warehousing concepts, and ETL frameworks
• Experience working with large-scale, distributed data systems
• Familiarity with CI/CD pipelines and version control tools (Git)
• Strong problem-solving and communication skills
Preferred / Nice to Have:
• Experience with Airflow or other workflow orchestration tools
• Knowledge of Kafka, Kinesis, or streaming data platforms
• Experience with Docker/Kubernetes
• Exposure to Delta Lake, Iceberg, or HuD.





