

AI Data Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for an AI Data Engineer in Philadelphia, PA, on a contract basis. Key skills include Python, AWS, ETL, and NLP. Experience with data ingestion, cleaning, and transformation for AI is required. Understanding of vector databases is a plus.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
-
ποΈ - Date discovered
September 4, 2025
π - Project duration
Unknown
-
ποΈ - Location type
On-site
-
π - Contract type
Unknown
-
π - Security clearance
Unknown
-
π - Location detailed
Philadelphia, PA
-
π§ - Skills detailed
#Data Ingestion #Libraries #"ETL (Extract #Transform #Load)" #AI (Artificial Intelligence) #Docker #Data Profiling #Databases #Python #Kubernetes #Data Engineering #AWS (Amazon Web Services) #NLP (Natural Language Processing) #Data Cleaning
Role description
Role: AI Data Engineer (AIRE)
Location: Philadelphia, PA (Onsite)
Duration: Contract
Job Description:
β’ Experience in Data Ingestion and integration from various sources and experience building automated ingestion pipelines using AWS
β’ Experience in Data cleaning, processing and transformation for unstructured and semi-structured data
β’ Experience preparing data for AI usage and experience with data profiling/qualityβidentify data gaps and inconsistencies.
β’ Tools - Strong Python, ETL libraries, SQLs, Docker/Kubernetes, Experience with some AWS Data/Pipeline Services, some exposure to NLP processing, experience with one vector DB
β’ Some Understanding of NLP processing - parsing, chunking, splitting, and tokenization
β’ Some understanding of vector databases and embedding
β’ Any experience working with a graph database is a plus
Role: AI Data Engineer (AIRE)
Location: Philadelphia, PA (Onsite)
Duration: Contract
Job Description:
β’ Experience in Data Ingestion and integration from various sources and experience building automated ingestion pipelines using AWS
β’ Experience in Data cleaning, processing and transformation for unstructured and semi-structured data
β’ Experience preparing data for AI usage and experience with data profiling/qualityβidentify data gaps and inconsistencies.
β’ Tools - Strong Python, ETL libraries, SQLs, Docker/Kubernetes, Experience with some AWS Data/Pipeline Services, some exposure to NLP processing, experience with one vector DB
β’ Some Understanding of NLP processing - parsing, chunking, splitting, and tokenization
β’ Some understanding of vector databases and embedding
β’ Any experience working with a graph database is a plus